Merge branch 'inactive-accounts' into staging
commit
02fe8bf380
|
@ -36,12 +36,27 @@ Quota.applyCustomLimits = function (Env) {
|
|||
return limits;
|
||||
}(Env.customLimits || {}));
|
||||
|
||||
Env.limits = Env.limits || {};
|
||||
Object.keys(customLimits).forEach(function (k) {
|
||||
if (!isLimit(customLimits[k])) { return; }
|
||||
Env.limits[k] = customLimits[k];
|
||||
});
|
||||
};
|
||||
|
||||
|
||||
/*
|
||||
Env = {
|
||||
myDomain,
|
||||
mySubdomain,
|
||||
adminEmail,
|
||||
Package.version,
|
||||
|
||||
};
|
||||
*/
|
||||
Quota.queryAccountServer = function (Env, cb) {
|
||||
cb = cb; // XXX
|
||||
};
|
||||
|
||||
Quota.updateCachedLimits = function (Env, cb) {
|
||||
Quota.applyCustomLimits(Env);
|
||||
if (Env.blockDailyCheck === true ||
|
||||
|
@ -80,6 +95,8 @@ Quota.updateCachedLimits = function (Env, cb) {
|
|||
var json = JSON.parse(str);
|
||||
Env.limits = json;
|
||||
Quota.applyCustomLimits(Env);
|
||||
//console.log('Env.customLimits', Env.customLimits);
|
||||
//console.log('Env.limits', Env.limits);
|
||||
cb(void 0);
|
||||
} catch (e) {
|
||||
cb(e);
|
||||
|
|
|
@ -0,0 +1,535 @@
|
|||
var nThen = require("nthen");
|
||||
var Bloom = require("@mcrowe/minibloom");
|
||||
var Util = require("../lib/common-util");
|
||||
var Pins = require("../lib/pins");
|
||||
|
||||
var getNewestTime = function (stats) {
|
||||
return stats[['atime', 'ctime', 'mtime'].reduce(function (a, b) {
|
||||
return stats[b] > stats[a]? b: a;
|
||||
})];
|
||||
};
|
||||
|
||||
/*
|
||||
|
||||
Env = {
|
||||
limits: {
|
||||
<unsafeKey>: <limit>,
|
||||
},
|
||||
archiveRetentionTime: <number of days>,
|
||||
accountRetentionTime: <number of days>,
|
||||
inactiveTime: <number of days>,
|
||||
paths: {
|
||||
pin: <path to pin storage>
|
||||
},
|
||||
store,
|
||||
pinStore,
|
||||
Log,
|
||||
blobStore,
|
||||
};
|
||||
|
||||
*/
|
||||
|
||||
module.exports = function (Env, cb) {
|
||||
var complete = Util.once(Util.mkAsync(cb));
|
||||
|
||||
// the administrator should have set an 'inactiveTime' in their config
|
||||
// if they didn't, just exit.
|
||||
if (!Env.inactiveTime || typeof(Env.inactiveTime) !== "number") {
|
||||
return void complete("NO_INACTIVE_TIME");
|
||||
}
|
||||
|
||||
// get a list of premium accounts on this instance
|
||||
// pre-converted to the 'safeKey' format so we can easily compare
|
||||
// them against ids we see on the filesystem
|
||||
var premiumSafeKeys = Object.keys(Env.limits || {})
|
||||
.filter(function (key) {
|
||||
return key.length === 44;
|
||||
})
|
||||
.map(function (unsafeKey) {
|
||||
return Util.escapeKeyCharacters(unsafeKey);
|
||||
});
|
||||
|
||||
// files which have not been changed since before this date can be considered inactive
|
||||
var inactiveTime = +new Date() - (Env.inactiveTime * 24 * 3600 * 1000);
|
||||
|
||||
// files which were archived before this date can be considered safe to remove
|
||||
var retentionTime = +new Date() - (Env.archiveRetentionTime * 24 * 3600 * 1000);
|
||||
|
||||
var store;
|
||||
var pinStore;
|
||||
var Log;
|
||||
var blobs;
|
||||
|
||||
/* It's fairly easy to know if a channel or blob is active
|
||||
but knowing whether it is pinned requires that we
|
||||
keep the set of pinned documents in memory.
|
||||
|
||||
Some users will share the same set of documents in their pin lists,
|
||||
so the representation of pinned documents should scale sub-linearly
|
||||
with the number of users and pinned documents.
|
||||
|
||||
That said, sub-linear isn't great...
|
||||
A Bloom filter is "a space-efficient probabilistic data structure"
|
||||
which lets us check whether an item is _probably_ or _definitely not_
|
||||
in a set. This is good enough for our purposes since we just want to
|
||||
know whether something can safely be removed and false negatives
|
||||
(not safe to remove when it actually is) are acceptable.
|
||||
|
||||
We set our capacity to some large number, and the error rate to whatever
|
||||
we think is acceptable.
|
||||
|
||||
TODO make this configurable ?
|
||||
*/
|
||||
var BLOOM_CAPACITY = (1 << 20) - 1; // over a million items
|
||||
var BLOOM_ERROR = 1 / 1000; // an error rate of one in a thousand
|
||||
|
||||
// we'll use one filter for the set of active documents
|
||||
var activeDocs = Bloom.optimalFilter(BLOOM_CAPACITY, BLOOM_ERROR);
|
||||
// and another one for the set of pinned documents
|
||||
var pinnedDocs = Bloom. optimalFilter(BLOOM_CAPACITY, BLOOM_ERROR);
|
||||
|
||||
var startTime = +new Date();
|
||||
var msSinceStart = function () {
|
||||
return (+new Date()) - startTime;
|
||||
};
|
||||
|
||||
var loadStorage = function () {
|
||||
store = Env.store;
|
||||
pinStore = Env.pinStore;
|
||||
Log = Env.Log;
|
||||
blobs = Env.blobStore;
|
||||
};
|
||||
|
||||
var removeArchivedChannels = function (w) {
|
||||
// this block will iterate over archived channels and removes them
|
||||
// if they've been in cold storage for longer than your configured archive time
|
||||
|
||||
// if the admin has not set an 'archiveRetentionTime', this block makes no sense
|
||||
// so just skip it
|
||||
if (typeof(Env.archiveRetentionTime) !== "number") { return; }
|
||||
|
||||
// count the number of files which have been removed in this run
|
||||
var removed = 0;
|
||||
var accounts = 0;
|
||||
|
||||
var handler = function (err, item, cb) {
|
||||
if (err) {
|
||||
Log.error('EVICT_ARCHIVED_CHANNEL_ITERATION', err);
|
||||
return void cb();
|
||||
}
|
||||
// don't mess with files that are freshly stored in cold storage
|
||||
// based on ctime because that's changed when the file is moved...
|
||||
if (+new Date(item.ctime) > retentionTime) {
|
||||
return void cb();
|
||||
}
|
||||
|
||||
// but if it's been stored for the configured time...
|
||||
// expire it
|
||||
store.removeArchivedChannel(item.channel, w(function (err) {
|
||||
if (err) {
|
||||
Log.error('EVICT_ARCHIVED_CHANNEL_REMOVAL_ERROR', {
|
||||
error: err,
|
||||
channel: item.channel,
|
||||
});
|
||||
return void cb();
|
||||
}
|
||||
Log.info('EVICT_ARCHIVED_CHANNEL_REMOVAL', item.channel);
|
||||
|
||||
if (item.channel.length === 32) {
|
||||
removed++;
|
||||
} else if (item.channel.length === 44) {
|
||||
accounts++;
|
||||
}
|
||||
|
||||
cb();
|
||||
}));
|
||||
};
|
||||
|
||||
// if you hit an error, log it
|
||||
// otherwise, when there are no more channels to process
|
||||
// log some stats about how many were removed
|
||||
var done = function (err) {
|
||||
if (err) {
|
||||
return Log.error('EVICT_ARCHIVED_FINAL_ERROR', err);
|
||||
}
|
||||
Log.info('EVICT_ARCHIVED_CHANNELS_REMOVED', removed);
|
||||
Log.info('EVICT_ARCHIVED_ACCOUNTS_REMOVED', accounts);
|
||||
};
|
||||
|
||||
store.listArchivedChannels(handler, w(done));
|
||||
};
|
||||
|
||||
var removeArchivedBlobProofs = function (w) {
|
||||
if (typeof(Env.archiveRetentionTime) !== "number") { return; }
|
||||
// Iterate over archive blob ownership proofs and remove them
|
||||
// if they are older than the specified retention time
|
||||
var removed = 0;
|
||||
blobs.list.archived.proofs(function (err, item, next) {
|
||||
if (err) {
|
||||
Log.error("EVICT_BLOB_LIST_ARCHIVED_PROOF_ERROR", err);
|
||||
return void next();
|
||||
}
|
||||
if (item && getNewestTime(item) > retentionTime) { return void next(); }
|
||||
blobs.remove.archived.proof(item.safeKey, item.blobId, (function (err) {
|
||||
if (err) {
|
||||
Log.error("EVICT_ARCHIVED_BLOB_PROOF_ERROR", item);
|
||||
return void next();
|
||||
}
|
||||
Log.info("EVICT_ARCHIVED_BLOB_PROOF", item);
|
||||
removed++;
|
||||
next();
|
||||
}));
|
||||
}, w(function () {
|
||||
Log.info('EVICT_ARCHIVED_BLOB_PROOFS_REMOVED', removed);
|
||||
}));
|
||||
};
|
||||
|
||||
var removeArchivedBlobs = function (w) {
|
||||
if (typeof(Env.archiveRetentionTime) !== "number") { return; }
|
||||
// Iterate over archived blobs and remove them
|
||||
// if they are older than the specified retention time
|
||||
var removed = 0;
|
||||
blobs.list.archived.blobs(function (err, item, next) {
|
||||
if (err) {
|
||||
Log.error("EVICT_BLOB_LIST_ARCHIVED_BLOBS_ERROR", err);
|
||||
return void next();
|
||||
}
|
||||
if (item && getNewestTime(item) > retentionTime) { return void next(); }
|
||||
blobs.remove.archived.blob(item.blobId, function (err) {
|
||||
if (err) {
|
||||
Log.error("EVICT_ARCHIVED_BLOB_ERROR", item);
|
||||
return void next();
|
||||
}
|
||||
Log.info("EVICT_ARCHIVED_BLOB", item);
|
||||
removed++;
|
||||
next();
|
||||
});
|
||||
}, w(function () {
|
||||
Log.info('EVICT_ARCHIVED_BLOBS_REMOVED', removed);
|
||||
}));
|
||||
};
|
||||
|
||||
var categorizeChannelsByActivity = function (w) {
|
||||
var channels = 0;
|
||||
var active = 0;
|
||||
var handler = function (err, item, cb) {
|
||||
channels++;
|
||||
if (err) {
|
||||
Log.error('EVICT_CHANNEL_CATEGORIZATION', err);
|
||||
return void cb();
|
||||
}
|
||||
|
||||
// if the channel has been modified recently
|
||||
// we don't use mtime because we don't want to count access to the file, just modifications
|
||||
if (+new Date(item.mtime) > inactiveTime) {
|
||||
// add it to the set of activeDocs
|
||||
activeDocs.add(item.channel);
|
||||
active++;
|
||||
return void cb();
|
||||
}
|
||||
|
||||
return void cb();
|
||||
};
|
||||
|
||||
var done = function () {
|
||||
Log.info('EVICT_CHANNELS_CATEGORIZED', {
|
||||
active: active,
|
||||
channels: channels,
|
||||
});
|
||||
};
|
||||
|
||||
store.listChannels(handler, w(done));
|
||||
};
|
||||
|
||||
var categorizeBlobsByActivity = function (w) {
|
||||
var n_blobs = 0;
|
||||
var active = 0;
|
||||
|
||||
blobs.list.blobs(function (err, item, next) {
|
||||
n_blobs++;
|
||||
if (err) {
|
||||
Log.error("EVICT_BLOB_CATEGORIZATION", err);
|
||||
return void next();
|
||||
}
|
||||
if (!item) {
|
||||
next();
|
||||
return void Log.error("EVICT_BLOB_CATEGORIZATION_INVALID", item);
|
||||
}
|
||||
if (getNewestTime(item) > inactiveTime) {
|
||||
activeDocs.add(item.blobId);
|
||||
active++;
|
||||
return void next();
|
||||
}
|
||||
next();
|
||||
}, w(function () {
|
||||
Log.info('EVICT_BLOBS_CATEGORIZED', {
|
||||
active: active,
|
||||
blobs: n_blobs,
|
||||
});
|
||||
}));
|
||||
};
|
||||
|
||||
var categorizeAccountsByActivity = function (w) {
|
||||
// iterate over all accounts
|
||||
var accounts = 0;
|
||||
var inactive = 0;
|
||||
|
||||
var accountRetentionTime;
|
||||
if (typeof(Env.accountRetentionTime) === 'number' && Env.accountRetentionTime > 0) {
|
||||
accountRetentionTime = +new Date() - (24 * 3600 * 1000 * Env.accountRetentionTime);
|
||||
} else {
|
||||
accountRetentionTime = -1;
|
||||
}
|
||||
|
||||
var pinAll = function (pinList) {
|
||||
pinList.forEach(function (docId) {
|
||||
pinnedDocs.add(docId);
|
||||
});
|
||||
};
|
||||
|
||||
var docIsActive = function (docId) {
|
||||
return activeDocs.test(docId);
|
||||
};
|
||||
|
||||
var accountIsActive = function (mtime, pinList, id) {
|
||||
// console.log("id [%s] in premiumSafeKeys", id, premiumSafeKeys.indexOf(id) !== -1);
|
||||
// if their pin log has changed recently then consider them active
|
||||
if (mtime && mtime > accountRetentionTime) {
|
||||
return true;
|
||||
}
|
||||
|
||||
// iterate over their pinned documents until you find one that has been active
|
||||
if (pinList.some(docIsActive)) {
|
||||
return true;
|
||||
}
|
||||
|
||||
// Finally, make sure it's not a premium account
|
||||
return premiumSafeKeys.indexOf(id) !== -1;
|
||||
};
|
||||
|
||||
var PRESERVE_INACTIVE_ACCOUNTS = accountRetentionTime <= 0;
|
||||
|
||||
// otherwise, we'll only retain data from active accounts
|
||||
// so we need more heuristics
|
||||
var handler = function (content, id, next) {
|
||||
accounts++;
|
||||
|
||||
var mtime = content.latest;
|
||||
var pinList = Object.keys(content.pins);
|
||||
|
||||
if (accountIsActive(mtime, pinList, id)) {
|
||||
// add active accounts' pinned documents to a second bloom filter
|
||||
pinAll(pinList);
|
||||
return void next();
|
||||
}
|
||||
|
||||
// Otherwise they are inactive.
|
||||
// We keep track of how many accounts are inactive whether or not
|
||||
// we plan to delete them, because it may be interesting information
|
||||
inactive++;
|
||||
if (PRESERVE_INACTIVE_ACCOUNTS) {
|
||||
pinAll(pinList);
|
||||
return void next();
|
||||
}
|
||||
|
||||
// remove the pin logs of inactive accounts if inactive account removal is configured
|
||||
pinStore.archiveChannel(id, function (err) {
|
||||
if (err) {
|
||||
Log.error('EVICT_INACTIVE_ACCOUNT_PIN_LOG', err);
|
||||
return void next();
|
||||
}
|
||||
Log.info('EVICT_INACTIVE_ACCOUNT_LOG', id);
|
||||
next();
|
||||
});
|
||||
};
|
||||
|
||||
var done = function () {
|
||||
var label = PRESERVE_INACTIVE_ACCOUNTS?
|
||||
"EVICT_COUNT_ACCOUNTS":
|
||||
"EVICT_INACTIVE_ACCOUNTS";
|
||||
|
||||
Log.info(label, {
|
||||
accounts: accounts,
|
||||
inactive: inactive,
|
||||
});
|
||||
};
|
||||
|
||||
Pins.load(w(done), {
|
||||
pinPath: Env.paths.pin,
|
||||
handler: handler,
|
||||
});
|
||||
};
|
||||
|
||||
var archiveInactiveBlobs = function (w) {
|
||||
// iterate over blobs and remove them
|
||||
// if they have not been accessed within the specified retention time
|
||||
var removed = 0;
|
||||
blobs.list.blobs(function (err, item, next) {
|
||||
if (err) {
|
||||
Log.error("EVICT_BLOB_LIST_BLOBS_ERROR", err);
|
||||
return void next();
|
||||
}
|
||||
if (!item) {
|
||||
next();
|
||||
return void Log.error('EVICT_BLOB_LIST_BLOBS_NO_ITEM', item);
|
||||
}
|
||||
if (pinnedDocs.test(item.blobId)) { return void next(); }
|
||||
if (activeDocs.test(item.blobId)) { return void next(); }
|
||||
|
||||
// This seems redundant because we're already checking the bloom filter
|
||||
// but we can't implement a 'fast mode' for the iterator
|
||||
// unless we address this race condition with this last-minute double-check
|
||||
if (getNewestTime(item) > inactiveTime) { return void next(); }
|
||||
|
||||
blobs.archive.blob(item.blobId, function (err) {
|
||||
if (err) {
|
||||
Log.error("EVICT_ARCHIVE_BLOB_ERROR", {
|
||||
error: err,
|
||||
item: item,
|
||||
});
|
||||
return void next();
|
||||
}
|
||||
Log.info("EVICT_ARCHIVE_BLOB", {
|
||||
item: item,
|
||||
});
|
||||
removed++;
|
||||
next();
|
||||
});
|
||||
}, w(function () {
|
||||
Log.info('EVICT_BLOBS_REMOVED', removed);
|
||||
}));
|
||||
};
|
||||
|
||||
var archiveInactiveBlobProofs = function (w) {
|
||||
// iterate over blob proofs and remove them
|
||||
// if they don't correspond to a pinned or active file
|
||||
var removed = 0;
|
||||
blobs.list.proofs(function (err, item, next) {
|
||||
if (err) {
|
||||
next();
|
||||
return void Log.error("EVICT_BLOB_LIST_PROOFS_ERROR", err);
|
||||
}
|
||||
if (!item) {
|
||||
next();
|
||||
return void Log.error('EVICT_BLOB_LIST_PROOFS_NO_ITEM', item);
|
||||
}
|
||||
if (pinnedDocs.test(item.blobId)) { return void next(); }
|
||||
if (getNewestTime(item) > inactiveTime) { return void next(); }
|
||||
nThen(function (w) {
|
||||
blobs.size(item.blobId, w(function (err, size) {
|
||||
if (err) {
|
||||
w.abort();
|
||||
next();
|
||||
return void Log.error("EVICT_BLOB_LIST_PROOFS_ERROR", err);
|
||||
}
|
||||
if (size !== 0) {
|
||||
w.abort();
|
||||
next();
|
||||
}
|
||||
}));
|
||||
}).nThen(function () {
|
||||
blobs.remove.proof(item.safeKey, item.blobId, function (err) {
|
||||
next();
|
||||
if (err) {
|
||||
return Log.error("EVICT_BLOB_PROOF_LONELY_ERROR", item);
|
||||
}
|
||||
removed++;
|
||||
return Log.info("EVICT_BLOB_PROOF_LONELY", item);
|
||||
});
|
||||
});
|
||||
}, w(function () {
|
||||
Log.info("EVICT_BLOB_PROOFS_REMOVED", removed);
|
||||
}));
|
||||
};
|
||||
|
||||
var archiveInactiveChannels = function (w) {
|
||||
var channels = 0;
|
||||
var archived = 0;
|
||||
|
||||
var handler = function (err, item, cb) {
|
||||
channels++;
|
||||
if (err) {
|
||||
Log.error('EVICT_CHANNEL_ITERATION', err);
|
||||
return void cb();
|
||||
}
|
||||
// check if the database has any ephemeral channels
|
||||
// if it does it's because of a bug, and they should be removed
|
||||
if (item.channel.length === 34) {
|
||||
return void store.removeChannel(item.channel, w(function (err) {
|
||||
if (err) {
|
||||
Log.error('EVICT_EPHEMERAL_CHANNEL_REMOVAL_ERROR', {
|
||||
error: err,
|
||||
channel: item.channel,
|
||||
});
|
||||
return void cb();
|
||||
}
|
||||
Log.info('EVICT_EPHEMERAL_CHANNEL_REMOVAL', item.channel);
|
||||
cb();
|
||||
}));
|
||||
}
|
||||
|
||||
// bail out if the channel is in the set of activeDocs
|
||||
if (activeDocs.test(item.channel)) { return void cb(); }
|
||||
|
||||
// ignore the channel if it's pinned
|
||||
if (pinnedDocs.test(item.channel)) { return void cb(); }
|
||||
|
||||
nThen(function (w) {
|
||||
// double check that the channel really is inactive before archiving it
|
||||
// because it might have been created after the initial activity scan
|
||||
store.getChannelStats(item.channel, w(function (err, newerItem) {
|
||||
if (err) { return; }
|
||||
if (item && getNewestTime(newerItem) > retentionTime) {
|
||||
// it's actually active, so don't archive it.
|
||||
w.abort();
|
||||
cb();
|
||||
}
|
||||
// else fall through to the archival
|
||||
}));
|
||||
}).nThen(function () {
|
||||
return void store.archiveChannel(item.channel, w(function (err) {
|
||||
if (err) {
|
||||
Log.error('EVICT_CHANNEL_ARCHIVAL_ERROR', {
|
||||
error: err,
|
||||
channel: item.channel,
|
||||
});
|
||||
return void cb();
|
||||
}
|
||||
Log.info('EVICT_CHANNEL_ARCHIVAL', item.channel);
|
||||
archived++;
|
||||
cb();
|
||||
}));
|
||||
});
|
||||
};
|
||||
|
||||
var done = function () {
|
||||
return void Log.info('EVICT_CHANNELS_ARCHIVED', archived);
|
||||
};
|
||||
|
||||
store.listChannels(handler, w(done), true); // using a hacky "fast mode" since we only need the channel id
|
||||
};
|
||||
|
||||
nThen(loadStorage)
|
||||
.nThen(removeArchivedChannels)
|
||||
.nThen(removeArchivedBlobProofs)
|
||||
.nThen(removeArchivedBlobs)
|
||||
|
||||
// iterate over all documents and add them to a bloom filter if they have been active
|
||||
.nThen(categorizeChannelsByActivity)
|
||||
.nThen(categorizeBlobsByActivity)
|
||||
|
||||
// iterate over all accounts and add them to a bloom filter if they are active
|
||||
.nThen(categorizeAccountsByActivity)
|
||||
|
||||
// iterate again and archive inactive unpinned documents
|
||||
// (documents which are not in either bloom filter)
|
||||
|
||||
.nThen(archiveInactiveBlobs)
|
||||
.nThen(archiveInactiveBlobProofs)
|
||||
.nThen(archiveInactiveChannels)
|
||||
.nThen(function () {
|
||||
Log.info("EVICT_TIME_TO_RUN_SCRIPT", msSinceStart());
|
||||
}).nThen(function () {
|
||||
complete();
|
||||
});
|
||||
};
|
|
@ -11,6 +11,7 @@ const Core = require("./commands/core");
|
|||
const Store = require("./storage/file");
|
||||
const BlobStore = require("./storage/blob");
|
||||
const Workers = require("./workers/index");
|
||||
//const Eviction = require("./eviction");
|
||||
|
||||
module.exports.create = function (config, cb) {
|
||||
const Log = config.log;
|
||||
|
@ -26,14 +27,15 @@ module.exports.create = function (config, cb) {
|
|||
|
||||
Log.silly('HK_LOADING', 'LOADING HISTORY_KEEPER MODULE');
|
||||
|
||||
// TODO populate Env with everything that you use from config
|
||||
// so that you can stop passing around your raw config
|
||||
// and more easily share state between historyKeeper and rpc
|
||||
const Env = {
|
||||
Log: Log,
|
||||
// store
|
||||
id: Crypto.randomBytes(8).toString('hex'),
|
||||
|
||||
inactiveTime: config.inactiveTime,
|
||||
archiveRetentionTime: config.archiveRetentionTime,
|
||||
accountRetentionTime: config.accountRetentionTime,
|
||||
|
||||
metadata_cache: {},
|
||||
channel_cache: {},
|
||||
queueStorage: WriteQueue(),
|
||||
|
@ -254,6 +256,10 @@ module.exports.create = function (config, cb) {
|
|||
verbose: config.verbose,
|
||||
openFileLimit: config.openFileLimit,
|
||||
|
||||
inactiveTime: config.inactiveTime,
|
||||
archiveRetentionTime: config.archiveRetentionTime,
|
||||
accountRetentionTime: config.accountRetentionTime,
|
||||
|
||||
maxWorkers: config.maxWorkers,
|
||||
}, w(function (err) {
|
||||
if (err) {
|
||||
|
@ -261,8 +267,8 @@ module.exports.create = function (config, cb) {
|
|||
}
|
||||
}));
|
||||
}).nThen(function () {
|
||||
if (config.disableIntegratedTasks) { return; }
|
||||
config.intervals = config.intervals || {};
|
||||
if (config.disableIntegratedTasks) { return; }
|
||||
|
||||
var tasks_running;
|
||||
config.intervals.taskExpiration = setInterval(function () {
|
||||
|
@ -275,6 +281,30 @@ module.exports.create = function (config, cb) {
|
|||
tasks_running = false;
|
||||
});
|
||||
}, 1000 * 60 * 5); // run every five minutes
|
||||
}).nThen(function () {
|
||||
if (config.disableIntegratedEviction) { return; }
|
||||
const ONE_DAY = 24 * 1000 * 60 * 60;
|
||||
// setting the time of the last eviction to "now"
|
||||
// effectively makes it so that we'll start evicting after the server
|
||||
// has been up for at least one day
|
||||
var last_eviction = +new Date();
|
||||
|
||||
var active = false;
|
||||
config.intervals.eviction = setInterval(function () {
|
||||
if (active) { return; }
|
||||
var now = +new Date();
|
||||
// evict inactive data once per day
|
||||
if (last_eviction && (now - ONE_DAY) < last_eviction) { return; }
|
||||
active = true;
|
||||
Env.evictInactive(function (err) {
|
||||
if (err) {
|
||||
// NO_INACTIVE_TIME
|
||||
Log.error('EVICT_INACTIVE_MAIN_ERROR', err);
|
||||
}
|
||||
active = false;
|
||||
last_eviction = now;
|
||||
});
|
||||
}, 60 * 1000);
|
||||
}).nThen(function () {
|
||||
RPC.create(Env, function (err, _rpc) {
|
||||
if (err) { throw err; }
|
||||
|
|
|
@ -421,6 +421,67 @@ var removeArchivedChannel = function (env, channelName, cb) {
|
|||
});
|
||||
};
|
||||
|
||||
var _getStats = function (metadataPath, filePath, channel, cb, isLonelyMetadata) {
|
||||
var metaStat, channelStat;
|
||||
var metaErr, channelErr;
|
||||
nThen(function (ww) {
|
||||
// get the stats for the metadata
|
||||
Fs.stat(metadataPath, ww(function (err, stats) {
|
||||
if (err) {
|
||||
metaErr = err;
|
||||
return;
|
||||
}
|
||||
metaStat = stats;
|
||||
}));
|
||||
|
||||
if (isLonelyMetadata) { return; }
|
||||
|
||||
Fs.stat(filePath, ww(function (err, stats) {
|
||||
if (err) {
|
||||
channelErr = err;
|
||||
return;
|
||||
}
|
||||
channelStat = stats;
|
||||
}));
|
||||
}).nThen(function () {
|
||||
if (channelErr && metaErr) {
|
||||
return void cb(channelErr);
|
||||
}
|
||||
|
||||
var data = {
|
||||
channel: channel,
|
||||
};
|
||||
|
||||
if (metaStat && channelStat) {
|
||||
// take max of times returned by either stat
|
||||
data.atime = Math.max(channelStat.atime, metaStat.atime);
|
||||
data.mtime = Math.max(channelStat.mtime, metaStat.mtime);
|
||||
data.ctime = Math.max(channelStat.ctime, metaStat.ctime);
|
||||
// return the sum of the size of the two files
|
||||
data.size = channelStat.size + metaStat.size;
|
||||
} else if (metaStat) {
|
||||
data.atime = metaStat.atime;
|
||||
data.mtime = metaStat.mtime;
|
||||
data.ctime = metaStat.ctime;
|
||||
data.size = metaStat.size;
|
||||
} else if (channelStat) {
|
||||
data.atime = channelStat.atime;
|
||||
data.mtime = channelStat.mtime;
|
||||
data.ctime = channelStat.ctime;
|
||||
data.size = channelStat.size;
|
||||
} else {
|
||||
return void cb('NO_DATA');
|
||||
}
|
||||
cb(void 0, data);
|
||||
});
|
||||
};
|
||||
|
||||
var getStats = function (env, channelName, cb) {
|
||||
var metadataPath = mkMetadataPath(env, channelName);
|
||||
var filePath = mkPath(env, channelName);
|
||||
_getStats(metadataPath, filePath, channelName, cb);
|
||||
};
|
||||
|
||||
// TODO use ../plan.js for a smaller memory footprint
|
||||
var listChannels = function (root, handler, cb, fast) {
|
||||
// do twenty things at a time
|
||||
|
@ -491,60 +552,12 @@ var listChannels = function (root, handler, cb, fast) {
|
|||
|
||||
var filePath = Path.join(nestedDirPath, channelName);
|
||||
var metadataPath = Path.join(nestedDirPath, metadataName);
|
||||
|
||||
var metaStat, channelStat;
|
||||
var metaErr, channelErr;
|
||||
nThen(function (ww) {
|
||||
// get the stats for the metadata
|
||||
Fs.stat(metadataPath, ww(function (err, stats) {
|
||||
if (err) {
|
||||
metaErr = err;
|
||||
return;
|
||||
}
|
||||
metaStat = stats;
|
||||
}));
|
||||
|
||||
if (isLonelyMetadata) { return; }
|
||||
|
||||
Fs.stat(filePath, ww(function (err, stats) {
|
||||
if (err) {
|
||||
channelErr = err;
|
||||
return;
|
||||
}
|
||||
channelStat = stats;
|
||||
}));
|
||||
}).nThen(function () {
|
||||
if (channelErr && metaErr) {
|
||||
return void handler(channelErr, void 0, next);
|
||||
return void _getStats(metadataPath, filePath, channel, function (err, data) {
|
||||
if (err) {
|
||||
return void handler(err, void 0, next);
|
||||
}
|
||||
|
||||
var data = {
|
||||
channel: channel,
|
||||
};
|
||||
|
||||
if (metaStat && channelStat) {
|
||||
// take max of times returned by either stat
|
||||
data.atime = Math.max(channelStat.atime, metaStat.atime);
|
||||
data.mtime = Math.max(channelStat.mtime, metaStat.mtime);
|
||||
data.ctime = Math.max(channelStat.ctime, metaStat.ctime);
|
||||
// return the sum of the size of the two files
|
||||
data.size = channelStat.size + metaStat.size;
|
||||
} else if (metaStat) {
|
||||
data.atime = metaStat.atime;
|
||||
data.mtime = metaStat.mtime;
|
||||
data.ctime = metaStat.ctime;
|
||||
data.size = metaStat.size;
|
||||
} else if (channelStat) {
|
||||
data.atime = channelStat.atime;
|
||||
data.mtime = channelStat.mtime;
|
||||
data.ctime = channelStat.ctime;
|
||||
data.size = channelStat.size;
|
||||
} else {
|
||||
return void handler('NO_DATA', void 0, next);
|
||||
}
|
||||
|
||||
handler(void 0, data, next);
|
||||
});
|
||||
}, isLonelyMetadata);
|
||||
});
|
||||
});
|
||||
})));
|
||||
|
@ -766,6 +779,11 @@ const messageBin = (env, chanName, msgBin, cb) => {
|
|||
chan.writeStream.write(msgBin, function () {
|
||||
chan.onError.splice(chan.onError.indexOf(complete), 1);
|
||||
complete();
|
||||
if (chan.onError.length) { return; }
|
||||
if (chan.delayClose && chan.delayClose.clear) {
|
||||
chan.delayClose.clear();
|
||||
delete env.channels[chanName];
|
||||
}
|
||||
});
|
||||
});
|
||||
};
|
||||
|
@ -1198,6 +1216,10 @@ module.exports.create = function (conf, _cb) {
|
|||
listChannels(Path.join(env.archiveRoot, 'datastore'), handler, cb, fastMode);
|
||||
},
|
||||
|
||||
getChannelStats: function (channelName, cb) {
|
||||
if (!isValidChannelId(channelName)) { return void cb(new Error('EINVAL')); }
|
||||
getStats(env, channelName, cb);
|
||||
},
|
||||
getChannelSize: function (channelName, cb) {
|
||||
if (!isValidChannelId(channelName)) { return void cb(new Error('EINVAL')); }
|
||||
// this method should be really fast and it probably doesn't matter much
|
||||
|
|
|
@ -13,6 +13,7 @@ const Saferphore = require("saferphore");
|
|||
const Logger = require("../log");
|
||||
const Tasks = require("../storage/tasks");
|
||||
const Nacl = require('tweetnacl/nacl-fast');
|
||||
const Eviction = require("../eviction");
|
||||
|
||||
const Env = {
|
||||
Log: {},
|
||||
|
@ -39,13 +40,21 @@ const init = function (config, _cb) {
|
|||
return void cb('E_INVALID_CONFIG');
|
||||
}
|
||||
|
||||
Env.paths = {
|
||||
pin: config.pinPath,
|
||||
};
|
||||
|
||||
Env.inactiveTime = config.inactiveTime;
|
||||
Env.archiveRetentionTime = config.archiveRetentionTime;
|
||||
Env.accountRetentionTime = config.accountRetentionTime;
|
||||
|
||||
nThen(function (w) {
|
||||
Store.create(config, w(function (err, _store) {
|
||||
if (err) {
|
||||
w.abort();
|
||||
return void cb(err);
|
||||
}
|
||||
store = _store;
|
||||
Env.store = store = _store;
|
||||
}));
|
||||
Store.create({
|
||||
filePath: config.pinPath,
|
||||
|
@ -54,7 +63,7 @@ const init = function (config, _cb) {
|
|||
w.abort();
|
||||
return void cb(err);
|
||||
}
|
||||
pinStore = _pinStore;
|
||||
Env.pinStore = pinStore = _pinStore;
|
||||
}));
|
||||
BlobStore.create({
|
||||
blobPath: config.blobPath,
|
||||
|
@ -66,7 +75,7 @@ const init = function (config, _cb) {
|
|||
w.abort();
|
||||
return void cb(err);
|
||||
}
|
||||
blobStore = blob;
|
||||
Env.blobStore = blobStore = blob;
|
||||
}));
|
||||
}).nThen(function (w) {
|
||||
Tasks.create({
|
||||
|
@ -436,6 +445,10 @@ const writeTask = function (data, cb) {
|
|||
Env.tasks.write(data.time, data.task_command, data.args, cb);
|
||||
};
|
||||
|
||||
const evictInactive = function (data, cb) {
|
||||
Eviction(Env, cb);
|
||||
};
|
||||
|
||||
const COMMANDS = {
|
||||
COMPUTE_INDEX: computeIndex,
|
||||
COMPUTE_METADATA: computeMetadata,
|
||||
|
@ -449,6 +462,7 @@ const COMMANDS = {
|
|||
REMOVE_OWNED_BLOB: removeOwnedBlob,
|
||||
RUN_TASKS: runTasks,
|
||||
WRITE_TASK: writeTask,
|
||||
EVICT_INACTIVE: evictInactive,
|
||||
};
|
||||
|
||||
COMMANDS.INLINE = function (data, cb) {
|
||||
|
|
|
@ -350,6 +350,12 @@ Workers.initialize = function (Env, config, _cb) {
|
|||
}, cb);
|
||||
};
|
||||
|
||||
Env.evictInactive = function (cb) {
|
||||
sendCommand({
|
||||
command: 'EVICT_INACTIVE',
|
||||
}, cb);
|
||||
};
|
||||
|
||||
Env.runTasks = function (cb) {
|
||||
sendCommand({
|
||||
command: 'RUN_TASKS',
|
||||
|
|
|
@ -1,513 +1,86 @@
|
|||
/* global process */
|
||||
|
||||
var Eviction = require("../lib/eviction");
|
||||
var nThen = require("nthen");
|
||||
|
||||
var Store = require("../lib/storage/file");
|
||||
var BlobStore = require("../lib/storage/blob");
|
||||
var Pins = require("../lib/pins");
|
||||
var Bloom = require("@mcrowe/minibloom");
|
||||
|
||||
var Quota = require("../lib/commands/quota");
|
||||
|
||||
var config = require("../lib/load-config");
|
||||
|
||||
// the administrator should have set an 'inactiveTime' in their config
|
||||
// if they didn't, just exit.
|
||||
if (!config.inactiveTime || typeof(config.inactiveTime) !== "number") { return; }
|
||||
|
||||
// files which have not been changed since before this date can be considered inactive
|
||||
var inactiveTime = +new Date() - (config.inactiveTime * 24 * 3600 * 1000);
|
||||
|
||||
// files which were archived before this date can be considered safe to remove
|
||||
var retentionTime = +new Date() - (config.archiveRetentionTime * 24 * 3600 * 1000);
|
||||
|
||||
var getNewestTime = function (stats) {
|
||||
return stats[['atime', 'ctime', 'mtime'].reduce(function (a, b) {
|
||||
return stats[b] > stats[a]? b: a;
|
||||
})];
|
||||
var Env = {
|
||||
inactiveTime: config.inactiveTime,
|
||||
archiveRetentionTime: config.archiveRetentionTime,
|
||||
accountRetentionTime: config.accountRetentionTime,
|
||||
paths: {
|
||||
pin: config.pinPath,
|
||||
},
|
||||
};
|
||||
|
||||
var store;
|
||||
var pinStore;
|
||||
var Log;
|
||||
var blobs;
|
||||
var prepareEnv = function (Env, cb) {
|
||||
Env.customLimits = config.customLimits;
|
||||
Quota.applyCustomLimits(Env);
|
||||
|
||||
/* It's fairly easy to know if a channel or blob is active
|
||||
but knowing whether it is pinned requires that we
|
||||
keep the set of pinned documents in memory.
|
||||
nThen(function (w) {
|
||||
/* Database adaptors
|
||||
*/
|
||||
|
||||
Some users will share the same set of documents in their pin lists,
|
||||
so the representation of pinned documents should scale sub-linearly
|
||||
with the number of users and pinned documents.
|
||||
|
||||
That said, sub-linear isn't great...
|
||||
A Bloom filter is "a space-efficient probabilistic data structure"
|
||||
which lets us check whether an item is _probably_ or _definitely not_
|
||||
in a set. This is good enough for our purposes since we just want to
|
||||
know whether something can safely be removed and false negatives
|
||||
(not safe to remove when it actually is) are acceptable.
|
||||
|
||||
We set our capacity to some large number, and the error rate to whatever
|
||||
we think is acceptable.
|
||||
*/
|
||||
var BLOOM_CAPACITY = (1 << 20) - 1; // over a million items
|
||||
var BLOOM_ERROR = 1 / 1000; // an error rate of one in a thousand
|
||||
|
||||
// we'll use one filter for the set of active documents
|
||||
var activeDocs = Bloom.optimalFilter(BLOOM_CAPACITY, BLOOM_ERROR);
|
||||
// and another one for the set of pinned documents
|
||||
var pinnedDocs = Bloom. optimalFilter(BLOOM_CAPACITY, BLOOM_ERROR);
|
||||
|
||||
var startTime = +new Date();
|
||||
var msSinceStart = function () {
|
||||
return (+new Date()) - startTime;
|
||||
};
|
||||
|
||||
var loadStorage = function (w) {
|
||||
// load the store which will be used for iterating over channels
|
||||
// and performing operations like archival and deletion
|
||||
Store.create(config, w(function (err, _) {
|
||||
if (err) {
|
||||
w.abort();
|
||||
throw err;
|
||||
}
|
||||
store = _;
|
||||
}));
|
||||
|
||||
Store.create({
|
||||
filePath: config.pinPath,
|
||||
}, w(function (err, _) {
|
||||
if (err) {
|
||||
w.abort();
|
||||
throw err;
|
||||
}
|
||||
pinStore = _;
|
||||
}));
|
||||
|
||||
// load the logging module so that you have a record of which
|
||||
// files were archived or deleted at what time
|
||||
var Logger = require("../lib/log");
|
||||
Logger.create(config, w(function (_) {
|
||||
Log = _;
|
||||
}));
|
||||
|
||||
config.getSession = function () {};
|
||||
BlobStore.create(config, w(function (err, _) {
|
||||
if (err) {
|
||||
w.abort();
|
||||
return console.error(err);
|
||||
}
|
||||
blobs = _;
|
||||
}));
|
||||
};
|
||||
|
||||
var removeArchivedChannels = function (w) {
|
||||
// this block will iterate over archived channels and removes them
|
||||
// if they've been in cold storage for longer than your configured archive time
|
||||
|
||||
// if the admin has not set an 'archiveRetentionTime', this block makes no sense
|
||||
// so just skip it
|
||||
if (typeof(config.archiveRetentionTime) !== "number") { return; }
|
||||
|
||||
// count the number of files which have been removed in this run
|
||||
var removed = 0;
|
||||
var accounts = 0;
|
||||
|
||||
var handler = function (err, item, cb) {
|
||||
if (err) {
|
||||
Log.error('EVICT_ARCHIVED_CHANNEL_ITERATION', err);
|
||||
return void cb();
|
||||
}
|
||||
// don't mess with files that are freshly stored in cold storage
|
||||
// based on ctime because that's changed when the file is moved...
|
||||
if (+new Date(item.ctime) > retentionTime) {
|
||||
return void cb();
|
||||
}
|
||||
|
||||
// but if it's been stored for the configured time...
|
||||
// expire it
|
||||
store.removeArchivedChannel(item.channel, w(function (err) {
|
||||
// load the store which will be used for iterating over channels
|
||||
// and performing operations like archival and deletion
|
||||
Store.create(config, w(function (err, _) {
|
||||
if (err) {
|
||||
Log.error('EVICT_ARCHIVED_CHANNEL_REMOVAL_ERROR', {
|
||||
error: err,
|
||||
channel: item.channel,
|
||||
});
|
||||
return void cb();
|
||||
w.abort();
|
||||
throw err;
|
||||
}
|
||||
Log.info('EVICT_ARCHIVED_CHANNEL_REMOVAL', item.channel);
|
||||
|
||||
if (item.channel.length === 32) {
|
||||
removed++;
|
||||
} else if (item.channel.length === 44) {
|
||||
accounts++;
|
||||
}
|
||||
|
||||
cb();
|
||||
Env.store = _;
|
||||
}));
|
||||
};
|
||||
|
||||
// if you hit an error, log it
|
||||
// otherwise, when there are no more channels to process
|
||||
// log some stats about how many were removed
|
||||
var done = function (err) {
|
||||
if (err) {
|
||||
return Log.error('EVICT_ARCHIVED_FINAL_ERROR', err);
|
||||
}
|
||||
Log.info('EVICT_ARCHIVED_CHANNELS_REMOVED', removed);
|
||||
Log.info('EVICT_ARCHIVED_ACCOUNTS_REMOVED', accounts);
|
||||
};
|
||||
|
||||
store.listArchivedChannels(handler, w(done));
|
||||
};
|
||||
|
||||
var removeArchivedBlobProofs = function (w) {
|
||||
if (typeof(config.archiveRetentionTime) !== "number") { return; }
|
||||
// Iterate over archive blob ownership proofs and remove them
|
||||
// if they are older than the specified retention time
|
||||
var removed = 0;
|
||||
blobs.list.archived.proofs(function (err, item, next) {
|
||||
if (err) {
|
||||
Log.error("EVICT_BLOB_LIST_ARCHIVED_PROOF_ERROR", err);
|
||||
return void next();
|
||||
}
|
||||
if (item && getNewestTime(item) > retentionTime) { return void next(); }
|
||||
blobs.remove.archived.proof(item.safeKey, item.blobId, (function (err) {
|
||||
Store.create({
|
||||
filePath: config.pinPath,
|
||||
}, w(function (err, _) {
|
||||
if (err) {
|
||||
Log.error("EVICT_ARCHIVED_BLOB_PROOF_ERROR", item);
|
||||
return void next();
|
||||
w.abort();
|
||||
throw err;
|
||||
}
|
||||
Log.info("EVICT_ARCHIVED_BLOB_PROOF", item);
|
||||
removed++;
|
||||
next();
|
||||
Env.pinStore = _;
|
||||
}));
|
||||
}, w(function () {
|
||||
Log.info('EVICT_ARCHIVED_BLOB_PROOFS_REMOVED', removed);
|
||||
}));
|
||||
};
|
||||
|
||||
var removeArchivedBlobs = function (w) {
|
||||
if (typeof(config.archiveRetentionTime) !== "number") { return; }
|
||||
// Iterate over archived blobs and remove them
|
||||
// if they are older than the specified retention time
|
||||
var removed = 0;
|
||||
blobs.list.archived.blobs(function (err, item, next) {
|
||||
if (err) {
|
||||
Log.error("EVICT_BLOB_LIST_ARCHIVED_BLOBS_ERROR", err);
|
||||
return void next();
|
||||
}
|
||||
if (item && getNewestTime(item) > retentionTime) { return void next(); }
|
||||
blobs.remove.archived.blob(item.blobId, function (err) {
|
||||
// load the logging module so that you have a record of which
|
||||
// files were archived or deleted at what time
|
||||
var Logger = require("../lib/log");
|
||||
Logger.create(config, w(function (_) {
|
||||
Env.Log = _;
|
||||
}));
|
||||
|
||||
config.getSession = function () {};
|
||||
BlobStore.create(config, w(function (err, _) {
|
||||
if (err) {
|
||||
Log.error("EVICT_ARCHIVED_BLOB_ERROR", item);
|
||||
return void next();
|
||||
w.abort();
|
||||
return console.error(err);
|
||||
}
|
||||
Log.info("EVICT_ARCHIVED_BLOB", item);
|
||||
removed++;
|
||||
next();
|
||||
});
|
||||
}, w(function () {
|
||||
Log.info('EVICT_ARCHIVED_BLOBS_REMOVED', removed);
|
||||
}));
|
||||
};
|
||||
|
||||
var categorizeChannelsByActivity = function (w) {
|
||||
var channels = 0;
|
||||
var active = 0;
|
||||
var handler = function (err, item, cb) {
|
||||
channels++;
|
||||
if (err) {
|
||||
Log.error('EVICT_CHANNEL_CATEGORIZATION', err);
|
||||
return void cb();
|
||||
}
|
||||
|
||||
// if the channel has been modified recently
|
||||
// we don't use mtime because we don't want to count access to the file, just modifications
|
||||
if (+new Date(item.mtime) > inactiveTime) {
|
||||
// add it to the set of activeDocs
|
||||
activeDocs.add(item.channel);
|
||||
active++;
|
||||
return void cb();
|
||||
}
|
||||
|
||||
return void cb();
|
||||
};
|
||||
|
||||
var done = function () {
|
||||
Log.info('EVICT_CHANNELS_CATEGORIZED', {
|
||||
active: active,
|
||||
channels: channels,
|
||||
});
|
||||
};
|
||||
|
||||
store.listChannels(handler, w(done));
|
||||
};
|
||||
|
||||
var categorizeBlobsByActivity = function (w) {
|
||||
var n_blobs = 0;
|
||||
var active = 0;
|
||||
|
||||
blobs.list.blobs(function (err, item, next) {
|
||||
n_blobs++;
|
||||
if (err) {
|
||||
Log.error("EVICT_BLOB_CATEGORIZATION", err);
|
||||
return void next();
|
||||
}
|
||||
if (!item) {
|
||||
next();
|
||||
return void Log.error("EVICT_BLOB_CATEGORIZATION_INVALID", item);
|
||||
}
|
||||
if (getNewestTime(item) > inactiveTime) {
|
||||
activeDocs.add(item.blobId);
|
||||
active++;
|
||||
return void next();
|
||||
}
|
||||
next();
|
||||
}, w(function () {
|
||||
Log.info('EVICT_BLOBS_CATEGORIZED', {
|
||||
active: active,
|
||||
blobs: n_blobs,
|
||||
});
|
||||
}));
|
||||
};
|
||||
|
||||
var categorizeAccountsByActivity = function (w) {
|
||||
// iterate over all accounts
|
||||
var accounts = 0;
|
||||
var inactive = 0;
|
||||
|
||||
var accountRetentionTime;
|
||||
if (typeof(config.accountRetentionTime) === 'number' && config.accountRetentionTime > 0) {
|
||||
accountRetentionTime = +new Date() - (24 * 3600 * 1000 * config.accountRetentionTime);
|
||||
} else {
|
||||
accountRetentionTime = -1;
|
||||
}
|
||||
|
||||
var pinAll = function (pinList) {
|
||||
pinList.forEach(function (docId) {
|
||||
pinnedDocs.add(docId);
|
||||
});
|
||||
};
|
||||
|
||||
var accountIsActive = function (mtime, pinList) {
|
||||
// if their pin log has changed recently then consider them active
|
||||
if (mtime && mtime > accountRetentionTime) {
|
||||
return true;
|
||||
}
|
||||
// otherwise iterate over their pinned documents until you find one that has been active
|
||||
return pinList.some(function (docId) {
|
||||
return activeDocs.test(docId);
|
||||
});
|
||||
};
|
||||
|
||||
var PRESERVE_INACTIVE_ACCOUNTS = accountRetentionTime <= 0;
|
||||
|
||||
// otherwise, we'll only retain data from active accounts
|
||||
// so we need more heuristics
|
||||
var handler = function (content, id, next) {
|
||||
accounts++;
|
||||
|
||||
var mtime = content.latest;
|
||||
var pinList = Object.keys(content.pins);
|
||||
|
||||
if (accountIsActive(mtime, pinList)) {
|
||||
// add active accounts' pinned documents to a second bloom filter
|
||||
pinAll(pinList);
|
||||
return void next();
|
||||
}
|
||||
|
||||
// Otherwise they are inactive.
|
||||
// We keep track of how many accounts are inactive whether or not
|
||||
// we plan to delete them, because it may be interesting information
|
||||
inactive++;
|
||||
if (PRESERVE_INACTIVE_ACCOUNTS) {
|
||||
pinAll(pinList);
|
||||
return void next();
|
||||
}
|
||||
|
||||
// remove the pin logs of inactive accounts if inactive account removal is configured
|
||||
pinStore.archiveChannel(id, function (err) {
|
||||
if (err) {
|
||||
Log.error('EVICT_INACTIVE_ACCOUNT_PIN_LOG', err);
|
||||
return void next();
|
||||
}
|
||||
Log.info('EVICT_INACTIVE_ACCOUNT_LOG', id);
|
||||
next();
|
||||
});
|
||||
};
|
||||
|
||||
var done = function () {
|
||||
var label = PRESERVE_INACTIVE_ACCOUNTS?
|
||||
"EVICT_COUNT_ACCOUNTS":
|
||||
"EVICT_INACTIVE_ACCOUNTS";
|
||||
|
||||
Log.info(label, {
|
||||
accounts: accounts,
|
||||
inactive: inactive,
|
||||
});
|
||||
};
|
||||
|
||||
Pins.load(w(done), {
|
||||
pinPath: config.pinPath,
|
||||
handler: handler,
|
||||
Env.blobStore = _;
|
||||
}));
|
||||
}).nThen(function () {
|
||||
cb();
|
||||
});
|
||||
};
|
||||
|
||||
var archiveInactiveBlobs = function (w) {
|
||||
// iterate over blobs and remove them
|
||||
// if they have not been accessed within the specified retention time
|
||||
var removed = 0;
|
||||
blobs.list.blobs(function (err, item, next) {
|
||||
if (err) {
|
||||
Log.error("EVICT_BLOB_LIST_BLOBS_ERROR", err);
|
||||
return void next();
|
||||
}
|
||||
if (!item) {
|
||||
next();
|
||||
return void Log.error('EVICT_BLOB_LIST_BLOBS_NO_ITEM', item);
|
||||
}
|
||||
if (pinnedDocs.test(item.blobId)) { return void next(); }
|
||||
if (activeDocs.test(item.blobId)) { return void next(); }
|
||||
|
||||
blobs.archive.blob(item.blobId, function (err) {
|
||||
if (err) {
|
||||
Log.error("EVICT_ARCHIVE_BLOB_ERROR", {
|
||||
error: err,
|
||||
item: item,
|
||||
});
|
||||
return void next();
|
||||
}
|
||||
Log.info("EVICT_ARCHIVE_BLOB", {
|
||||
item: item,
|
||||
});
|
||||
removed++;
|
||||
next();
|
||||
});
|
||||
}, w(function () {
|
||||
Log.info('EVICT_BLOBS_REMOVED', removed);
|
||||
}));
|
||||
};
|
||||
|
||||
var archiveInactiveBlobProofs = function (w) {
|
||||
// iterate over blob proofs and remove them
|
||||
// if they don't correspond to a pinned or active file
|
||||
var removed = 0;
|
||||
blobs.list.proofs(function (err, item, next) {
|
||||
if (err) {
|
||||
next();
|
||||
return void Log.error("EVICT_BLOB_LIST_PROOFS_ERROR", err);
|
||||
}
|
||||
if (!item) {
|
||||
next();
|
||||
return void Log.error('EVICT_BLOB_LIST_PROOFS_NO_ITEM', item);
|
||||
}
|
||||
if (pinnedDocs.test(item.blobId)) { return void next(); }
|
||||
if (getNewestTime(item) > inactiveTime) { return void next(); }
|
||||
nThen(function (w) {
|
||||
blobs.size(item.blobId, w(function (err, size) {
|
||||
if (err) {
|
||||
w.abort();
|
||||
next();
|
||||
return void Log.error("EVICT_BLOB_LIST_PROOFS_ERROR", err);
|
||||
}
|
||||
if (size !== 0) {
|
||||
w.abort();
|
||||
next();
|
||||
}
|
||||
}));
|
||||
}).nThen(function () {
|
||||
blobs.remove.proof(item.safeKey, item.blobId, function (err) {
|
||||
next();
|
||||
if (err) {
|
||||
return Log.error("EVICT_BLOB_PROOF_LONELY_ERROR", item);
|
||||
}
|
||||
removed++;
|
||||
return Log.info("EVICT_BLOB_PROOF_LONELY", item);
|
||||
});
|
||||
});
|
||||
}, w(function () {
|
||||
Log.info("EVICT_BLOB_PROOFS_REMOVED", removed);
|
||||
}));
|
||||
};
|
||||
|
||||
var archiveInactiveChannels = function (w) {
|
||||
var channels = 0;
|
||||
var archived = 0;
|
||||
|
||||
var handler = function (err, item, cb) {
|
||||
channels++;
|
||||
if (err) {
|
||||
Log.error('EVICT_CHANNEL_ITERATION', err);
|
||||
return void cb();
|
||||
}
|
||||
// check if the database has any ephemeral channels
|
||||
// if it does it's because of a bug, and they should be removed
|
||||
if (item.channel.length === 34) {
|
||||
return void store.removeChannel(item.channel, w(function (err) {
|
||||
if (err) {
|
||||
Log.error('EVICT_EPHEMERAL_CHANNEL_REMOVAL_ERROR', {
|
||||
error: err,
|
||||
channel: item.channel,
|
||||
});
|
||||
return void cb();
|
||||
}
|
||||
Log.info('EVICT_EPHEMERAL_CHANNEL_REMOVAL', item.channel);
|
||||
cb();
|
||||
}));
|
||||
}
|
||||
|
||||
// bail out if the channel is in the set of activeDocs
|
||||
if (activeDocs.test(item.channel)) { return void cb(); }
|
||||
|
||||
// ignore the channel if it's pinned
|
||||
if (pinnedDocs.test(item.channel)) { return void cb(); }
|
||||
|
||||
return void store.archiveChannel(item.channel, w(function (err) {
|
||||
if (err) {
|
||||
Log.error('EVICT_CHANNEL_ARCHIVAL_ERROR', {
|
||||
error: err,
|
||||
channel: item.channel,
|
||||
});
|
||||
return void cb();
|
||||
}
|
||||
Log.info('EVICT_CHANNEL_ARCHIVAL', item.channel);
|
||||
archived++;
|
||||
cb();
|
||||
}));
|
||||
};
|
||||
|
||||
var done = function () {
|
||||
return void Log.info('EVICT_CHANNELS_ARCHIVED', archived);
|
||||
};
|
||||
|
||||
store.listChannels(handler, w(done), true); // using a hacky "fast mode" since we only need the channel id
|
||||
};
|
||||
|
||||
nThen(loadStorage)
|
||||
.nThen(function () {
|
||||
Log.info("EVICT_TIME_TO_LOAD_PINS", msSinceStart());
|
||||
})
|
||||
.nThen(removeArchivedChannels)
|
||||
.nThen(removeArchivedBlobProofs)
|
||||
.nThen(removeArchivedBlobs)
|
||||
|
||||
// iterate over all documents and add them to a bloom filter if they have been active
|
||||
.nThen(categorizeChannelsByActivity)
|
||||
.nThen(categorizeBlobsByActivity)
|
||||
|
||||
// iterate over all accounts and add them to a bloom filter if they are active
|
||||
.nThen(categorizeAccountsByActivity)
|
||||
|
||||
// iterate again and archive inactive unpinned documents
|
||||
// (documents which are not in either bloom filter)
|
||||
|
||||
.nThen(archiveInactiveBlobs)
|
||||
.nThen(archiveInactiveBlobProofs)
|
||||
.nThen(archiveInactiveChannels)
|
||||
.nThen(function () {
|
||||
Log.info("EVICT_TIME_TO_RUN_SCRIPT", msSinceStart());
|
||||
}).nThen(function () {
|
||||
var shutdown = function (Env) {
|
||||
// the store will keep this script running if you don't shut it down
|
||||
store.shutdown();
|
||||
Log.shutdown();
|
||||
pinStore.shutdown();
|
||||
process.exit();
|
||||
//Env.store.shutdown();
|
||||
//Env.Log.shutdown();
|
||||
//Env.pinStore.shutdown();
|
||||
};
|
||||
|
||||
nThen(function (w) {
|
||||
// load database adaptors and configuration values into the environment
|
||||
prepareEnv(Env, w(function () {
|
||||
|
||||
|
||||
}));
|
||||
}).nThen(function (w) {
|
||||
Eviction(Env, w(function () {
|
||||
|
||||
}));
|
||||
}).nThen(function () {
|
||||
// shut down database adaptors
|
||||
shutdown(Env);
|
||||
});
|
||||
|
||||
|
|
Loading…
Reference in New Issue