Merge branch 'staging' of github.com:xwiki-labs/cryptpad into staging
commit
a6a40c3f6a
@ -1,47 +0,0 @@
|
||||
/* jshint esversion: 6, node: true */
|
||||
const Fs = require("fs");
|
||||
const nThen = require("nthen");
|
||||
const Saferphore = require("saferphore");
|
||||
const PinnedData = require('./pinneddata');
|
||||
const config = require("../lib/load-config");
|
||||
|
||||
if (!config.inactiveTime || typeof(config.inactiveTime) !== "number") { return; }
|
||||
|
||||
/* Instead of this script you should probably use
|
||||
evict-inactive.js which moves things to an archive directory
|
||||
in case the data that would have been deleted turns out to be important.
|
||||
it also handles removing that archived data after a set period of time
|
||||
|
||||
it only works for channels at the moment, though, and nothing else.
|
||||
*/
|
||||
|
||||
let inactiveTime = +new Date() - (config.inactiveTime * 24 * 3600 * 1000);
|
||||
let inactiveConfig = {
|
||||
unpinned: true,
|
||||
olderthan: inactiveTime,
|
||||
blobsolderthan: inactiveTime,
|
||||
|
||||
filePath: config.filePath,
|
||||
blobPath: config.blobPath,
|
||||
pinPath: config.pinPath,
|
||||
};
|
||||
let toDelete;
|
||||
nThen(function (waitFor) {
|
||||
PinnedData.load(inactiveConfig, waitFor(function (err, data) {
|
||||
if (err) {
|
||||
waitFor.abort();
|
||||
throw new Error(err);
|
||||
}
|
||||
toDelete = data;
|
||||
}));
|
||||
}).nThen(function () {
|
||||
var sem = Saferphore.create(10);
|
||||
toDelete.forEach(function (f) {
|
||||
sem.take(function (give) {
|
||||
Fs.unlink(f.filename, give(function (err) {
|
||||
if (err) { return void console.error(err + " " + f.filename); }
|
||||
console.log(f.filename + " " + f.size + " " + (+f.mtime) + " " + (+new Date()));
|
||||
}));
|
||||
});
|
||||
});
|
||||
});
|
@ -1,267 +0,0 @@
|
||||
/* jshint esversion: 6, node: true */
|
||||
const Fs = require('fs');
|
||||
const Semaphore = require('saferphore');
|
||||
const nThen = require('nthen');
|
||||
const Path = require('path');
|
||||
const Pins = require('../lib/pins');
|
||||
|
||||
/*
|
||||
takes an array of pinned file names
|
||||
and a global map of stats indexed by public keys
|
||||
returns the sum of the size of those pinned files
|
||||
*/
|
||||
const sizeForHashes = (hashes, dsFileStats) => {
|
||||
let sum = 0;
|
||||
hashes.forEach((h) => {
|
||||
const s = dsFileStats[h];
|
||||
if (typeof(s) !== 'object' || typeof(s.size) !== 'number') {
|
||||
//console.log('missing ' + h + ' ' + typeof(s));
|
||||
} else {
|
||||
sum += s.size;
|
||||
}
|
||||
});
|
||||
return sum;
|
||||
};
|
||||
|
||||
// do twenty things at a time
|
||||
const sema = Semaphore.create(20);
|
||||
|
||||
let dirList;
|
||||
const fileList = []; // array which we reuse for a lot of things
|
||||
const dsFileStats = {}; // map of stats
|
||||
const out = []; // what we return at the end
|
||||
const pinned = {}; // map of pinned files
|
||||
|
||||
// define a function: 'load' which takes a config
|
||||
// and a callback
|
||||
module.exports.load = function (config, cb) {
|
||||
var filePath = config.filePath || './datastore';
|
||||
var blobPath = config.blobPath || './blob';
|
||||
var pinPath = config.pinPath || './pins';
|
||||
|
||||
|
||||
nThen((waitFor) => {
|
||||
// read the subdirectories in the datastore
|
||||
Fs.readdir(filePath, waitFor((err, list) => {
|
||||
if (err) { throw err; }
|
||||
dirList = list;
|
||||
}));
|
||||
}).nThen((waitFor) => {
|
||||
// iterate over all subdirectories
|
||||
dirList.forEach((f) => {
|
||||
// process twenty subdirectories simultaneously
|
||||
sema.take((returnAfter) => {
|
||||
// get the list of files in every subdirectory
|
||||
// and push them to 'fileList'
|
||||
Fs.readdir(Path.join(filePath, f), waitFor(returnAfter((err, list2) => {
|
||||
if (err) { throw err; }
|
||||
list2.forEach((ff) => { fileList.push(Path.join(filePath, f, ff)); });
|
||||
})));
|
||||
});
|
||||
});
|
||||
}).nThen((waitFor) => {
|
||||
// read the subdirectories in 'blob'
|
||||
Fs.readdir(blobPath, waitFor((err, list) => {
|
||||
if (err) { throw err; }
|
||||
// overwrite dirList
|
||||
dirList = list;
|
||||
}));
|
||||
}).nThen((waitFor) => {
|
||||
// iterate over all subdirectories
|
||||
dirList.forEach((f) => {
|
||||
// process twenty subdirectories simultaneously
|
||||
sema.take((returnAfter) => {
|
||||
// get the list of files in every subdirectory
|
||||
// and push them to 'fileList'
|
||||
Fs.readdir(Path.join(blobPath, f), waitFor(returnAfter((err, list2) => {
|
||||
if (err) { throw err; }
|
||||
list2.forEach((ff) => { fileList.push(Path.join(blobPath, f, ff)); });
|
||||
})));
|
||||
});
|
||||
});
|
||||
}).nThen((waitFor) => {
|
||||
// iterate over the fileList
|
||||
fileList.forEach((f) => {
|
||||
// process twenty files simultaneously
|
||||
sema.take((returnAfter) => {
|
||||
// get the stats of each files
|
||||
Fs.stat(f, waitFor(returnAfter((err, st) => {
|
||||
if (err) { throw err; }
|
||||
st.filename = f;
|
||||
// push them to a big map of stats
|
||||
dsFileStats[f.replace(/^.*\/([^\/\.]*)(\.ndjson)?$/, (all, a) => (a))] = st;
|
||||
})));
|
||||
});
|
||||
});
|
||||
}).nThen((waitFor) => {
|
||||
// read the subdirectories in the pinstore
|
||||
Fs.readdir(pinPath, waitFor((err, list) => {
|
||||
if (err) { throw err; }
|
||||
dirList = list;
|
||||
}));
|
||||
}).nThen((waitFor) => {
|
||||
// set file list to an empty array
|
||||
// fileList = [] ??
|
||||
fileList.splice(0, fileList.length);
|
||||
dirList.forEach((f) => {
|
||||
// process twenty directories at a time
|
||||
sema.take((returnAfter) => {
|
||||
// get the list of files in every subdirectory
|
||||
// and push them to 'fileList' (which is empty because we keep reusing it)
|
||||
Fs.readdir(Path.join(pinPath, f), waitFor(returnAfter((err, list2) => {
|
||||
if (err) { throw err; }
|
||||
list2.forEach((ff) => { fileList.push(Path.join(pinPath, f, ff)); });
|
||||
})));
|
||||
});
|
||||
});
|
||||
}).nThen((waitFor) => {
|
||||
// iterate over the list of pin logs
|
||||
fileList.forEach((f) => {
|
||||
// twenty at a time
|
||||
sema.take((returnAfter) => {
|
||||
// read the full content
|
||||
Fs.readFile(f, waitFor(returnAfter((err, content) => {
|
||||
if (err) { throw err; }
|
||||
// get the list of channels pinned by this log
|
||||
const hashes = Pins.calculateFromLog(content.toString('utf8'), f);
|
||||
if (config.unpinned) {
|
||||
hashes.forEach((x) => { pinned[x] = 1; });
|
||||
} else {
|
||||
// get the size of files pinned by this log
|
||||
// but only if we're gonna use it
|
||||
let size = sizeForHashes(hashes, dsFileStats);
|
||||
// we will return a list of values
|
||||
// [user_public_key, size_of_files_they_have_pinned]
|
||||
out.push([f, Math.floor(size / (1024 * 1024))]);
|
||||
}
|
||||
})));
|
||||
});
|
||||
});
|
||||
}).nThen(() => {
|
||||
// handle all the information you've processed so far
|
||||
if (config.unpinned) {
|
||||
// the user wants data about what has not been pinned
|
||||
|
||||
// by default we concern ourselves with pads and files older than infinity (everything)
|
||||
let before = Infinity;
|
||||
|
||||
// but you can override this with config
|
||||
if (config.olderthan) {
|
||||
before = config.olderthan;
|
||||
// FIXME validate inputs before doing the heavy lifting
|
||||
if (isNaN(before)) { // make sure the supplied value is a number
|
||||
return void cb('--olderthan error [' + config.olderthan + '] not a valid date');
|
||||
}
|
||||
}
|
||||
|
||||
// you can specify a different time for blobs...
|
||||
let blobsbefore = before;
|
||||
if (config.blobsolderthan) {
|
||||
// use the supplied date if it exists
|
||||
blobsbefore = config.blobsolderthan;
|
||||
if (isNaN(blobsbefore)) {
|
||||
return void cb('--blobsolderthan error [' + config.blobsolderthan + '] not a valid date');
|
||||
}
|
||||
}
|
||||
let files = [];
|
||||
// iterate over all the stats that you've saved
|
||||
Object.keys(dsFileStats).forEach((f) => {
|
||||
// we only care about files which are not in the pin map
|
||||
if (!(f in pinned)) {
|
||||
// check if it's a blob or a 'pad'
|
||||
const isBlob = dsFileStats[f].filename.indexOf('.ndjson') === -1;
|
||||
|
||||
// if the mtime is newer than the specified value for its file type, ignore this file
|
||||
|
||||
if ((+dsFileStats[f].mtime) >= ((isBlob) ? blobsbefore : before)) { return; }
|
||||
|
||||
// otherwise push it to the list of files, with its filename, size, and mtime
|
||||
files.push({
|
||||
filename: dsFileStats[f].filename,
|
||||
size: dsFileStats[f].size,
|
||||
mtime: dsFileStats[f].mtime
|
||||
});
|
||||
}
|
||||
});
|
||||
|
||||
// return the list of files
|
||||
cb(null, files);
|
||||
} else {
|
||||
// if you're not in 'unpinned' mode, sort by size (ascending)
|
||||
out.sort((a,b) => (a[1] - b[1]));
|
||||
// and return the sorted data
|
||||
cb(null, out.slice());
|
||||
}
|
||||
});
|
||||
};
|
||||
|
||||
|
||||
// This script can be called directly on its own
|
||||
// or required as part of another script
|
||||
if (!module.parent) {
|
||||
// if no parent, it is being invoked directly
|
||||
let config = {}; // build the config from command line arguments...
|
||||
|
||||
var Config = require("../lib/load-config");
|
||||
|
||||
config.filePath = Config.filePath;
|
||||
config.blobPath = Config.blobPath;
|
||||
config.pinPath = Config.pinPath;
|
||||
|
||||
// --unpinned gets the list of unpinned files
|
||||
// if you don't pass this, it will list the size of pinned data per user
|
||||
if (process.argv.indexOf('--unpinned') > -1) { config.unpinned = true; }
|
||||
|
||||
// '--olderthan' must be used in conjunction with '--unpinned'
|
||||
// if you pass '--olderthan' with a string date or number, it will limit
|
||||
// results only to pads older than the supplied time
|
||||
// it defaults to 'infinity', or no filter at all
|
||||
const ot = process.argv.indexOf('--olderthan');
|
||||
if (ot > -1) {
|
||||
config.olderthan = Number(process.argv[ot+1]) ? new Date(Number(process.argv[ot+1]))
|
||||
: new Date(process.argv[ot+1]);
|
||||
}
|
||||
|
||||
// '--blobsolderthan' must be used in conjunction with '--unpinned'
|
||||
// if you pass '--blobsolderthan with a string date or number, it will limit
|
||||
// results only to blobs older than the supplied time
|
||||
// it defaults to using the same value passed '--olderthan'
|
||||
const bot = process.argv.indexOf('--blobsolderthan');
|
||||
if (bot > -1) {
|
||||
config.blobsolderthan = Number(process.argv[bot+1]) ? new Date(Number(process.argv[bot+1]))
|
||||
: new Date(process.argv[bot+1]);
|
||||
}
|
||||
|
||||
// call our big function directly
|
||||
// pass our constructed configuration and a callback
|
||||
module.exports.load(config, function (err, data) {
|
||||
if (err) { throw new Error(err); } // throw errors
|
||||
if (!Array.isArray(data)) { return; } // if the returned value is not an array, you're done
|
||||
if (config.unpinned) {
|
||||
// display the list of unpinned files with their size and mtime
|
||||
data.forEach((f) => { console.log(f.filename + " " + f.size + " " + (+f.mtime)); });
|
||||
} else {
|
||||
// display the list of public keys and the size of the data they have pinned in megabytes
|
||||
data.forEach((x) => { console.log(x[0] + ' ' + x[1] + ' MB'); });
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
|
||||
/* Example usage of this script...
|
||||
|
||||
# display the list of public keys and the size of the data the have pinned in megabytes
|
||||
node pinneddata.js
|
||||
|
||||
# display the list of unpinned pads and blobs with their size and mtime
|
||||
node pinneddata.js --unpinned
|
||||
|
||||
# display the list of unpinned pads and blobs older than 12345 with their size and mtime
|
||||
node pinneddata.js --unpinned --olderthan 12345
|
||||
|
||||
|
||||
# display the list of unpinned pads older than 12345 and unpinned blobs older than 123
|
||||
# each with their size and mtime
|
||||
node pinneddata.js --unpinned --olderthan 12345 --blobsolderthan 123
|
||||
|
||||
*/
|
@ -1,97 +0,0 @@
|
||||
define([
|
||||
'/bower_components/tweetnacl/nacl-fast.min.js',
|
||||
], function () {
|
||||
var Nacl = window.nacl;
|
||||
var Curve = {};
|
||||
|
||||
var concatenateUint8s = function (A) {
|
||||
var len = 0;
|
||||
var offset = 0;
|
||||
A.forEach(function (uints) {
|
||||
len += uints.length || 0;
|
||||
});
|
||||
var c = new Uint8Array(len);
|
||||
A.forEach(function (x) {
|
||||
c.set(x, offset);
|
||||
offset += x.length;
|
||||
});
|
||||
return c;
|
||||
};
|
||||
|
||||
var encodeBase64 = Nacl.util.encodeBase64;
|
||||
var decodeBase64 = Nacl.util.decodeBase64;
|
||||
var decodeUTF8 = Nacl.util.decodeUTF8;
|
||||
var encodeUTF8 = Nacl.util.encodeUTF8;
|
||||
|
||||
Curve.encrypt = function (message, secret) {
|
||||
var buffer = decodeUTF8(message);
|
||||
var nonce = Nacl.randomBytes(24);
|
||||
var box = Nacl.box.after(buffer, nonce, secret);
|
||||
return encodeBase64(nonce) + '|' + encodeBase64(box);
|
||||
};
|
||||
|
||||
Curve.decrypt = function (packed, secret) {
|
||||
var unpacked = packed.split('|');
|
||||
var nonce = decodeBase64(unpacked[0]);
|
||||
var box = decodeBase64(unpacked[1]);
|
||||
var message = Nacl.box.open.after(box, nonce, secret);
|
||||
if (message === false) { return null; }
|
||||
return encodeUTF8(message);
|
||||
};
|
||||
|
||||
Curve.signAndEncrypt = function (msg, cryptKey, signKey) {
|
||||
var packed = Curve.encrypt(msg, cryptKey);
|
||||
return encodeBase64(Nacl.sign(decodeUTF8(packed), signKey));
|
||||
};
|
||||
|
||||
Curve.openSigned = function (msg, cryptKey /*, validateKey STUBBED*/) {
|
||||
var content = decodeBase64(msg).subarray(64);
|
||||
return Curve.decrypt(encodeUTF8(content), cryptKey);
|
||||
};
|
||||
|
||||
Curve.deriveKeys = function (theirs, mine) {
|
||||
try {
|
||||
var pub = decodeBase64(theirs);
|
||||
var secret = decodeBase64(mine);
|
||||
|
||||
var sharedSecret = Nacl.box.before(pub, secret);
|
||||
var salt = decodeUTF8('CryptPad.signingKeyGenerationSalt');
|
||||
|
||||
// 64 uint8s
|
||||
var hash = Nacl.hash(concatenateUint8s([salt, sharedSecret]));
|
||||
var signKp = Nacl.sign.keyPair.fromSeed(hash.subarray(0, 32));
|
||||
var cryptKey = hash.subarray(32, 64);
|
||||
|
||||
return {
|
||||
cryptKey: encodeBase64(cryptKey),
|
||||
signKey: encodeBase64(signKp.secretKey),
|
||||
validateKey: encodeBase64(signKp.publicKey)
|
||||
};
|
||||
} catch (e) {
|
||||
console.error('invalid keys or other problem deriving keys');
|
||||
console.error(e);
|
||||
return null;
|
||||
}
|
||||
};
|
||||
|
||||
Curve.createEncryptor = function (keys) {
|
||||
if (!keys || typeof(keys) !== 'object') {
|
||||
return void console.error("invalid input for createEncryptor");
|
||||
}
|
||||
|
||||
var cryptKey = decodeBase64(keys.cryptKey);
|
||||
var signKey = decodeBase64(keys.signKey);
|
||||
var validateKey = decodeBase64(keys.validateKey);
|
||||
|
||||
return {
|
||||
encrypt: function (msg) {
|
||||
return Curve.signAndEncrypt(msg, cryptKey, signKey);
|
||||
},
|
||||
decrypt: function (packed) {
|
||||
return Curve.openSigned(packed, cryptKey, validateKey);
|
||||
}
|
||||
};
|
||||
};
|
||||
|
||||
return Curve;
|
||||
});
|
Loading…
Reference in New Issue