mirror of
https://github.com/zadam/trilium.git
synced 2025-03-01 14:22:32 +01:00
153 lines
3.8 KiB
JavaScript
153 lines
3.8 KiB
JavaScript
"use strict";
|
|
|
|
const sql = require('./sql');
|
|
const utils = require('./utils');
|
|
const log = require('./log');
|
|
const eventLogService = require('./event_log');
|
|
const messagingService = require('./messaging');
|
|
|
|
function getHash(rows) {
|
|
let hash = '';
|
|
|
|
for (const row of rows) {
|
|
hash = utils.hash(hash + JSON.stringify(row));
|
|
}
|
|
|
|
return hash;
|
|
}
|
|
|
|
async function getHashes() {
|
|
const startTime = new Date();
|
|
|
|
const hashes = {
|
|
notes: getHash(await sql.getRows(`
|
|
SELECT
|
|
noteId,
|
|
title,
|
|
content,
|
|
type,
|
|
dateModified,
|
|
isProtected,
|
|
isDeleted
|
|
FROM notes
|
|
ORDER BY noteId`)),
|
|
|
|
branches: getHash(await sql.getRows(`
|
|
SELECT
|
|
branchId,
|
|
noteId,
|
|
parentNoteId,
|
|
notePosition,
|
|
dateModified,
|
|
isDeleted,
|
|
prefix
|
|
FROM branches
|
|
ORDER BY branchId`)),
|
|
|
|
note_revisions: getHash(await sql.getRows(`
|
|
SELECT
|
|
noteRevisionId,
|
|
noteId,
|
|
title,
|
|
content,
|
|
dateModifiedFrom,
|
|
dateModifiedTo
|
|
FROM note_revisions
|
|
ORDER BY noteRevisionId`)),
|
|
|
|
recent_notes: getHash(await sql.getRows(`
|
|
SELECT
|
|
branchId,
|
|
notePath,
|
|
dateAccessed,
|
|
isDeleted
|
|
FROM recent_notes
|
|
ORDER BY notePath`)),
|
|
|
|
options: getHash(await sql.getRows(`
|
|
SELECT
|
|
name,
|
|
value
|
|
FROM options
|
|
WHERE isSynced = 1
|
|
ORDER BY name`)),
|
|
|
|
// we don't include image data on purpose because they are quite large, checksum is good enough
|
|
// to represent the data anyway
|
|
images: getHash(await sql.getRows(`
|
|
SELECT
|
|
imageId,
|
|
format,
|
|
checksum,
|
|
name,
|
|
isDeleted,
|
|
dateModified,
|
|
dateCreated
|
|
FROM images
|
|
ORDER BY imageId`)),
|
|
|
|
note_images: getHash(await sql.getRows(`
|
|
SELECT
|
|
noteImageId,
|
|
noteId,
|
|
imageId,
|
|
isDeleted,
|
|
dateModified,
|
|
dateCreated
|
|
FROM note_images
|
|
ORDER BY noteImageId`)),
|
|
|
|
labels: getHash(await sql.getRows(`
|
|
SELECT
|
|
labelId,
|
|
noteId,
|
|
name,
|
|
value,
|
|
dateModified,
|
|
dateCreated
|
|
FROM labels
|
|
ORDER BY labelId`)),
|
|
|
|
api_tokens: getHash(await sql.getRows(`
|
|
SELECT
|
|
apiTokenId,
|
|
token,
|
|
dateCreated,
|
|
isDeleted
|
|
FROM api_tokens
|
|
ORDER BY apiTokenId`))
|
|
};
|
|
|
|
const elapseTimeMs = new Date().getTime() - startTime.getTime();
|
|
|
|
log.info(`Content hash computation took ${elapseTimeMs}ms`);
|
|
|
|
return hashes;
|
|
}
|
|
|
|
async function checkContentHashes(otherHashes) {
|
|
const hashes = await getHashes();
|
|
let allChecksPassed = true;
|
|
|
|
for (const key in hashes) {
|
|
if (hashes[key] !== otherHashes[key]) {
|
|
allChecksPassed = false;
|
|
|
|
await eventLogService.addEvent(`Content hash check for ${key} FAILED. Local is ${hashes[key]}, remote is ${resp.hashes[key]}`);
|
|
|
|
if (key !== 'recent_notes') {
|
|
// let's not get alarmed about recent notes which get updated often and can cause failures in race conditions
|
|
await messagingService.sendMessageToAllClients({type: 'sync-hash-check-failed'});
|
|
}
|
|
}
|
|
}
|
|
|
|
if (allChecksPassed) {
|
|
log.info("Content hash checks PASSED");
|
|
}
|
|
}
|
|
|
|
module.exports = {
|
|
getHashes,
|
|
checkContentHashes
|
|
}; |