mirror of
https://github.com/zadam/trilium.git
synced 2025-03-01 14:22:32 +01:00
383 lines
12 KiB
JavaScript
383 lines
12 KiB
JavaScript
"use strict";
|
|
|
|
const Attribute = require('../../entities/attribute');
|
|
const Link = require('../../entities/link');
|
|
const utils = require('../../services/utils');
|
|
const log = require('../../services/log');
|
|
const repository = require('../../services/repository');
|
|
const noteService = require('../../services/notes');
|
|
const Branch = require('../../entities/branch');
|
|
const tar = require('tar-stream');
|
|
const stream = require('stream');
|
|
const path = require('path');
|
|
const commonmark = require('commonmark');
|
|
const mimeTypes = require('mime-types');
|
|
|
|
async function importTar(fileBuffer, importRootNote) {
|
|
// maps from original noteId (in tar file) to newly generated noteId
|
|
const noteIdMap = {};
|
|
const attributes = [];
|
|
const links = [];
|
|
// path => noteId
|
|
const createdPaths = { '/': importRootNote.noteId, '\\': importRootNote.noteId };
|
|
const mdReader = new commonmark.Parser();
|
|
const mdWriter = new commonmark.HtmlRenderer();
|
|
let metaFile = null;
|
|
let firstNote = null;
|
|
|
|
const extract = tar.extract();
|
|
|
|
function getNewNoteId(origNoteId) {
|
|
// in case the original noteId is empty. This probably shouldn't happen, but still good to have this precaution
|
|
if (!origNoteId.trim()) {
|
|
return "";
|
|
}
|
|
|
|
// we allow references to root and they don't need translation
|
|
if (origNoteId === 'root') {
|
|
return origNoteId;
|
|
}
|
|
|
|
if (!noteIdMap[origNoteId]) {
|
|
noteIdMap[origNoteId] = utils.newEntityId();
|
|
}
|
|
|
|
return noteIdMap[origNoteId];
|
|
}
|
|
|
|
function getMeta(filePath) {
|
|
if (!metaFile) {
|
|
return {};
|
|
}
|
|
|
|
const pathSegments = filePath.split(/[\/\\]/g);
|
|
|
|
let cursor = {
|
|
isImportRoot: true,
|
|
children: metaFile.files
|
|
};
|
|
|
|
let parent;
|
|
|
|
for (const segment of pathSegments) {
|
|
if (!cursor || !cursor.children || cursor.children.length === 0) {
|
|
return {};
|
|
}
|
|
|
|
parent = cursor;
|
|
cursor = cursor.children.find(file => file.dataFileName === segment || file.dirFileName === segment);
|
|
}
|
|
|
|
return {
|
|
parentNoteMeta: parent,
|
|
noteMeta: cursor
|
|
};
|
|
}
|
|
|
|
function getParentNoteId(filePath, parentNoteMeta) {
|
|
let parentNoteId;
|
|
|
|
if (parentNoteMeta) {
|
|
parentNoteId = parentNoteMeta.isImportRoot ? importRootNote.noteId : getNewNoteId(parentNoteMeta.noteId);
|
|
}
|
|
else {
|
|
const parentPath = path.dirname(filePath);
|
|
|
|
if (parentPath === '.') {
|
|
parentNoteId = importRootNote.noteId;
|
|
}
|
|
else if (parentPath in createdPaths) {
|
|
parentNoteId = createdPaths[parentPath];
|
|
}
|
|
else {
|
|
throw new Error(`Could not find existing path ${parentPath} for ${filePath}.`);
|
|
}
|
|
}
|
|
|
|
return parentNoteId;
|
|
}
|
|
|
|
function getNoteTitle(filePath, noteMeta) {
|
|
if (noteMeta) {
|
|
return noteMeta.title;
|
|
}
|
|
else {
|
|
const basename = path.basename(filePath);
|
|
|
|
return getTextFileWithoutExtension(basename);
|
|
}
|
|
}
|
|
|
|
function getNoteId(noteMeta, filePath) {
|
|
if (noteMeta) {
|
|
return getNewNoteId(noteMeta.noteId);
|
|
}
|
|
else {
|
|
const filePathNoExt = getTextFileWithoutExtension(filePath);
|
|
|
|
if (filePathNoExt in createdPaths) {
|
|
return createdPaths[filePathNoExt];
|
|
}
|
|
else {
|
|
return utils.newEntityId();
|
|
}
|
|
}
|
|
}
|
|
|
|
function detectFileTypeAndMime(filePath) {
|
|
const mime = mimeTypes.lookup(filePath);
|
|
let type = 'file';
|
|
|
|
if (mime) {
|
|
if (mime === 'text/html' || mime === 'text/markdown') {
|
|
type = 'text';
|
|
}
|
|
else if (mime.startsWith('image/')) {
|
|
type = 'image';
|
|
}
|
|
}
|
|
|
|
return { type, mime };
|
|
}
|
|
|
|
async function saveAttributesAndLinks(note, noteMeta) {
|
|
if (!noteMeta) {
|
|
return;
|
|
}
|
|
|
|
for (const attr of noteMeta.attributes) {
|
|
attr.noteId = note.noteId;
|
|
|
|
if (attr.type === 'relation') {
|
|
attr.value = getNewNoteId(attr.value);
|
|
}
|
|
|
|
attributes.push(attr);
|
|
}
|
|
|
|
for (const link of noteMeta.links) {
|
|
link.noteId = note.noteId;
|
|
link.targetNoteId = getNewNoteId(link.targetNoteId);
|
|
|
|
links.push(link);
|
|
}
|
|
}
|
|
|
|
async function saveDirectory(filePath) {
|
|
const { parentNoteMeta, noteMeta } = getMeta(filePath);
|
|
|
|
const noteId = getNoteId(noteMeta, filePath);
|
|
const noteTitle = getNoteTitle(filePath, noteMeta);
|
|
const parentNoteId = getParentNoteId(filePath, parentNoteMeta);
|
|
|
|
let note = await repository.getNote(noteId);
|
|
|
|
if (note) {
|
|
return;
|
|
}
|
|
|
|
({note} = await noteService.createNote(parentNoteId, noteTitle, '', {
|
|
noteId,
|
|
type: noteMeta ? noteMeta.type : 'text',
|
|
mime: noteMeta ? noteMeta.mime : 'text/html',
|
|
prefix: noteMeta ? noteMeta.prefix : '',
|
|
isExpanded: noteMeta ? noteMeta.isExpanded : false
|
|
}));
|
|
|
|
await saveAttributesAndLinks(note, noteMeta);
|
|
|
|
if (!firstNote) {
|
|
firstNote = note;
|
|
}
|
|
|
|
createdPaths[filePath] = noteId;
|
|
}
|
|
|
|
function getTextFileWithoutExtension(filePath) {
|
|
const extension = path.extname(filePath).toLowerCase();
|
|
|
|
if (extension === '.md' || extension === '.html') {
|
|
return filePath.substr(0, filePath.length - extension.length);
|
|
}
|
|
else {
|
|
return filePath;
|
|
}
|
|
}
|
|
|
|
async function saveNote(filePath, content) {
|
|
const {parentNoteMeta, noteMeta} = getMeta(filePath);
|
|
|
|
const noteId = getNoteId(noteMeta, filePath);
|
|
const parentNoteId = getParentNoteId(filePath, parentNoteMeta);
|
|
|
|
if (noteMeta && noteMeta.isClone) {
|
|
await new Branch({
|
|
noteId,
|
|
parentNoteId,
|
|
isExpanded: noteMeta.isExpanded,
|
|
prefix: noteMeta.prefix,
|
|
notePosition: noteMeta.notePosition
|
|
}).save();
|
|
|
|
return;
|
|
}
|
|
|
|
const {type, mime} = noteMeta ? noteMeta : detectFileTypeAndMime(filePath);
|
|
|
|
if (type !== 'file' && type !== 'image') {
|
|
content = content.toString("UTF-8");
|
|
|
|
if (noteMeta) {
|
|
// this will replace all internal links (<a> and <img>) inside the body
|
|
// links pointing outside the export will be broken and changed (ctx.getNewNoteId() will still assign new noteId)
|
|
for (const link of noteMeta.links || []) {
|
|
// no need to escape the regexp find string since it's a noteId which doesn't contain any special characters
|
|
content = content.replace(new RegExp(link.targetNoteId, "g"), getNewNoteId(link.targetNoteId));
|
|
}
|
|
}
|
|
}
|
|
|
|
if ((noteMeta && noteMeta.format === 'markdown') || (!noteMeta && mime === 'text/markdown')) {
|
|
const parsed = mdReader.parse(content);
|
|
content = mdWriter.render(parsed);
|
|
}
|
|
|
|
let note = await repository.getNote(noteId);
|
|
|
|
if (note) {
|
|
note.content = content;
|
|
await note.save();
|
|
}
|
|
else {
|
|
const noteTitle = getNoteTitle(filePath, noteMeta);
|
|
|
|
({note} = await noteService.createNote(parentNoteId, noteTitle, content, {
|
|
noteId,
|
|
type,
|
|
mime,
|
|
prefix: noteMeta ? noteMeta.prefix : '',
|
|
isExpanded: noteMeta ? noteMeta.isExpanded : false,
|
|
notePosition: noteMeta ? noteMeta.notePosition : false
|
|
}));
|
|
|
|
await saveAttributesAndLinks(note, noteMeta);
|
|
|
|
if (!noteMeta && (type === 'file' || type === 'image')) {
|
|
attributes.push({
|
|
noteId,
|
|
type: 'label',
|
|
name: 'originalFileName',
|
|
value: path.basename(filePath)
|
|
});
|
|
|
|
attributes.push({
|
|
noteId,
|
|
type: 'label',
|
|
name: 'fileSize',
|
|
value: content.byteLength
|
|
});
|
|
}
|
|
|
|
if (!firstNote) {
|
|
firstNote = note;
|
|
}
|
|
|
|
if (type === 'text') {
|
|
filePath = getTextFileWithoutExtension(filePath);
|
|
}
|
|
|
|
createdPaths[filePath] = noteId;
|
|
}
|
|
}
|
|
|
|
/** @return path without leading or trailing slash and backslashes converted to forward ones*/
|
|
function normalizeFilePath(filePath) {
|
|
filePath = filePath.replace(/\\/g, "/");
|
|
|
|
if (filePath.startsWith("/")) {
|
|
filePath = filePath.substr(1);
|
|
}
|
|
|
|
if (filePath.endsWith("/")) {
|
|
filePath = filePath.substr(0, filePath.length - 1);
|
|
}
|
|
|
|
return filePath;
|
|
}
|
|
|
|
extract.on('entry', function(header, stream, next) {
|
|
const chunks = [];
|
|
|
|
stream.on("data", function (chunk) {
|
|
chunks.push(chunk);
|
|
});
|
|
|
|
// header is the tar header
|
|
// stream is the content body (might be an empty stream)
|
|
// call next when you are done with this entry
|
|
|
|
stream.on('end', async function() {
|
|
let filePath = normalizeFilePath(header.name);
|
|
|
|
const content = Buffer.concat(chunks);
|
|
|
|
if (filePath === '!!!meta.json') {
|
|
metaFile = JSON.parse(content.toString("UTF-8"));
|
|
}
|
|
else if (header.type === 'directory') {
|
|
await saveDirectory(filePath);
|
|
}
|
|
else if (header.type === 'file') {
|
|
await saveNote(filePath, content);
|
|
}
|
|
else {
|
|
log.info("Ignoring tar import entry with type " + header.type);
|
|
}
|
|
|
|
next(); // ready for next entry
|
|
});
|
|
|
|
stream.resume(); // just auto drain the stream
|
|
});
|
|
|
|
return new Promise(resolve => {
|
|
extract.on('finish', async function() {
|
|
const createdNoteIds = {};
|
|
|
|
for (const path in createdPaths) {
|
|
createdNoteIds[createdPaths[path]] = true;
|
|
}
|
|
|
|
// we're saving attributes and links only now so that all relation and link target notes
|
|
// are already in the database (we don't want to have "broken" relations, not even transitionally)
|
|
for (const attr of attributes) {
|
|
if (attr.type !== 'relation' || attr.value in createdNoteIds) {
|
|
await new Attribute(attr).save();
|
|
}
|
|
else {
|
|
log.info("Relation not imported since target note doesn't exist: " + JSON.stringify(attr));
|
|
}
|
|
}
|
|
|
|
for (const link of links) {
|
|
if (link.targetNoteId in createdNoteIds) {
|
|
await new Link(link).save();
|
|
}
|
|
else {
|
|
log.info("Link not imported since target note doesn't exist: " + JSON.stringify(link));
|
|
}
|
|
}
|
|
|
|
resolve(firstNote);
|
|
});
|
|
|
|
const bufferStream = new stream.PassThrough();
|
|
bufferStream.end(fileBuffer);
|
|
|
|
bufferStream.pipe(extract);
|
|
});
|
|
}
|
|
|
|
module.exports = {
|
|
importTar
|
|
}; |