mirror of
https://github.com/zadam/trilium.git
synced 2025-03-01 14:22:32 +01:00
fix import
This commit is contained in:
parent
4d5a17583f
commit
56e2b44c25
@ -126,10 +126,11 @@ async function exportToTar(branch, format, res) {
|
||||
meta.dirFileName = getUniqueFilename(existingFileNames, baseFileName);
|
||||
meta.children = [];
|
||||
|
||||
// namespace is shared by children in the same note
|
||||
const childExistingNames = {};
|
||||
|
||||
for (const childBranch of childBranches) {
|
||||
const note = await getNote(childBranch, existingFileNames);
|
||||
const note = await getNote(childBranch, childExistingNames);
|
||||
|
||||
// can be undefined if export is disabled for this note
|
||||
if (note) {
|
||||
@ -160,27 +161,27 @@ async function exportToTar(branch, format, res) {
|
||||
if (noteMeta.isClone) {
|
||||
const content = "Note is present at " + notePaths[noteMeta.noteId];
|
||||
|
||||
pack.entry({name: path + '/' + noteMeta.dataFileName, size: content.length}, content);
|
||||
pack.entry({name: path + noteMeta.dataFileName, size: content.length}, content);
|
||||
return;
|
||||
}
|
||||
|
||||
const note = await repository.getNote(noteMeta.noteId);
|
||||
|
||||
notePaths[note.noteId] = path + '/' + (noteMeta.dataFileName || noteMeta.dirFileName);
|
||||
notePaths[note.noteId] = path + (noteMeta.dataFileName || noteMeta.dirFileName);
|
||||
|
||||
if (noteMeta.dataFileName) {
|
||||
const content = prepareContent(note, noteMeta.format);
|
||||
|
||||
pack.entry({name: path + '/' + noteMeta.dataFileName, size: content.length}, content);
|
||||
pack.entry({name: path + noteMeta.dataFileName, size: content.length}, content);
|
||||
}
|
||||
|
||||
if (noteMeta.children && noteMeta.children.length > 0) {
|
||||
const directoryPath = path + '/' + noteMeta.dirFileName;
|
||||
const directoryPath = path + noteMeta.dirFileName;
|
||||
|
||||
pack.entry({name: directoryPath, type: 'directory'});
|
||||
|
||||
for (const childMeta of noteMeta.children) {
|
||||
await saveNote(childMeta, directoryPath);
|
||||
await saveNote(childMeta, directoryPath + '/');
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -3,6 +3,7 @@
|
||||
const Attribute = require('../../entities/attribute');
|
||||
const Link = require('../../entities/link');
|
||||
const utils = require('../../services/utils');
|
||||
const log = require('../../services/log');
|
||||
const repository = require('../../services/repository');
|
||||
const noteService = require('../../services/notes');
|
||||
const Branch = require('../../entities/branch');
|
||||
@ -15,6 +16,8 @@ const mimeTypes = require('mime-types');
|
||||
async function importTar(fileBuffer, importRootNote) {
|
||||
// maps from original noteId (in tar file) to newly generated noteId
|
||||
const noteIdMap = {};
|
||||
const attributes = [];
|
||||
const links = [];
|
||||
// path => noteId
|
||||
const createdPaths = { '/': importRootNote.noteId, '\\': importRootNote.noteId };
|
||||
const mdReader = new commonmark.Parser();
|
||||
@ -134,7 +137,7 @@ async function importTar(fileBuffer, importRootNote) {
|
||||
return { type, mime };
|
||||
}
|
||||
|
||||
async function saveAttributes(note, noteMeta) {
|
||||
async function saveAttributesAndLinks(note, noteMeta) {
|
||||
if (!noteMeta) {
|
||||
return;
|
||||
}
|
||||
@ -146,14 +149,14 @@ async function importTar(fileBuffer, importRootNote) {
|
||||
attr.value = getNewNoteId(attr.value);
|
||||
}
|
||||
|
||||
await new Attribute(attr).save();
|
||||
attributes.push(attr);
|
||||
}
|
||||
|
||||
for (const link of noteMeta.links) {
|
||||
link.noteId = note.noteId;
|
||||
link.targetNoteId = getNewNoteId(link.targetNoteId);
|
||||
|
||||
await new Link(link).save();
|
||||
links.push(link);
|
||||
}
|
||||
}
|
||||
|
||||
@ -178,14 +181,12 @@ async function importTar(fileBuffer, importRootNote) {
|
||||
isExpanded: noteMeta ? noteMeta.isExpanded : false
|
||||
}));
|
||||
|
||||
await saveAttributes(note, noteMeta);
|
||||
await saveAttributesAndLinks(note, noteMeta);
|
||||
|
||||
if (!firstNote) {
|
||||
firstNote = note;
|
||||
}
|
||||
|
||||
console.log(filePath);
|
||||
|
||||
createdPaths[filePath] = noteId;
|
||||
}
|
||||
|
||||
@ -207,8 +208,6 @@ async function importTar(fileBuffer, importRootNote) {
|
||||
const parentNoteId = getParentNoteId(filePath, parentNoteMeta);
|
||||
|
||||
if (noteMeta && noteMeta.isClone) {
|
||||
console.log(`Creating branch for ${noteId} in ${parentNoteId}`);
|
||||
|
||||
await new Branch({
|
||||
noteId,
|
||||
parentNoteId,
|
||||
@ -247,7 +246,7 @@ async function importTar(fileBuffer, importRootNote) {
|
||||
isExpanded: noteMeta ? noteMeta.isExpanded : false
|
||||
}));
|
||||
|
||||
await saveAttributes(note, noteMeta);
|
||||
await saveAttributesAndLinks(note, noteMeta);
|
||||
|
||||
if (!firstNote) {
|
||||
firstNote = note;
|
||||
@ -257,7 +256,6 @@ async function importTar(fileBuffer, importRootNote) {
|
||||
filePath = getTextFileWithoutExtension(filePath);
|
||||
}
|
||||
|
||||
console.log(filePath);
|
||||
createdPaths[filePath] = noteId;
|
||||
}
|
||||
}
|
||||
@ -273,6 +271,7 @@ async function importTar(fileBuffer, importRootNote) {
|
||||
if (filePath.endsWith("/")) {
|
||||
filePath = filePath.substr(0, filePath.length - 1);
|
||||
}
|
||||
|
||||
return filePath;
|
||||
}
|
||||
|
||||
@ -298,9 +297,12 @@ async function importTar(fileBuffer, importRootNote) {
|
||||
else if (header.type === 'directory') {
|
||||
await saveDirectory(filePath);
|
||||
}
|
||||
else {
|
||||
else if (header.type === 'file') {
|
||||
await saveNote(filePath, content);
|
||||
}
|
||||
else {
|
||||
log.info("Ignoring tar import entry with type " + header.type);
|
||||
}
|
||||
|
||||
next(); // ready for next entry
|
||||
});
|
||||
@ -309,7 +311,27 @@ async function importTar(fileBuffer, importRootNote) {
|
||||
});
|
||||
|
||||
return new Promise(resolve => {
|
||||
extract.on('finish', function() {
|
||||
extract.on('finish', async function() {
|
||||
const createdNoteIds = {};
|
||||
|
||||
for (const path in createdPaths) {
|
||||
createdNoteIds[createdPaths[path]] = true;
|
||||
}
|
||||
|
||||
// we're saving attributes and links only now so that all relation and link target notes
|
||||
// are already in the database (we don't want to have "broken" relations, not even transitionally)
|
||||
for (const attr of attributes) {
|
||||
if (attr.value in createdNoteIds) {
|
||||
await new Attribute(attr).save();
|
||||
}
|
||||
}
|
||||
|
||||
for (const link of links) {
|
||||
if (link.targetNoteId in createdNoteIds) {
|
||||
await new Link(link).save();
|
||||
}
|
||||
}
|
||||
|
||||
resolve(firstNote);
|
||||
});
|
||||
|
||||
|
@ -327,11 +327,11 @@ eventService.subscribe(eventService.ENTITY_CHANGED, async ({entityName, entity})
|
||||
|
||||
if (attribute.type === 'label' && attribute.name === 'archived') {
|
||||
// we're not using label object directly, since there might be other non-deleted archived label
|
||||
const hideLabel = await repository.getEntity(`SELECT * FROM attributes WHERE isDeleted = 0 AND type = 'label'
|
||||
const archivedLabel = await repository.getEntity(`SELECT * FROM attributes WHERE isDeleted = 0 AND type = 'label'
|
||||
AND name = 'archived' AND noteId = ?`, [attribute.noteId]);
|
||||
|
||||
if (hideLabel) {
|
||||
archived[attribute.noteId] = hideLabel.isInheritable ? 1 : 0;
|
||||
if (archivedLabel) {
|
||||
archived[attribute.noteId] = archivedLabel.isInheritable ? 1 : 0;
|
||||
}
|
||||
else {
|
||||
delete archived[attribute.noteId];
|
||||
|
@ -107,6 +107,7 @@ async function updateEntity(entity) {
|
||||
|
||||
// it seems to be better to handle deletion and update separately
|
||||
await eventService.emit(entity.isDeleted ? eventService.ENTITY_DELETED : eventService.ENTITY_CHANGED, eventPayload);
|
||||
|
||||
}
|
||||
});
|
||||
}
|
||||
|
Loading…
x
Reference in New Issue
Block a user