mirror of
https://github.com/Kong/insomnia
synced 2024-11-08 23:00:30 +00:00
549ce23ce8
* All projects into monorepo * Update CI * More CI updates * Extracted a bunch of things into packages * Publish - insomnia-plugin-base64@1.0.1 - insomnia-plugin-default-headers@1.0.2 - insomnia-plugin-file@1.0.1 - insomnia-plugin-hash@1.0.1 - insomnia-plugin-now@1.0.1 - insomnia-plugin-request@1.0.1 - insomnia-plugin-response@1.0.1 - insomnia-plugin-uuid@1.0.1 - insomnia-cookies@0.0.2 - insomnia-importers@1.5.2 - insomnia-prettify@0.0.3 - insomnia-url@0.0.2 - insomnia-xpath@0.0.2 * A bunch of small fixes * Improved build script * Fixed * Merge dangling files * Usability refactor * Handle duplicate plugin names
612 lines
16 KiB
JavaScript
612 lines
16 KiB
JavaScript
// @flow
|
|
import type {BaseModel} from '../models/index';
|
|
import * as models from '../models/index';
|
|
import electron from 'electron';
|
|
import NeDB from 'nedb';
|
|
import fsPath from 'path';
|
|
import {DB_PERSIST_INTERVAL} from './constants';
|
|
import uuid from 'uuid';
|
|
|
|
export const CHANGE_INSERT = 'insert';
|
|
export const CHANGE_UPDATE = 'update';
|
|
export const CHANGE_REMOVE = 'remove';
|
|
|
|
const database = {};
|
|
const db = {
|
|
_empty: true
|
|
};
|
|
|
|
// ~~~~~~~ //
|
|
// HELPERS //
|
|
// ~~~~~~~ //
|
|
|
|
function allTypes () {
|
|
return Object.keys(db);
|
|
}
|
|
|
|
function getDBFilePath (modelType) {
|
|
// NOTE: Do not EVER change this. EVER!
|
|
const {app} = electron.remote || electron;
|
|
const basePath = app.getPath('userData');
|
|
return fsPath.join(basePath, `insomnia.${modelType}.db`);
|
|
}
|
|
|
|
export async function initClient () {
|
|
electron.ipcRenderer.on('db.changes', async (e, changes) => {
|
|
for (const fn of changeListeners) {
|
|
await fn(changes);
|
|
}
|
|
});
|
|
console.log('[db] Initialized DB client');
|
|
}
|
|
|
|
export async function init (
|
|
types: Array<string>,
|
|
config: Object = {},
|
|
forceReset: boolean = false
|
|
) {
|
|
if (forceReset) {
|
|
changeListeners = [];
|
|
for (const attr of Object.keys(db)) {
|
|
if (attr === '_empty') {
|
|
continue;
|
|
}
|
|
|
|
delete db[attr];
|
|
}
|
|
}
|
|
|
|
// Fill in the defaults
|
|
for (const modelType of types) {
|
|
if (db[modelType]) {
|
|
console.log(`[db] Already initialized DB.${modelType}`);
|
|
continue;
|
|
}
|
|
|
|
const filePath = getDBFilePath(modelType);
|
|
const collection = new NeDB(Object.assign({
|
|
autoload: true,
|
|
filename: filePath
|
|
}, config));
|
|
|
|
collection.persistence.setAutocompactionInterval(DB_PERSIST_INTERVAL);
|
|
|
|
db[modelType] = collection;
|
|
}
|
|
|
|
delete db._empty;
|
|
|
|
electron.ipcMain.on('db.fn', async (e, fnName, replyChannel, ...args) => {
|
|
const result = await database[fnName](...args);
|
|
e.sender.send(replyChannel, result);
|
|
});
|
|
|
|
// NOTE: Only repair the DB if we're not running in memory. Repairing here causes tests to
|
|
// hang indefinitely for some reason.
|
|
// TODO: Figure out why this makes tests hang
|
|
if (!config.inMemoryOnly) {
|
|
await _repairDatabase();
|
|
}
|
|
|
|
if (!config.inMemoryOnly) {
|
|
console.log(`[db] Initialized DB at ${getDBFilePath('$TYPE')}`);
|
|
}
|
|
}
|
|
|
|
// ~~~~~~~~~~~~~~~~ //
|
|
// Change Listeners //
|
|
// ~~~~~~~~~~~~~~~~ //
|
|
|
|
let bufferingChanges = false;
|
|
let changeBuffer = [];
|
|
let changeListeners = [];
|
|
|
|
export function onChange (callback: Function): void {
|
|
changeListeners.push(callback);
|
|
}
|
|
|
|
export function offChange (callback: Function): void {
|
|
changeListeners = changeListeners.filter(l => l !== callback);
|
|
}
|
|
|
|
export const bufferChanges = database.bufferChanges = async function (millis: number = 1000): Promise<void> {
|
|
if (db._empty) return _send('bufferChanges', ...arguments);
|
|
|
|
bufferingChanges = true;
|
|
setTimeout(database.flushChanges, millis);
|
|
};
|
|
|
|
export const flushChangesAsync = database.flushChangesAsync = async function (): Promise<void> {
|
|
process.nextTick(async () => {
|
|
await flushChanges();
|
|
});
|
|
};
|
|
|
|
export const flushChanges = database.flushChanges = async function (): Promise<void> {
|
|
if (db._empty) return _send('flushChanges', ...arguments);
|
|
|
|
bufferingChanges = false;
|
|
const changes = [...changeBuffer];
|
|
changeBuffer = [];
|
|
|
|
if (changes.length === 0) {
|
|
// No work to do
|
|
return;
|
|
}
|
|
|
|
// Notify local listeners too
|
|
for (const fn of changeListeners) {
|
|
await fn(changes);
|
|
}
|
|
|
|
// Notify remote listeners
|
|
const windows = electron.BrowserWindow.getAllWindows();
|
|
for (const window of windows) {
|
|
window.webContents.send('db.changes', changes);
|
|
}
|
|
};
|
|
|
|
async function notifyOfChange (event: string, doc: BaseModel, fromSync: boolean): Promise<void> {
|
|
changeBuffer.push([event, doc, fromSync]);
|
|
|
|
// Flush right away if we're not buffering
|
|
if (!bufferingChanges) {
|
|
await database.flushChanges();
|
|
}
|
|
}
|
|
|
|
// ~~~~~~~ //
|
|
// Helpers //
|
|
// ~~~~~~~ //
|
|
|
|
export const getMostRecentlyModified = database.getMostRecentlyModified = async function (
|
|
type: string,
|
|
query: Object = {}
|
|
): Promise<BaseModel | null> {
|
|
if (db._empty) return _send('getMostRecentlyModified', ...arguments);
|
|
|
|
const docs = await database.findMostRecentlyModified(type, query, 1);
|
|
return docs.length ? docs[0] : null;
|
|
};
|
|
|
|
export const findMostRecentlyModified = database.findMostRecentlyModified = async function (
|
|
type: string,
|
|
query: Object = {},
|
|
limit: number | null = null
|
|
): Promise<Array<BaseModel>> {
|
|
if (db._empty) return _send('findMostRecentlyModified', ...arguments);
|
|
|
|
return new Promise(resolve => {
|
|
db[type].find(query).sort({modified: -1}).limit(limit).exec(async (err, rawDocs) => {
|
|
if (err) {
|
|
console.warn('[db] Failed to find docs', err);
|
|
resolve([]);
|
|
return;
|
|
}
|
|
|
|
const docs = [];
|
|
for (const rawDoc of rawDocs) {
|
|
docs.push(await models.initModel(type, rawDoc));
|
|
}
|
|
|
|
resolve(docs);
|
|
});
|
|
});
|
|
};
|
|
|
|
export const find = database.find = async function <T: BaseModel> (
|
|
type: string,
|
|
query: Object = {},
|
|
sort: Object = {created: 1}
|
|
): Promise<Array<T>> {
|
|
if (db._empty) return _send('find', ...arguments);
|
|
|
|
return new Promise((resolve, reject) => {
|
|
db[type].find(query).sort(sort).exec(async (err, rawDocs) => {
|
|
if (err) {
|
|
return reject(err);
|
|
}
|
|
|
|
const docs = [];
|
|
for (const rawDoc of rawDocs) {
|
|
docs.push(await models.initModel(type, rawDoc));
|
|
}
|
|
|
|
resolve(docs);
|
|
});
|
|
});
|
|
};
|
|
|
|
export const all = database.all = async function <T: BaseModel> (type: string): Promise<Array<T>> {
|
|
if (db._empty) return _send('all', ...arguments);
|
|
|
|
return database.find(type);
|
|
};
|
|
|
|
export const getWhere = database.getWhere = async function <T: BaseModel> (
|
|
type: string,
|
|
query: Object
|
|
): Promise<T | null> {
|
|
if (db._empty) return _send('getWhere', ...arguments);
|
|
|
|
const docs = await database.find(type, query);
|
|
return docs.length ? docs[0] : null;
|
|
};
|
|
|
|
export const get = database.get = async function <T: BaseModel> (
|
|
type: string,
|
|
id: string
|
|
): Promise<T | null> {
|
|
if (db._empty) return _send('get', ...arguments);
|
|
|
|
// Short circuit IDs used to represent nothing
|
|
if (!id || id === 'n/a') {
|
|
return null;
|
|
} else {
|
|
return database.getWhere(type, {_id: id});
|
|
}
|
|
};
|
|
|
|
export const count = database.count = async function (
|
|
type: string,
|
|
query: Object = {}
|
|
): Promise<number> {
|
|
if (db._empty) return _send('count', ...arguments);
|
|
|
|
return new Promise((resolve, reject) => {
|
|
db[type].count(query, (err, count) => {
|
|
if (err) {
|
|
return reject(err);
|
|
}
|
|
|
|
resolve(count);
|
|
});
|
|
});
|
|
};
|
|
|
|
export const upsert = database.upsert = async function (
|
|
doc: BaseModel,
|
|
fromSync: boolean = false
|
|
): Promise<BaseModel> {
|
|
if (db._empty) return _send('upsert', ...arguments);
|
|
|
|
const existingDoc = await database.get(doc.type, doc._id);
|
|
if (existingDoc) {
|
|
return database.update(doc, fromSync);
|
|
} else {
|
|
return database.insert(doc, fromSync);
|
|
}
|
|
};
|
|
|
|
export const insert = database.insert = async function <T: BaseModel> (
|
|
doc: T,
|
|
fromSync: boolean = false
|
|
): Promise<T> {
|
|
if (db._empty) return _send('insert', ...arguments);
|
|
|
|
return new Promise(async (resolve, reject) => {
|
|
const docWithDefaults = await models.initModel(doc.type, doc);
|
|
db[doc.type].insert(docWithDefaults, (err, newDoc) => {
|
|
if (err) {
|
|
return reject(err);
|
|
}
|
|
|
|
resolve(newDoc);
|
|
|
|
// NOTE: This needs to be after we resolve
|
|
notifyOfChange(CHANGE_INSERT, newDoc, fromSync);
|
|
});
|
|
});
|
|
};
|
|
|
|
export const update = database.update = async function <T: BaseModel> (
|
|
doc: T,
|
|
fromSync: boolean = false
|
|
): Promise<T> {
|
|
if (db._empty) return _send('update', ...arguments);
|
|
|
|
return new Promise(async (resolve, reject) => {
|
|
const docWithDefaults = await models.initModel(doc.type, doc);
|
|
db[doc.type].update({_id: docWithDefaults._id}, docWithDefaults, err => {
|
|
if (err) {
|
|
return reject(err);
|
|
}
|
|
|
|
resolve(docWithDefaults);
|
|
|
|
// NOTE: This needs to be after we resolve
|
|
notifyOfChange(CHANGE_UPDATE, docWithDefaults, fromSync);
|
|
});
|
|
});
|
|
};
|
|
|
|
export const remove = database.remove = async function <T: BaseModel> (
|
|
doc: T,
|
|
fromSync: boolean = false
|
|
): Promise<void> {
|
|
if (db._empty) return _send('remove', ...arguments);
|
|
|
|
await database.bufferChanges();
|
|
|
|
const docs = await database.withDescendants(doc);
|
|
const docIds = docs.map(d => d._id);
|
|
const types = [...new Set(docs.map(d => d.type))];
|
|
|
|
// Don't really need to wait for this to be over;
|
|
types.map(t => db[t].remove({_id: {$in: docIds}}, {multi: true}));
|
|
|
|
docs.map(d => notifyOfChange(CHANGE_REMOVE, d, fromSync));
|
|
|
|
await database.flushChanges();
|
|
};
|
|
|
|
export const removeWhere = database.removeWhere = async function (
|
|
type: string,
|
|
query: Object
|
|
): Promise<void> {
|
|
if (db._empty) return _send('removeWhere', ...arguments);
|
|
|
|
await database.bufferChanges();
|
|
|
|
for (const doc of await database.find(type, query)) {
|
|
const docs = await database.withDescendants(doc);
|
|
const docIds = docs.map(d => d._id);
|
|
const types = [...new Set(docs.map(d => d.type))];
|
|
|
|
// Don't really need to wait for this to be over;
|
|
types.map(t => db[t].remove({_id: {$in: docIds}}, {multi: true}));
|
|
|
|
docs.map(d => notifyOfChange(CHANGE_REMOVE, d, false));
|
|
}
|
|
|
|
await database.flushChanges();
|
|
};
|
|
|
|
// ~~~~~~~~~~~~~~~~~~~ //
|
|
// DEFAULT MODEL STUFF //
|
|
// ~~~~~~~~~~~~~~~~~~~ //
|
|
|
|
export async function docUpdate<T: BaseModel> (originalDoc: T, ...patches: Array<Object>): Promise<T> {
|
|
const doc = await models.initModel(
|
|
originalDoc.type,
|
|
originalDoc,
|
|
|
|
// NOTE: This is before `patch` because we want `patch.modified` to win if it has it
|
|
{modified: Date.now()},
|
|
|
|
...patches
|
|
);
|
|
|
|
return database.update(doc);
|
|
}
|
|
|
|
export async function docCreate<T: BaseModel> (
|
|
type: string,
|
|
...patches: Array<Object>
|
|
): Promise<T> {
|
|
const doc = await models.initModel(
|
|
type,
|
|
...patches,
|
|
|
|
// Fields that the user can't touch
|
|
{type: type}
|
|
);
|
|
|
|
return database.insert(doc);
|
|
}
|
|
|
|
// ~~~~~~~ //
|
|
// GENERAL //
|
|
// ~~~~~~~ //
|
|
|
|
export const withDescendants = database.withDescendants = async function (
|
|
doc: BaseModel | null,
|
|
stopType: string | null = null
|
|
): Promise<Array<BaseModel>> {
|
|
if (db._empty) return _send('withDescendants', ...arguments);
|
|
|
|
let docsToReturn = doc ? [doc] : [];
|
|
|
|
async function next (docs: Array<BaseModel | null>): Promise<Array<BaseModel>> {
|
|
let foundDocs = [];
|
|
|
|
for (const d of docs) {
|
|
if (stopType && d && d.type === stopType) {
|
|
continue;
|
|
}
|
|
|
|
for (const type of allTypes()) {
|
|
// If the doc is null, we want to search for parentId === null
|
|
const parentId = d ? d._id : null;
|
|
const more = await database.find(type, {parentId});
|
|
foundDocs = [...foundDocs, ...more];
|
|
}
|
|
}
|
|
|
|
if (foundDocs.length === 0) {
|
|
// Didn't find anything. We're done
|
|
return docsToReturn;
|
|
}
|
|
|
|
// Continue searching for children
|
|
docsToReturn = [...docsToReturn, ...foundDocs];
|
|
return next(foundDocs);
|
|
}
|
|
|
|
return next([doc]);
|
|
};
|
|
|
|
export const withAncestors = database.withAncestors = async function (
|
|
doc: BaseModel | null,
|
|
types: Array<string> = allTypes()
|
|
): Promise<Array<BaseModel>> {
|
|
if (db._empty) return _send('withAncestors', ...arguments);
|
|
|
|
if (!doc) {
|
|
return [];
|
|
}
|
|
|
|
let docsToReturn = doc ? [doc] : [];
|
|
|
|
async function next (docs: Array<BaseModel>): Promise<Array<BaseModel>> {
|
|
let foundDocs = [];
|
|
for (const d: BaseModel of docs) {
|
|
for (const type of types) {
|
|
// If the doc is null, we want to search for parentId === null
|
|
const another = await database.get(type, d.parentId);
|
|
another && foundDocs.push(another);
|
|
}
|
|
}
|
|
|
|
if (foundDocs.length === 0) {
|
|
// Didn't find anything. We're done
|
|
return docsToReturn;
|
|
}
|
|
|
|
// Continue searching for children
|
|
docsToReturn = [...docsToReturn, ...foundDocs];
|
|
return next(foundDocs);
|
|
}
|
|
|
|
return next([doc]);
|
|
};
|
|
|
|
export const duplicate = database.duplicate = async function <T: BaseModel> (
|
|
originalDoc: T,
|
|
patch: Object = {}
|
|
): Promise<T> {
|
|
if (db._empty) return _send('duplicate', ...arguments);
|
|
|
|
await database.bufferChanges();
|
|
|
|
async function next<T: BaseModel> (docToCopy: T, patch: Object): Promise<T> {
|
|
// 1. Copy the doc
|
|
const newDoc = Object.assign({}, docToCopy, patch);
|
|
delete newDoc._id;
|
|
delete newDoc.created;
|
|
delete newDoc.modified;
|
|
|
|
const createdDoc = await docCreate(newDoc.type, newDoc);
|
|
|
|
// 2. Get all the children
|
|
for (const type of allTypes()) {
|
|
// Note: We never want to duplicate a response
|
|
if (!models.canDuplicate(type)) {
|
|
continue;
|
|
}
|
|
|
|
const parentId = docToCopy._id;
|
|
const children = await database.find(type, {parentId});
|
|
for (const doc of children) {
|
|
await next(doc, {parentId: createdDoc._id});
|
|
}
|
|
}
|
|
|
|
return createdDoc;
|
|
}
|
|
|
|
const createdDoc = await next(originalDoc, patch);
|
|
|
|
await database.flushChanges();
|
|
|
|
return createdDoc;
|
|
};
|
|
|
|
// ~~~~~~~ //
|
|
// Helpers //
|
|
// ~~~~~~~ //
|
|
|
|
async function _send<T> (fnName: string, ...args: Array<any>): Promise<T> {
|
|
return new Promise((resolve, reject) => {
|
|
const replyChannel = `db.fn.reply:${uuid.v4()}`;
|
|
electron.ipcRenderer.send('db.fn', fnName, replyChannel, ...args);
|
|
electron.ipcRenderer.once(replyChannel, (e, result) => {
|
|
resolve(result);
|
|
});
|
|
});
|
|
}
|
|
|
|
/**
|
|
* Run various database repair scripts
|
|
*/
|
|
export async function _repairDatabase () {
|
|
console.log(`[fix] Running database repairs`);
|
|
for (const workspace of await find(models.workspace.type)) {
|
|
await _repairBaseEnvironments(workspace);
|
|
await _fixMultipleCookieJars(workspace);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* This function repairs workspaces that have multiple base environments. Since a workspace
|
|
* can only have one, this function walks over all base environments, merges the data, and
|
|
* moves all children as well.
|
|
*/
|
|
async function _repairBaseEnvironments (workspace) {
|
|
const baseEnvironments = await find(models.environment.type, {parentId: workspace._id});
|
|
|
|
// Nothing to do here
|
|
if (baseEnvironments.length <= 1) {
|
|
return;
|
|
}
|
|
|
|
const chosenBase = baseEnvironments[0];
|
|
for (const baseEnvironment of baseEnvironments) {
|
|
if (baseEnvironment._id === chosenBase._id) {
|
|
continue;
|
|
}
|
|
|
|
chosenBase.data = Object.assign(baseEnvironment.data, chosenBase.data);
|
|
const subEnvironments = await find(models.environment.type, {parentId: baseEnvironment._id});
|
|
|
|
for (const subEnvironment of subEnvironments) {
|
|
await docUpdate(subEnvironment, {parentId: chosenBase._id});
|
|
}
|
|
|
|
// Remove unnecessary base env
|
|
await remove(baseEnvironment);
|
|
}
|
|
|
|
// Update remaining base env
|
|
await update(chosenBase);
|
|
|
|
console.log(`[fix] Merged ${baseEnvironments.length} base environments under ${workspace.name}`);
|
|
}
|
|
|
|
/**
|
|
* This function repairs workspaces that have multiple cookie jars. Since a workspace
|
|
* can only have one, this function walks over all jars and merges them and their cookies
|
|
* together.
|
|
*/
|
|
async function _fixMultipleCookieJars (workspace) {
|
|
const cookieJars = await find(models.cookieJar.type, {parentId: workspace._id});
|
|
|
|
// Nothing to do here
|
|
if (cookieJars.length <= 1) {
|
|
return;
|
|
}
|
|
|
|
const chosenJar = cookieJars[0];
|
|
for (const cookieJar of cookieJars) {
|
|
if (cookieJar._id === chosenJar._id) {
|
|
continue;
|
|
}
|
|
|
|
for (const cookie of cookieJar.cookies) {
|
|
if (chosenJar.cookies.find(c => c.id === cookie.id)) {
|
|
continue;
|
|
}
|
|
|
|
chosenJar.cookies.push(cookie);
|
|
}
|
|
|
|
// Remove unnecessary jar
|
|
await remove(cookieJar);
|
|
}
|
|
|
|
// Update remaining jar
|
|
await update(chosenJar);
|
|
|
|
console.log(`[fix] Merged ${cookieJars.length} cookie jars under ${workspace.name}`);
|
|
}
|