mirror of
https://github.com/Kong/insomnia
synced 2024-11-08 14:49:53 +00:00
902 lines
31 KiB
JavaScript
902 lines
31 KiB
JavaScript
// @flow
|
|
import type {ResponseHeader, ResponseTimelineEntry} from '../models/response';
|
|
import type {Request, RequestHeader} from '../models/request';
|
|
import type {Workspace} from '../models/workspace';
|
|
import type {Settings} from '../models/settings';
|
|
import type {RenderedRequest} from '../common/render';
|
|
import {getRenderedRequest, getRenderedRequestAndContext, RENDER_PURPOSE_SEND} from '../common/render';
|
|
import mkdirp from 'mkdirp';
|
|
import clone from 'clone';
|
|
import {parse as urlParse, resolve as urlResolve} from 'url';
|
|
import {Curl} from 'insomnia-libcurl';
|
|
import {join as pathJoin} from 'path';
|
|
import uuid from 'uuid';
|
|
import * as electron from 'electron';
|
|
import * as models from '../models';
|
|
import {AUTH_AWS_IAM, AUTH_BASIC, AUTH_DIGEST, AUTH_NETRC, AUTH_NTLM, CONTENT_TYPE_FORM_DATA, CONTENT_TYPE_FORM_URLENCODED, getAppVersion, getTempDir, STATUS_CODE_PLUGIN_ERROR} from '../common/constants';
|
|
import {delay, describeByteSize, getContentTypeHeader, getHostHeader, getLocationHeader, getSetCookieHeaders, hasAcceptEncodingHeader, hasAcceptHeader, hasAuthHeader, hasContentTypeHeader, hasUserAgentHeader, waitForStreamToFinish} from '../common/misc';
|
|
import {buildQueryStringFromParams, joinUrlAndQueryString, setDefaultProtocol, smartEncodeUrl} from 'insomnia-url';
|
|
import fs from 'fs';
|
|
import * as db from '../common/database';
|
|
import * as CACerts from './cacert';
|
|
import * as plugins from '../plugins/index';
|
|
import * as pluginContexts from '../plugins/context/index';
|
|
import {getAuthHeader} from './authentication';
|
|
import {cookiesFromJar, jarFromCookies} from 'insomnia-cookies';
|
|
import {urlMatchesCertHost} from './url-matches-cert-host';
|
|
import aws4 from 'aws4';
|
|
import {buildMultipart} from './multipart';
|
|
|
|
const {app} = electron.remote || electron;
|
|
|
|
export type ResponsePatch = {
|
|
statusMessage?: string,
|
|
error?: string,
|
|
url?: string,
|
|
statusCode?: number,
|
|
bytesContent?: number,
|
|
bodyPath?: string,
|
|
bodyCompression?: 'zip' | null,
|
|
message?: string,
|
|
httpVersion?: string,
|
|
headers?: Array<ResponseHeader>,
|
|
elapsedTime?: number,
|
|
contentType?: string,
|
|
bytesRead?: number,
|
|
parentId?: string,
|
|
settingStoreCookies?: boolean,
|
|
settingSendCookies?: boolean,
|
|
timeline?: Array<ResponseTimelineEntry>
|
|
};
|
|
|
|
// Time since user's last keypress to wait before making the request
|
|
const MAX_DELAY_TIME = 1000;
|
|
|
|
let cancelRequestFunction = null;
|
|
let lastUserInteraction = Date.now();
|
|
|
|
export function cancelCurrentRequest () {
|
|
if (typeof cancelRequestFunction === 'function') {
|
|
cancelRequestFunction();
|
|
}
|
|
}
|
|
|
|
export async function _actuallySend (
|
|
renderedRequest: RenderedRequest,
|
|
workspace: Workspace,
|
|
settings: Settings
|
|
): Promise<ResponsePatch> {
|
|
return new Promise(async resolve => {
|
|
let timeline: Array<ResponseTimelineEntry> = [];
|
|
|
|
// Initialize the curl handle
|
|
const curl = new Curl();
|
|
|
|
/** Helper function to respond with a success */
|
|
function respond (patch: ResponsePatch, bodyPath: ?string): void {
|
|
const response = Object.assign(({
|
|
parentId: renderedRequest._id,
|
|
bodyCompression: null, // Will default to .zip otherwise
|
|
timeline: timeline,
|
|
bodyPath: bodyPath || '',
|
|
settingSendCookies: renderedRequest.settingSendCookies,
|
|
settingStoreCookies: renderedRequest.settingStoreCookies
|
|
}: ResponsePatch), patch);
|
|
|
|
resolve(response);
|
|
|
|
// Apply plugin hooks and don't wait for them and don't throw from them
|
|
process.nextTick(async () => {
|
|
try {
|
|
await _applyResponsePluginHooks(response);
|
|
} catch (err) {
|
|
// TODO: Better error handling here
|
|
console.warn('Response plugin failed', err);
|
|
}
|
|
});
|
|
}
|
|
|
|
/** Helper function to respond with an error */
|
|
function handleError (err: Error): void {
|
|
respond({
|
|
url: renderedRequest.url,
|
|
parentId: renderedRequest._id,
|
|
error: err.message,
|
|
elapsedTime: 0,
|
|
statusMessage: 'Error',
|
|
settingSendCookies: renderedRequest.settingSendCookies,
|
|
settingStoreCookies: renderedRequest.settingStoreCookies
|
|
});
|
|
}
|
|
|
|
/** Helper function to set Curl options */
|
|
function setOpt (opt: number, val: any, optional: boolean = false) {
|
|
const name = Object.keys(Curl.option).find(name => Curl.option[name] === opt);
|
|
try {
|
|
curl.setOpt(opt, val);
|
|
} catch (err) {
|
|
if (!optional) {
|
|
throw new Error(`${err.message} (${opt} ${name || 'n/a'})`);
|
|
} else {
|
|
console.warn(`Failed to set optional Curl opt (${opt} ${name || 'n/a'})`);
|
|
}
|
|
}
|
|
}
|
|
|
|
function enable (feature: number) {
|
|
curl.enable(feature);
|
|
}
|
|
|
|
try {
|
|
// Setup the cancellation logic
|
|
cancelRequestFunction = () => {
|
|
respond({
|
|
elapsedTime: curl.getInfo(Curl.info.TOTAL_TIME) * 1000,
|
|
bytesRead: curl.getInfo(Curl.info.SIZE_DOWNLOAD),
|
|
url: curl.getInfo(Curl.info.EFFECTIVE_URL),
|
|
statusMessage: 'Cancelled',
|
|
error: 'Request was cancelled'
|
|
});
|
|
|
|
// Kill it!
|
|
curl.close();
|
|
};
|
|
|
|
// Set all the basic options
|
|
setOpt(Curl.option.FOLLOWLOCATION, settings.followRedirects);
|
|
setOpt(Curl.option.TIMEOUT_MS, settings.timeout); // 0 for no timeout
|
|
setOpt(Curl.option.VERBOSE, true); // True so debug function works
|
|
setOpt(Curl.option.NOPROGRESS, false); // False so progress function works
|
|
setOpt(Curl.option.ACCEPT_ENCODING, ''); // Auto decode everything
|
|
enable(Curl.feature.NO_HEADER_PARSING);
|
|
enable(Curl.feature.NO_DATA_PARSING);
|
|
|
|
// Set maximum amount of redirects allowed
|
|
// NOTE: Setting this to -1 breaks some versions of libcurl
|
|
if (settings.maxRedirects > 0) {
|
|
setOpt(Curl.option.MAXREDIRS, settings.maxRedirects);
|
|
}
|
|
|
|
// Only set CURLOPT_CUSTOMREQUEST if not HEAD or GET. This is because Curl
|
|
// See https://curl.haxx.se/libcurl/c/CURLOPT_CUSTOMREQUEST.html
|
|
switch (renderedRequest.method.toUpperCase()) {
|
|
case 'HEAD':
|
|
// This is how you tell Curl to send a HEAD request
|
|
setOpt(Curl.option.NOBODY, 1);
|
|
break;
|
|
case 'POST':
|
|
// This is how you tell Curl to send a POST request
|
|
setOpt(Curl.option.POST, 1);
|
|
break;
|
|
default:
|
|
// IMPORTANT: Only use CUSTOMREQUEST for all but HEAD and POST
|
|
setOpt(Curl.option.CUSTOMREQUEST, renderedRequest.method);
|
|
break;
|
|
}
|
|
|
|
// Setup debug handler
|
|
setOpt(Curl.option.DEBUGFUNCTION, (infoType: string, content: string) => {
|
|
const name = Object.keys(Curl.info.debug).find(k => Curl.info.debug[k] === infoType) || '';
|
|
|
|
if (
|
|
infoType === Curl.info.debug.SSL_DATA_IN ||
|
|
infoType === Curl.info.debug.SSL_DATA_OUT
|
|
) {
|
|
return 0;
|
|
}
|
|
|
|
// Ignore the possibly large data messages
|
|
if (infoType === Curl.info.debug.DATA_OUT) {
|
|
if (content.length === 0) {
|
|
// Sometimes this happens, but I'm not sure why. Just ignore it.
|
|
} else if (content.length < 1000) {
|
|
timeline.push({name, value: content});
|
|
} else {
|
|
timeline.push({name, value: `(${describeByteSize(content.length)} hidden)`});
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
if (infoType === Curl.info.debug.DATA_IN) {
|
|
timeline.push({
|
|
name: 'TEXT',
|
|
value: `Received ${describeByteSize(content.length)} chunk`
|
|
});
|
|
return 0;
|
|
}
|
|
|
|
// Don't show cookie setting because this will display every domain in the jar
|
|
if (infoType === Curl.info.debug.TEXT && content.indexOf('Added cookie') === 0) {
|
|
return 0;
|
|
}
|
|
|
|
timeline.push({name, value: content});
|
|
|
|
return 0; // Must be here
|
|
});
|
|
|
|
// Set the headers (to be modified as we go)
|
|
const headers = clone(renderedRequest.headers);
|
|
|
|
let lastPercent = 0;
|
|
// NOTE: This option was added in 7.32.0 so make it optional
|
|
setOpt(Curl.option.XFERINFOFUNCTION, (dltotal, dlnow, ultotal, ulnow) => {
|
|
if (dltotal === 0) {
|
|
return 0;
|
|
}
|
|
|
|
const percent = Math.round(dlnow / dltotal * 100);
|
|
if (percent !== lastPercent) {
|
|
// console.log(`[network] Request downloaded ${percent}%`);
|
|
lastPercent = percent;
|
|
}
|
|
|
|
return 0;
|
|
}, true);
|
|
|
|
// Set the URL, including the query parameters
|
|
const qs = buildQueryStringFromParams(renderedRequest.parameters);
|
|
const url = joinUrlAndQueryString(renderedRequest.url, qs);
|
|
const isUnixSocket = url.match(/https?:\/\/unix:\//);
|
|
const finalUrl = smartEncodeUrl(url, renderedRequest.settingEncodeUrl);
|
|
if (isUnixSocket) {
|
|
// URL prep will convert "unix:/path" hostname to "unix/path"
|
|
const match = finalUrl.match(/(https?:)\/\/unix:?(\/[^:]+):\/(.+)/);
|
|
const protocol = (match && match[1]) || '';
|
|
const socketPath = (match && match[2]) || '';
|
|
const socketUrl = (match && match[3]) || '';
|
|
curl.setUrl(`${protocol}//${socketUrl}`);
|
|
setOpt(Curl.option.UNIX_SOCKET_PATH, socketPath);
|
|
} else {
|
|
curl.setUrl(finalUrl);
|
|
}
|
|
timeline.push({name: 'TEXT', value: 'Preparing request to ' + finalUrl});
|
|
timeline.push({name: 'TEXT', value: `Using ${Curl.getVersion()}`});
|
|
|
|
// log some things
|
|
if (renderedRequest.settingEncodeUrl) {
|
|
timeline.push({name: 'TEXT', value: 'Enable automatic URL encoding'});
|
|
} else {
|
|
timeline.push({name: 'TEXT', value: 'Disable automatic URL encoding'});
|
|
}
|
|
|
|
// SSL Validation
|
|
if (settings.validateSSL) {
|
|
timeline.push({name: 'TEXT', value: 'Enable SSL validation'});
|
|
} else {
|
|
setOpt(Curl.option.SSL_VERIFYHOST, 0);
|
|
setOpt(Curl.option.SSL_VERIFYPEER, 0);
|
|
timeline.push({name: 'TEXT', value: 'Disable SSL validation'});
|
|
}
|
|
|
|
// Setup CA Root Certificates if not on Mac. Thanks to libcurl, Mac will use
|
|
// certificates form the OS.
|
|
if (process.platform !== 'darwin') {
|
|
const baseCAPath = getTempDir();
|
|
const fullCAPath = pathJoin(baseCAPath, CACerts.filename);
|
|
|
|
try {
|
|
fs.statSync(fullCAPath);
|
|
} catch (err) {
|
|
// Doesn't exist yet, so write it
|
|
mkdirp.sync(baseCAPath);
|
|
fs.writeFileSync(fullCAPath, CACerts.blob);
|
|
console.log('[net] Set CA to', fullCAPath);
|
|
}
|
|
|
|
setOpt(Curl.option.CAINFO, fullCAPath);
|
|
}
|
|
|
|
// Set cookies from jar
|
|
if (renderedRequest.settingSendCookies) {
|
|
// Tell Curl to store cookies that it receives. This is only important if we receive
|
|
// a cookie on a redirect that needs to be sent on the next request in the chain.
|
|
curl.setOpt(Curl.option.COOKIEFILE, '');
|
|
|
|
const cookies = renderedRequest.cookieJar.cookies || [];
|
|
for (const cookie of cookies) {
|
|
let expiresTimestamp = 0;
|
|
if (cookie.expires) {
|
|
const expiresDate = new Date(cookie.expires);
|
|
expiresTimestamp = Math.round(expiresDate.getTime() / 1000);
|
|
}
|
|
|
|
setOpt(Curl.option.COOKIELIST, [
|
|
cookie.httpOnly ? `#HttpOnly_${cookie.domain}` : cookie.domain,
|
|
cookie.hostOnly ? 'FALSE' : 'TRUE',
|
|
cookie.path,
|
|
cookie.secure ? 'TRUE' : 'FALSE',
|
|
expiresTimestamp,
|
|
cookie.key,
|
|
cookie.value
|
|
].join('\t'));
|
|
}
|
|
|
|
for (const {name, value} of renderedRequest.cookies) {
|
|
setOpt(Curl.option.COOKIE, `${name}=${value}`);
|
|
}
|
|
|
|
timeline.push({
|
|
name: 'TEXT',
|
|
value: 'Enable cookie sending with jar of ' +
|
|
`${cookies.length} cookie${cookies.length !== 1 ? 's' : ''}`
|
|
});
|
|
} else {
|
|
timeline.push({
|
|
name: 'TEXT',
|
|
value: 'Disable cookie sending due to user setting'
|
|
});
|
|
}
|
|
|
|
// Set proxy settings if we have them
|
|
if (settings.proxyEnabled) {
|
|
const {protocol} = urlParse(renderedRequest.url);
|
|
const {httpProxy, httpsProxy, noProxy} = settings;
|
|
const proxyHost = protocol === 'https:' ? httpsProxy : httpProxy;
|
|
const proxy = proxyHost ? setDefaultProtocol(proxyHost) : null;
|
|
timeline.push({name: 'TEXT', value: `Enable network proxy for ${protocol || ''}`});
|
|
if (proxy) {
|
|
setOpt(Curl.option.PROXY, proxy);
|
|
setOpt(Curl.option.PROXYAUTH, Curl.auth.ANY);
|
|
}
|
|
if (noProxy) {
|
|
setOpt(Curl.option.NOPROXY, noProxy);
|
|
}
|
|
} else {
|
|
setOpt(Curl.option.PROXY, '');
|
|
}
|
|
|
|
// Set client certs if needed
|
|
const clientCertificates = await models.clientCertificate.findByParentId(workspace._id);
|
|
for (const certificate of clientCertificates) {
|
|
if (certificate.disabled) {
|
|
continue;
|
|
}
|
|
|
|
const cHostWithProtocol = setDefaultProtocol(certificate.host, 'https:');
|
|
|
|
if (urlMatchesCertHost(cHostWithProtocol, renderedRequest.url)) {
|
|
const ensureFile = blobOrFilename => {
|
|
try {
|
|
fs.statSync(blobOrFilename);
|
|
} catch (err) {
|
|
// Certificate file not found!
|
|
// LEGACY: Certs used to be stored in blobs (not as paths), so let's write it to
|
|
// the temp directory first.
|
|
const fullBase = getTempDir();
|
|
const name = `${renderedRequest._id}_${renderedRequest.modified}`;
|
|
const fullPath = pathJoin(fullBase, name);
|
|
fs.writeFileSync(fullPath, Buffer.from(blobOrFilename, 'base64'));
|
|
|
|
// Set filename to the one we just saved
|
|
blobOrFilename = fullPath;
|
|
}
|
|
|
|
return blobOrFilename;
|
|
};
|
|
|
|
const {passphrase, cert, key, pfx} = certificate;
|
|
|
|
if (cert) {
|
|
setOpt(Curl.option.SSLCERT, ensureFile(cert));
|
|
setOpt(Curl.option.SSLCERTTYPE, 'PEM');
|
|
timeline.push({name: 'TEXT', value: 'Adding SSL PEM certificate'});
|
|
}
|
|
|
|
if (pfx) {
|
|
setOpt(Curl.option.SSLCERT, ensureFile(pfx));
|
|
setOpt(Curl.option.SSLCERTTYPE, 'P12');
|
|
timeline.push({name: 'TEXT', value: 'Adding SSL P12 certificate'});
|
|
}
|
|
|
|
if (key) {
|
|
setOpt(Curl.option.SSLKEY, ensureFile(key));
|
|
timeline.push({name: 'TEXT', value: 'Adding SSL KEY certificate'});
|
|
}
|
|
|
|
if (passphrase) {
|
|
setOpt(Curl.option.KEYPASSWD, passphrase);
|
|
}
|
|
}
|
|
}
|
|
|
|
// Build the body
|
|
let noBody = false;
|
|
let requestBody = null;
|
|
const expectsBody = ['POST', 'PUT', 'PATCH'].includes(renderedRequest.method.toUpperCase());
|
|
if (renderedRequest.body.mimeType === CONTENT_TYPE_FORM_URLENCODED) {
|
|
requestBody = buildQueryStringFromParams(renderedRequest.body.params || [], false);
|
|
} else if (renderedRequest.body.mimeType === CONTENT_TYPE_FORM_DATA) {
|
|
const params = renderedRequest.body.params || [];
|
|
const {filePath: multipartBodyPath, boundary, contentLength} = await buildMultipart(params);
|
|
|
|
// Extend the Content-Type header
|
|
const contentTypeHeader = getContentTypeHeader(headers);
|
|
if (contentTypeHeader) {
|
|
contentTypeHeader.value = `multipart/form-data; boundary=${boundary}`;
|
|
} else {
|
|
headers.push({
|
|
name: 'Content-Type',
|
|
value: `multipart/form-data; boundary=${boundary}`
|
|
});
|
|
}
|
|
|
|
const fd = fs.openSync(multipartBodyPath, 'r+');
|
|
|
|
setOpt(Curl.option.INFILESIZE_LARGE, contentLength);
|
|
setOpt(Curl.option.UPLOAD, 1);
|
|
setOpt(Curl.option.READDATA, fd);
|
|
|
|
// We need this, otherwise curl will send it as a PUT
|
|
setOpt(Curl.option.CUSTOMREQUEST, renderedRequest.method);
|
|
|
|
const fn = () => {
|
|
fs.closeSync(fd);
|
|
fs.unlink(multipartBodyPath, () => {
|
|
});
|
|
};
|
|
|
|
curl.on('end', fn);
|
|
curl.on('error', fn);
|
|
} else if (renderedRequest.body.fileName) {
|
|
const {size} = fs.statSync(renderedRequest.body.fileName);
|
|
const fileName = renderedRequest.body.fileName || '';
|
|
const fd = fs.openSync(fileName, 'r+');
|
|
|
|
setOpt(Curl.option.INFILESIZE_LARGE, size);
|
|
setOpt(Curl.option.UPLOAD, 1);
|
|
setOpt(Curl.option.READDATA, fd);
|
|
|
|
// We need this, otherwise curl will send it as a POST
|
|
setOpt(Curl.option.CUSTOMREQUEST, renderedRequest.method);
|
|
|
|
const fn = () => fs.closeSync(fd);
|
|
curl.on('end', fn);
|
|
curl.on('error', fn);
|
|
} else if (typeof renderedRequest.body.mimeType === 'string' || expectsBody) {
|
|
requestBody = renderedRequest.body.text || '';
|
|
} else {
|
|
// No body
|
|
noBody = true;
|
|
}
|
|
|
|
if (!noBody) {
|
|
// Don't chunk uploads
|
|
headers.push({name: 'Expect', value: ''});
|
|
headers.push({name: 'Transfer-Encoding', value: ''});
|
|
}
|
|
|
|
// If we calculated the body within Insomnia (ie. not computed by Curl)
|
|
if (requestBody !== null) {
|
|
setOpt(Curl.option.POSTFIELDS, requestBody);
|
|
}
|
|
|
|
// Handle Authorization header
|
|
if (!hasAuthHeader(headers) && !renderedRequest.authentication.disabled) {
|
|
if (renderedRequest.authentication.type === AUTH_BASIC) {
|
|
const {username, password} = renderedRequest.authentication;
|
|
setOpt(Curl.option.HTTPAUTH, Curl.auth.BASIC);
|
|
setOpt(Curl.option.USERNAME, username || '');
|
|
setOpt(Curl.option.PASSWORD, password || '');
|
|
} else if (renderedRequest.authentication.type === AUTH_DIGEST) {
|
|
const {username, password} = renderedRequest.authentication;
|
|
setOpt(Curl.option.HTTPAUTH, Curl.auth.DIGEST);
|
|
setOpt(Curl.option.USERNAME, username || '');
|
|
setOpt(Curl.option.PASSWORD, password || '');
|
|
} else if (renderedRequest.authentication.type === AUTH_NTLM) {
|
|
const {username, password} = renderedRequest.authentication;
|
|
setOpt(Curl.option.HTTPAUTH, Curl.auth.NTLM);
|
|
setOpt(Curl.option.USERNAME, username || '');
|
|
setOpt(Curl.option.PASSWORD, password || '');
|
|
} else if (renderedRequest.authentication.type === AUTH_AWS_IAM) {
|
|
if (!noBody && !requestBody) {
|
|
return handleError(
|
|
new Error('AWS authentication not supported for provided body type'));
|
|
}
|
|
const credentials = {
|
|
accessKeyId: renderedRequest.authentication.accessKeyId || '',
|
|
secretAccessKey: renderedRequest.authentication.secretAccessKey || '',
|
|
sessionToken: renderedRequest.authentication.sessionToken || ''
|
|
};
|
|
const extraHeaders = _getAwsAuthHeaders(
|
|
credentials,
|
|
headers,
|
|
requestBody || '',
|
|
finalUrl,
|
|
renderedRequest.method
|
|
);
|
|
for (const header of extraHeaders) {
|
|
headers.push(header);
|
|
}
|
|
} else if (renderedRequest.authentication.type === AUTH_NETRC) {
|
|
setOpt(Curl.option.NETRC, Curl.netrc.REQUIRED);
|
|
} else {
|
|
const authHeader = await getAuthHeader(
|
|
renderedRequest._id,
|
|
finalUrl,
|
|
renderedRequest.method,
|
|
renderedRequest.authentication
|
|
);
|
|
|
|
if (authHeader) {
|
|
headers.push({
|
|
name: authHeader.name,
|
|
value: authHeader.value
|
|
});
|
|
}
|
|
}
|
|
}
|
|
|
|
// Send a default Accept headers of anything
|
|
if (!hasAcceptHeader(headers)) {
|
|
headers.push({name: 'Accept', value: '*/*'}); // Default to anything
|
|
}
|
|
|
|
// Don't auto-send Accept-Encoding header
|
|
if (!hasAcceptEncodingHeader(headers)) {
|
|
headers.push({name: 'Accept-Encoding', value: ''});
|
|
}
|
|
|
|
// Set User-Agent if it't not already in headers
|
|
if (!hasUserAgentHeader(headers)) {
|
|
setOpt(Curl.option.USERAGENT, `insomnia/${getAppVersion()}`);
|
|
}
|
|
|
|
// Prevent curl from adding default content-type header
|
|
if (!hasContentTypeHeader(headers)) {
|
|
headers.push({name: 'content-type', value: ''});
|
|
}
|
|
|
|
// NOTE: This is last because headers might be modified multiple times
|
|
const headerStrings = headers
|
|
.filter(h => h.name)
|
|
.map(h => `${(h.name || '').trim()}: ${h.value}`);
|
|
setOpt(Curl.option.HTTPHEADER, headerStrings);
|
|
|
|
let responseBodyBytes = 0;
|
|
const responsesDir = pathJoin(app.getPath('userData'), 'responses');
|
|
mkdirp.sync(responsesDir);
|
|
const responseBodyPath = pathJoin(responsesDir, uuid.v4() + '.response');
|
|
const responseBodyWriteStream = fs.createWriteStream(responseBodyPath);
|
|
curl.on('end', () => responseBodyWriteStream.end());
|
|
curl.on('error', () => responseBodyWriteStream.end());
|
|
setOpt(Curl.option.WRITEFUNCTION, (buff: Buffer) => {
|
|
responseBodyBytes += buff.length;
|
|
responseBodyWriteStream.write(buff);
|
|
return buff.length;
|
|
});
|
|
|
|
// Handle the response ending
|
|
curl.on('end', async (_1, _2, rawHeaders) => {
|
|
const allCurlHeadersObjects = _parseHeaders(rawHeaders);
|
|
// Headers are an array (one for each redirect)
|
|
const lastCurlHeadersObject = allCurlHeadersObjects[allCurlHeadersObjects.length - 1];
|
|
|
|
// Collect various things
|
|
const httpVersion = lastCurlHeadersObject.version || '';
|
|
const statusCode = lastCurlHeadersObject.code || -1;
|
|
const statusMessage = lastCurlHeadersObject.reason || '';
|
|
|
|
// Collect the headers
|
|
const headers = lastCurlHeadersObject.headers;
|
|
|
|
// Calculate the content type
|
|
const contentTypeHeader = getContentTypeHeader(headers);
|
|
const contentType = contentTypeHeader ? contentTypeHeader.value : '';
|
|
|
|
// Update Cookie Jar
|
|
let currentUrl = finalUrl;
|
|
let setCookieStrings: Array<string> = [];
|
|
const jar = jarFromCookies(renderedRequest.cookieJar.cookies);
|
|
|
|
for (const {headers} of allCurlHeadersObjects) {
|
|
// Collect Set-Cookie headers
|
|
const setCookieHeaders = getSetCookieHeaders(headers);
|
|
setCookieStrings = [...setCookieStrings, ...setCookieHeaders.map(h => h.value)];
|
|
|
|
// Pull out new URL if there is a redirect
|
|
const newLocation = getLocationHeader(headers);
|
|
if (newLocation !== null) {
|
|
currentUrl = urlResolve(currentUrl, newLocation.value);
|
|
}
|
|
}
|
|
|
|
// Update jar with Set-Cookie headers
|
|
for (const setCookieStr of setCookieStrings) {
|
|
try {
|
|
jar.setCookieSync(setCookieStr, currentUrl);
|
|
} catch (err) {
|
|
timeline.push({name: 'TEXT', value: `Rejected cookie: ${err.message}`});
|
|
}
|
|
}
|
|
|
|
// Update cookie jar if we need to and if we found any cookies
|
|
if (renderedRequest.settingStoreCookies && setCookieStrings.length) {
|
|
const cookies = await cookiesFromJar(jar);
|
|
models.cookieJar.update(renderedRequest.cookieJar, {cookies});
|
|
}
|
|
|
|
// Print informational message
|
|
if (setCookieStrings.length > 0) {
|
|
const n = setCookieStrings.length;
|
|
if (renderedRequest.settingStoreCookies) {
|
|
timeline.push({name: 'TEXT', value: `Saved ${n} cookie${n === 1 ? '' : 's'}`});
|
|
} else {
|
|
timeline.push({name: 'TEXT', value: `Ignored ${n} cookie${n === 1 ? '' : 's'}`});
|
|
}
|
|
}
|
|
|
|
// Return the response data
|
|
const responsePatch = {
|
|
headers,
|
|
contentType,
|
|
statusCode,
|
|
httpVersion,
|
|
statusMessage,
|
|
elapsedTime: curl.getInfo(Curl.info.TOTAL_TIME) * 1000,
|
|
bytesRead: curl.getInfo(Curl.info.SIZE_DOWNLOAD),
|
|
bytesContent: responseBodyBytes,
|
|
url: curl.getInfo(Curl.info.EFFECTIVE_URL)
|
|
};
|
|
|
|
// Close the request
|
|
curl.close();
|
|
|
|
// Make sure the response body has been fully written first
|
|
await waitForStreamToFinish(responseBodyWriteStream);
|
|
|
|
respond(responsePatch, responseBodyPath);
|
|
});
|
|
|
|
curl.on('error', function (err, code) {
|
|
let error = err + '';
|
|
let statusMessage = 'Error';
|
|
|
|
if (code === Curl.code.CURLE_ABORTED_BY_CALLBACK) {
|
|
error = 'Request aborted';
|
|
statusMessage = 'Abort';
|
|
}
|
|
|
|
respond({statusMessage, error});
|
|
});
|
|
|
|
curl.perform();
|
|
} catch (err) {
|
|
handleError(err);
|
|
}
|
|
});
|
|
}
|
|
|
|
export async function sendWithSettings (
|
|
requestId: string,
|
|
requestPatch: Object
|
|
): Promise<ResponsePatch> {
|
|
const request = await models.request.getById(requestId);
|
|
if (!request) {
|
|
throw new Error(`Failed to find request: ${requestId}`);
|
|
}
|
|
|
|
const settings = await models.settings.getOrCreate();
|
|
const ancestors = await db.withAncestors(request, [
|
|
models.request.type,
|
|
models.requestGroup.type,
|
|
models.workspace.type
|
|
]);
|
|
|
|
const workspaceDoc = ancestors.find(doc => doc.type === models.workspace.type);
|
|
const workspaceId = workspaceDoc ? workspaceDoc._id : 'n/a';
|
|
const workspace = await models.workspace.getById(workspaceId);
|
|
if (!workspace) {
|
|
throw new Error(`Failed to find workspace for: ${requestId}`);
|
|
}
|
|
|
|
const workspaceMeta = await models.workspaceMeta.getOrCreateByParentId(workspace._id);
|
|
const environmentId: string = workspaceMeta.activeEnvironmentId || 'n/a';
|
|
|
|
const newRequest: Request = await models.initModel(models.request.type, requestPatch, {
|
|
_id: request._id + '.other',
|
|
parentId: request._id
|
|
});
|
|
|
|
let renderedRequest: RenderedRequest;
|
|
try {
|
|
renderedRequest = await getRenderedRequest(newRequest, environmentId);
|
|
} catch (err) {
|
|
throw new Error(`Failed to render request: ${requestId}`);
|
|
}
|
|
|
|
return _actuallySend(renderedRequest, workspace, settings);
|
|
}
|
|
|
|
export async function send (
|
|
requestId: string,
|
|
environmentId: string
|
|
): Promise<ResponsePatch> {
|
|
// HACK: wait for all debounces to finish
|
|
/*
|
|
* TODO: Do this in a more robust way
|
|
* The following block adds a "long" delay to let potential debounces and
|
|
* database updates finish before making the request. This is done by tracking
|
|
* the time of the user's last keypress and making sure the request is sent a
|
|
* significant time after the last press.
|
|
*/
|
|
const timeSinceLastInteraction = Date.now() - lastUserInteraction;
|
|
const delayMillis = Math.max(0, MAX_DELAY_TIME - timeSinceLastInteraction);
|
|
if (delayMillis > 0) {
|
|
await delay(delayMillis);
|
|
}
|
|
|
|
// Fetch some things
|
|
const request = await models.request.getById(requestId);
|
|
const settings = await models.settings.getOrCreate();
|
|
const ancestors = await db.withAncestors(request, [
|
|
models.request.type,
|
|
models.requestGroup.type,
|
|
models.workspace.type
|
|
]);
|
|
|
|
if (!request) {
|
|
throw new Error(`Failed to find request to send for ${requestId}`);
|
|
}
|
|
|
|
const renderResult = await getRenderedRequestAndContext(
|
|
request,
|
|
environmentId,
|
|
RENDER_PURPOSE_SEND
|
|
);
|
|
|
|
const renderedRequestBeforePlugins = renderResult.request;
|
|
const renderedContextBeforePlugins = renderResult.context;
|
|
|
|
const workspaceDoc = ancestors.find(doc => doc.type === models.workspace.type);
|
|
const workspace = await models.workspace.getById(workspaceDoc ? workspaceDoc._id : 'n/a');
|
|
if (!workspace) {
|
|
throw new Error(`Failed to find workspace for request: ${requestId}`);
|
|
}
|
|
|
|
let renderedRequest: RenderedRequest;
|
|
try {
|
|
renderedRequest = await _applyRequestPluginHooks(
|
|
renderedRequestBeforePlugins,
|
|
renderedContextBeforePlugins
|
|
);
|
|
} catch (err) {
|
|
return {
|
|
response: {
|
|
url: renderedRequestBeforePlugins.url,
|
|
parentId: renderedRequestBeforePlugins._id,
|
|
error: err.message,
|
|
statusCode: STATUS_CODE_PLUGIN_ERROR,
|
|
statusMessage: err.plugin ? `Plugin ${err.plugin.name}` : 'Plugin',
|
|
settingSendCookies: renderedRequestBeforePlugins.settingSendCookies,
|
|
settingStoreCookies: renderedRequestBeforePlugins.settingStoreCookies
|
|
},
|
|
bodyBuffer: null
|
|
};
|
|
}
|
|
|
|
return _actuallySend(renderedRequest, workspace, settings);
|
|
}
|
|
|
|
async function _applyRequestPluginHooks (
|
|
renderedRequest: RenderedRequest,
|
|
renderedContext: Object
|
|
): Promise<RenderedRequest> {
|
|
let newRenderedRequest = renderedRequest;
|
|
for (const {plugin, hook} of await plugins.getRequestHooks()) {
|
|
newRenderedRequest = clone(newRenderedRequest);
|
|
|
|
const context = {
|
|
...pluginContexts.app.init(),
|
|
...pluginContexts.request.init(newRenderedRequest, renderedContext)
|
|
};
|
|
|
|
try {
|
|
await hook(context);
|
|
} catch (err) {
|
|
err.plugin = plugin;
|
|
throw err;
|
|
}
|
|
}
|
|
|
|
return newRenderedRequest;
|
|
}
|
|
|
|
async function _applyResponsePluginHooks (
|
|
response: ResponsePatch
|
|
): Promise<void> {
|
|
for (const {plugin, hook} of await plugins.getResponseHooks()) {
|
|
const context = {
|
|
...pluginContexts.app.init(),
|
|
...pluginContexts.response.init(response)
|
|
};
|
|
|
|
try {
|
|
await hook(context);
|
|
} catch (err) {
|
|
err.plugin = plugin;
|
|
throw err;
|
|
}
|
|
}
|
|
}
|
|
|
|
export function _parseHeaders (
|
|
buffer: Buffer
|
|
): Array<{headers: Array<ResponseHeader>, version: string, code: number, reason: string}> {
|
|
const results = [];
|
|
|
|
const lines = buffer.toString('utf8').split(/\r?\n|\r/g);
|
|
|
|
for (let i = 0, currentResult = null; i < lines.length; i++) {
|
|
const line = lines[i];
|
|
const isEmptyLine = line.trim() === '';
|
|
|
|
// If we hit an empty line, start parsing the next response
|
|
if (isEmptyLine && currentResult) {
|
|
results.push(currentResult);
|
|
currentResult = null;
|
|
continue;
|
|
}
|
|
|
|
if (!currentResult) {
|
|
const [version, code, ...other] = line.split(/ +/g);
|
|
currentResult = {
|
|
version,
|
|
code: parseInt(code, 10),
|
|
reason: other.join(' '),
|
|
headers: []
|
|
};
|
|
} else {
|
|
const [name, value] = line.split(/:\s(.+)/);
|
|
const header: ResponseHeader = {name, value: value || ''};
|
|
currentResult.headers.push(header);
|
|
}
|
|
}
|
|
|
|
return results;
|
|
}
|
|
|
|
// exported for unit tests only
|
|
export function _getAwsAuthHeaders (
|
|
credentials: Object,
|
|
headers: Array<RequestHeader>,
|
|
body: string,
|
|
url: string,
|
|
method: string
|
|
) {
|
|
const parsedUrl = urlParse(url);
|
|
const contentTypeHeader = getContentTypeHeader(headers);
|
|
|
|
// AWS uses host header for signing so prioritize that if the user set it manually
|
|
const hostHeader = getHostHeader(headers);
|
|
const host = hostHeader ? hostHeader.value : parsedUrl.host;
|
|
|
|
const awsSignOptions = {
|
|
body,
|
|
method,
|
|
host,
|
|
path: parsedUrl.path,
|
|
headers: {
|
|
'content-type': contentTypeHeader ? contentTypeHeader.value : ''
|
|
}
|
|
};
|
|
|
|
const signature = aws4.sign(awsSignOptions, credentials);
|
|
|
|
return Object.keys(signature.headers)
|
|
.filter(name => name !== 'content-type') // Don't add this because we already have it
|
|
.map(name => ({name, value: signature.headers[name]}));
|
|
}
|
|
|
|
document.addEventListener('keydown', (e: KeyboardEvent) => {
|
|
if (e.ctrlKey || e.metaKey || e.altKey) {
|
|
return;
|
|
}
|
|
|
|
lastUserInteraction = Date.now();
|
|
});
|
|
|
|
document.addEventListener('paste', e => {
|
|
lastUserInteraction = Date.now();
|
|
});
|