mirror of
https://github.com/meteor/meteor.git
synced 2026-05-02 03:01:46 -04:00
1637 lines
49 KiB
JavaScript
1637 lines
49 KiB
JavaScript
var _ = require('underscore');
|
|
var path = require('path');
|
|
var fs = require('fs');
|
|
var files = require('./files.js');
|
|
var utils = require('./utils.js');
|
|
var parseStack = require('./parse-stack.js');
|
|
var release = require('./release.js');
|
|
var catalog = require('./catalog.js');
|
|
var archinfo = require('./archinfo.js');
|
|
var packageLoader = require('./package-loader.js');
|
|
var Future = require('fibers/future');
|
|
var uniload = require('./uniload.js');
|
|
var config = require('./config.js');
|
|
var buildmessage = require('./buildmessage.js');
|
|
var util = require('util');
|
|
var child_process = require('child_process');
|
|
var webdriver = require('browserstack-webdriver');
|
|
var phantomjs = require('phantomjs');
|
|
|
|
var Package = uniload.load({ packages: ["ejson"] });
|
|
|
|
var toolPackageName = "meteor-tool";
|
|
|
|
// Exception representing a test failure
|
|
var TestFailure = function (reason, details) {
|
|
var self = this;
|
|
self.reason = reason;
|
|
self.details = details || {};
|
|
self.stack = (new Error).stack;
|
|
};
|
|
|
|
// Use this to decorate functions that throw TestFailure. Decorate the
|
|
// first function that should not be included in the call stack shown
|
|
// to the user.
|
|
var markStack = function (f) {
|
|
return parseStack.markTop(f);
|
|
};
|
|
|
|
// Call from a test to throw a TestFailure exception and bail out of the test
|
|
var fail = markStack(function (reason) {
|
|
throw new TestFailure(reason);
|
|
});
|
|
|
|
// Call from a test to assert that 'actual' is equal to 'expected',
|
|
// with 'actual' being the value that the test got and 'expected'
|
|
// being the expected value
|
|
var expectEqual = markStack(function (actual, expected) {
|
|
if (! Package.ejson.EJSON.equals(actual, expected)) {
|
|
throw new TestFailure("not-equal", {
|
|
expected: expected,
|
|
actual: actual
|
|
});
|
|
}
|
|
});
|
|
|
|
var expectThrows = markStack(function (f) {
|
|
var threw = false;
|
|
try {
|
|
f();
|
|
} catch (e) {
|
|
threw = true;
|
|
}
|
|
|
|
if (! threw)
|
|
throw new TestFailure("expected-exception");
|
|
});
|
|
|
|
var getToolsPackage = function () {
|
|
buildmessage.assertInCapture();
|
|
// Rebuild the tool package --- necessary because we don't actually
|
|
// rebuild the tool in the cached version every time.
|
|
if (catalog.complete.rebuildLocalPackages([toolPackageName]) !== 1) {
|
|
throw Error("didn't rebuild meteor-tool?");
|
|
}
|
|
var loader = new packageLoader.PackageLoader({
|
|
versions: null,
|
|
catalog: catalog.complete
|
|
});
|
|
return loader.getPackage(toolPackageName);
|
|
};
|
|
|
|
// Execute a command synchronously, discarding stderr.
|
|
var execFileSync = function (binary, args, opts) {
|
|
return Future.wrap(function(cb) {
|
|
var cb2 = function(err, stdout, stderr) { cb(err, stdout); };
|
|
child_process.execFile(binary, args, opts, cb2);
|
|
})().wait();
|
|
};
|
|
|
|
var doOrThrow = function (f) {
|
|
var ret;
|
|
var messages = buildmessage.capture(function () {
|
|
ret = f();
|
|
});
|
|
if (messages.hasMessages()) {
|
|
throw Error(messages.formatMessages());
|
|
}
|
|
return ret;
|
|
};
|
|
|
|
///////////////////////////////////////////////////////////////////////////////
|
|
// Matcher
|
|
///////////////////////////////////////////////////////////////////////////////
|
|
|
|
// Handles the job of waiting until text is seen that matches a
|
|
// regular expression.
|
|
|
|
var Matcher = function (run) {
|
|
var self = this;
|
|
self.buf = "";
|
|
self.ended = false;
|
|
self.matchPattern = null;
|
|
self.matchFuture = null;
|
|
self.matchStrict = null;
|
|
self.run = run; // used only to set a field on exceptions
|
|
};
|
|
|
|
_.extend(Matcher.prototype, {
|
|
write: function (data) {
|
|
var self = this;
|
|
self.buf += data;
|
|
self._tryMatch();
|
|
},
|
|
|
|
match: function (pattern, timeout, strict) {
|
|
var self = this;
|
|
if (self.matchFuture)
|
|
throw new Error("already have a match pending?");
|
|
self.matchPattern = pattern;
|
|
self.matchStrict = strict;
|
|
var f = self.matchFuture = new Future;
|
|
self._tryMatch(); // could clear self.matchFuture
|
|
|
|
var timer = null;
|
|
if (timeout) {
|
|
timer = setTimeout(function () {
|
|
self.matchPattern = null;
|
|
self.matchStrict = null;
|
|
self.matchFuture = null;
|
|
f['throw'](new TestFailure('match-timeout', { run: self.run }));
|
|
}, timeout * 1000);
|
|
}
|
|
|
|
try {
|
|
return f.wait();
|
|
} finally {
|
|
if (timer)
|
|
clearTimeout(timer);
|
|
}
|
|
},
|
|
|
|
end: function () {
|
|
var self = this;
|
|
self.ended = true;
|
|
self._tryMatch();
|
|
self.buf = '';
|
|
},
|
|
|
|
matchEmpty: function () {
|
|
var self = this;
|
|
|
|
if (self.buf.length > 0)
|
|
throw new TestFailure('junk-at-end', { run: self.run });
|
|
},
|
|
|
|
_tryMatch: function () {
|
|
var self = this;
|
|
|
|
var f = self.matchFuture;
|
|
if (! f)
|
|
return;
|
|
|
|
var ret = null;
|
|
|
|
if (self.matchPattern instanceof RegExp) {
|
|
var m = self.buf.match(self.matchPattern);
|
|
if (m) {
|
|
if (self.matchStrict && m.index !== 0) {
|
|
self.matchFuture = null;
|
|
self.matchStrict = null;
|
|
self.matchPattern = null;
|
|
f['throw'](new TestFailure('junk-before', { run: self.run }));
|
|
return;
|
|
}
|
|
ret = m;
|
|
self.buf = self.buf.slice(m.index + m[0].length);
|
|
}
|
|
} else {
|
|
var i = self.buf.indexOf(self.matchPattern);
|
|
if (i !== -1) {
|
|
if (self.matchStrict && i !== 0) {
|
|
self.matchFuture = null;
|
|
self.matchStrict = null;
|
|
self.matchPattern = null;
|
|
f['throw'](new TestFailure('junk-before', { run: self.run }));
|
|
return;
|
|
}
|
|
ret = self.matchPattern;
|
|
self.buf = self.buf.slice(i + self.matchPattern.length);
|
|
}
|
|
}
|
|
|
|
if (ret !== null) {
|
|
self.matchFuture = null;
|
|
self.matchStrict = null;
|
|
self.matchPattern = null;
|
|
f['return'](ret);
|
|
return;
|
|
}
|
|
|
|
if (self.ended) {
|
|
self.matchFuture = null;
|
|
self.matchStrict = null;
|
|
self.matchPattern = null;
|
|
f['throw'](new TestFailure('no-match', { run: self.run }));
|
|
return;
|
|
}
|
|
}
|
|
});
|
|
|
|
|
|
///////////////////////////////////////////////////////////////////////////////
|
|
// OutputLog
|
|
///////////////////////////////////////////////////////////////////////////////
|
|
|
|
// Maintains a line-by-line merged log of multiple output channels
|
|
// (eg, stdout and stderr).
|
|
|
|
var OutputLog = function (run) {
|
|
var self = this;
|
|
|
|
// each entry is an object with keys 'channel', 'text', and if it is
|
|
// the last entry and there was no newline terminator, 'bare'
|
|
self.lines = [];
|
|
|
|
// map from a channel name to an object representing a partially
|
|
// read line of text on that channel. That object has keys 'text'
|
|
// (text read), 'offset' (cursor position, equal to text.length
|
|
// unless a '\r' has been read).
|
|
self.buffers = {};
|
|
|
|
// a Run, exclusively for inclusion in exceptions
|
|
self.run = run;
|
|
};
|
|
|
|
_.extend(OutputLog.prototype, {
|
|
write: function (channel, text) {
|
|
var self = this;
|
|
|
|
if (! _.has(self.buffers, 'channel'))
|
|
self.buffers[channel] = { text: '', offset: 0};
|
|
var b = self.buffers[channel];
|
|
|
|
while (text.length) {
|
|
var m = text.match(/^[^\n\r]+/);
|
|
if (m) {
|
|
// A run of non-control characters.
|
|
b.text = b.text.substr(0, b.offset) +
|
|
m[0] + b.text.substr(b.offset + m[0].length);
|
|
b.offset += m[0].length;
|
|
text = text.substr(m[0].length);
|
|
continue;
|
|
}
|
|
|
|
if (text[0] === '\r') {
|
|
b.offset = 0;
|
|
text = text.substr(1);
|
|
continue;
|
|
}
|
|
|
|
if (text[0] === '\n') {
|
|
self.lines.push({ channel: channel, text: b.text });
|
|
b.text = '';
|
|
b.offset = 0;
|
|
text = text.substr(1);
|
|
continue;
|
|
}
|
|
|
|
throw new Error("conditions should have been exhaustive?");
|
|
}
|
|
},
|
|
|
|
end: function () {
|
|
var self = this;
|
|
|
|
_.each(_.keys(self.buffers), function (channel) {
|
|
if (self.buffers[channel].text.length) {
|
|
self.lines.push({ channel: channel,
|
|
text: self.buffers[channel].text,
|
|
bare: true });
|
|
self.buffers[channel] = { text: '', offset: 0};
|
|
}
|
|
});
|
|
},
|
|
|
|
forbid: function (pattern, channel) {
|
|
var self = this;
|
|
_.each(self.lines, function (line) {
|
|
if (channel && channel !== line.channel)
|
|
return;
|
|
|
|
var match = (pattern instanceof RegExp) ?
|
|
(line.text.match(pattern)) : (line.text.indexOf(pattern) !== -1);
|
|
if (match)
|
|
throw new TestFailure('forbidden-string-present', { run: self.run });
|
|
});
|
|
},
|
|
|
|
get: function () {
|
|
var self = this;
|
|
return self.lines;
|
|
}
|
|
});
|
|
|
|
|
|
///////////////////////////////////////////////////////////////////////////////
|
|
// Sandbox
|
|
///////////////////////////////////////////////////////////////////////////////
|
|
|
|
// Represents an install of the tool. Creating this creates a private
|
|
// sandbox with its own state, separate from the state of the current
|
|
// meteor install or checkout, from the user's homedir, and from the
|
|
// state of any other sandbox. It also creates an empty directory
|
|
// which will be, by default, the cwd for runs created inside the
|
|
// sandbox (you can change this with the cd() method).
|
|
//
|
|
// This will throw TestFailure if it has to build packages to set up
|
|
// the sandbox and the build fails. So, only call it from inside
|
|
// tests.
|
|
//
|
|
// options:
|
|
// - warehouse: set to sandbox the warehouse too. If you don't do
|
|
// this, the tests are run in the same context (checkout or
|
|
// warehouse) as the actual copy of meteor you're running (the
|
|
// meteor in 'meteor self-test'. This may only be set when you're
|
|
// running 'meteor self-test' from a checkout. If it is set, it
|
|
// should look something like this:
|
|
// {
|
|
// version1: { tools: 'tools1', notices: (...) },
|
|
// version2: { tools: 'tools2', upgraders: ["a"],
|
|
// notices: (...), latest: true }
|
|
// }
|
|
// This would set up a simulated warehouse with two releases in it,
|
|
// one called 'version1' and having a tools version of 'tools1', and
|
|
// similarly with 'version2'/'tools2', with the latter being marked
|
|
// as the latest release, and the latter also having a single
|
|
// upgrader named "a". The releases are made by building the
|
|
// checkout into a release, and are identical except for their
|
|
// version names. If you pass 'notices' (which is optional), set it
|
|
// to the verbatim contents of the notices.json file for the
|
|
// release, as an object.
|
|
// - fakeMongo: if set, set an environment variable that causes our
|
|
// 'fake-mongod' stub process to be started instead of 'mongod'. The
|
|
// tellMongo method then becomes available on Runs for controlling
|
|
// the stub.
|
|
// - clients
|
|
// - browserstack: true if browserstack clients should be used
|
|
// - port: the port that the clients should run on
|
|
|
|
var Sandbox = function (options) {
|
|
var self = this;
|
|
// default options
|
|
options = _.extend({ clients: {} }, options);
|
|
|
|
self.root = files.mkdtemp();
|
|
self.warehouse = null;
|
|
|
|
self.home = path.join(self.root, 'home');
|
|
fs.mkdirSync(self.home, 0755);
|
|
self.cwd = self.home;
|
|
self.env = {};
|
|
self.fakeMongo = options.fakeMongo;
|
|
|
|
// By default, tests use the package server that this meteor binary is built
|
|
// with. If a test is tagged 'test-package-server', it uses the test
|
|
// server. Tests that publish packages should have this flag; tests that
|
|
// assume that the release's packages can be found on the server should not.
|
|
// Note that this only affects subprocess meteor runs, not direct invocation
|
|
// of packageClient!
|
|
if (runningTest.tags['test-package-server']) {
|
|
self.set('METEOR_PACKAGE_SERVER_URL', exports.testPackageServerUrl);
|
|
}
|
|
|
|
if (_.has(options, 'warehouse')) {
|
|
if (!files.inCheckout())
|
|
throw Error("make only use a fake warehouse in a checkout");
|
|
self.warehouse = path.join(self.root, 'tropohouse');
|
|
self._makeWarehouse(options.warehouse);
|
|
}
|
|
|
|
self.clients = [new PhantomClient({
|
|
host: 'localhost',
|
|
port: options.clients.port || 3000
|
|
})];
|
|
|
|
if (options.clients && options.clients.browserstack) {
|
|
var browsers = [
|
|
{ browserName: 'firefox' },
|
|
{ browserName: 'chrome' },
|
|
{ browserName: 'internet explorer',
|
|
browserVersion: '11' },
|
|
{ browserName: 'internet explorer',
|
|
browserVersion: '8',
|
|
timeout: 60 },
|
|
{ browserName: 'safari' },
|
|
{ browserName: 'android' }
|
|
];
|
|
|
|
_.each(browsers, function (browser) {
|
|
self.clients.push(new BrowserStackClient({
|
|
host: 'localhost',
|
|
port: 3000,
|
|
browserName: browser.browserName,
|
|
browserVersion: browser.browserVersion,
|
|
timeout: browser.timeout
|
|
}));
|
|
});
|
|
}
|
|
|
|
// Figure out the 'meteor' to run
|
|
if (self.warehouse)
|
|
self.execPath = path.join(self.warehouse, 'meteor');
|
|
else
|
|
self.execPath = path.join(files.getCurrentToolsDir(), 'meteor');
|
|
};
|
|
|
|
_.extend(Sandbox.prototype, {
|
|
// Create a new test run of the tool in this sandbox.
|
|
run: function (/* arguments */) {
|
|
var self = this;
|
|
|
|
return new Run(self.execPath, {
|
|
sandbox: self,
|
|
args: _.toArray(arguments),
|
|
cwd: self.cwd,
|
|
env: self._makeEnv(),
|
|
fakeMongo: self.fakeMongo
|
|
});
|
|
},
|
|
|
|
// Tests a set of clients with the argument function. Each call to f(run)
|
|
// instantiates a Run with a different client.
|
|
// Use:
|
|
// sandbox.testWithAllClients(function (run) {
|
|
// // pre-connection checks
|
|
// run.connectClient();
|
|
// // post-connection checks
|
|
// });
|
|
testWithAllClients: function (f) {
|
|
var self = this;
|
|
var argsArray = _.compact(_.toArray(arguments).slice(1));
|
|
|
|
console.log("running test with " + self.clients.length + " client(s).");
|
|
|
|
_.each(self.clients, function (client) {
|
|
console.log("testing with " + client.name + "...");
|
|
var run = new Run(self.execPath, {
|
|
sandbox: self,
|
|
args: argsArray,
|
|
cwd: self.cwd,
|
|
env: self._makeEnv(),
|
|
fakeMongo: self.fakeMongo,
|
|
client: client
|
|
});
|
|
run.baseTimeout = client.timeout;
|
|
f(run);
|
|
});
|
|
},
|
|
|
|
// Copy an app from a template into the current directory in the
|
|
// sandbox. 'to' is the subdirectory to put the app in, and
|
|
// 'template' is a subdirectory of tools/tests/apps to copy.
|
|
//
|
|
// Note that the arguments are the opposite order from 'cp'. That
|
|
// seems more intuitive to me -- if you disagree, my apologies.
|
|
//
|
|
// For example:
|
|
// s.createApp('myapp', 'empty');
|
|
// s.cd('myapp');
|
|
createApp: function (to, template) {
|
|
var self = this;
|
|
files.cp_r(path.join(__dirname, 'tests', 'apps', template),
|
|
path.join(self.cwd, to),
|
|
{ ignore: [/^local$/] });
|
|
// If the test isn't explicitly managing a mock warehouse, ensure that apps
|
|
// run with our release by default.
|
|
if (!self.warehouse && release.current.isProperRelease()) {
|
|
self.write(path.join(to, '.meteor/release'), release.current.name);
|
|
}
|
|
},
|
|
|
|
// Same as createApp, but with a package.
|
|
//
|
|
// For example:
|
|
// s.createPackage('mypack', 'empty');
|
|
// s.cd('mypack');
|
|
createPackage: function (to, template) {
|
|
var self = this;
|
|
files.cp_r(path.join(__dirname, 'tests', 'packages', template),
|
|
path.join(self.cwd, to));
|
|
},
|
|
|
|
// Change the cwd to be used for subsequent runs. For example:
|
|
// s.run('create', 'myapp').expectExit(0);
|
|
// s.cd('myapp');
|
|
// s.run('add', 'somepackage') ...
|
|
// If you provide a callback, it will invoke the callback and then
|
|
// change the cwd back to the previous value. eg:
|
|
// s.cd('app1', function () {
|
|
// s.run('add', 'somepackage');
|
|
// });
|
|
// s.cd('app2', function () {
|
|
// s.run('add', 'somepackage');
|
|
// });
|
|
cd: function (relativePath, callback) {
|
|
var self = this;
|
|
var previous = self.cwd;
|
|
self.cwd = path.resolve(self.cwd, relativePath);
|
|
if (callback) {
|
|
callback();
|
|
self.cwd = previous;
|
|
}
|
|
},
|
|
|
|
// Set an environment variable for subsequent runs.
|
|
set: function (name, value) {
|
|
var self = this;
|
|
self.env[name] = value;
|
|
},
|
|
|
|
// Undo set().
|
|
unset: function (name) {
|
|
var self = this;
|
|
delete self.env[name];
|
|
},
|
|
|
|
// Write to a file in the sandbox, overwriting its current contents
|
|
// if any. 'filename' is a path intepreted relative to the Sandbox's
|
|
// cwd. 'contents' is a string (utf8 is assumed).
|
|
write: function (filename, contents) {
|
|
var self = this;
|
|
fs.writeFileSync(path.join(self.cwd, filename), contents, 'utf8');
|
|
},
|
|
|
|
// Reads a file in the sandbox as a utf8 string. 'filename' is a
|
|
// path intepreted relative to the Sandbox's cwd. Returns null if
|
|
// file does not exist.
|
|
read: function (filename) {
|
|
var self = this;
|
|
var file = path.join(self.cwd, filename);
|
|
if (!fs.existsSync(file))
|
|
return null;
|
|
else
|
|
return fs.readFileSync(path.join(self.cwd, filename), 'utf8');
|
|
},
|
|
|
|
// Copy the contents of one file to another. In these series of tests, we often
|
|
// want to switch contents of package.js files. It is more legible to copy in
|
|
// the backup file rather than trying to write into it manually.
|
|
cp: function(from, to) {
|
|
var self = this;
|
|
var contents = self.read(from);
|
|
if (!contents) {
|
|
throw new Error("File " + from + " does not exist.");
|
|
};
|
|
self.write(to, contents);
|
|
},
|
|
|
|
// Delete a file in the sandbox. 'filename' is as in write().
|
|
unlink: function (filename) {
|
|
var self = this;
|
|
fs.unlinkSync(path.join(self.cwd, filename));
|
|
},
|
|
|
|
// Make a directory in the sandbox. 'filename' is as in write().
|
|
mkdir: function (dirname) {
|
|
var self = this;
|
|
var dirPath = path.join(self.cwd, dirname);
|
|
if (! fs.existsSync(dirPath)) {
|
|
fs.mkdirSync(dirPath);
|
|
}
|
|
},
|
|
|
|
// Rename something in the sandbox. 'oldName' and 'newName' are as in write().
|
|
rename: function (oldName, newName) {
|
|
var self = this;
|
|
fs.renameSync(path.join(self.cwd, oldName),
|
|
path.join(self.cwd, newName));
|
|
},
|
|
|
|
// Return the current contents of .meteorsession in the sandbox.
|
|
readSessionFile: function () {
|
|
var self = this;
|
|
return fs.readFileSync(path.join(self.root, '.meteorsession'), 'utf8');
|
|
},
|
|
|
|
// Overwrite .meteorsession in the sandbox with 'contents'. You
|
|
// could use this in conjunction with readSessionFile to save and
|
|
// restore authentication states.
|
|
writeSessionFile: function (contents) {
|
|
var self = this;
|
|
return fs.writeFileSync(path.join(self.root, '.meteorsession'),
|
|
contents, 'utf8');
|
|
},
|
|
|
|
_makeEnv: function () {
|
|
var self = this;
|
|
var env = _.clone(self.env);
|
|
env.METEOR_SESSION_FILE = path.join(self.root, '.meteorsession');
|
|
|
|
if (self.warehouse) {
|
|
// Tell it where the warehouse lives.
|
|
env.METEOR_WAREHOUSE_DIR = self.warehouse;
|
|
// Don't ever try to refresh the stub catalog we made.
|
|
env.METEOR_OFFLINE_CATALOG = "t";
|
|
}
|
|
|
|
// By default (ie, with no mock warehouse and no --release arg) we should be
|
|
// testing the actual release this is built in, so we pretend that it is the
|
|
// latest release.
|
|
if (!self.warehouse && release.current.isProperRelease())
|
|
env.METEOR_TEST_LATEST_RELEASE = release.current.name;
|
|
return env;
|
|
},
|
|
|
|
// Writes a stub warehouse (really a tropohouse) to the directory
|
|
// self.warehouse. This warehouse only contains a meteor-tool package and some
|
|
// releases containing that tool only (and no packages).
|
|
//
|
|
// packageServerUrl indicates which package server we think we are using. Use
|
|
// the default, if we do not pass this in; you should pass it in any case that
|
|
// you will be specifying $METEOR_PACKAGE_SERVER_URL in the environment of a
|
|
// command you are running in this sandbox.
|
|
_makeWarehouse: function (releases) {
|
|
var self = this;
|
|
var serverUrl = self.env.METEOR_PACKAGE_SERVER_URL;
|
|
var packagesDirectoryName = config.getPackagesDirectoryName(serverUrl);
|
|
files.mkdir_p(path.join(self.warehouse, packagesDirectoryName), 0755);
|
|
files.mkdir_p(path.join(self.warehouse, 'package-metadata', 'v1'), 0755);
|
|
|
|
var stubCatalog = {
|
|
syncToken: {},
|
|
formatVersion: "1.0",
|
|
collections: {
|
|
packages: [],
|
|
versions: [],
|
|
builds: [],
|
|
releaseTracks: [],
|
|
releaseVersions: []
|
|
}
|
|
};
|
|
|
|
// Build all packages and symlink them into the warehouse. Remember
|
|
// their versions (which happen to contain build IDs).
|
|
// XXX Not sure where this comment comes from. We should presumably
|
|
// be building some packages besides meteor-tool (so that we can
|
|
// build apps that contain core packages).
|
|
|
|
var toolPackage, toolPackageDirectory;
|
|
doOrThrow(function () {
|
|
toolPackage = getToolsPackage();
|
|
toolPackageDirectory = '.' + toolPackage.version + '.XXX++'
|
|
+ toolPackage.buildArchitectures();
|
|
toolPackage.saveToPath(path.join(self.warehouse, packagesDirectoryName,
|
|
toolPackageName, toolPackageDirectory),
|
|
{ elideBuildInfo: true });
|
|
});
|
|
|
|
fs.symlinkSync(toolPackageDirectory,
|
|
path.join(self.warehouse, packagesDirectoryName,
|
|
toolPackageName, toolPackage.version));
|
|
stubCatalog.collections.packages.push({
|
|
name: toolPackageName,
|
|
_id: utils.randomToken()
|
|
});
|
|
var toolVersionId = utils.randomToken();
|
|
stubCatalog.collections.versions.push({
|
|
packageName: toolPackageName,
|
|
version: toolPackage.version,
|
|
earliestCompatibleVersion: toolPackage.version,
|
|
containsPlugins: false,
|
|
description: toolPackage.metadata.summary,
|
|
dependencies: {},
|
|
compilerVersion: require('./compiler.js').BUILT_BY,
|
|
_id: toolVersionId
|
|
});
|
|
|
|
self.toolPackageVersion = toolPackage.version;
|
|
|
|
stubCatalog.collections.builds.push({
|
|
architecture: toolPackage.buildArchitectures(),
|
|
versionId: toolVersionId,
|
|
_id: utils.randomToken()
|
|
});
|
|
stubCatalog.collections.releaseTracks.push({
|
|
name: catalog.DEFAULT_TRACK,
|
|
_id: utils.randomToken()
|
|
});
|
|
|
|
// Now create each requested release.
|
|
_.each(releases, function (configuration, releaseName) {
|
|
// Release info
|
|
stubCatalog.collections.releaseVersions.push({
|
|
track: catalog.DEFAULT_TRACK,
|
|
version: releaseName,
|
|
orderKey: releaseName,
|
|
description: "test release " + releaseName,
|
|
recommended: !!configuration.recommended,
|
|
// XXX support multiple tools packages for springboard tests
|
|
tool: toolPackageName + "@" + toolPackage.version,
|
|
packages: {}
|
|
});
|
|
});
|
|
|
|
// XXX: This is an incremental hack to be able to create apps from the
|
|
// warehouse. We need the constraint solver that runs are create-time to be
|
|
// able to solve for the starting app packages (standrd-app-packages,
|
|
// insecure & autopublish). But the solution doesn't have to be
|
|
// accurate. Later, when we care about the solution making sense, we should
|
|
// consider actually importing real data.
|
|
|
|
// XXXX: HACK. We are going to cheat and assume that these are already
|
|
// in the official catalog. Since we don't care about the contents, we
|
|
// should be OK.
|
|
var oldOffline = catalog.official.offline;
|
|
catalog.official.offline = true;
|
|
doOrThrow(function () {
|
|
catalog.official.refresh();
|
|
});
|
|
_.each(
|
|
['autopublish', 'meteor-platform', 'insecure'],
|
|
function (name) {
|
|
var versionRec = doOrThrow(function () {
|
|
return catalog.official.getLatestMainlineVersion(name);
|
|
});
|
|
if (!versionRec) {
|
|
catalog.official.offline = false;
|
|
doOrThrow(function () {
|
|
catalog.official.refresh();
|
|
});
|
|
catalog.official.offline = true;
|
|
versionRec = doOrThrow(function () {
|
|
return catalog.official.getLatestMainlineVersion(name);
|
|
});
|
|
if (!versionRec) {
|
|
throw new Error(" hack fails for " + name);
|
|
}
|
|
}
|
|
var buildRec = doOrThrow(function () {
|
|
return catalog.official.getAllBuilds(name, versionRec.version)[0];
|
|
});
|
|
|
|
// Insert into packages.
|
|
stubCatalog.collections.packages.push({
|
|
name: name,
|
|
_id: utils.randomToken()
|
|
});
|
|
|
|
// Insert into versions. We are making up a lot of this data.
|
|
var versionId = utils.randomToken();
|
|
stubCatalog.collections.versions.push({
|
|
packageName: name,
|
|
version: versionRec.version,
|
|
earliestCompatibleVersion: versionRec.earliestCompatibleVersion,
|
|
containsPlugins: false,
|
|
description: "warehouse test",
|
|
dependencies: {},
|
|
compilerVersion: require('./compiler.js').BUILT_BY,
|
|
_id: versionRec._id
|
|
});
|
|
|
|
// Insert into builds. Assume the package is available for all
|
|
// architectures.
|
|
stubCatalog.collections.builds.push({
|
|
buildArchitectures: "web.browser+os",
|
|
versionId: versionRec._id,
|
|
build: buildRec.build,
|
|
_id: utils.randomToken()
|
|
});
|
|
});
|
|
catalog.official.offline = oldOffline;
|
|
|
|
var dataFile = config.getLocalPackageCacheFilename(serverUrl);
|
|
fs.writeFileSync(
|
|
path.join(self.warehouse, 'package-metadata', 'v1', dataFile),
|
|
JSON.stringify(stubCatalog, null, 2));
|
|
|
|
// And a cherry on top
|
|
fs.symlinkSync(path.join(packagesDirectoryName,
|
|
toolPackageName, toolPackage.version,
|
|
'meteor-tool-' + archinfo.host(), 'meteor'),
|
|
path.join(self.warehouse, 'meteor'));
|
|
}
|
|
});
|
|
|
|
///////////////////////////////////////////////////////////////////////////////
|
|
// Client
|
|
///////////////////////////////////////////////////////////////////////////////
|
|
|
|
var Client = function (options) {
|
|
var self = this;
|
|
|
|
self.host = options.host;
|
|
self.port = options.port;
|
|
self.url = "http://" + self.host + ":" + self.port + '/' +
|
|
(Math.random() * 0x100000000 + 1).toString(36);
|
|
self.timeout = options.timeout || 40;
|
|
|
|
if (! self.connect || ! self.stop) {
|
|
console.log("Missing methods in subclass of Client.");
|
|
}
|
|
};
|
|
|
|
// PhantomClient
|
|
var PhantomClient = function (options) {
|
|
var self = this;
|
|
Client.apply(this, arguments);
|
|
|
|
self.name = "phantomjs";
|
|
self.process = null;
|
|
};
|
|
|
|
util.inherits(PhantomClient, Client);
|
|
|
|
_.extend(PhantomClient.prototype, {
|
|
connect: function () {
|
|
var self = this;
|
|
|
|
var phantomScript = "require('webpage').create().open('" + self.url + "');";
|
|
var phantomPath = phantomjs.path;
|
|
self.process = child_process.execFile(
|
|
'/bin/bash',
|
|
['-c',
|
|
("exec " + phantomPath + " --load-images=no /dev/stdin <<'END'\n" +
|
|
phantomScript + "END\n")]);
|
|
},
|
|
stop: function() {
|
|
var self = this;
|
|
self.process && self.process.kill();
|
|
self.process = null;
|
|
}
|
|
});
|
|
|
|
// BrowserStackClient
|
|
var browserStackKey = null;
|
|
|
|
var BrowserStackClient = function (options) {
|
|
var self = this;
|
|
Client.apply(this, arguments);
|
|
|
|
self.tunnelProcess = null;
|
|
self.driver = null;
|
|
|
|
self.browserName = options.browserName;
|
|
self.browserVersion = options.browserVersion;
|
|
|
|
self.name = "BrowserStack - " + self.browserName;
|
|
if (self.browserVersion) {
|
|
self.name += " " + self.browserVersion;
|
|
}
|
|
};
|
|
|
|
util.inherits(BrowserStackClient, Client);
|
|
|
|
_.extend(BrowserStackClient.prototype, {
|
|
connect: function () {
|
|
var self = this;
|
|
|
|
// memoize the key
|
|
if (browserStackKey === null)
|
|
browserStackKey = self._getBrowserStackKey();
|
|
if (! browserStackKey)
|
|
throw new Error("BrowserStack key not found. Ensure that you " +
|
|
"have installed your S3 credentials.");
|
|
|
|
var capabilities = {
|
|
'browserName' : self.browserName,
|
|
'browserstack.user' : 'meteor',
|
|
'browserstack.local' : 'true',
|
|
'browserstack.key' : browserStackKey
|
|
};
|
|
|
|
if (self.browserVersion) {
|
|
capabilities.browserVersion = self.browserVersion;
|
|
}
|
|
|
|
self._launchBrowserStackTunnel(function (error) {
|
|
if (error)
|
|
throw error;
|
|
|
|
self.driver = new webdriver.Builder().
|
|
usingServer('http://hub.browserstack.com/wd/hub').
|
|
withCapabilities(capabilities).
|
|
build();
|
|
self.driver.get(self.url);
|
|
});
|
|
},
|
|
|
|
stop: function() {
|
|
var self = this;
|
|
self.tunnelProcess && self.tunnelProcess.kill();
|
|
self.tunnelProcess = null;
|
|
|
|
self.driver && self.driver.quit();
|
|
self.driver = null;
|
|
},
|
|
|
|
_getBrowserStackKey: function () {
|
|
var outputDir = path.join(files.mkdtemp(), "key");
|
|
|
|
try {
|
|
execFileSync("s3cmd", ["get",
|
|
"s3://meteor-browserstack-keys/browserstack-key",
|
|
outputDir
|
|
]);
|
|
|
|
return fs.readFileSync(outputDir, "utf8").trim();
|
|
} catch (e) {
|
|
return null;
|
|
}
|
|
},
|
|
|
|
_launchBrowserStackTunnel: function (callback) {
|
|
var self = this;
|
|
var browserStackPath =
|
|
path.join(files.getDevBundle(), 'bin', 'BrowserStackLocal');
|
|
fs.chmodSync(browserStackPath, 0755);
|
|
|
|
var args = [
|
|
browserStackPath,
|
|
browserStackKey,
|
|
[self.host, self.port, 0].join(','),
|
|
// Disable Live Testing and Screenshots, just test with Automate.
|
|
'-onlyAutomate',
|
|
// Do not wait for the server to be ready to spawn the process.
|
|
'-skipCheck'
|
|
];
|
|
self.tunnelProcess = child_process.execFile(
|
|
'/bin/bash',
|
|
['-c', args.join(' ')]
|
|
);
|
|
|
|
// Called when the SSH tunnel is established.
|
|
self.tunnelProcess.stdout.on('data', function(data) {
|
|
if (data.toString().match(/You can now access your local server/))
|
|
callback();
|
|
});
|
|
}
|
|
});
|
|
|
|
///////////////////////////////////////////////////////////////////////////////
|
|
// Run
|
|
///////////////////////////////////////////////////////////////////////////////
|
|
|
|
// Represents a test run of the tool. Typically created through the
|
|
// run() method on Sandbox, but can also be created directly, say if
|
|
// you want to do something other than invoke the 'meteor' command in
|
|
// a nice sandbox.
|
|
//
|
|
// Options: args, cwd, env
|
|
var Run = function (execPath, options) {
|
|
var self = this;
|
|
|
|
self.execPath = execPath;
|
|
self.cwd = options.cwd || process.cwd();
|
|
self.env = options.env || {};
|
|
self._args = [];
|
|
self.proc = null;
|
|
self.baseTimeout = 20;
|
|
self.extraTime = 0;
|
|
self.client = options.client;
|
|
|
|
self.stdoutMatcher = new Matcher(self);
|
|
self.stderrMatcher = new Matcher(self);
|
|
self.outputLog = new OutputLog(self);
|
|
|
|
self.exitStatus = undefined; // 'null' means failed rather than exited
|
|
self.exitFutures = [];
|
|
|
|
var opts = options.args || [];
|
|
self.args.apply(self, opts || []);
|
|
|
|
self.fakeMongoPort = null;
|
|
self.fakeMongoConnection = null;
|
|
if (options.fakeMongo) {
|
|
self.fakeMongoPort = require('./utils.js').randomPort();
|
|
self.env.METEOR_TEST_FAKE_MONGOD_CONTROL_PORT = self.fakeMongoPort;
|
|
}
|
|
|
|
runningTest.onCleanup(function () {
|
|
self._stopWithoutWaiting();
|
|
});
|
|
};
|
|
|
|
_.extend(Run.prototype, {
|
|
// Set command-line arguments. This may be called multiple times as
|
|
// long as the run has not yet started (the run starts after the
|
|
// first call to a function that requires it, like match()).
|
|
//
|
|
// Pass as many arguments as you want. Non-object values will be
|
|
// cast to string, and object values will be treated as maps from
|
|
// option names to values.
|
|
args: function (/* arguments */) {
|
|
var self = this;
|
|
|
|
if (self.proc)
|
|
throw new Error("already started?");
|
|
|
|
_.each(_.toArray(arguments), function (a) {
|
|
if (typeof a !== "object") {
|
|
self._args.push('' + a);
|
|
} else {
|
|
_.each(a, function (value, key) {
|
|
self._args.push("--" + key);
|
|
self._args.push('' + value);
|
|
});
|
|
}
|
|
});
|
|
},
|
|
|
|
connectClient: function () {
|
|
var self = this;
|
|
if (! self.client)
|
|
throw new Error("Must create Run with a client to use connectClient().");
|
|
|
|
self._ensureStarted();
|
|
self.client.connect();
|
|
},
|
|
|
|
_exited: function (status) {
|
|
var self = this;
|
|
|
|
if (self.exitStatus !== undefined)
|
|
throw new Error("already exited?");
|
|
|
|
self.client && self.client.stop();
|
|
|
|
self.exitStatus = status;
|
|
var exitFutures = self.exitFutures;
|
|
self.exitFutures = null;
|
|
_.each(exitFutures, function (f) {
|
|
f['return']();
|
|
});
|
|
|
|
self.stdoutMatcher.end();
|
|
self.stderrMatcher.end();
|
|
},
|
|
|
|
_ensureStarted: function () {
|
|
var self = this;
|
|
|
|
if (self.proc)
|
|
return;
|
|
|
|
var env = _.clone(process.env);
|
|
_.extend(env, self.env);
|
|
|
|
var child_process = require('child_process');
|
|
self.proc = child_process.spawn(self.execPath, self._args, {
|
|
cwd: self.cwd,
|
|
env: env
|
|
});
|
|
|
|
self.proc.on('close', function (code, signal) {
|
|
if (self.exitStatus === undefined)
|
|
self._exited({ code: code, signal: signal });
|
|
});
|
|
|
|
self.proc.on('error', function (err) {
|
|
if (self.exitStatus === undefined)
|
|
self._exited(null);
|
|
});
|
|
|
|
self.proc.stdout.setEncoding('utf8');
|
|
self.proc.stdout.on('data', function (data) {
|
|
self.outputLog.write('stdout', data);
|
|
self.stdoutMatcher.write(data);
|
|
});
|
|
|
|
self.proc.stderr.setEncoding('utf8');
|
|
self.proc.stderr.on('data', function (data) {
|
|
self.outputLog.write('stderr', data);
|
|
self.stderrMatcher.write(data);
|
|
});
|
|
},
|
|
|
|
// Wait until we get text on stdout that matches 'pattern', which
|
|
// may be a regular expression or a string. Consume stdout up to
|
|
// that point. If this pattern does not appear after a timeout (or
|
|
// the program exits before emitting the pattern), fail.
|
|
match: markStack(function (pattern, _strict) {
|
|
var self = this;
|
|
self._ensureStarted();
|
|
|
|
var timeout = self.baseTimeout + self.extraTime;
|
|
timeout *= utils.timeoutScaleFactor;
|
|
self.extraTime = 0;
|
|
return self.stdoutMatcher.match(pattern, timeout, _strict);
|
|
}),
|
|
|
|
// As expect(), but for stderr instead of stdout.
|
|
matchErr: markStack(function (pattern, _strict) {
|
|
var self = this;
|
|
self._ensureStarted();
|
|
|
|
var timeout = self.baseTimeout + self.extraTime;
|
|
timeout *= utils.timeoutScaleFactor;
|
|
self.extraTime = 0;
|
|
return self.stderrMatcher.match(pattern, timeout, _strict);
|
|
}),
|
|
|
|
// Like match(), but won't skip ahead looking for a match. It must
|
|
// follow immediately after the last thing we matched or read.
|
|
read: markStack(function (pattern) {
|
|
return this.match(pattern, true);
|
|
}),
|
|
|
|
// As read(), but for stderr instead of stdout.
|
|
readErr: markStack(function (pattern) {
|
|
return this.matchErr(pattern, true);
|
|
}),
|
|
|
|
// Assert that 'pattern' (again, a regexp or string) has not
|
|
// occurred on stdout at any point so far in this run. Currently
|
|
// this works on complete lines, so unlike match() and read(),
|
|
// 'pattern' cannot span multiple lines, and furthermore if it is
|
|
// called before the end of the program, it may not see text on a
|
|
// partially read line. We could lift these restrictions easily, but
|
|
// there may not be any benefit since the usual way to use this is
|
|
// to call it after expectExit or expectEnd.
|
|
forbid: markStack(function (pattern) {
|
|
this._ensureStarted();
|
|
this.outputLog.forbid(pattern, 'stdout');
|
|
}),
|
|
|
|
// As forbid(), but for stderr instead of stdout.
|
|
forbidErr: markStack(function (pattern) {
|
|
this._ensureStarted();
|
|
this.outputLog.forbid(pattern, 'stderr');
|
|
}),
|
|
|
|
// Combination of forbid() and forbidErr(). Forbids the pattern on
|
|
// both stdout and stderr.
|
|
forbidAll: markStack(function (pattern) {
|
|
this._ensureStarted();
|
|
this.outputLog.forbid(pattern);
|
|
}),
|
|
|
|
// Expect the program to exit without anything further being
|
|
// printed on either stdout or stderr.
|
|
expectEnd: markStack(function () {
|
|
var self = this;
|
|
self._ensureStarted();
|
|
|
|
var timeout = self.baseTimeout + self.extraTime;
|
|
timeout *= utils.timeoutScaleFactor;
|
|
self.extraTime = 0;
|
|
self.expectExit();
|
|
|
|
self.stdoutMatcher.matchEmpty();
|
|
self.stderrMatcher.matchEmpty();
|
|
}),
|
|
|
|
// Expect the program to exit with the given (numeric) exit
|
|
// status. Fail if the process exits with a different code, or if
|
|
// the process does not exit after a timeout. You can also omit the
|
|
// argument to simply wait for the program to exit.
|
|
expectExit: markStack(function (code) {
|
|
var self = this;
|
|
self._ensureStarted();
|
|
|
|
if (self.exitStatus === undefined) {
|
|
var timeout = self.baseTimeout + self.extraTime;
|
|
timeout *= utils.timeoutScaleFactor;
|
|
self.extraTime = 0;
|
|
|
|
var fut = new Future;
|
|
self.exitFutures.push(fut);
|
|
var timer = setTimeout(function () {
|
|
self.exitFutures = _.without(self.exitFutures, fut);
|
|
fut['throw'](new TestFailure('exit-timeout', { run: self }));
|
|
}, timeout * 1000);
|
|
|
|
try {
|
|
fut.wait();
|
|
} finally {
|
|
clearTimeout(timer);
|
|
}
|
|
}
|
|
|
|
if (! self.exitStatus)
|
|
throw new TestFailure('spawn-failure', { run: self });
|
|
if (code !== undefined && self.exitStatus.code !== code) {
|
|
throw new TestFailure('wrong-exit-code', {
|
|
expected: { code: code },
|
|
actual: self.exitStatus,
|
|
run: self
|
|
});
|
|
}
|
|
}),
|
|
|
|
// Extend the timeout for the next operation by 'secs' seconds.
|
|
waitSecs: function (secs) {
|
|
var self = this;
|
|
self.extraTime += secs;
|
|
},
|
|
|
|
// Send 'string' to the program on its stdin.
|
|
write: function (string) {
|
|
var self = this;
|
|
self._ensureStarted();
|
|
self.proc.stdin.write(string);
|
|
},
|
|
|
|
// Kill the program and then wait for it to actually exit.
|
|
stop: markStack(function () {
|
|
var self = this;
|
|
if (self.exitStatus === undefined) {
|
|
self._ensureStarted();
|
|
self.client && self.client.stop();
|
|
self.proc.kill();
|
|
self.expectExit();
|
|
}
|
|
}),
|
|
|
|
// Like stop, but doesn't wait for it to exit.
|
|
_stopWithoutWaiting: function () {
|
|
var self = this;
|
|
if (self.exitStatus === undefined && self.proc) {
|
|
self.client && self.client.stop();
|
|
self.proc.kill();
|
|
}
|
|
},
|
|
|
|
// If the fakeMongo option was set, sent a command to the stub
|
|
// mongod. Available commands currently are:
|
|
//
|
|
// - { stdout: "xyz" } to make fake-mongod write "xyz" to stdout
|
|
// - { stderr: "xyz" } likewise for stderr
|
|
// - { exit: 123 } to make fake-mongod exit with code 123
|
|
//
|
|
// Blocks until a connection to fake-mongod can be
|
|
// established. Throws a TestFailure if it cannot be established.
|
|
tellMongo: markStack(function (command) {
|
|
var self = this;
|
|
|
|
if (! self.fakeMongoPort)
|
|
throw new Error("fakeMongo option on sandbox must be set");
|
|
|
|
self._ensureStarted();
|
|
|
|
// If it's the first time we've called tellMongo on this sandbox,
|
|
// open a connection to fake-mongod. Wait up to 2 seconds for it
|
|
// to accept the connection, retrying every 100ms.
|
|
//
|
|
// XXX we never clean up this connection. Hopefully once
|
|
// fake-mongod has dropped its end of the connection, and we hold
|
|
// no reference to our end, it will get gc'd. If not, that's not
|
|
// great, but it probably doesn't actually create any practical
|
|
// problems since this is only for testing.
|
|
if (! self.fakeMongoConnection) {
|
|
var net = require('net');
|
|
|
|
var lastStartTime = 0;
|
|
for (var attempts = 0; ! self.fakeMongoConnection && attempts < 50;
|
|
attempts ++) {
|
|
// Throttle attempts to one every 100ms
|
|
utils.sleepMs((lastStartTime + 100) - (+ new Date));
|
|
lastStartTime = +(new Date);
|
|
|
|
// Use an anonymous function so that each iteration of the
|
|
// loop gets its own values of 'fut' and 'conn'.
|
|
(function () {
|
|
var fut = new Future;
|
|
var conn = net.connect(self.fakeMongoPort, function () {
|
|
if (fut)
|
|
fut['return'](true);
|
|
});
|
|
conn.setNoDelay();
|
|
conn.on('error', function () {
|
|
if (fut)
|
|
fut['return'](false);
|
|
});
|
|
setTimeout(function () {
|
|
if (fut)
|
|
fut['return'](false); // 100ms connection timeout
|
|
}, 100);
|
|
|
|
// This is all arranged so that if a previous attempt
|
|
// belatedly succeeds, somehow, we ignore it.
|
|
if (fut.wait())
|
|
self.fakeMongoConnection = conn;
|
|
fut = null;
|
|
})();
|
|
}
|
|
|
|
if (! self.fakeMongoConnection)
|
|
throw new TestFailure("mongo-not-running", { run: self });
|
|
}
|
|
|
|
self.fakeMongoConnection.write(JSON.stringify(command) + "\n");
|
|
// If we told it to exit, then we should close our end and connect again if
|
|
// asked to send more.
|
|
if (command.exit) {
|
|
self.fakeMongoConnection.end();
|
|
self.fakeMongoConnection = null;
|
|
}
|
|
})
|
|
});
|
|
|
|
|
|
///////////////////////////////////////////////////////////////////////////////
|
|
// Defining tests
|
|
///////////////////////////////////////////////////////////////////////////////
|
|
|
|
var Test = function (options) {
|
|
var self = this;
|
|
self.name = options.name;
|
|
self.file = options.file;
|
|
self.fileHash = options.fileHash;
|
|
self.tags = options.tags || {};
|
|
self.f = options.func;
|
|
self.cleanupHandlers = [];
|
|
};
|
|
|
|
_.extend(Test.prototype, {
|
|
onCleanup: function (cleanupHandler) {
|
|
this.cleanupHandlers.push(cleanupHandler);
|
|
},
|
|
cleanup: function () {
|
|
var self = this;
|
|
_.each(self.cleanupHandlers, function (cleanupHandler) {
|
|
cleanupHandler();
|
|
});
|
|
self.cleanupHandlers = [];
|
|
}
|
|
});
|
|
|
|
var allTests = null;
|
|
var fileBeingLoaded = null;
|
|
var fileBeingLoadedHash = null;
|
|
var runningTest = null;
|
|
var getAllTests = function () {
|
|
if (allTests)
|
|
return allTests;
|
|
allTests = [];
|
|
|
|
// Load all files in the 'tests' directory that end in .js. They
|
|
// are supposed to then call define() to register their tests.
|
|
var testdir = path.join(__dirname, 'tests');
|
|
var filenames = fs.readdirSync(testdir);
|
|
_.each(filenames, function (n) {
|
|
if (! n.match(/^[^.].*\.js$/)) // ends in '.js', doesn't start with '.'
|
|
return;
|
|
try {
|
|
if (fileBeingLoaded)
|
|
throw new Error("called recursively?");
|
|
fileBeingLoaded = path.basename(n, '.js');
|
|
|
|
var fullPath = path.join(testdir, n);
|
|
var contents = fs.readFileSync(fullPath, 'utf8');
|
|
fileBeingLoadedHash =
|
|
require('crypto').createHash('sha1').update(contents).digest('hex');
|
|
|
|
require(path.join(testdir, n));
|
|
} finally {
|
|
fileBeingLoaded = null;
|
|
fileBeingLoadedHash = null;
|
|
}
|
|
});
|
|
|
|
return allTests;
|
|
};
|
|
|
|
var define = function (name, tagsList, f) {
|
|
if (typeof tagsList === "function") {
|
|
// tagsList is optional
|
|
f = tagsList;
|
|
tagsList = [];
|
|
}
|
|
|
|
var tags = {};
|
|
_.each(tagsList, function (tag) {
|
|
tags[tag] = true;
|
|
});
|
|
|
|
allTests.push(new Test({
|
|
name: name,
|
|
tags: tags,
|
|
file: fileBeingLoaded,
|
|
fileHash: fileBeingLoadedHash,
|
|
func: f
|
|
}));
|
|
};
|
|
|
|
|
|
///////////////////////////////////////////////////////////////////////////////
|
|
// Running tests
|
|
///////////////////////////////////////////////////////////////////////////////
|
|
|
|
var tagDescriptions = {
|
|
checkout: 'can only run from checkouts',
|
|
net: 'require an internet connection',
|
|
slow: 'take quite a long time',
|
|
// these last two are not actually test tags; they reflect the use of
|
|
// --changed and --tests
|
|
unchanged: 'unchanged since last pass',
|
|
'non-matching': "don't match specified pattern"
|
|
};
|
|
|
|
// options: onlyChanged, offline, includeSlowTests, historyLines, testRegexp
|
|
// clients:
|
|
// - browserstack (need s3cmd credentials)
|
|
var runTests = function (options) {
|
|
var failureCount = 0;
|
|
|
|
var tests = getAllTests();
|
|
|
|
if (! tests.length) {
|
|
process.stderr.write("No tests defined.\n");
|
|
return 0;
|
|
}
|
|
|
|
var testStateFile = path.join(process.env.HOME, '.meteortest');
|
|
var testState;
|
|
if (fs.existsSync(testStateFile))
|
|
testState = JSON.parse(fs.readFileSync(testStateFile, 'utf8'));
|
|
if (! testState || testState.version !== 1)
|
|
testState = { version: 1, lastPassedHashes: {} };
|
|
var currentHashes = {};
|
|
|
|
// _.keys(skipCounts) is the set of tags to skip
|
|
var skipCounts = {};
|
|
if (! files.inCheckout())
|
|
skipCounts['checkout'] = 0;
|
|
|
|
if (options.offline)
|
|
skipCounts['net'] = 0;
|
|
|
|
if (! options.includeSlowTests)
|
|
skipCounts['slow'] = 0;
|
|
|
|
if (options.testRegexp) {
|
|
var lengthBeforeTestRegexp = tests.length;
|
|
// Filter out tests whose name doesn't match.
|
|
tests = _.filter(tests, function (test) {
|
|
return options.testRegexp.test(test.name);
|
|
});
|
|
skipCounts['non-matching'] = lengthBeforeTestRegexp - tests.length;
|
|
}
|
|
|
|
if (options.onlyChanged) {
|
|
var lengthBeforeOnlyChanged = tests.length;
|
|
// Filter out tests that haven't changed since they last passed.
|
|
tests = _.filter(tests, function (test) {
|
|
return test.fileHash !== testState.lastPassedHashes[test.file];
|
|
});
|
|
skipCounts.unchanged = lengthBeforeOnlyChanged - tests.length;
|
|
}
|
|
|
|
var failuresInFile = {};
|
|
var skipsInFile = {};
|
|
var totalRun = 0;
|
|
_.each(tests, function (test) {
|
|
currentHashes[test.file] = test.fileHash;
|
|
// Is this a test we're supposed to skip?
|
|
var shouldSkip = false;
|
|
_.each(_.keys(test.tags), function (tag) {
|
|
if (_.has(skipCounts, tag)) {
|
|
shouldSkip = true;
|
|
skipCounts[tag] ++;
|
|
}
|
|
});
|
|
if (shouldSkip) {
|
|
skipsInFile[test.file] = true;
|
|
return;
|
|
}
|
|
|
|
totalRun++;
|
|
process.stderr.write(test.name + "... ");
|
|
|
|
var failure = null;
|
|
try {
|
|
runningTest = test;
|
|
test.f(options);
|
|
} catch (e) {
|
|
if (e instanceof TestFailure) {
|
|
failure = e;
|
|
} else {
|
|
process.stderr.write("exception\n\n");
|
|
throw e;
|
|
}
|
|
} finally {
|
|
runningTest = null;
|
|
test.cleanup();
|
|
}
|
|
|
|
if (failure) {
|
|
process.stderr.write("fail!\n");
|
|
failureCount++;
|
|
var frames = parseStack.parse(failure);
|
|
var relpath = path.relative(files.getCurrentToolsDir(),
|
|
frames[0].file);
|
|
process.stderr.write(" => " + failure.reason + " at " +
|
|
relpath + ":" + frames[0].line + "\n");
|
|
if (failure.reason === 'no-match') {
|
|
}
|
|
if (failure.reason === "wrong-exit-code") {
|
|
var s = function (status) {
|
|
return status.signal || ('' + status.code) || "???";
|
|
};
|
|
|
|
process.stderr.write(" => Expected: " + s(failure.details.expected) +
|
|
"; actual: " + s(failure.details.actual) + "\n");
|
|
}
|
|
if (failure.reason === 'expected-exception') {
|
|
}
|
|
if (failure.reason === 'not-equal') {
|
|
process.stderr.write(
|
|
" => Expected: " + JSON.stringify(failure.details.expected) +
|
|
"; actual: " + JSON.stringify(failure.details.actual) + "\n");
|
|
}
|
|
|
|
if (failure.details.run) {
|
|
failure.details.run.outputLog.end();
|
|
var lines = failure.details.run.outputLog.get();
|
|
if (! lines.length) {
|
|
process.stderr.write(" => No output\n");
|
|
} else {
|
|
var historyLines = options.historyLines || 100;
|
|
|
|
process.stderr.write(" => Last " + historyLines + " lines:\n");
|
|
_.each(lines.slice(-historyLines), function (line) {
|
|
process.stderr.write(" " +
|
|
(line.channel === "stderr" ? "2| " : "1| ") +
|
|
line.text +
|
|
(line.bare ? "%" : "") + "\n");
|
|
});
|
|
}
|
|
}
|
|
|
|
if (failure.details.messages) {
|
|
process.stderr.write(" => Errors while building:\n");
|
|
process.stderr.write(failure.details.messages.formatMessages());
|
|
}
|
|
|
|
failuresInFile[test.file] = true;
|
|
} else {
|
|
process.stderr.write("ok\n");
|
|
}
|
|
});
|
|
|
|
_.each(_.keys(currentHashes), function (f) {
|
|
if (failuresInFile[f])
|
|
delete testState.lastPassedHashes[f];
|
|
else if (! skipsInFile[f])
|
|
testState.lastPassedHashes[f] = currentHashes[f];
|
|
});
|
|
|
|
if (tests.length)
|
|
fs.writeFileSync(testStateFile, JSON.stringify(testState), 'utf8');
|
|
|
|
if (totalRun > 0)
|
|
process.stderr.write("\n");
|
|
|
|
var totalSkipCount = 0;
|
|
_.each(skipCounts, function (count, tag) {
|
|
totalSkipCount += count;
|
|
if (count) {
|
|
process.stderr.write("Skipped " + count + " " + tag + " test" +
|
|
(count > 1 ? "s" : "") + " (" +
|
|
tagDescriptions[tag] + ")\n");
|
|
}
|
|
});
|
|
|
|
if (tests.length === 0) {
|
|
process.stderr.write("No tests run.\n");
|
|
return 0;
|
|
} else if (failureCount === 0) {
|
|
var disclaimers = '';
|
|
if (totalSkipCount > 0)
|
|
disclaimers += " other";
|
|
process.stderr.write("All" + disclaimers + " tests passed.\n");
|
|
return 0;
|
|
} else {
|
|
process.stderr.write(failureCount + " failure" +
|
|
(failureCount > 1 ? "s" : "") + ".\n");
|
|
return 1;
|
|
}
|
|
};
|
|
|
|
// To create self-tests:
|
|
//
|
|
// Create a new .js file in the tests directory. It will be picked
|
|
// up automatically.
|
|
//
|
|
// Start your file with something like:
|
|
// var selftest = require('../selftest.js');
|
|
// var Sandbox = selftest.Sandbox;
|
|
//
|
|
// Define tests with:
|
|
// selftest.define("test-name", ['tag1', 'tag2'], function () {
|
|
// ...
|
|
// });
|
|
//
|
|
// The tags are used to group tests. Currently used tags:
|
|
// - 'checkout': should only be run when we're running from a
|
|
// checkout as opposed to a released copy.
|
|
// - 'net': test requires an internet connection. Not going to work
|
|
// if you're on a plane; will be skipped if we appear to be
|
|
// offline unless run with 'self-test --force-online'.
|
|
// - 'slow': test is slow enough that you don't want to run it
|
|
// except on purpose. Won't run unless you say 'self-test --slow'.
|
|
//
|
|
// If you don't want to set any tags, you can omit that parameter
|
|
// entirely.
|
|
//
|
|
// Inside your test function, first create a Sandbox object, then call
|
|
// the run() method on the sandbox to set up a new run of meteor with
|
|
// arguments of your choice, and then use functions like match(),
|
|
// write(), and expectExit() to script that run.
|
|
|
|
_.extend(exports, {
|
|
runTests: runTests,
|
|
markStack: markStack,
|
|
define: define,
|
|
Sandbox: Sandbox,
|
|
Run: Run,
|
|
fail: fail,
|
|
expectEqual: expectEqual,
|
|
expectThrows: expectThrows,
|
|
getToolsPackage: getToolsPackage,
|
|
execFileSync: execFileSync,
|
|
doOrThrow: doOrThrow,
|
|
testPackageServerUrl: config.getTestPackageServerUrl()
|
|
});
|