WARNING: this repository and the information is outdated. The research was done in July of 2013. This repository is no longer maintained.
This document represents some research done in order to find the best solution in client-side application testing, while meeting a set of requirements.
Many of the things described are implemented in this baseplate project.
Please feel free to comment e.g. by opening tickets, or comment on this README.
Some definitions of test type used here:
This document focuses on unit and behavior testing for client-side single-page applications (SPA), leaving functional, integration and acceptance testing out of scope.
Test frameworks provide the interface to write tests, run the tests, and report results.
These frameworks support BDD style tests.
Assertion libraries provide the interface to make "assertions" (in TDD). They're called "expectations" in BDD.
Mocha leaves this open, and Jasmine ships with Matchers (expect()
style methods). All libraries can be extended with custom assertions/matchers.
Note that Chai is a more popular choice and has more features, but does not support oldIE (while expect.js supports IE6+).
Since unit testing means testing isolated and dependency-free units, dependencies need to be mocked or stubbed out (i.e. "test doubles"). For instance, method calls to modules the unit depends on, and XHR calls to services.
Test runners drive the test frameworks. Currently, popular solutions include:
Buster is a great effort and looks very interesting.
Configuration requires to load all files that might be needed. E.g. sources: ["src/**/*.js", "lib/**/*.js"]
. This might include a lot of unnecessary files (libraries may come with a large number of files that are not needed to use it), so it's not great performance wise. But the actual issue is that some matching .js
files are Node executables, and thus start with a hash-bang (i.e. #!/usr/bin/env node
), causing a fatal error for Buster ([Fatal] Syntax error
).
Filed issue #362.
The project has an explicit "beta" stamp on it. Project activity seems pretty low (Community, Twitter, Issue Tracker).
Overall, this doesn't seem the right time to jump on this train (especially if things don't work straight away).
Of course I would love to see things working!
Intern is the new kid on the block, and it's quite impressive and complete. Many features come out of the box or are in the works, including Sauce Labs and Travis CI integration, functional testing (WebDriver API), and various code coverage reporters (istanbul).
Karma is a testacular effort from the AngularJS team, powered by Node and Socket.IO. It is fairly easy to set up. Configuration may result in some headaches for AMD setups (you want to reuse existing application and test configuration files). However, once it's done it doesn't need to be looked after anymore.
Karma comes with a decent set of reporters, and custom reporters can be plugged in.
For code coverage, istanbul is included (they pair up great). This is easy to configure with various output formats (including html
, lcov
, and cobertura
).
Mocha is a test runner built with Node. It can run tests from CLI, but does not steer browsers.
However, it could still complement a setup where Mocha is the client test framework, and the test runner doesn't support running tests in Node. It could be complementary, if the Mocha test runner would support AMD. And this could be solved by using PhantomJS, if Mocha didn't have this issue since v1.10. However, things work using mocha-phantomjs and Mocha v1.9.
Mocha supports code coverage output out of the box. It can also be configured to work with:
Testem is built with Node (read more about it in Testem: Interactive JS Test Runner).
While configuration is not without hurdles in the other test runners, Testem has an advantage by the ability to hook into a custom test page. When a page is configured for testing (i.e. with Jasmine or Mocha), that page needs two extra lines of code (to inject testem.js
). Then, if Testem is driving the show, it automatically detects the framework in the page, and hooks into it (i.e. with a custom reporter for Jasmine, or monkey-patching Mocha's Runner.prototype.emit
).
In short, if Jasmine or Mocha runs, Testem runs.
Additionally, a big plus of this strategy is that it's easy to configure functional testing (see below).
A downside of Testem is that it doesn't have support for code coverage (at least not built-in). Since Testem hooks into the client side frameworks, those can be configured to use coverage reporters (but after generating coverage files from sources). One working effort includes Testem Coverage Sandbox.
Based on the given definition ("verify correct behavior of isolated functions based on user interaction, while faking events etc. in JavaScript"), the challenge is to initialize a function of the application, and then simulate user interaction with it. Examples include:
Things are actually quite feasible, even simple, when extending the workflow and tools for unit testing to functional tests. What is needed on top of this:
There are a couple of caveats when programmatically simulating user interactions:
$('input[name=firstName]').val('John')
) won't trigger a "change" event. This can be simulated by appending .trigger('change')
, or use something like jQuery.autotype (untested).It's actually trivial to set this up using Testem, Jasmine (or Mocha) and jQuery. Driven by Testem, the functional tests can easily be ran cross-browser and/or quickly using only PhantomJS. Here is how it might look like in a working example.
It should be feasible to do this with Karma as well (untested).
Features | BusterJS | Intern | Mocha (1) | Karma | Testem |
---|---|---|---|---|---|
AMD | Y (2) | Y | N (3) | Y (2) | Y |
BDD | Y | Y | Y | Y | Y |
Jasmine | N (3) | N | N (2) | Y (2) | Y |
Mocha | N | N | Y | Y (2) | Y |
CLI: Node | Y | Y | N (4) | ||
CLI: PhantomJS | Y | Y (2) | N (3) | Y | Y |
CLI: Run tests in real browser | Y | Y | N | Y | Y |
CLI: Open/close browser | N | Y (5) | N | Y | Y |
Code coverage | Y | Y | N | ||
Behavior Tests | Y | N | Y | Y | |
Functional Tests | N | Y | N | N |
Running a single test, including opening and closing 4 different browsers is fast in Karma and Testem. Testem has the option to run browsers in parallel, while Karma seems to do this automatically. The (same) test is run using minimal, default settings; Jasmine; and in Chrome, Chrome Canary, Firefox and PhantomJS:
Command | Time (real) |
---|---|
time testem ci -P 4 |
2.8s |
time karma start --single-run |
3.8s |
Not sure how things stack up on other systems, in other browsers and with more tests.
For functional testing, when the browser itself needs to be driven by the tests (e.g. for page navigation, file uploads), solutions like the following might be interesting to check out:
Note: The WebDriver API is a W3C standard (which in turn is based on Selenium WebDriver).
Source code can be written in CommonJS, and be wrapped in the AMD transporter format for the browser. In doing so, some things around testing would probably get easier. Especially configuration and running stuff in Node (think Mocha) come to mind. The Mocha test runner drives CommonJS-style modules/specs. Interesting if real browser testing is out of scope, and e.g. speed and code coverage reporting options are more important.
Switching to CommonJS source and spec files doesn't seem of much benefit to Karma and Testem (apart from ease of configuration). Both Karma and Testem can pre-process source files if needed (i.e. convert to AMD).
Please, draw your own! YMMV.