The benefit of the ubiquity of the Web is also its pain point when it comes to accurate testing. It is common to see handwavy statements for browser compatibility, for example "IE9+", as if every other OS/platform combination will be just fine if IE9 "works"! For those who need more realistic data, this module makes it much easier to test code earlier on in the development lifecycle, even in realtime.
Snapshot of Test Results for Ripple v0.3 on latest Chrome, Firefox, IE, Android and iOS
.popper.yml
file for persisting test config per reponpm test
run popper
which will return exit cleanly if the tests pass in all defined browsersPOPPER_TIMEOUT=milliseconds
to help investigate why a CI test is failiing (defaults to 20s)See roadmap issues label for upcoming features/idea.
# since this uses lots and lots of tiny libs, I recommend using npm3
npm i -g popper # install globally
npm i -D popper # install locally
# to run
popper
# to also see logs from each browser in terminal window
NODE_ENV=debug popper
# if you are using browserstack
export BROWSERSTACK_USERNAME=...
export BROWSERSTACK_KEY=...
Once running, open a browser tab to localhost:1945
(or the external ngrok URL) to run the tests, and keep open localhost:1945/dashboard
to see the results as you continue to make changes. If you specified any browsers
, they will be launched on BrowserStack and pointed to the test page.
When you run popper
in a folder:
popper.js
file, it will run that (example)popper.yml
file, it will use options from that (example).yml
configpopper
You can set all the following options using via the CLI, YAML or JS API:
usage: popper
options:
-b, --browsers: browser to spawn and run tests on, defaults to none
-t, --test: command to generate test bundle, defaults to "browserify test.js"
-p, --port: port to run on, defaults to 1945
-w, --watch: files to watch for changes, defaults to .
-n, --notunnel: disable opening tunnel, defaults to open
-l, --timeout: maximum time to wait in ci mode for results, defaults to POPPER_TIMEOUT or 20000
-r, --runner: the runner to use, either mocha or tape, defatuls to mocha
-f, --farm: the remote browser farm to spawn browsers in, defaults to browserstack
If any of the options are missing from the local YAML config or CLI arguments, they will default to:
none
none
browserify test.js
1945
.
false
mocha
process.env.POPPER_TIMEOUT || 20000
browserstack
# these will be added to the head
globals:
- <script src="https://cdn.polyfill.io/v1/polyfill.min.js"></script>
- <script src="https://cdnjs.cloudflare.com/ajax/libs/d3/3.5.5/d3.min.js" charset="utf-8"></script>
- <script src="https://cdnjs.cloudflare.com/ajax/libs/chai/3.0.0/chai.min.js"></script>
- <script src="https://cdnjs.cloudflare.com/ajax/libs/moment.js/2.10.3/moment.min.js"></script>
# this is the command to generate the tests bundle on startup and after a file change detected
tests: browserify ./node_modules/*/test.js
-i moment
-i colors
-i jsdom
-i chai
-i d3
-i ./node_modules/pause/test.js
-i ./node_modules/send/test.js
-i ./node_modules/file/test.js
-i ./node_modules/via/test.js
| sed -E "s/require\('moment'\)/window.moment/"
| sed -E "s/require\('chai'\)/window.chai/"
| sed -E "s/require\('d3'\)/window.d3/"
| uglifyjs
# browsers to spawn in browserstack/sauce
# can be wd capabilities object to specify os, device, version, etc: https://www.browserstack.com/automate/capabilities
browsers:
- ie9
- android
- iphone
- opera
- safari
# port to run on locally
port: 1945
# glob(s) to watch for file changes
watch: ./node_modules/*/index.js
In this case, the test command will rebuild the project before bundling the tests after each file change.
globals:
- <script src="https://cdn.polyfill.io/v1/polyfill.min.js"></script>
- <script src="https://cdnjs.cloudflare.com/ajax/libs/chai/3.0.0/chai.min.js"></script>
tests: (npm run build > /dev/null) && browserify ./test.js
-i colors
-i chai
| sed -E "s/require\('chai'\)/window.chai/"
| uglifyjs
watch:
- src
- test.js
popper = require('popper')
popper = popper({
watch: ['src', 'test']
, port: 19450
, tests: stream // function that returns stream to be piped to the test bundle file
, globals: string // string of global script tags to add
, browsers: array // array of browsers to spawn
})
Popper uses Ripple under the hood. The JS API is particularly useful if you need to extend the available resources. For example, for testing Ripple itself and it's server/client synchronisation module, I use the following to reset test resources before each test:
popper.io.on('connection', function(socket){
socket.on('beforeEach', function(){
popper('foo' , 'bar', headers())
popper('object' , { a:0 , b:1, c:2 }, headers())
popper('array' , [{i:0}, {i:1},{i:2}], headers())
popper({ name: 'proxy', body: [{i:0}, {i:1},{i:2}], headers: { to: to, from: from, 'cache-control': 'no-cache', silent: true, reactive: false }})
popper('my-component' , component, headers())
popper.sync(socket)()
socket.emit('done')
})
})
If you'd like to add a new remote browser farm, you just need to add a folder under /farms
and then you can set the farm
option (via JS, YAML or CLI) to the name of your farm. The folder should contain two things (see browserstack for an example):
browsers.json
: Mapping of shorthands (e.g. ie9
) to browser capability objects. Since there is different properties across services (e.g. browser
vs browserName
), you should also specify the properties _name
, _version
, _os
and _os_version
where applicable, which is what will show up in the popper logs.
connect
: A connect function, which will take in a wd
instance and using the relevant environment variables, connect to the service. If it failed to connect, it should return false.
If you'd like to add a new test runner, you just need to add a folder /client
and then you can set the runner
option (via JS, YAML or CLI) to the name of your runner. The folder should contain two things (see mocha for an example):
index.html
: This is the HTML file that will be run on the test agent. You will also need a script (see client.js
) that streams back the results of the test.
logs.html
: A HTML file that will be used from the dashboard (when you click "View Results") to format the read-only results of the tests run on a specific agent.
Thanks to BrowserStack and SauceLabs for providing a free open source account for testing.