Closed dervism closed 5 years ago
Those three tools have completely different rulesets with different priorities. axe-core focuses on accessibility support and no false positives; I can't speak to the priorities for the others. It is expected that there would be differences, as they are developed by totally different teams.
To read more about how we make decisions on accessibility support, here's a blog post: https://www.deque.com/blog/weve-got-your-back-with-accessibility-supported-in-axe/
@dervism The simplest answer to your question of: "Why do tools produce different results", is that there is no standard for accessibility tools. All tools behave different in different situations.
One thing I can tell you is that, I don't know how you produced that report, but it doesn't have a correct implementation of axe-core. The frames-tested rule will only return a warning when axe-core isn't properly executing. This means none of the content in those two frames was tested for accessibility. If you're interested, you can find a list of axe-core implementations here: https://github.com/dequelabs/axe-core/blob/develop/doc/projects.md
Lastly, there is work under way to actually standardise how automated accessibility tools work. You can find information about this here: https://www.w3.org/community/auto-wcag/
This might be just a question, not necessarily a bug. I tested Total Validator and Pa11y and compared their output with axe-core. I ran a test against a Norwegian news website: http://www.vg.no
Why is there such a big difference in the number of detected violations between these tools? According to the AlphaGov audit, HTML CodeSniffer (used by Pa11y), is finding a lower amount of errors compared to axe. But this tests shows the opposite.
Could you please help me understand the difference in the numbers? See output from each tool below.
Total Validator:
axe
Pa11y