A high-throughput tokenizer and parser (soon™️) for the Zig programming language.
So far, a tokenizer implementation is provided. The mainline Zig tokenizer uses a deterministic finite state machine. Those are pretty good for some applications, but tokenizing can often employ the use of other techniques for added speed.
Click here to see my latest work.
Currently the utf8 validator is turned off! I did a lot of performance optimization the past few days and did not finish porting my changes over yet.
The test bench fully reads in all of the Zig files under the folders in the src/files_to_parse
folder. In my test I installed the Zig compiler, ZLS, and a few other Zig projects in my src/files_to_parse
folder. The test bench iterates over the source bytes from each Zig file (with added sentinels) and calls the tokenization function on each with the utf8 validator turned off.
To tokenize 3,215 Zig files with 1,298,139 newlines, the original tokenizer and my new tokenizer have the following characteristics:
memory (megabytes) | |
---|---|
raw source files | 59.162811MB |
original (tokens) | 46.08376MB |
this (tokens) | 18.50587MB |
That's 2.49x less memory!
Please keep in mind that comparing to the legacy tokenizer's speed is not necessarily straightforward. It is not difficult for me to see the legacy tokenizer's performance change by ~15% by making a trivial change in my code. It varies heavily depending on the particular compile. That said, here are some numbers I am seeing on my machine (with the utf8 validator turned off on my implementation):
Currently the utf8 validator is turned off! I did a lot of performance optimization the past few days and did not finish porting my changes over yet.
run-time (milliseconds) | throughput (megabytes per second) | throughput (million lines of code per second) | |
---|---|---|---|
read files (baseline) | 35.269ms | 1677.45 MB/s | 35.01M loc/s |
original | 235.293ms | 251.44 MB/s | 5.52M loc/s |
this | 78.525ms | 753.42 MB/s | 16.53M loc/s |
That's ~3.00x faster! Currently the utf8 validator is turned off! I did a lot of performance optimization the past few days and did not finish porting my changes over yet.
Currently the utf8 validator is turned off! I did a lot of performance optimization the past few days and did not finish porting my changes over yet.
run-time (milliseconds) | throughput (megabytes per second) | throughput (million lines of code per second) | |
---|---|---|---|
read files (baseline) | 318.989ms | 185.47 MB/s | 4.07M loc/s |
original | 2.206s | 26.81 MB/s | 0.59M loc/s |
this | 894.963ms | 66.11 MB/s | 1.45M loc/s |
That's ~2.47x faster! Currently the utf8 validator is turned off! I did a lot of performance optimization the past few days and did not finish porting my changes over yet.
Oddly enough, I think some of this code is more maintainable too, as adding an operator or keyword to the tokenizer is literally just adding another string into the relevant array. All of the assumptions and tricks I use are explicitly checked for in compile-time assertions (grep
for comptime assert
), so violating any of those invariants will result in compile errors that tell you why you can't change certain things.
However, I do have a bunch of weird SWAR tricks that the compiler will hopefully perform automatically one day.
In the delicate balancing act that is performance optimization, you generally want:
I try to achieve each of these in the following ways:
SIMD, i.e. single-instruction, multiple data. Instead of operating on a single element at a time, you can operate on 16, 32, or 64 elements simultaneously. Instead of going character-by-character, we use SIMD to check for the length of identifiers/keywords, the length of quotes, the length of whitespace, and the length of comments or single-line quotes. This allows us to move quicker than one byte at a time. We also use a SIMD technique to validate proper utf8 conformance, ported from simdjson by travisstaloch for use in simdjzon. Please note that that particular code is licensed under the Apache license, included at the bottom of the main.zig
file.
I do not actually use SIMD to find "character literals" of the form 'a'
because these are generally extremely short and did not actually give much benefit in tests.
We can't and don't want to use SIMD for absolutely everything because:
SWAR, i.e., SIMD within a register. This is where we read multiple bytes into a 4 or 8 byte register and use conventional arithmetic and logical instructions to operate on multiple bytes simultaneously.
SWAR fallbacks are provided for machines which lack proper SIMD instructions.
0xaabbccddeeffgghh
^ 0xcccccccccccccccc
--------------------
0x~~~~00~~~~~~~~~~
cc
). We can then add a broadcasted 0x7F
.
0x~~~~00~~~~~~~~~~
0x8~8~7F8~8~8~8~8~
Example with 32 bit integers:
We want to concentrate the upper bits of each byte into a single nibble.
Doing the gradeschool multiplication algorithm, we can see that each 1 bit
in the bottom multiplicand shifts the upper multiplicand, and then we add all these
shifted bitstrings together. (Note `.` represents a 0)
a.......b.......c.......d.......
a.......b.......c.......d....... .b.......c.......d.............. ..c.......d.....................
abcd....bcd.....cd......d.......
Then we simply shift to the right by 32 - 4
(bitstring size minus the number of relevant
bits) to isolate the desired abcd
bits in the least significant byte!
Reducing unpredictable branches through:
Using SIMD/SWAR. Using a conventional while loop to capture a completely unpredictable number of characters in the aforementioned categories all but guarantees a branch mispredict every time we exit the loop, and possibly multiple throughout the loop if the branch predictor is having a bad day. Using SIMD/SWAR, we can instead produce a bitstring with 0's marked in the place corresponding to target characters like the matching "
, shift the bitstring according to our cursor's position, and count the trailing ones (the reason the bits are the inverse of what you might expect is because when we shift the bitstring it will be filled with 0's). In most cases, a single "count trailing one's" operation is all we need to find the position we are supposed to go to next. No need for a totally unpredictable while loop that goes character-by-character!
Using perfect hash functions. Specifically, keywords like var
and const
are mapped into a 7 bit address space by a perfect hash function. Identifiers can be checked against the list of keywords by applying the perfect hash function to each identifier and doing a table lookup to find what keyword it may match, then doing a single 16-byte vs 16-byte comparison to see if the identifier matches that keyword. The keywords are padded in memory to be 16 bytes and have a len
stored in the final byte so we can check that the incoming identifier has the same length as the prospective keyword. We also use Phil Bagwell's array-mapped trie compression technique, meaning we have a 128-bit bitmap and find which position to check using the bitmap, thereby enabling us to have a packed buffer that need not have 128 slots. We do a similar trick for operators.
I use a trick where I just allocate the upper-bound amount of memory for tokens per-file, and use the resize
facility of the allocator to reclaim the space I did not fill. The nice thing about this trick is I can always assume there is sufficient space, which eliminates the need to check that such a thing is safe.
I place sentinels at the end of files (and I place a newline at the front) to make the rest of the code simpler. This allows us to safely go back a character at any point if the perfect hash function wants us to grab the last two characters from an identifier with only one character, and allows us to safely go past the end of the source file as well. By placing "
and '
characters at the end of our buffer, we can eliminate bounds-checking in the code that searches for those characters, and simply check whether we hit the sentinel node after the hot loop finishes. We currently don't break out of these for newlines though, which we should probably do. All other validation for these should occur when actually trying to allocate the string or character they are supposed to represent.
Some things we do unconditionally that could be hidden behind a branch, but are very inexpensive so there is no point. Other things we hide behind a branch when it's expensive and generally predictable. E.g. utf8 validation is typically just making sure all bytes are less than 128, i.e. 0x80. Once we see some non-ascii sequences, then we have to do the more computationally expensive work of making sure the byte sequence is valid.
Table lookups. I consolidate the SIMD/SWAR code into one so that we go down the exact same codepaths to find how many non_newline/identifier/non_unescaped_quote/space characters to jump over. This is probably much more efficient than having 4 separate copies of the same hot loop.
Inlining the SIMD/SWAR loop, even on machines that need to unroll 8 times. This turns out to be worth it in my tests, probably because it is an extremely hot loop!
We reduce memory consumption by not storing start indices explicitly, which typically need to match the address space of the source length. In the case of Zig, where source files are constrained to be at most ~4GiB, only 32 bits of address space is needed for any given file. Thus the goal becomes reducing 32-bit start indices to something smaller. Quasi-succinct schemes for reducing the space consumption of monotonically increasing integer sequences immediately spring to mind, such as Elias-Fano encoding. However, we can achieve good space compression by simply storing the length of each token rather than the start index. Because tokens almost always have a length that can fit in a byte, we try to store all lengths in a byte. In the event that the length is too large to be stored in a byte, we store a 0
instead and make the next 4 bytes the true length. This works because tokens cannot have a length of 0, else they would not exist, therefore we can use lengths of 0
to trigger special behavior. We also know that this idea does not affect the upper bound on the number of Token elements we need to allocate because in order for a token to take up 3 times as much space as a typical token, it needs to have a length of at least 256, which the astute reader may note is significantly larger than 3.
Use fewer variables where possible. While machines nowadays have a lot more registers than they used to, you still only have access to 16 or 32 general purpose registers! If you have more variables than that, you have to spill to the stack (it's actually worse than this, because intermediate values in expressions temporarily need their own registers too). While machines do have extra registers they can use under the hood, you do not! Therefore, we can get better performance by
non_newlines
bitstrings. Instead of storing all of the bitstrings I get from the SIMD/SWAR code on the stack in a [4]u64
(on 64 bit machines), and then writing separately to a non_newlines
pointer, I write all the bitstrings into the memory allocated for the non_newlines
bitstrings. Each time, I increment the place we are writing in the allocation by the width of a single bitstring, i.e. 8 bytes on 64 bit machines. Since I always write the non_newlines
into the current position in the allocated memory and the other bitstrings are written after it, we will be left at the end with only non_newlines
bitstrings. The only downside is we need to overallocate an extra 3 u64's than we otherwise would, but that's hardly any trouble. Here is a diagram of how this strategy looks in memory:|0|1|2|3|4|5|6|7|8|9| <- slots
|a|b|c|d| <- We write our bitstrings to 4 slots. (`a` is `non_newlines`)
|a|b|c|d| <- Each time, we move one slot forward
|a|b|c|d|
|a|b|c|d|
|a|b|c|d|
|a|b|c|d|
|a|b|c|d|
|a|a|a|a|a|a|a|b|c|d| <- In the end, we are left with this
Aside from the to-do's listed in the main.zig
file, the plan with this is to rewrite the Zig parser which produces the Abstract Syntax Tree as well. I have a number of ideas on how to dramatically improve efficiency there as well. Stay tuned!
My ultimate goal is that this repository will be integrated with the Zig compiler.
git clone https://github.com/Validark/Accelerated-Zig-Parser.git
Next, install one or more Zig projects under the src/files_to_parse
folder.
cd Zig-Parser-Experiment/src/files_to_parse
git clone https://github.com/ziglang/zig.git
git clone https://github.com/zigtools/zls.git
Then run it!
cd ../..
zig build -Doptimize=ReleaseFast run
In the last few days, I have:
Disabled loop unrolling for quote parsing loop on SWAR-enabled machines. Only a 1% uplift on my Sifive U74, but considering that's about 9 milliseconds at the moment, I'll take it.
Updated the keyword lookup algorithm for non-vector architectures to use aligned loads where possible. There still could be room for improvement but today I saw a ~5% performance uplift.
Updated to a slightly better optimized version of the escape-detection algorithm.
Made toggles so that "
and '
can be moved between the SIMD/SWAR section and a naïve scalar version very easily. It appears that for machines which have to use SWAR, it is faster to do the naïve scalar version (almost ~8% uplift on my RISC-V SiFive U74). On the other hand, it's still more efficient on my desktop to do quote classification in SIMD, but for other less-powerful devices, it may not be worth it.
There is also a trade-off to be made on big-endian hardware. The SIMD '
/"
escape detection algorithm currently has to be done in little-endian, so there necessarily has to be a reversal somewhere (or byte-reversal on the vectors) if we want to not use the naïve scalar version.
With SIMD, we need to do a vector reversal, unless we have fast machine-word bit-reverse instructions. At the moment I am not aware of any ISA's supported by the Zig compiler with a fast bit-reverse instruction besides arm/aarch64.
rbit
) so that we can use clz
directly in our bitstring queries, rather than rbit
+clz
. On little-endian machines, we do the flip after the escape-detection algorithm. On big-endian, we can do it before, but then we can just bitreverse the backslashes before and after the escape detection algorithm. arm is typically little-endian these days, but who knows, maybe a future ISA can take advantage of the flexibility.mips64 and powerpc64 have built-in clz
instructions, and emulate ctz
via @bitSizeOf(usize) - clz(~x & (x -% 1))
. Therefore, if we wanted to do quotes in SIMD and use clz
, we would have to flip our bitstrings twice! Ouch! Hopefully I or someone else figures out how to make a big-endian equivalent of the escape character bitstring generation algorithm.
some sparc64 machines have popc
(e.g. niagara 2 and up), which can emulate ctz
via popc(~x & (x -% 1))
. To do clz
we have to do x |= x >> 1; x |= x >> 2; x |= x >> 4; x |= x >> 8; x |= x >> 16; x |= x >> 32;
to spread the most-significant bit all the way right-wards, then we can invert the bitstring to get a mask of the leading zeroes, and popcount that. So on big-endian sparc64 machines, we WANT to do a bitreverse. Also, LLVM's Vectorization currently does not work on sparc64 (or powerpc) machines, so we probably have to use the SWAR algorithm for the time being.
machines which do not have clz
builtin can probably emulate a ctz
faster than a clz
.
With SWAR, we can do either a @byteSwap
on the register being treated as a vector or we can reverse the bits with a bitreversing movmask algorithm. The problem with the latter is that we have to do an extra shift right on the bitstring before multiplying because the most significant bit of the most significant byte has to be moved to the lowest bit position of its byte. We can avoid this extra shift by instead using a multiply-high operation and concentrating the bits in the upper half of a double machine-word, and maintaining a 3 instruction movmask. However, the problem with this idea is that multiply-high might be a function call or have terrible throughput / port constraints, whereas multiplies typically have a throughput of 1 per cycle. Is the throughput actually a problem in practice though? Unsure. We do have quite a lot of other work to do in between multiplies.
'
/"
in SWAR). Therefore, if we can do a machine-word byte-reverse faster than 3 shifts and/or in fewer than 3 instructions, it would be smarter to do the byte-reverse. Alternatively, if we have fast multiply-high's somehow, we could use that to eliminate the 3 extra shifts (per native subchunk).For now, I think the bitstring escape sequence algorithm is best left disabled on big-endian hardware without a fast bit-reverse (so just arm atm).
Since sparc64 and powerpc have to use SWAR until LLVM improves, they should do quote/escape logic in scalar code, not vector code. sparc64 machines lack bit-reverse and byte-reverse, so we can use the movemask-reversed function on sparc64.
powerpc can stay in big-endian land and use clz
.
mips64 has an extension to the ISA which adds vector instructions, although I'm not sure if it has gone into real hardware yet or whether those vectors are actually useful for the kind of SIMD we are doing here.
clz
.Partially added some control character bans but there is still more to be done. Still, as of yet, incomplete.
Replaced the SWAR movmask algorithm with one significantly better on typical hardware. Before, we were using an algorithm from Wojciech Muła which for 64 bit operand x
would basically do: (@as(u128, x) * constant) >> 64
. Now, we can stay within the lower 64 bits by concentrating the target bits in the most significant byte, so no widening is necessary. This is really good news for basically every machine I could find info on for the difference between mulhi
vs mul
. Typically mulhi
instructions have much higher latency and signicantly worse throughput, and some machines do not even have a mulhi
instruction at all. My algorithm modifies Wojciech Muła's algorithm to use only the lower 64 bits of the product of the multiplication:
Example with 32 bit integers:
We want to concentrate the upper bits of each byte into a single nibble.
Doing the gradeschool multiplication algorithm, we can see that each 1 bit
in the bottom multiplicand shifts the upper multiplicand, and then we add all these
shifted bitstrings together. (Note `.` represents a 0)
a.......b.......c.......d.......
* ..........1......1......1......1
-------------------------------------------------------------------------
a.......b.......c.......d.......
.b.......c.......d..............
..c.......d.....................
+ ...d............................
-------------------------------------------------------------------------
abcd....bcd.....cd......d.......
Then we simply shift to the right by `32 - 4` (bitstring size minus the number of relevant
bits) to isolate the desired `abcd` bits in the least significant byte!
Laid groundwork for exporting non_newline bitmaps, that way we can use it later on in the compiler to figure out what line we are on without needing to go byte-by-byte later on in the pipeline.
|0|1|2|3|4|5|6|7|8|9| <- slots
|a|b|c|d| <- We write our bitstrings to 4 slots. (`a` is `non_newlines`)
|a|b|c|d| <- Each time, we move one slot forward
|a|b|c|d|
|a|b|c|d|
|a|b|c|d|
|a|b|c|d|
|a|b|c|d|
|a|a|a|a|a|a|a|b|c|d| <- In the end, we are left with this
Fixed random performance issues, like the compiler not realizing that our SIMD/SWAR chunks are always aligned loads. (It matters a lot on less-mainstream machines!)
Made the SIMD/SWAR code go chunk by chunk in lockstep rather than have each individual component load its 64 (on 64-bit machines) bytes separately. I am assuming that LLVM was able to reuse loaded vectors on some occasions, but in practice I saw a massive speedup in the last week. Granted, the utf8 validator was turned off temporarily while it is being reworked. However, on my Zen 3 machine I typically saw basically no performance difference between running the utf8 validator versus not. The reason for this is because we can almost always exit early (when the entire chunk is ascii). Due to alignment/cache/happenstance, I typically saw my tokenization times go down with the utf8 validator turned on, so I don't think I am unfairly advantaging my most recent measurements.
Turned off the utf8 validator. I need to fix the types for it so it can be re-enabled. We also need to port a SWAR version. simdjson or Golang might have some tricks we can use.
Added an option to enable or disable the folding of comments into adjacent nodes (FOLD_COMMENTS_INTO_ADJACENT_NODES
). This should make it a little easier to change my mind on the particulars of the AST implementation.
Added more tests and compile-time assertions. We're getting there!