Closed mgilbir closed 2 years ago
Honestly I've never really stress tested Zod with payloads that large. I suspect there is lots of low-hanging fruit that would fastly improve performance.
I'm a little confused though. Zod is taking 500+ms to parse each object? How big is each object? Is it possible you're hitting memory limits of the Nodejs process?
Without a sample data set it's hard for me to duplicate this. Are you able to share?
Thanks for taking the time to think about this.
I was also very confused by the slowness I was seeing. I first managed to narrow it down to the parse
or check
call, and then confirmed with the profiler. I would not be surprised if there are other effects at play here.
I usually work with go and was ready for performance to be lower but as I don't have experience of what is considered "normal" I decided to ask.
I'm going to try to prepare a dataset and extract a code example that allows to reproduce.
Yeah all of Zod's complex logic is in .parse
, and .check
just delegates to .parse
. I don't think this is normal.
@mgilbir Here is a github that tracks benchmarks for different validation libraries: https://github.com/moltar/typescript-runtime-type-benchmarks if performance is an issue.
@davidmdm Thanks for pointing to those benchmarks!
I took a look with the profiler and found that makeError was allocating, and doing a lot of work upfront, on every call to the parse function even when no error occurred. I'm new to typescript, so I'm not sure this is the best way to solve it, but it may point you towards some low-hanging fruit. Code is available in PR #218
My team was using zod to validate JSON messages coming from IoT devices via a queue. We received hundreds of messages per minute (so not exactly high volume). Validating 150 messages with zod takes more than 6 seconds. Using ajv we can validate the same amount of messages in 120 ms with a fraction of the CPU usage.
We liked the idea of zod, but we have to move away from it until we can get a decent performance. :-(
I saw this test results and it actually came after I had already looked into a issue with slow code where I discovered that zod was taking up 600 ms to parse 1000 json elements.
moltar/typescript-runtime-type-benchmarks
Zod is one of the slowest in the overview.
Normally speed is not a problem since I use zod on user input, but if you run a busy api endpoint this could be something important.
@dgg I maintain myzod, which is basically just a lightweight reimplementation of zod. It doesn't have a big community like zod does, but it is much much faster, and has an extremely similar api.
I am here because I am actually punched by the slow performance of zod. I use zod to validate http api response. It takes seconds to parse 2000+ rows of data, sad :(
Interesting things are there are so many zod-like libraries here with tens of stars just because of the performance. If zod could improve the performance of itself, the world would be cleaner since zod is absolutely with the most features and supports.
This is forcing me to abandon zod
for the time being, unfortunately. I'm seeing it take seconds to parse even small chunks of data. I was hoping to see an improvement with v3 but it's actually worse for me. It's really unfortunate because I'm having a hard time finding a decent replacement. I was hoping to do a drop-in replacement with myzod, but that doesn't work quite the same and has some really unusual behaviors. Every other library I've tried has severe limitations as well, it seems. And none of them match the level of documentation found here. It's a really excellent library otherwise, and I appreciate it very much -- I hope I can use it again soon.
I have parsed 10k objects base on the schema provided by @mgilbir with version '3.0.0-alpha.29' and '3.1.0', the output is high improved by the team! @colinhacks
the result is:
const bigObject = { a: 'hello', b: 'hello', id: 'Welcome', creation: 'Today is Tuesday', content: { id: 'id1234', title: 'Software Engineer', version: 100, }, labels: ['2342l42342344', '24234234', '234243242345', '2342342342342'], answers: [ { key: 334242, value: { name: 'hello', age: 20, }, }, ], result: [ { key: '324234324234', value: { name: 'hello', age: 20, a: 'hello', b: 'hello', id: 'Welcome', creation: 'Today is Tuesday', content: { id: 'id1234', title: 'Software Engineer', version: 100, }, }, }, ], };
the issue has been fixed!
@zhenhao18 What sort of results were you seeing before 3.0.0-alpha.29
that we can use as a comparison? Because I'm still seeing very slow parsing performance in all versions, including the latest (3.1.0
).
Unfortunately I can't really share my dataset and schemas at the moment because I'm using some real code and data to benchmark (if I have time, I'll try to generate some mock data and schemas to share), but these are my latest internal results:
library | parse time (ms) |
---|---|
zod 3.1.0 | 2200.53 |
zod 3.0.0 | 2102.35 |
zod 1.11.17 | 656.82 |
myzod 1.7.4 | 19.19 |
We've switched to myzod for now, which is why it is included here. And as you can see it's significantly faster (and the schemas used in each case are virtually identical - nothing has been slimmed down or removed). But myzod is not without issues and concerns in itself. Also interesting to note how much performance has decreased in zod v3 from v1.
Edit: I wanted to add a note that these results are from parsing in NodeJS, not in a browser. So they really can only be used as a relative measure to each other and would probably vary significantly in other environments. For example, a subset of this data in a browser was taking upwards of 10 seconds in some cases for us.
I still not do a test on the version before 3.0.0-alpha.29, will do that today or tomorrow, and let you know the result. myzod looks great!
On Fri, 11 Jun 2021 at 06:35, Joe Krill @.***> wrote:
@zhenhao18 https://github.com/zhenhao18 What sort of results were you seeing before 3.0.0-alpha.29 that we can use as a comparison? Because I'm still seeing very slow parsing performance in all versions, including the latest (3.1.0).
Unfortunately I can't really share my dataset and schemas at the moment because I'm using some real code and data to benchmark (if I have time, I'll try to generate some mock data and schemas to share), but these are my latest internal results: library parse time (ms) zod 3.1.0 2200.53 zod 3.0.0 2102.35 zod 1.11.17 656.82 myzod 1.7.4 19.19
We've switched to myzod https://github.com/davidmdm/myzod for now, which is why it is included here. And as you can see it's significantly faster (and the schemas used in each case are virtually identical - nothing has been slimmed down or removed). But myzod is not without issues and concerns in itself. Also interesting to note how much performance has decreased in zod v3 from v1.
— You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/colinhacks/zod/issues/205#issuecomment-859038556, or unsubscribe https://github.com/notifications/unsubscribe-auth/AT6BEI543N6N2UE2E4W7NBDTSEOZPANCNFSM4TECTSOQ .
This morning I decided to check in on zod and see if there's been any changes (it's been a few weeks since last I checked), and WOW! HUGE improvements since v3.1! Here is the results of my latest benchmark on our internal dataset adding Zod 3.5.1:
library | parse time (ms) |
---|---|
zod 3.5.1 | 76.74 |
zod 3.1.0 | 2453.06 |
zod 3.0.0 | 2080.56 |
zod 1.11.17 | 723.87 |
myzod 1.11.17 | 26.72 |
So just a massive improvement. This is so awesome. Thank you so much to @milankinen, @colinhacks, and everyone who has been contributing - I really appreciate you all. This is really amazing and definitely makes zod a usable option for me again - I'll definitely be looking into switching back.
Side note: I did run into some minor tweaks I had to make to my schema where I was using passthrough
with intersection
which is causing an error in 3.5.1, but that was pretty straightforward to clear up.
Just went through versions on our end too. Seems like there's a huge performance regression from 3.9 -> 3.10. One of our functions went from 192ms to 660ms between the two.
3.11: 3.9:
I really love this lib and anything typescript, but taking 3x to 4x to parse things compared to myzod. Performance should be a one of the highest priority as the ecosystem starting to build around zod.
@mmahalwy Do you have an example of something that has gone up in time? I think that would help debug and maybe a benchmark suite could be added
Recently migrated from ajv to zod then to myzod, so there is no baseline for me to compare between zod versions. Just compared between zod & myzod and zod overall making my task to take around 50% to 80% more time to complete. As the task itself has lot's mini validations individually it won't look much but they add up by the time task completes. I also suspect there could be some GC pressure too.
I've implemented #1023, #1021, and #1022 to explore different vectors at making zod faster. I think updating the target
is the highest payoff for the least amount of change, but the other PRs also have some notable perf improvements. I'll keep looking - I think it's possible to make zod as fast as the alternatives.
@tmcw
Incredible work and thanks so much for digging in! I'll spend some time tomorrow and this weekend trying to think through any implications that need to be addressed.
The point about switching our target is an important one: we have mostly treated compatibility as a best-effort thing, but this would be a bit of a line in the sand. I am not involved in enough open source projects of this size to have a good feel for the general approach the community usually takes here, but I feel like that might warrant a major bump?
@tmcw I’m not sure if I understand the current caching correctly. Are you implying that the cache is permanent? This would explain why parsing a 150 MiB JSON with a complex schema uses up over 4 GiB memory (not sure how much over, just that it’s less than 8 as I had to manually bump node’s memory limit) on one of my side-projects.
Do you know if the current cache gets freed over time? Also, do you know how I could try out this change to see what the performance difference is with my real-world data? The schema is insanely complex as the data structure it has to parse is ancient and organically grew. There’s lots of .optional()
s etc.. So, would be really interesting to see the impact of your PRs.
Also, thanks a lot for starting to work on zod performance. I’m really looking forward to this as it currently takes >4 GiB of memory and around 2 minutes to parse the data. So, let me know if there’s a way to try this out.
I originally commented this on the PR, but I think let’s keep the PR’s comments for the code itself
I managed to test-out the es2018 version (#1022), and this change alone reduces my real-world 150 MiB complex JSON schema parsing time from 100s to 67s. Nice 33% reduction, just from adjusting a single line.
Trying out the cache-related changes from #1023, I only see an improvement from 100s to 96s. Not bad, but also not life-changing haha
Both PRs combined via cherry-picking locally reduce the total time to 64s. 36% better performance is a very very nice start. I hope there’s still some more aspects that can be optimized, but I’d gladly take even a 0.5% improvement at this point 😄
With both PRs combined, the minimum memory required to succeed is between 3.75-3.875 GiB. Down from the (now tested 4.4-4.5 GiB) which is a nice 15% reduction
I've focused mostly on the primitive types because they've been pretty straightforward to tweak & get some little wins. Should probably take a break to work on my product 😆 but in the meantime, some thoughts:
.optional().nullable()
on a lot of types. That has come with a performance penalty, which #1028 reduces, but could be reduced more by making optional & nullable more like 'modifiers' (like maxLength) than their own types. But that would make the codebase less elegant. Anyway, this is all at the micro-level with values and probably isn't the biggest performance drag.Fwiw, if folks want to benchmark their things, what I've been relying on is the npm run benchmark
script in this repo combined with
node --prof bench.js
node --prof-process --preprocess -j isolate*.log > profile.v8log.json
And then dumping the results into speedscope.
I ran the moltar benchmarks on the latest zod release and it's 3.6x faster than the previous benchmarked release (3.11.x).
I've skimmed a few competing modules and I think the main remaining difference is in Zod's approach to objects and arrays: zod truly parses, doesn't validate, and it creates copies of input datastructures. Probably if that approach changed - or we optionally allowed for a non-copying implementation - we'd be able to close more of the gap.
@tmcw that is extremely cool!
I gotta say @tmcw, your work is seriously appreciated. I migrated away from Zod because its real-world performance was too slow but this focus on performance makes it viable again!
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions.
Not stale. Re-opening due to importance of performance.
I think a fresh issue about performance should probably be more specific than this one, otherwise it's not really something that could ever be closed. Performance could always be improved, it's a theme not a task.
My guess in terms of further performance tweaks is that zod would need an option to turn off cloning, which would also need to disable methods like transform
or require a new fancy implementation of those methods. Cloning is basically the main overhead, afaict, and the modules that do well in the moltar benchmarks are the ones that do not clone.
@tmcw That was insightful! There has only been one other ticket mentioning "cloning" which was closed. https://github.com/colinhacks/zod/issues/922 What do you think? 😄
Yep, I think that's the direction (though obviously it'd be useful to do a prototype and early benchmarks to confirm that intuition). Parse, don't validate is a great principle and probably the right default, but I think it might be the cause for a sort of performance ceiling.
@tmcw I’m not sure if that fits into the scope of zod
though. For me, this library has always been about type safety and security (e.g. by disallowing unexpected stuff in API parameters) by ensuring that data is always recreated rather than cloned or copied. I’m not sure if it’s possible to guarantee avoiding all the possible issues of prototype pollution and otherwise modified objects without recreating the data from scratch
Sure, so - a no-clone option, as far as I can see it, would mean:
.strict
or .passthrough
, not .strip
This is, fwiw, the same tradeoff you'll see in any of the faster modules on moltar's benchmarks. I would reckon that people are generally okay with these tradeoffs because those modules are heavily used, and if zod doesn't support a no-cloning mode, then it'll never get close to their performance.
Anyway, cloning would be a choice, and you could use the cloning (default) mode and get the guarantee of fresh object prototypes, or use the no-cloning mode and accept that you might get a modified object prototype. At least for my application, I think no-cloning would be very useful and would probably be a lot faster, and modified object prototypes - in my experience - are not especially concerning.
This affects array
and object
schemas mostly, correct? I could see adding different "no-clone" versions of those schema types as a potential optimization that you have to opt into explicitly. Definitely would feel a little unergonomic in some cases since if you have nested schemas you'd have to ensure they are also of this same type if you wanted maximum performance in that case.
Might be worth opening a Discussion about this.
I strongly agree with the idea about "parse, don't validate" semantics which is (imho) one of the biggest reasons to choose zod over those high performance alternatives. We need to keep in mind that those libraries have been developed from the scratch focusing explicitly on the performance, whereas zod has had more focus on developer friendliness and fail-proof semantics. Because of this, the "legacy" of the former design and implementation choices (e.g. decision to support asynchronous parsing) will have a moderate effect on the performance even though the API and semantics are changed. As a result, zod loses one of it's main advantages (developer experience) and gets a little bit better performance but is still not able to compete with the top ones in terms of performance.
Luckily there still seems to be some room for performance improvements in zod's internals. 🙂 I quickly skimmed through the codebase and at least following parts could give some nice speed improvements without changing the existing API and semantics:
The current implementation seems to be using contexts with "lazy" parsing path where path is concatenated upon request. In theory, this is a nice idea in theory, but if we take look at the usage, path is read for each object property. That's O(k*n) complexity where k = length(path) and n = length(objectKeys). In addition, the whole parse context get "cloned" for each property because the path is different for each key. Same applies to arrays. This is a huge amount of internal cloning, object allocations and GC pressure that must be done regardless of the public API semantics.
I blame myself of this design choice. 😞 When I did my performance improvement PR, I left the path management in such state that the path is constructed prematurely which was definitely a mistake. If we think about where the path is really needed, the only places are issues that are returned to the user. By "returned", I mean issues that are visible to the user of zod APIs - as long as the parsing is in progress, the parse path isn't relevant. In practice this means that the path must be generated only if the (sub)-parser returns any issues. In pseudo:
ZodObject._parse = (ctx, object) =>
const issues = []
const parsed = {}
shapeKeys.forEachKey key =>
const keyResult = keyParser._parse(ctx, object[key])
if (isInvalid(keyResult))
keyResult.issues.forEach issue => issue.path.unshift(key)
issues.push(...keyResult.issues)
else
parsed[key] = keyResult.value
return issues.length > 0 ? INVALID(issues) : OK(parsed)
Even the ctx
can be passed as it is because there are no dynamic parts in it anymore. The only required "cloning" is when pushing key issues to the object's issues. In fact, this could also be eliminated by using linked list (linking two lists is an O(1) operation). Compared to the current implementation, the difference should be significant.
The current implementation has roughly O(n^2) complexity. And in addition those extra keys are not even needed in the default mode. Fortunately the fix for the default case (strip) should be trivial, just use shapeKeys
and don't even try to find extras. For passthrough, strict and catchall, the most efficient implementation that comes to my mind is to merge both shape and data object into single work-in-progress clone and use it:
// same error handling as in default "strip" case
const issues = []
const wip = Object.assign({}, shapeObject, dataObject)
for const key of keys(wip)
if key in shapeObject
wip[key] = shapeObject[key]._parse(ctx, dataObject[key])
else
wip[key] = handleExtraKey(ctx, key, dataObject[key])
return issues.length > 0 ? INVALID(issues) : OK(wip)
There are also some other smaller improvements that could be implemented, e.g using function composition for number/string refinements and simplifying the ok/aborted/invalid result handling but I'm not sure whether they're worth changing or not.
the only places are issues that are returned to the user
Now that I looked the codebase again, it seems that the path is actually needed in the refinement context and that might be the reason why I left the implementation as it is... 😬
The .path
property is not documented in the README or in the error handling documents at all, so how big deal it would be to remove it? Technically it's a breaking change but not to the official documented feature... 😇
@milankinen isn't path the only reasonable way to programmatically use zod errors? It'd be a pretty big deal if they're not machine-readable anymore
Path is definitely documented in multiple places including the error handling guide.
Yes, path indeed is an important property when processing the received parsing/validation issues and the issue contract should not be changed. 👍 But I'm not talking about issue paths. I'm talking about refinement context path
property that is not officially document (or at least I didn't find any references to it) but still accessible through superRefine
, e.g.
import { z } from "zod";
const notTsers = z.string().superRefine((s, ctx) => {
console.log("ctx.path", ctx.path); // <<< this
if (s === "tsers") {
ctx.addIssue({
code: z.ZodIssueCode.custom,
message: "can't be 'tsers'"
});
}
});
const person = z.object({
name: z.object({
first: notTsers,
last: notTsers
})
});
console.log(
"result",
person.safeParse({ name: { first: "matti", last: "tsers" } })
);
If you run the snippet, you can see that the ctx.path
prints ['name', 'first']
and ['name', 'last']
to the console. However, the parse issue (from safeParse
) includes the same path although I don't explicitly set it to the added issue.
My question is: how important the read-only access to the ctx.path
during the superRefine
callback execution is, and could that be removed? That is the only functionality in the public API contract that forces the path to be created prematurely and preventing the performance optimizations I presented earlier. Removing the ctx.path
doesn't affect the issues returned from safeParse
, they'll continue to include the path like before.
I've implemented the suggested tweak to the default object parsing - not collecting extraKeys, in #1393. It's a very modest performance bump, seems like around 5% in the realworld testcase. Unfortunately I think that's what we're looking at with the cloning strategy - there's maybe another 10-20% optimization left, but systematically zod will be magnitudes slower than non-cloning modules.
Hello,
We at Twilio are considering using your library to validate messages sent over WebSocket. The API is probably the most fantastic thing I've seen over any validators/parses I've checked. The only question that remains is performance. Our WebSocket protocol requires to be as optimized as possible and Zod hasn't really been great in that regard. I've read this discussion about adding an option that disables cloning. Do you know whether that's feasible in near future and if that will have a significant impact on performance?
@InfiniteRain Would you be willing to contribute to the performance improvement effort :)?
@kibertoad If adding a no-clone option is something that the maintainers would want and if it brings significant performance improvements, then I would love to contribute.
@colinhacks Would you accept a PR for no-clone option?
Oh wow a toggle for no clone mode would be incredible! Literally best of both worlds depending on scenario...
I'm definitely open to this! Slightly concerned about bundle-size bloat, and I'm not sure how things like transforms will work (I suppose they would just throw). But ultimately it's super worth it if perf is a blocker on adoption. @InfiniteRain don't hesistate to get in touch about any implementational questions - Twitter is probably best 👍👍
My first instinct is that this should be a separate method, probably .validate{Async}
. (Yeah, yeah, I know.) I like the idea of having the return type as a type predicate (input is T
) since there's no need to re-return the value you just passed in. This API also makes it structurally apparent that the method isn't returning a clone of the input (since only a boolean gets returned.)
const mySchema = z.object({ name: z.string() });
const data = { ... };
const isValid = mySchema.validate(data);
if(isValid){
data; // {name: string}
}else{
data; // unknown
}
The downside is that a "safe" version of this method isn't really possible, because type predicates can't be nested inside objects.
@colinhacks I'm very interested in working on this! I'm on vacation right now, but once I'm back I'd love to get in contact with you to discuss the details further.
@colinhacks if zod
supports tree shaking, I wouldn’t worry much about bundle size. It hardly matters in that case
Just to add an additional data point, since Slonik adopted Zod for all query validation, our CPU usage doubled. Avoiding cloning would dramatically improve things.
Typeform.com is also considering using zod
for form validations and API request payloads in comparison to yup
👍
I did a very fast experimental refactoring last weekend based on the descriptions in my previous comments (diff here, very WIPWIP) and got approximately 50% performance improvement in "real world" and "object" suites and 100% in primitive suites. I'd be nice to compare this with the no-clone approach. 🙂
We are currently parsing API responses with Immutable.js Record and looking to move to zod. Anyone know how performance compares?
I've just started using this library to check that parsed JSON objects conform to the types I define. I'm really enjoying it, but I'm afraid I've hit a performance snag.
Everything works fine with small tests but when I ran it on a slightly bigger input file (~10MB) I noticed that it was really slow.
I've tried the latest version and the beta, and
check
vsparse
, with similar results. The beta is faster, but I don't see a big difference betweencheck
andparse
.After profiling the
check
version in the beta, I'm seeing calls to.check
taking in the 500ms-1100ms range per object.Are those numbers typical, or am I doing something wrong with the schema definitions?
My schema definitions look like:
I'm really hoping that there is a way to speed it up, as this is too expensive for the use case where I'll have to process files with 100k+ objects.
Thanks!