Closed fblissjr closed 4 months ago
@fblissjr we currently do not support windows and have not optimized for windows yet, please see https://docs.modular.com/engine/get-started#requirements for supported platforms and OSes.
Please try another machine if you can as we've primarily focused on X86 Data Center hardware! You can see from our Performance Dashboard. Thank you!
Bug description
Sharing the report for visibility! Great to see this stuff working!
~/mojo/max/examples/performance-showcase$ python3 run.py -m roberta
Doing some one time setup. This takes 5 minutes or so, depending on the model. Get a cup of coffee and we'll see you in a minute!
Done! [100%]
Starting inference throughput comparison
----------------------------------------System Info---------------------------------------- CPU: 13th Gen Intel(R) Core(TM) i9-13900K Arch: X86_64 Clock speed: 2.9952 GHz Cores: 30
Running with TensorFlow .......................................................................................... QPS: 46.26
Running with PyTorch .......................................................................................... QPS: 22.21
Running with MAX Engine Compiling model. Done! .......................................................................................... QPS: 25.92
====== Speedup Summary ======
MAX Engine vs TensorFlow: Oh, darn that's only 0.56x stock performance. MAX Engine vs PyTorch: That's about 1.17x faster.
Hold on a tick... We normally see speedups of roughly 2.50x on TensorFlow and 1.20x on PyTorch for roberta on X86_64. Honestly, we would love to hear from you to learn more about the system you're running on! (https://github.com/modularml/max/issues/new/choose)
Steps to reproduce
System information