triton-inference-server / server

The Triton Inference Server provides an optimized cloud and edge inferencing solution.
https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/index.html
BSD 3-Clause "New" or "Revised" License
8.38k stars 1.49k forks source link

Removing caching on windows. #7717

Closed mc-nv closed 3 weeks ago

mc-nv commented 1 month ago

related to: https://github.com/triton-inference-server/onnxruntime_backend/pull/276

mc-nv commented 1 month ago

related to: https://github.com/triton-inference-server/onnxruntime_backend/pull/276

nv-kmcgill53 commented 4 weeks ago

Overall, I like the direction this is going and the solution, just needs some fine tuning.