mit-han-lab / tinyengine

[NeurIPS 2020] MCUNet: Tiny Deep Learning on IoT Devices; [NeurIPS 2021] MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep Learning; [NeurIPS 2022] MCUNetV3: On-Device Training Under 256KB Memory
https://mcunet.mit.edu
MIT License
757 stars 127 forks source link

Higher peak memory usage for patch wise inference #104

Open rahulvigneswaran opened 2 months ago

rahulvigneswaran commented 2 months ago

While the paper suggests that patchwise inference generally decreases peak memory usage across various models, our experimentation reveals that the peak memory comparisons between patchwise and non-patchwise inference models such as mcunet-vww2, mcunet-in2, and mcunet-in3 exhibit higher peak memory usage with patchwise inference.

peakperformance