huawei-noah / VanillaNet

MIT License
811 stars 56 forks source link

About the width of the model #8

Open wsy-yjys opened 1 year ago

wsy-yjys commented 1 year ago

Your work is exciting! but I have a question about the width of the model,Have you tried to reduce the width of the model, why is the width of the smallest Vanillanet_5 need [1284, 2564, 5124, 10244]? thank you~

HantingChen commented 1 year ago

Thank you for your question! Our aim is to maximize the parallelism of the network and reduce the number of layers as much as possible to achieve better speed on the GPU. The current number of channels represents a good trade-off. Reducing the channels may cause a significant decrease in accuracy with relatively little gain in speed, hence our choice of these specific channel numbers. Of course, you're welcome to try different channel numbers and train the network yourself to explore further.

wsy-yjys commented 1 year ago

Thank you for your reply! as you say “maximize the parallelism of the network”, so if the optimal channels of layer in model is related to the hardware performance, whether the optimal number of channels is different on different hardware? Especially the edge end Gpus?

HantingChen commented 1 year ago

Absolutely, the optimal number of channels is indeed influenced by the specific hardware in use. Particularly for edge-end GPUs, which may have less parallel processing capability compared to more robust systems, a smaller number of channels could be more suitable. It's always crucial to take into account the specific hardware characteristics when seeking to optimize for the best performance.

wsy-yjys commented 1 year ago

I see. Thank you very much for your reply~