-
When calculating absorber and distillation columns, it's assumed that the complete vapor-liquid equilibrium is achieved at each tray section, or that the user have knowledge about the efficiency of th…
-
Hello, author, you directly use CLIP model to classify the class-agnostic binary mask during the testing phase. This seems to violate the principle of zero-shot learning, because CLIP already has the …
-
Hello
How are you?
Thanks for contributing to this project.
I think that u have introduced only Pixel-wise distillation in the paper "Structured Knowledge Distillation for Dense Prediction".
What …
-
Hello,
When reading your paper entitled "Making Monolingual Sentence Embeddings Multilingual using Knowledge Distillation", I found that the results of LASER on wuu, yue and jpn are different from …
-
Hi,
I want to do the experiment using other dataset such as VOC dataset. What should I do before executing the examples/object_detection.py script?
I converted the VOC annotation to COCO f…
-
Dear authors,
I didn't see the code for transformer distill. Could you please tell me where you did it?
I looked into the code. Below is what I learned.
- `model/` containts the model definitio…
-
![image](https://user-images.githubusercontent.com/48303408/133541228-9d9a2916-76d0-4f59-bbd7-8c376bc62f89.png)
怎么能让两个小模型都保存下来,这个模型保存路径应该怎么设置
-
从这篇论文"KDnet-RUL: A Knowledge Distillation Framework to Compress Deep Neural Networks for Machine Remaining Useful Life Prediction"。 FD002 14.47. FD004 15.96. 它的FD002和FD004效果很好,你可以复现类似的结果吗。我看你复现的这个论文结果…
-
能讲讲你蒸馏的思路么,比如采用什么方法,参考文献?
-
Thanks for your excellent work!
So, as I investigated from the code, currently, your pipeline has a quality bottleneck: DALL-E can predict latent code of the VQVAE only 1024 tokens length. Vanilla …