-
### Links
- Paper : https://arxiv.org/abs/2111.11418
- Github : https://github.com/sail-sg/poolformer
### 한 줄 요약
![image](https://user-images.githubusercontent.com/39791467/197780827-b34eb533-e2…
-
## CLIP
* [\[Blog\]](https://openai.com/blog/clip/)
* [\[Paper\]](https://arxiv.org/abs/2103.00020)
* [\[code\]](https://github.com/openai/CLIP)
* [\[Model Card\]](https://github.com/openai/CL…
-
# Interesting papers
- [Davison 2018 - FutureMapping: The Computational Structure of Spatial AI Systems](https://arxiv.org/abs/1803.11288)
- Imperial College London의 Dyson Robotics Lab 교수님이신 A…
-
How can we use it for NLP transformers
-
Dear Unsloth,
Based on the initialization used in [your sample notebook](https://colab.research.google.com/drive/1Ys44kVvmeZtnICzWz0xgpRnrIOjZAuxp?usp=sharing) I created [my own notebook](https://c…
-
this is the detailed error:
(VIT) llb@raypc:~/codes/vision_transformer$ python3 -m vit_jax.train --name ViT-B_16-cifar10_`date +%F_%H%M%S` --model ViT-B_16 --logdir /tmp/vit_logs --dataset cifar10
2…
-
Eg https://huggingface.co/laion/CLIP-ViT-H-14-frozen-xlm-roberta-large-laion5B-s13B-b90k/tree/main
Need more config, adapting the weights and also changing the model at https://github.com/huggingfa…
-
Excuse me.I notice your work.It gives me a lot of encourage to explore more work about neighborhood-attention transformer.I read your python file,but have problem in relative position encoding.Can you…
-
This is what I'm getting:
```
load checkpoint from https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_caption_capfilt_large.pth
100%|██████████████████████████████…
-
Hi, in `momentum_clip`, there defined an optimizer that stores state using half-precision. I guess it is for quantification reasons. But this optimizer is nothing like what i am familiar with. and in …