-
Just an idea, if possbile to somehow enable `vim` while in REPL mode? Chats can be pretty long and jumping and modifying own text would be much faster.
-
Hello,
I want to create a semantic search functionality. The model is accurate in calculating the similarity between the word synonyms. Here are a few excerpts of the demo:
Supposedly, my dataset ha…
rut00 updated
9 months ago
-
it supports Chinese, while does it also work for Korea and Japanese?
-
Hi, I'm interested in running the example found in `biencoder/nli_msmarco/scripts/train_bloom7b1.slurm`. Is it possible to execute these using accelerate and deepspeed?
I'm planning to experiment wit…
-
The original run.py saves the model in pytorch_model.bin, which cannot be loaded directly using the code provided in this repository. After replacing line 422 `trainer.save_model()` in training/run.py…
-
I need to input a length greater than the maximum sequence length.Could I need change the max_seq_length?
-
![maim-region-20240220-114838](https://github.com/TheR1D/shell_gpt/assets/92385307/f2a5ce77-c750-4be8-bb45-97b53ea5aa19)
-
Token indices sequence length is longer than the specified maximum sequence length for this model (nnnn > 1024). Running this sequence through the model will result in indexing errors.
-
Hi @Muennighoff,
I would like to use your cross encoder with different GPT models.
I have noticed that this [script](https://github.com/Muennighoff/sgpt#cross-encoder) is different from the code in…
-
Hi!
I saw that you have used lower chunksizes (2-4) in training of models, may I know why? I am sure 40GB of RAM in a GPU can handle more? Does it give better empirical results?
Thanks!