-
Dear community,
I'm excited to share Colossal-AI, a deep learning framework for training and inference tasks in large language models (LLMs).
Colossal-AI stands out for its exceptional speed, su…
-
-
# Prerequisites
Please answer the following questions for yourself before submitting an issue.
- [ Yes] I am running the latest code. Development is very rapid so there are no tagged versions as…
-
2024-02-14 21:01 INFO 2048692:root - Downloaded https://dl.fbaipublicfiles.com/laser/CCMatrix/v1.0.0/2020-10_0278.tsv.gz [200] took 8s (5766.4kB/s)
2024-02-14 21:01 INFO 2048692:root - Starting downl…
-
**Is your feature request related to a problem? Please describe.**
LLM training is expensive, allowing sample packing is a more efficient way of training.
**Describe the use case**
I am trying to…
-
Dear @flowersteam,
trying to reproduce your results for coursework.
I find there are a number of issues in running the code. Here is a list of what I found so far.
## Importing
Several files…
-
#### Description
I am retraining a LLaMA3 model. Due to the limited size of my dataset, I attempted to use `freeze_updates` as referenced in the [NVIDIA NeMo documentation](https://docs.nvidia.com/…
-
Thank you for your great work!
In the stage 1 training mentioned in the paper, is the input of llm images and text,because the description ‘After the pretraining stage, the model is capable of genera…
-
### Description
Hi, I am using the latest version of LLamaSharp and my model is Llama-3 70b gguf version, when the number of GpuLayerCount is 0 to 5, although it is not very fast, I get the answer, b…
-
### Contact Details
github
### What happened?
I came here to report the issue / bug / my incompetence around the error of: `llama_model_load: error loading model: done_getting_tensors: wrong numbe…