-
The idea of a data exchange platform for regulatory complaints and issues, integrated with an LLM, relates to financial institutions in several significant ways:
### 1. **Regulatory Compliance Manage…
-
![image](https://github.com/NVIDIA/nccl-tests/assets/79137028/46df9e5a-fc8c-4a7e-9dee-425de5b60165)
when i run nccl-test with sharp, i meet the error, what cause this
I tested using the NGC 24.05 ve…
-
-
We are from a small research group of a big tech company working on some LLM post training methods. As described in the agreement of llama2/3, we are bound by the Additional Commercial Terms and are n…
-
I have used an instance of Azure Ai gpt-4
and i got the error
I have initialized my model like this :
```
azure_model = AzureChatOpenAI(
openai_api_base=config.OPENAI_API_BASE,
openai_a…
-
Hello everyone, practitioner here,
I am looking to train a very serious non-LLM model, and the training is expected to be very hard, so I am looking for maximum speed.
I know that Google's TPUs …
-
Hello! Thank you for the clean + user friendly codebase!
I'm trying to finetune the VQ-VAE tokenizer and noticed some keys might be missing from the pretrained checkpoint listed on huggingface: `"o…
-
### Prerequisites
- [X] I am running the latest code. Mention the version if possible as well.
- [X] I carefully followed the [README.md](https://github.com/ggerganov/llama.cpp/blob/master/README.…
-
### Describe the feature
I want to continue pre-training llama 2 70b using my own data. My data is about 1b tokens. I have read [Fine-tuning Llama 2 70B using PyTorch FSDP ](https://huggingface.co/bl…
-
[meta engineering blog post](https://engineering.fb.com/2024/06/12/data-infrastructure/training-large-language-models-at-scale-meta/)
- Meta requires massive computational power to train large lang…