-
Here are some topic suggestions for the presentations.
Please comment the topic you want to work on!
## Tools and Frameworks
- GitHub Copilot
- Langchain 👉🏼 @Alex-J-W
- Grammarly 👉🏼 @Leama…
-
This issue concerns the dataset description at https://github.com/awslabs/open-data-registry/blob/main/datasets/software-heritage.yaml
Due to the recent rise in the quest of data for LLM training, we…
-
### Jan version
0.5.7
### Describe the Bug
Using Jan v0.5.7 on a Mac with an M1 processor, running Llama 3.2 3B instruct q8 via the API. Occasionally, the server stops responding to POST requ…
-
# Background
Dlrover is an elastic deep learning framework, with fault-tolerance of processes failure, POD losting etc. Since the LLM training is at large scale and always span for a long time, many …
-
Is there any successful story about LLM training using GGML?
-
### Feature request
Integrate Liger (Linkedin GPU Efficient Runtime) Kernel to HuggingFace Trainer, user could decide whether to enable kernel with a simple flag
### Motivation
Liger (Linkedi…
-
bother you again。
what is the loss in two stage in the experiments?
with my own coupus and chinese llm, the loss for alignment stage is 1.5; the learn to rank is still training with now loss=60.
-
Create a basic data preprocessing pipeline for a specific bioinformatics dataset to prepare it for LLM training. The pipeline should include steps for data cleaning, tokenization, and formatting
-
-
Great job!
Could you share your Wikipedia corpus for retrieval?
I'm curious about the data amount and the method for calculating top-n recall metrics.