issues
search
erfanzar
/
EasyDeL
Accelerate your training with this open-source library. Optimize performance with streamlined training and serving options with JAX. 🚀
https://easydel.readthedocs.io/en/latest/
Apache License 2.0
167
stars
19
forks
source link
issues
Newest
Newest
Most commented
Recently updated
Oldest
Least commented
Least recently updated
Fix unexpected indent error
#107
sr5434
closed
4 months ago
2
Exception while running any model - einops.EinopsError: Error while processing rearrange-reduction pattern "b (c n) d -> b c n d".
#106
jchauhan
closed
4 months ago
1
Example shown on https://pypi.org/project/EasyDeL/ to finetune tinyllama raise exception on kaggle
#105
jchauhan
closed
4 months ago
3
Error while finetuning Tinyllama on Kaggle TPU
#104
jchauhan
closed
4 months ago
5
QLoRA Finetune Example
#103
sr5434
closed
4 months ago
11
Add label smoothing, z_loss and ignore <=0 tokens in loss calculation
#102
yhavinga
closed
4 months ago
13
Optimize mean loss and accuracy calculation
#100
yhavinga
closed
4 months ago
0
Potential regression causing resource exhausted after recent commit
#99
yhavinga
closed
5 months ago
3
Error while training GPT2 on the kaggle
#98
jchauhan
closed
4 months ago
3
ValueError: `params` cannot be accessed from model when the model is created with `_do_init=False`.
#96
jchauhan
closed
5 months ago
2
What is the hardware spec are you using to tran a LLAMA model with 7B params
#95
jchauhan
closed
5 months ago
1
Error running remote model that has custom code
#94
jchauhan
closed
5 months ago
2
Error while training a Phi2 model
#93
jchauhan
closed
5 months ago
1
ValueError: Dict key mismatch; expected keys: ['transformer'];
#92
jchauhan
closed
5 months ago
2
Text Generation with Mixtral fails
#91
clintg6
closed
5 months ago
2
Step time increasing as training progresses
#90
yhavinga
closed
5 months ago
8
While training Gpt2 model - Exception - TypeError: in_shardings leaf specifications are expected to be PartitionSpec instances or None, but got *
#89
jchauhan
closed
5 months ago
1
AMD Hardware Support
#88
ThePerfectComputer
closed
5 months ago
14
Error while running a GPT2 model
#87
jchauhan
closed
5 months ago
3
Output of Tiny Llama using Easydel vs hugging face transformer api differs
#86
jchauhan
closed
5 months ago
4
Training on TPU Using Flash Attention
#83
IvoryTower800
closed
5 months ago
9
Inference on Single Node multiple GPUs
#82
clintg6
closed
5 months ago
10
Mixtral 8x7B support?
#81
clintg6
closed
5 months ago
1
Error while serving model as per documentation,
#80
jchauhan
closed
5 months ago
2
Will this project support lora?
#79
IvoryTower800
closed
5 months ago
2
RESOURCE_EXHAUSTED: XLA:TPU compile permanent error
#78
jchauhan
closed
5 months ago
1
Failing at the end launch gradio server
#77
jchauhan
closed
5 months ago
2
Running every model gives an error Shapes must be 1D sequences of concrete values of integer type
#76
jchauhan
closed
5 months ago
1
The function call llama_from_pretrained expects a device param that is missing while call it
#75
jchauhan
closed
5 months ago
2
Update Beta
#72
erfanzar
closed
5 months ago
0
Help train on tpu v3-32
#71
StableFluffy
closed
5 months ago
12
Dependency Error on latest version
#70
StableFluffy
closed
6 months ago
1
The model generate repeated words.
#69
IvoryTower800
closed
5 months ago
15
4D Mesh now supported for all the Models and BITs improved
#64
erfanzar
closed
6 months ago
0
don't let ai write docs for you <3
#63
erfanzar
closed
6 months ago
0
Updating Beta
#62
erfanzar
closed
6 months ago
0
Loss increases randomly
#61
infamix
closed
6 months ago
10
Changing Mesh
#60
erfanzar
closed
6 months ago
0
Kaggle issue
#59
infamix
closed
6 months ago
4
Updating Beta Branch
#58
erfanzar
closed
6 months ago
0
Updating Mistral and Llama Models
#57
erfanzar
closed
7 months ago
0
Update Beta Branch
#56
erfanzar
closed
7 months ago
0
Update Beta Branch
#55
erfanzar
closed
7 months ago
0
Updating Beta Branch
#54
erfanzar
closed
7 months ago
0
Support Sphinx Docstring Format
#53
w11wo
closed
7 months ago
0
Update V0.0.40 Beta (Adding Flash Attention, Adding 8,6,4 Bit models ,improving Documentations)
#52
erfanzar
closed
7 months ago
0
Fix eval batch loop (beta branch)
#51
w11wo
closed
7 months ago
0
Question: Does low-bit config reduce TPU HBM memory usage when training?
#50
Beomi
closed
6 months ago
2
Fix eval batch loop
#49
w11wo
closed
7 months ago
1
Update Beta Branch
#48
erfanzar
closed
7 months ago
0
Previous
Next