issues
search
NVIDIA
/
RULER
This repo contains the source code for RULER: What’s the Real Context Size of Your Long-Context Language Models?
Apache License 2.0
738
stars
47
forks
source link
issues
Newest
Newest
Most commented
Recently updated
Oldest
Least commented
Least recently updated
DOCKER_BUILDKIT=1 docker build -f Dockerfile -t cphsieh/ruler:0.2.0 . excute wrong
#73
yuanhang110
opened
2 days ago
1
Bump vllm from 0.5.4 to 0.5.5 in /docker
#72
dependabot[bot]
opened
1 month ago
0
Detailed scores of Phi-3-mini-128k
#71
huangyuxiang03
closed
1 month ago
2
Unable to reproduce result for Llama3.1(8B)
#70
muhangao
closed
1 month ago
1
Evaluate on Jamba-1.5-Mini
#69
coranholmes
opened
2 months ago
17
Fix citation typo in README.
#68
gmlwns2000
closed
2 months ago
1
Note about Mistral models
#67
inflatebot
opened
2 months ago
0
OOM issue during evaluation
#66
mengniwang95
opened
2 months ago
3
About Mistral-Small-Instruct-2409
#65
showgood163
opened
2 months ago
3
Issue with installation: huggingface-hub
#64
SimJeg
closed
1 month ago
1
GPT-4-1106-preview
#63
yxgcsq
closed
2 months ago
3
docs: update README.md
#62
eltociear
closed
2 months ago
0
Request for permissions
#61
ChenAlmagor
closed
1 month ago
4
New Command R 08-2024 and Command R+ 08-2024 models
#60
jukofyork
closed
2 months ago
2
--accidental PR--
#59
montehoover
closed
2 months ago
0
datasets where
#58
yxgcsq
closed
2 months ago
1
removed unnecessary imports
#57
prateekvellala
closed
3 months ago
1
Performance Differences in Qwen2-72B-Instruct-131k
#56
lwang2070
closed
3 months ago
2
Fix paul grahams order
#55
Wangmerlyn
closed
3 months ago
1
Performance discrepancy of Llama3.1-8b-instruct
#54
zhenyuhe00
closed
3 months ago
8
Can't reproduce results of meta-llama/Meta-Llama-3.1-8B-Instruct
#53
PiotrNawrot
closed
3 months ago
4
Any chance of testing ` Mistral-Large-Instruct-2407`?
#52
jukofyork
closed
3 months ago
2
Add the template of Phi3 in template.py
#51
Luc1an0-H3
closed
3 months ago
0
Add the template of Phi3 in template.py
#50
Luc1an0-H3
closed
3 months ago
0
Test results for the June sneaky update of the Phi 3 models ?
#49
bhugueney
closed
1 month ago
1
A mistral long context - MegaBeam-Mistral-512K
#48
chenwuperth
closed
3 months ago
2
About InterLM2.5
#47
showgood163
closed
3 months ago
2
hope add qwen2-7b-chat result
#46
Chandler-Bing
closed
3 months ago
2
fix: correct batch output list shape when using pipeline
#45
Wangmerlyn
closed
3 months ago
1
Cache the prepared data from multiple runs
#44
Ying1123
closed
3 months ago
1
Gemini flash 1.5 results
#43
augusto-rehfeldt
opened
3 months ago
1
Add multi-thread for process_batch
#42
Ying1123
closed
3 months ago
1
RULER with Mamba
#41
Andron00e
closed
4 months ago
1
Implemented batching
#40
ViktorooReps
closed
4 months ago
3
Is there a particular reason to not support batch processing?
#39
ViktorooReps
closed
4 months ago
1
The one-shot example of CWE task
#38
guanzhchen
closed
4 months ago
0
Update variable_tracking.py
#37
Chandler-Bing
closed
4 months ago
1
Reproducing results 4k (LLaMA-2 7B chat, Mistral 7B Instruct v0.2)
#36
ThomasSURF
closed
4 months ago
3
Error in hugging face links in README
#35
etienneasln
closed
4 months ago
1
128K sequence length means 131072 or 128000
#34
syp1997
closed
4 months ago
1
Qwen2 and DeepSeek-V2 results?
#33
hijkzzz
closed
1 month ago
1
Add SGLang backend
#32
Ying1123
closed
5 months ago
0
Base vs Chat prompt question.
#31
karansaxena
closed
5 months ago
3
Prediction format during evals
#30
karansaxena
closed
5 months ago
5
pre_sample in qa code
#29
vkaul11
closed
5 months ago
1
request for evaluating GLM4-9B-chat(-1M)
#28
yucc-leon
closed
5 months ago
2
questions about ICL code for variable tracking
#27
vkaul11
closed
4 months ago
1
Is there any issue in extending context length to 1 million using your script
#26
vkaul11
closed
4 months ago
1
What is the need for is_icl parameter?
#25
vkaul11
closed
4 months ago
1
lost in the middle problem
#24
vkaul11
closed
4 months ago
2
Next