issues
search
mistralai
/
mistral-inference
Official inference library for Mistral models
https://mistral.ai/
Apache License 2.0
9.37k
stars
817
forks
source link
issues
Newest
Newest
Most commented
Recently updated
Oldest
Least commented
Least recently updated
Update README.md: Fix page not found for link to guardrailing
#105
martin0258
opened
7 months ago
0
Fix typo
#104
moritztng
opened
7 months ago
0
What is the best way for the inference process in LORA in PEFT approach
#103
pradeepdev-1995
opened
7 months ago
0
Mistral input context length limitation
#102
DanYoto
opened
7 months ago
0
Minor typos
#101
sethupavan12
closed
2 days ago
0
Which is the actual way to store the Adapter after PEFT finetuning
#100
pradeepdev-1995
opened
7 months ago
0
vLLM Build Issue using the provided Dockerfile
#99
Good-Coffee
closed
7 months ago
4
Create Issue templates
#98
adityaraute
opened
7 months ago
0
Docs: Add tutorials for using Python client to generating embeddings and chat completion
#97
m-newhauser
opened
7 months ago
0
Fixing typos in MD
#96
Cassini-chris
opened
7 months ago
0
Has any thought been given to using LoRA to increase the number of experts (100x) with minimal memory?
#95
sixChar
opened
7 months ago
8
Fix typo/spelling in README.md
#94
GilesBathgate
opened
7 months ago
1
Mixtral Feedbacks
#93
titouandk
opened
7 months ago
0
Incomplete Output even with max_new_tokens
#92
pradeepdev-1995
opened
7 months ago
0
Building Mistral docker container results in OOM kill of the entire system
#91
codevbus
opened
7 months ago
0
wrong link in documentation
#90
Frank-Buss
opened
7 months ago
1
Adds attention mask with `model.forward(..., cache=None)`.
#89
andsteing
opened
7 months ago
0
Why does `cache=None` produce different outputs?
#88
andsteing
opened
7 months ago
0
Is the code up to date? Is the code the same for different model versions?
#87
zysNLP
opened
7 months ago
0
Inquiry on Implementing Sliding Window Attention for Custom Sequence Lengths
#86
yihong1120
opened
7 months ago
0
fix minor typo in README.md
#85
nheagy
opened
7 months ago
0
Fix link to official documentation in README.md
#84
webchick
opened
7 months ago
0
Add MoE and pipelining support
#83
diegolascasas
closed
7 months ago
0
Update classifier.ipynb
#82
eltociear
opened
7 months ago
0
Fix Dockerfile
#81
nicholasjpaterno
opened
7 months ago
0
on Jetson ORIN, Xformer, Memory-efficient attention, SwiGLU, sparse and more won't be available.
#80
cj401
opened
7 months ago
0
Is window attention technology also used during the training phase?
#79
peiyingxin
opened
7 months ago
0
How to process batch input in mistral-src/model.py ?
#78
NLPwoods
opened
7 months ago
0
repeated build failure
#77
juanmf
opened
7 months ago
0
The detected CUDA version (11.8) mismatches the version that was used to compile
#76
juanmf
closed
7 months ago
2
Fix: no system prompt in request
#75
michel-ds
opened
7 months ago
0
No safetensors in HF model card?
#74
EricLBuehler
closed
7 months ago
2
What is the difference between the files you publish on GitHub and Hugging Face
#73
zhzfight
opened
8 months ago
0
Unabled to load to GPU with 24 GB vRAM with quantization
#72
fangzhouli
opened
8 months ago
1
How is The 131K Attention Span Achieved?
#71
ThePerfectComputer
opened
8 months ago
0
Update README
#70
luv-bansal
opened
8 months ago
0
How to train mistral?
#69
mihalt
opened
8 months ago
0
Was Mistral Pretrained with Dropout Enabled?
#68
zaptrem
opened
8 months ago
0
Question about finetune mistral 7B (data format)
#67
xihajun
opened
8 months ago
1
model is giving answer in russian
#66
Sanchit-404
opened
8 months ago
4
how to explain Attention that input QKV tensor # xformers requires (B=1, S, H, D)
#65
dhcode-cpp
closed
8 months ago
1
Does mistral-instruct-7b support fast transformer deployment
#64
lebronjamesking
opened
9 months ago
0
Update README.md
#63
VinayKokate22
opened
9 months ago
1
Embedding model and Engine??
#62
muhtalhakhan
opened
9 months ago
6
More language support?
#61
OnceJune
opened
9 months ago
6
sliding window size in prefill and decode stage
#60
ofhwei
opened
9 months ago
0
Can't load xFormers because of PyTorch 2.1.0+cu121
#59
russ22cox
opened
9 months ago
2
Feature: Adding contributors section to the README.md file.
#58
Kalyanimhala
opened
9 months ago
2
Code complete?
#57
zhoumengbo
opened
9 months ago
3
Update README.md
#56
eltociear
opened
9 months ago
0
Previous
Next