issues
search
premAI-io
/
prem-services
AI Services Maintained by Prem Core Team.
MIT License
6
stars
8
forks
source link
issues
Newest
Newest
Most commented
Recently updated
Oldest
Least commented
Least recently updated
metal (gpu) support for llama_cpp services
#149
biswaroop1547
closed
11 months ago
0
Prompt Template as arg
#148
filopedraz
closed
11 months ago
0
Fixed general errors about Mistral model
#147
filopedraz
closed
11 months ago
1
Mistra Instruct weird error
#146
tiero
closed
11 months ago
0
Mistral 128k: does not support more than 512 tokens
#145
tiero
closed
11 months ago
0
Chat: uplift errors of model on user requests
#144
tiero
opened
11 months ago
0
Mistral 7B Instruct - Tokens <|im_end|> <|im_start|> in response
#143
Janaka-Steph
opened
11 months ago
0
Mistral 7B Instruct - "cannot parse response" after one or two response
#142
Janaka-Steph
opened
11 months ago
5
fixes stable beluga generation stopping midway
#141
biswaroop1547
closed
11 months ago
4
drop `./ml/models`
#140
casperdcl
opened
11 months ago
0
Stable Beluga generation issue
#139
filopedraz
closed
11 months ago
1
fix: redis script bug unbound variable
#138
biswaroop1547
closed
11 months ago
0
add: redis script fix for mac
#137
biswaroop1547
closed
11 months ago
0
Embedding binary enhancements
#136
nsosio
closed
11 months ago
0
vec-redis docker
#135
casperdcl
opened
11 months ago
0
Check CI not building binaries
#134
tiero
opened
11 months ago
0
fix: petals script uuid -> uuidgen cmd
#133
biswaroop1547
closed
11 months ago
0
added main.py to data
#132
nsosio
closed
11 months ago
0
script for setup of redis binary service
#131
biswaroop1547
closed
11 months ago
1
Binary support for ebd-all-minilm
#130
nsosio
closed
11 months ago
0
binary sentence transformers (embedding model) support
#129
biswaroop1547
closed
11 months ago
1
llama-2-13b-chat does not expose a /docs endpoint
#128
tiero
closed
11 months ago
1
Streaming support for Petals services
#127
biswaroop1547
closed
11 months ago
0
[pre-commit.ci] pre-commit autoupdate
#126
pre-commit-ci[bot]
opened
11 months ago
0
cht-petals: minor edits on paths and default parameters
#125
biswaroop1547
closed
11 months ago
1
adds chatml prompt template as string to maintain configurability
#124
biswaroop1547
closed
12 months ago
1
cht-llama-cpp: bump version
#123
casperdcl
closed
12 months ago
1
added configurable port for mistral models server
#122
biswaroop1547
closed
12 months ago
0
cht-llama-cpp/build-darwin: use venv <= virtualenv
#121
casperdcl
closed
12 months ago
2
make internal port configurable
#120
casperdcl
opened
1 year ago
0
cht-petals: new script and changes for managing installation and starting service without docker
#119
biswaroop1547
closed
12 months ago
0
[pre-commit.ci] pre-commit autoupdate
#118
pre-commit-ci[bot]
closed
12 months ago
0
Mistral-7B_q5, mac M1/M2 binaries
#117
filopedraz
closed
1 year ago
0
ci: fix schedule
#116
casperdcl
closed
1 year ago
0
Compile Petals services
#115
filopedraz
opened
1 year ago
0
Enable streaming response for Petals services
#114
filopedraz
closed
11 months ago
0
Download Petals models from Torrent
#113
filopedraz
opened
1 year ago
2
ci: fix PR cancellations
#112
casperdcl
closed
1 year ago
0
docker build speedups & memory reduction
#111
casperdcl
closed
1 year ago
3
Using float32 on all archs for petals clients
#110
filopedraz
closed
12 months ago
1
Serve Xwin LM
#109
filopedraz
closed
1 year ago
0
Fix petals on amd
#108
filopedraz
closed
1 year ago
0
CI: check PRs against base branch
#107
casperdcl
closed
1 year ago
0
ci: debug external trigger
#106
casperdcl
closed
1 year ago
2
ci: fix PR builds
#105
casperdcl
closed
1 year ago
0
Added petals models
#104
filopedraz
closed
1 year ago
7
Serve uncensored Models
#103
filopedraz
closed
1 year ago
0
Serve Stable Audio
#102
filopedraz
closed
1 year ago
0
Vicuna and GPT4All not working on arm
#101
filopedraz
closed
1 year ago
0
[pre-commit.ci] pre-commit autoupdate
#100
pre-commit-ci[bot]
closed
1 year ago
0
Next