issues
search
neuralmagic
/
deepsparse
Sparsity-aware deep learning inference runtime for CPUs
https://neuralmagic.com/deepsparse/
Other
3.01k
stars
176
forks
source link
issues
Newest
Newest
Most commented
Recently updated
Oldest
Least commented
Least recently updated
[deepsparse.benchmark] enable benchmarking internal kv cache
#1308
bfineran
closed
1 year ago
0
Add support for pulling down ARM and MacOS artifacts
#1307
mgoin
closed
1 year ago
0
Fixes for GHA quality
#1306
mgoin
closed
1 year ago
0
Add basic LLM docs
#1305
mgoin
closed
1 year ago
0
Add `deepsparse[llm]` extras alias
#1304
mgoin
closed
1 year ago
0
[bugfix] enforce that all transformer deployement files are downloaded by pipeline
#1303
bfineran
closed
1 year ago
0
Change to requiring the vanilla transformers package
#1302
mgoin
closed
1 year ago
2
Support for SentenceTransformers with `deepsparse.sentence_transformers.SentenceTransformer`
#1301
mgoin
closed
1 year ago
0
[server] Update OpenAI Model Support
#1300
dsikka
closed
1 year ago
0
Decrease prompt_sequence_length default to 16
#1299
mgoin
closed
1 year ago
0
Make `Pipeline` printable
#1298
mgoin
closed
1 year ago
0
Llm docs
#1297
robertgshaw2-neuralmagic
closed
1 year ago
0
[bugfix] model_to_path should also download .data files from zoo
#1296
bfineran
closed
1 year ago
0
Make `generate_random_inputs` use zeros for integers
#1295
mgoin
closed
1 year ago
0
[TextGeneration][Bug] Fix timer
#1294
dsikka
closed
1 year ago
0
Revert #1263
#1293
SageMoore
closed
1 year ago
0
[TexgGeneration][Bug] streaming skips generated token added from last prompt token
#1292
dsikka
closed
1 year ago
0
fix top_p bug
#1291
horheynm
closed
1 year ago
0
remove NM_ALLOW_DARWIN
#1290
andy-neuma
closed
1 year ago
0
Set text-generation default values to match performance and UX needs
#1289
mgoin
closed
1 year ago
0
[server] Refactor + OpenAI Chat Completion Support
#1288
dsikka
closed
1 year ago
1
Pipelines Refactor - Initial Impl
#1287
bfineran
closed
1 year ago
1
[TextGeneration] Fix max_length = 1, repeat token
#1286
dsikka
closed
1 year ago
0
Inference UX, accept input data
#1285
horheynm
closed
1 year ago
0
[TextGeneration] Add new `text_generation.md` with examples and tables with `generation_config` attributes
#1284
dsikka
closed
1 year ago
0
[BugFix] Error with streaming in Chat Pipeline
#1283
rahul-tuli
closed
1 year ago
0
"input_shape" should be "input_shapes"
#1282
Satrat
closed
1 year ago
0
[feature] Add-evaluate-function
#1281
rahul-tuli
closed
11 months ago
2
Fix headers in transformers/README.md
#1280
dbarbuzzi
closed
1 year ago
0
Replace Dead SparseZoo Stubs in Documentation
#1279
Satrat
closed
1 year ago
0
[TextGeneration] Add kwarg support for generation config attributes
#1278
dsikka
closed
1 year ago
0
Use nullcontext in nl_decoder_engine when timer_manager.current is None
#1277
mgoin
closed
1 year ago
1
[Chat pipeline] session context manager
#1276
bfineran
closed
1 year ago
0
[Cherry pick] #1262 explicitly set ORT providers on all InferenceSessions
#1275
rahul-tuli
closed
1 year ago
0
[Text Gen UX] top level constructor aliases + code gen subclass
#1274
bfineran
closed
1 year ago
0
Remove Dead Code
#1273
rahul-tuli
closed
1 year ago
0
[Feat] Add streaming to chatbot
#1272
rahul-tuli
closed
1 year ago
0
[Text Gen UX] default pipeline input to value
#1271
bfineran
closed
1 year ago
1
[Text Gen UX] Add `prompt` as input param alias
#1270
bfineran
closed
1 year ago
1
Dial back GenerationConfig logging and add from_dict
#1269
mgoin
closed
1 year ago
0
Default Config for Benchmark Pipeline
#1268
Satrat
closed
1 year ago
0
Make `test_pipeline_call_is_async` check greater than speedup
#1267
mgoin
closed
1 year ago
0
[Text Generation][KVCacheStorage] `ChatPipeline` implementation
#1266
dbogunowicz
closed
1 year ago
1
Add minor updates
#1265
rahul-tuli
closed
1 year ago
0
Raise limit of `test_pipeline_call_is_async` further
#1264
mgoin
closed
1 year ago
0
[Temporary Patch] ignore `empty` do not pass empty kv cache to runtime
#1263
bfineran
closed
1 year ago
0
explicitly set ORT providers on all InferenceSession creations
#1262
bfineran
closed
1 year ago
1
infer num_classes for yolov8-seg pipeline
#1261
bfineran
closed
1 year ago
1
Support loading HF repos with `hf:` stubs
#1260
mgoin
closed
1 year ago
0
benchmarker API
#1259
horheynm
closed
1 year ago
0
Previous
Next