-
Hi,
Thank you for your amazing work!
We'd like to add an embedding template for users to deploy on RunPod, and we're deciding between Infinity and [HF's Text Embedding Inference](https://github…
-
### Descripe the feature and motivation
We develop a newSelf-Calibrated Illumination (SCI) learning framework for fast, flexible, androbust brightening images in real-world low-light scenarios. To be…
-
### 🚀 The feature, motivation and pitch
I just stumbled on https://twitter.com/DrJimFan/status/1615018393601716224, there is https://github.com/NVlabs/tiny-cuda-nn which fuses small MLPs for fast tra…
-
**Background**:
The KubeEdge SIG AI is chartered to facilitate Edge AI applications with KubeEdge. An overview of SIG AI activities can be found in this [charter](https://github.com/kubeedge/communit…
-
Can you provide a inference script.
-
Many tabular datasets contain an "ID" integer column that is useless as a predictive feature. Such features are often actively harmful to include as predictors (eg. because they take larger values in…
-
The current implementation of `DalleBart._sample()` performs "super conditioning" by running `DalleBart.decode()` twice in a row:
https://github.com/borisdayma/dalle-mini/blob/a72705f46ce29a45d1c56…
-
We have currently received several requests (#112, #110, #97) to run the SPHINX inference on GPUs with smaller memory. We also believe that fitting it into the 24GB memory bar benefits a broad range o…
-
Hi, can you tell me the way you calculate the latency of the model? Do you use the benchmark.py or other codes? Thanks.
-
When I perform the command "sh slurm_inference.sh test_video mp4 24 smpler_x_h32" which use python command
`python inference.py \
--num_gpus ${GPUS_PER_NODE} \
--exp_name output/demo_${JOB…