-
@LaurentMazare
How can I use candle for a cross-encoder from sentence-transformers models (msmarco models: e.g. msmarco-distilroberta-base-v3)?
Does it require differents stack of implementation …
-
I finished rendering and when I was ready to train nerf, I only used 20 data sets and found out that I needed quite a lot of memory. What happened? I need your help。
(instantmesh1) mrguanglei@guang…
-
I'm encountering an issue with the dimensions of the text encoder output in a fine-tuned CLIP model. The fine-tuning output of my CLIP model based on RN50 is (1, 1024), whereas the output from CLIPTex…
-
(base) root@autodl-container-aa9a42a072-5939b670:~/autodl-tmp# conda activate llama-omni
(llama-omni) root@autodl-container-aa9a42a072-5939b670:~/autodl-tmp# python -m omni_speech.serve.model_worker…
-
# 🌟 New model addition
## Model description
Recently Google is published paper titled ["Beyond 512 Tokens: Siamese Multi-depth Transformer-based Hierarchical Encoder for Long-Form Document Matchin…
-
Hi!
After replacing an eight-layer Transformer encoder with Mamba, the training loss fails to decrease. Could it be that Mamba doesn't perform as effectively as the Transformer in the diffusion model…
-
Hi,
I'm interested in understanding what the code does.
```
**easyblock("model.diffusion_model.output_blocks.6.0", "P_bg208","P_bg209"),
**conv("model.diffusion_model.output_blocks.6.0…
-
### Describe the bug
I am using the training script documented here https://github.com/huggingface/diffusers/blob/main/examples/dreambooth/README_flux.md to train a LORA on my dataset.
here is th…
-
### Current Behavior
Using the default onnx model,
**Score function**
```
def get_score(a, b):
return evaluation.evaluation(
{
'question': a
},
{
…
-
I am trying to import the aforesaid model using the following command:
```
eland_import_hub_model --url --hub-model-id dunzhang/stella_en_400M_v5 \
--task-type text_embedding --es-username e…
ivssh updated
11 hours ago