-
SentenceVAE/
│
├── encoder.py
│ ```python
│ import torch
│ from torch import nn
│
│ class SentenceEncoder(nn.Module):
│ '''Sentence Encoder with byte-level BPE tokenization, lear…
-
Why do I encounter a situation where the sentence list does not match the encoding list when I use start_rultisprocess_pool() to start the process pool and then start Python multithreading
eg:
batch…
-
concurent request number:1
code
`import math
from typing import List
import time
import threading
import ctypes
import collections
import numpy as np
import tensorrt as trt
import pycu…
-
## 🐛 Bug
Hi, I tried to test nllb for translating some English sentences to Chinese, and all my sentences are less than 60 tokens. However, most of sentences which more than 30 tokens cannot be gen…
-
Hello.. Highly appreciate this library for NLP related tasks.
However I am trying to encode long sentences and it seems to take more than 24 hours and longer for encoding 2Million records using mo…
-
Currently I need to load a tokenizer from HuggingFace, and use it for simply encoding and decoding sentences. While doing that from Transformers.jl interface is awkward already (I had to go `tok = Tra…
-
embedding model前向推理时,不同batchsize,相同文本,输出embedding小数点后几位不同,请问这是什么原因导致的?理论上相同的输入应该会有完全相同的输出才对?
使用的是示例代码:
```
sentences_1 = ["样例数据"]
sentences_2 = ["样例数据", "样例数据", "样例数据", "样例数据"]
model_path='…
-
I'd like to use GPT to encode my dataset and use the representations further for the task of question generation. I have problems with understanding the code and the name of the arguments in the train…
-
When running the test program, I encountered the following error message. It seems to be related to the use of `multiprocessing` in the [sentence_transformers](https://huggingface.co/sentence-transfor…
-
The SentenceTransformers [Documentation](https://www.sbert.net/examples/applications/computing-embeddings/README.html?highlight=model%20encode) says that multiple sentences are also accepted while enc…