-
### Your current environment
The output of `python collect_env.py`
```text
Your output of `python collect_env.py` here
```
### 🐛 Describe the bug
(base) bob@test-ESC8000A-E11:~$ python…
-
> > 我也是bert跑出来准确率50左右
> > 为什么我按照源代码 原数据集训练出来的准确率 和你的差很多呢 而且时间也很长
> > 我检查了下是词表出现了乱码,你可以看看bert里vocab.txt这个文件打开是不是正常的汉字,erine是正常的
> > 我看了一下 我bert里vocab.txt这个文件中的文字跟作者放的链接里面的是一样的 但是我觉得这个中文有点儿奇怪
>
> …
-
### Error Message
INFO:root:converting to Gluon checkpoint ...
Traceback (most recent call last):
File "convert_tf_model.py", line 159, in
assert len(tf_config) == len(tf_config_names_to_gl…
-
Hi,
I am doing some task,that is very similar to your task, expect its inputs and outputs are chinese . The frame is also seq2seq. I write code as same as yours. When I run the code, the train accura…
-
- UBUNTU16.04
- PYTHON3.6
```
# of word in train: 55304:
# of n-gram in memory: 71499
Traceback (most recent call last):
File "wmseg_main.py", line 677, in
main()
File "wmseg_main…
-
I wonder what the vocab size of spiece.model (seems 32k)? I am trying to improve this part, could anyone share the vocab size of spiece.model?
Besides, what is the data size which trained to get spi…
-
when i want to generate text i get a memory error
`Traceback (most recent call last):
File "rnn_tf.py", line 300, in
main()
File "rnn_tf.py", line 221, in main
data, vocab = load_da…
-
我用笔记本电脑的GTX1070(显存8G),训练了一个虎嗅新闻语料,约110MB,也能跑,每一步保存模型,时间间隔是5分钟。想看看大家的机器及训练效率,欢迎在下方留言探讨。我的参数设置如下:
{
"architectures": [
"GPT2LMHeadModel"
],
"attn_pdrop": 0.1,
"bos_token_id": 0,
"do…
-
https://github.com/golang/go/issues/5763#issue-51284151 observes “It is very strange to use, say Z成本 or Jぶつける as identifiers.” In that issue we discussed potentially changing the default export rule, …
-
And can this model be helpful on Chinese dataset?