-
I trained the model with my own data. My dataset size is 5M. After I trained the model, I ran the reconstruction code.
```python text.py --dataset [dataset] --decode_from [pretrained model path] -…
-
What would be the advisable way to set a minimum/maximum constraint for technologies per country?
I was thinking about using a `.csv` with the same structure as the `costs.csv`. This could than be …
-
Would be any chance that you will release the scripts for generating the blocks tower dataset?
-
hello,could you please tell me where can I get the data set?
-
While running CElegans - Multiple Channel Example.ipynb
Also had a quick look at the source, and could not track tfdecon down, but could be blind.
-
Multilingual documents are common in the computer age of today. Plethora of these
documents exist in the form of translations, books, operational manuals, etc. The
abundance of these multilingual do…
-
I have a dataset which has around 150.000 entries. Exploring SMOTHE sampling seems to be pretty slow as only a single core is used to perform calculations.
Am I missing a configuration property? How e…
-
The taxi ride dataset used in tests and benchmarks is starting to outlive its usefulness. Some issues:
- the dataset is unwieldy and takes up a bunch of disk space
- loading even a subset of 100M re…
-
-
Script identification is a crucial step in digitizing multilingual documents. In addition
to simplifying the OCR process, it could be used for many other tasks like document
indexing, routing, and c…