Closed adam-funk closed 1 year ago
I managed to get the correct versions by creating a new conda environment and just doing pip install -r requirements.txt
, but I still get exactly the same dataloader error.
Hi, you're using the --fast option which expects a "pickled" dataset. Try without this option, then the images should be loaded separately from disk.
Hi
I did conda run -n simplehtr1 python create_lmdb.py --data_dir "${training_directory}"
first (where $training_directory
is the one containing gt/words.txt
and img
). I thought that should allow --fast
to work?
(Anyway, I'm getting another error further along without --fast
but I'll see what I can do with it.)
Thanks Adam
check if the lmdb was actually created: in the IAM dataset folder there are the subfolders gt, img and lmdb, and lmdb should contain a (quite large) data.mdb file.
Further, it might be due to broken images, as OpenCV loads broken images as None (instead of throwing an exception).
In create_lmdb.py you can put an assert img is not None
after the cv2.imread to check if this is the case.
Hi, I'm trying to train SimpleHTR on our own dataset of handwritten two-digit numbers. I think I have set the training data files up correctly:
and the
words.txt
file looks like thisFor the initial test on my laptop (to move onto an HPC server later) I have 255 images. I'm using the following command:
The console shows output from
0 255
to254 255
, so I think it is finding the training data files, but it fails with this exception:I admit that the environment has >= versions of the packages in requirements.txt because I can't get conda to install the exact versions.
Am I doing something wrong? Any idea how to fix this?
Thanks Adam