Open tomruarol opened 5 years ago
Thank you for your post. We noticed you have not filled out the following field in the issue template. Could you update them if they are relevant in your case, or leave them as N/A? Thanks. What is the top-level directory of the model you are using Have I written custom code OS Platform and Distribution TensorFlow installed from TensorFlow version Bazel version CUDA/cuDNN version GPU model and memory Exact command to reproduce
Ok, I managed to move on a bit by skipping the use of the glob
library in the dataset_builder.py
script under the route research/object_detection/builders/
. It is not working as it should, so by just removing the use of it the script runs a bit ahead, but it still throws and error:
NotFoundError (see above for traceback): /tensorflow/models/research/tmp/mscoco/coco_train.record-00001-of-00010; No such file or directory
[[node IteratorGetNext (defined at object_detection/model_main.py:105) = IteratorGetNext[output_shapes=[[128], [128,300,300,3], [128,2], [128,3], [128,100], [128,100,4], [128,100,2], [128,100,2], [128,100], [128,100], [128,100], [128]],
output_types=[DT_INT32, DT_FLOAT, DT_INT32, DT_INT32, DT_FLOAT, DT_FLOAT, DT_FLOAT, DT_FLOAT, DT_INT32, DT_BOOL, DT_FLOAT, DT_INT32], _device="/job:localhost/replica:0/task:0/device:CPU:0"](IteratorV2)]]
I have not figured out how to move on from here.
I am trying to train a whole model based on the COCO dataset using this scripts provided but reducing the number of classes to only 6.
I run the
download_and_preprocess_coco.sh
script which downloads the dataset and calls thecreate_coco_tf_record.py
script which creates the TFRecords from the dataset previously downloaded. After that steps (successfully achieved) I try to run theretrain_detection_model.sh
as it is described in the tutorial, but modifying the labels .pdtxt file in order to take into account only 6 clases and modifying thepipeline.config
file in order to achieve the same (with a v2 net and training the whole model option).The first error that came out was:
RuntimeError: Did not find any input files matching the glob pattern [u'/tensorflow/models/research/tmp/mscoco/coco_train.record-00001-of-00010']
When I do have a file under:
/tensorflow/models/research/tmp/mscoco/
which contains files of the following format:Being the first set of 5 numbers after the record part numbers that go from 00000 to 00099.
So I do have those files that the error reports I do not have, and I have the PATH specified in the pipeline.config file.
I have narrowed down the problem to this part of the
/research/object_detection/utils/config_util.py
script:It crashes in the Merge call. print(pipeline_config) returns nothing. I guess the object is empty or it does not merge and crash if it doesn't find
num_layers
in the pipeline_config object.I paste my
pipeline.config
file: