Open phanchutoan opened 5 years ago
This error seem to be solved after i add K.clear_session() before return
of method detectCustomObjectsFromImage
in /imageai/Detection/__init__.py
The error is occurring because for each inference, you load the detection model. Although , the K.clear_session() function does reset the detection instance and prevents the error, there will still be latency for each time the model loads.
What you can do is create the detection instance as a global value and load the model outside the function.
Thanks for your respond.
I tried to put the detection instance outside of the funtion, but it raise another error Ensure you specified correct input image, input type, output type and/or output image path
.
I tried to check the input file with
check = os.path.isfile(image_path)
, it returned true so may be it is not the problem of input not correct.
Traceback (most recent call last):
File "/Users/ptoan/Desktop/object-detection/venv/lib/python3.6/site-packages/flask/app.py", line 2309, in __call__
return self.wsgi_app(environ, start_response)
File "/Users/ptoan/Desktop/object-detection/venv/lib/python3.6/site-packages/flask/app.py", line 2295, in wsgi_app
response = self.handle_exception(e)
File "/Users/ptoan/Desktop/object-detection/venv/lib/python3.6/site-packages/flask/app.py", line 1741, in handle_exception
reraise(exc_type, exc_value, tb)
File "/Users/ptoan/Desktop/object-detection/venv/lib/python3.6/site-packages/flask/_compat.py", line 35, in reraise
raise value
File "/Users/ptoan/Desktop/object-detection/venv/lib/python3.6/site-packages/flask/app.py", line 2292, in wsgi_app
response = self.full_dispatch_request()
File "/Users/ptoan/Desktop/object-detection/venv/lib/python3.6/site-packages/flask/app.py", line 1815, in full_dispatch_request
rv = self.handle_user_exception(e)
File "/Users/ptoan/Desktop/object-detection/venv/lib/python3.6/site-packages/flask/app.py", line 1718, in handle_user_exception
reraise(exc_type, exc_value, tb)
File "/Users/ptoan/Desktop/object-detection/venv/lib/python3.6/site-packages/flask/_compat.py", line 35, in reraise
raise value
File "/Users/ptoan/Desktop/object-detection/venv/lib/python3.6/site-packages/flask/app.py", line 1813, in full_dispatch_request
rv = self.dispatch_request()
File "/Users/ptoan/Desktop/object-detection/venv/lib/python3.6/site-packages/flask/app.py", line 1799, in dispatch_request
return self.view_functions[rule.endpoint](**req.view_args)
File "/Users/ptoan/Desktop/object-detection/app.py", line 15, in hello
func()
File "/Users/ptoan/Desktop/object-detection/app.py", line 32, in func
detections = detector.detectCustomObjectsFromImage(custom_objects=custom_objects, input_image=image_path)
File "/Users/ptoan/Desktop/object-detection/venv/lib/python3.6/site-packages/imageai/Detection/__init__.py", line 860, in detectCustomObjectsFromImage
"Ensure you specified correct input image, input type, output type and/or output image path ")
ValueError: Ensure you specified correct input image, input type, output type and/or output image path
I am also trying to use imageai for object detection and facing same issue when creating detection instance as global variable. If I use K.clear_session()
then time taken for object detection is increased.
I seem that it is not working on multiple thread.
I am having the same issue between Flask and imageAI when creating the ObjectDetection
object as a global variable. It raises
"ValueError: Ensure you specified correct input image, input type, output type and/or output image"
when calling detectCustomObjectsFromImage()
with input_type=array
, output_type=array
, the custom object and passing a numpy array as the image.
Interestingly, the model I'm loading is yolov3
, so it must not be model dependant.
Upon closer inspection, it appears to raise this exception when self.sess.run
is called with the feed_dict
, which appears to raise a TypeError exception
specifying that the feed_dict key can not be interpreted as a tensor.
As far as my understanding goes, when run in debug mode (which is how i'm running flask) it is run on a single thread by default. So I don't know if multi-threading is the issue, although there does seem to be a weird interaction between flask and ImageAI
I have the same problem, has someone solved it or found a workaround?
I had the same problem a week ago. and I had solved it by setting the flask run in just one thread. the following is my example:
from flask import Flask, Response, jsonify
app = Flask(__name__)
import os
from imageai.Detection import ObjectDetection
import time
import json
execution_path = os.getcwd()
st = time.time()
detector = ObjectDetection()
detector.setModelTypeAsRetinaNet()
detector.setModelPath(os.path.join(execution_path, "model", "resnet50_coco_best_v2.0.1.h5"))
# detector.setModelTypeAsTinyYOLOv3()
# detector.setModelPath(os.path.join(execution_path, "model", "yolo-tiny.h5"))
detector.loadModel()
# detector.loadModel(detection_speed="fastest")
print(f'Init Timer: {time.time()-st}')
@app.route('/detect/<pic_name>')
def boat_detection(pic_name):
st = time.time()
results = getDetections(pic_name)
print(f'Sum Timer: {time.time()-st}')
msg = {}
for i, result in enumerate(results, 1):
result['percentage_probability'] = float(result['percentage_probability'])
result['box_points'] = list(result['box_points'])
for index in range(len(result['box_points'])):
result['box_points'][index] = int(result['box_points'][index])
result['box_points'] = tuple(result['box_points'])
msg[str(i)] = json.dumps(result)
return jsonify(msg)
def getDetections(file_name):
start = time.time()
image_folder = os.path.join(execution_path, 'data\\ship2\\')
output_folder = os.path.join(execution_path, 'data\\output\\')
st1 = time.time()
image_file = os.path.join(image_folder, file_name)
new_image_file = os.path.join(output_folder, file_name)
print(image_file, "-->", new_image_file)
if not os.path.exists(image_file):
print("not exist.")
return
# global detector
custom_objects = detector.CustomObjects(boat=True)
detections = detector.detectCustomObjectsFromImage(
custom_objects=custom_objects,
input_image=image_file,
output_image_path=new_image_file,
minimum_percentage_probability=30)
print(f'[Info]识别到 boat{len(detections)}艘')
for eachObject in detections:
print(eachObject.items())
end = time.time()
print(f'Excute Timer: {end-st1}')
print ("耗时: ",end-start)
return detections
if __name__ == '__main__':
app.run(threaded=False) # ban the Threaded mode
the most important points is set the app.run(threaded=False). my packages version: Flask==1.1.1 h5py==2.9.0 Keras==2.2.4 opencv-python==4.1.0.25 tensorflow==1.14.0 imageai==2.1.0
I have another solution to solve this problem. Recently, I find a param which is thread_safe, it can share the model with thread by using k.session.graph in source code. We need to set it to True. So we can use the threading actually in flask or in other place with threading. And my before code can change like the following:
from flask import Flask, Response, jsonify
app = Flask(__name__)
import os
from imageai.Detection import ObjectDetection
import time
import json
execution_path = os.getcwd()
st = time.time()
detector = ObjectDetection()
detector.setModelTypeAsRetinaNet()
detector.setModelPath(os.path.join(execution_path, "model", "resnet50_coco_best_v2.0.1.h5"))
# detector.setModelTypeAsTinyYOLOv3()
# detector.setModelPath(os.path.join(execution_path, "model", "yolo-tiny.h5"))
detector.loadModel()
# detector.loadModel(detection_speed="fastest")
print(f'Init Timer: {time.time()-st}')
@app.route('/detect/<pic_name>')
def boat_detection(pic_name):
st = time.time()
results = getDetections(pic_name)
print(f'Sum Timer: {time.time()-st}')
msg = {}
for i, result in enumerate(results, 1):
result['percentage_probability'] = float(result['percentage_probability'])
result['box_points'] = list(result['box_points'])
for index in range(len(result['box_points'])):
result['box_points'][index] = int(result['box_points'][index])
result['box_points'] = tuple(result['box_points'])
msg[str(i)] = json.dumps(result)
return jsonify(msg)
def getDetections(file_name):
start = time.time()
image_folder = os.path.join(execution_path, 'data\\ship2\\')
output_folder = os.path.join(execution_path, 'data\\output\\')
st1 = time.time()
image_file = os.path.join(image_folder, file_name)
new_image_file = os.path.join(output_folder, file_name)
print(image_file, "-->", new_image_file)
if not os.path.exists(image_file):
print("not exist.")
return
# global detector
custom_objects = detector.CustomObjects(boat=True)
detections = detector.detectCustomObjectsFromImage(
custom_objects=custom_objects,
input_image=image_file,
output_image_path=new_image_file,
minimum_percentage_probability=30,
thread_safe=True) # **notice this param**
print(f'[Info]识别到 boat{len(detections)}艘')
for eachObject in detections:
print(eachObject.items())
end = time.time()
print(f'Excute Timer: {end-st1}')
print ("耗时: ",end-start)
return detections
if __name__ == '__main__':
app.run(threaded=True) # we can use the threading to detect
the most important points is set the thread_safe=True. If you do that, you will find we can load model just one time and can use in many threads.
@xiluzi Both solution worked.
Hi, do you know a solution for the prediction procedure? I'm trying to do the same as you do but with the prediction function and its throwing me the same error.
@dntzbgh Both of the above mentioned solutions by xiluzi solves it.
@Akashtyagi08 yes, it solves it for the detection function. I was asking if anyone tried to do the same for a prediction since I'm trying to use both in a flask app.
Yes, when using prediction and Detection collectively in the same api, the issue still persist. I tried few workaround but were unsuccessful.
@xiluzi can you please help me in understanding the issue why its throwing error when I try to call detectObjectsFromImage method from flask api method without loading model again.
If I load the model in api post method and call for detectObjectsFromImage method, its giving results with a delay of loading model time.. but I m trying to reduce this delay of loading model time in each API call.
Please refer https://github.com/OlafenwaMoses/ImageAI/issues/531 to see my code.
please help in understand why its not working in my code but working fine in your above code.
thanks in advance..
I am having an issue when trying to implement ObjectDetection as an api with Flask framework
My source code is just the simple use of ObjectDetection
Problem is the api only success on the first request. From second request, it comes with following error:
TypeError: Cannot interpret feed_dict key as Tensor: Tensor Tensor("Placeholder:0", shape=(3, 3, 3, 16), dtype=float32) is not an element of this graph.
Full stack trace: