In this project, we developed a lightweight web service app wherein different modules -integrated with existing ones- detect the available faces in the input image. Later they will evaluate the input image and compare it with the user's provided images stored in the database. The output includes the name, coordinations, sentiment of each extracted face, and besides a brief description of the scene's context.
To work with this app, right off the bat, the user should create an account so, he can upload some images that he wants to be compared with the image he wishes to describe. Then, whenever he wishes to evaluate or describe an image, he can easily input the image's URL and wait for the result to appear on the screen. In order to manipulate the faces, I used the deepface framework. Besides, to predict and describe the context of the image I made use of CA⫶TR: Image Captioning with Transformers implemented in this repository.
You can also find a short clip illustrating the appearance of the app here.
git clone https://github.com/behzadshomali/Image-Describe-Pipe.git
cd Image-Describe-Pipe/
python3 -m venv <VIRTUALENV_NAME> ; source <VIRTUALENV_NAME>/bin/activate
pip3 install -r requirements.txt
Run the container associated with the database
Run the flask server on your local machine:
python3 main.py
127.0.0.1:5000/home