This time I decided to create something bigger, with more depth to it in order to learn more.
My main concept was to create a product (mock-up SaaS) that would allow users to send their captcha images and receive solutions in simple text format.
In order to achieve this I had to solve several problems and divide my project into 4 parts.
Working on this project was very fun and fulfilling. I had many opportunities to work closer and explore more deeply things like creating GUI, design patterns, asynchronous programming and finally Docker and docker-compose with AWS and EC2.
Firstly I selected 2 example captcha types and with Scrapy I needed to gather enough data to have a learning dataset for my model.
After solving some problems with monkey patching MediaPipeline it went smoothly.
After I gathered captcha images I had to come up with a better way of renaming them than the ones that are built in operating systems because of dataset sizes.
I decided to build an application with the help of Tkinter and Pillow. The app had several tools to make renaming captcha images easier and faster. Amongst others, it helped user with faster switching between images after assigning a new name, had options to increase or decrease size and contrast or convert captcha to black and white mode.
When all captcha images were renamed with proper solutions, it was time to actually build and save Machine Learning models with Tensorflow and Keras.
I needed three models, one to distinguish between captcha types and two for predicting proper solutions of those captcha types.
In this task, the book "Neural Network Projects with Python" written by James Loy and website keras.io with articles by A_K_Nain were really helpful.
The final part of the whole project. I decided to go with Fast API for this task to also explore deeper asynchronous programming and PostgreSQL with SQLAlchemy for database tasks. For security, I used oauth2. I wanted to create an API so every potential "user" could get their captcha solved. I created several endpoints for things like, solving captcha, adding, deleting, viewing info about users and adding credits.
API was accepting CURL requests formdata with captcha images, which were processed by saved ML models from previous step and returning a solution. I also decided to write tests using pytest to make my life easier.
I Dockerized the whole API alongside PostgreSQL and PgAdmin using docker-compose.
As a final touch, I decided to host it on AWS with EC2.
Example captcha images to solve:
Test user: Username - test
Password - test
Admin user: Username - admin
Password - admin
(for deleting users and adding more credits)
Solve Captcha:
curl -X POST '[URL]/upload_captcha' --header 'Authorization: Bearer [YOUR_TOKEN]' -F captcha_image=@/directory/file.jpeg"
Add User:
curl -X POST '[URL]/add_user' -H 'accept: application/json' -H 'Content-Type: application/json' -d '{\"username\": \"[YOUR_USERNAME]\", \"email\": \"[YOUR_EMAIL]\", \"password\": \"[YOUR_PASSWORD]\"}'
Get Token:
curl -X POST '[URL]/get_token' -F username=[YOUR_USERNAME] -F password=[YOUR_PASSWORD]
Add Credits:
curl -X POST '[URL]/add_credit_balance' -H 'accept: application/json' -H 'Content-Type: application/json' --header 'Authorization: Bearer [ADMIN_TOKEN]' -d '{\"user_id\": [ID], \"credit_amount\": [CREDIT_AMOUNT]}'
Get Account Info:
curl -X POST '[URL]/get_account_info' --header 'Authorization: Bearer [YOUR_TOKEN]'
Delete User:
curl -X POST '[URL]/delete_user?delete_user_id=[ID]' --header 'Authorization: Bearer [ADMIN_TOKEN]'
(you need to have docker and docker-compose installed on your machine)
docker-compose-development.yml
.env
cd
into download directorydocker-compose -f docker-compose-deployment.yml --env-file env up
command to build and start imagesPreviously, API was also available online, hosted on AWS EC2.
Some of the technologies used - Scrapy, Tkinter, Pillow, Py2app, Tensorflow, Keras, FastAPI, PostgreSQL, SQLAlchemy, pytest, Docker, docker-compose, AWS, EC2
It was really fun to work on this.
Thank You :-)