Image recognition services provide different labels for the same images, with differences broadly either being different word choices for the same concept or not seeing same objects on both images. Cross-service Laber Agreement Score (COSLAB) is a method to evaluate the semantic similarity on image labels across several image recognition systems. It first produces each image a set of labels, one for each service. Then it uses word embedding models to find the closest match across label sets to see how much cross-service agreement there are for each label.
Our work further highlights that image recognition services differ on what they see on images. Due to this, we suggest approaching them as interprentations instead of objective truths. In response, we propose two different strategies which you can employ to use them in your research:
from coslab import aws
from coslab import googlecloud
from coslab import azure_vision
from coslab import taggerresults
from coslab import tag_comparator
## establishing a container for all results
results = taggerresults.TaggerResults()
## establish classifiers
amazon = aws.AWS(api_id="", api_key="", api_region="")
google = googlecloud.GoogleCloud(service_account_info="")
azure = azure_vision.Azure(subscription_key="", endpoint="")
amazon.process_local( results, "image.png")
google.process_local( results, "image.png")
azure.process_local( results, "image.png")
results.export_pickle("image.pickle")
tag_comparator.compare_data( results )
COSLAB is available via Pypi, so you can
python3 -m pip install coslab-core
We generiously thank C. V. Åkerlund Foundation for supporting this work.