As these scripts are used more and more by a researcher, there will be duplicate projects saved to lgtm.com. lgtm.com can properly handle duplicate projects (returning an error stating the project is already saved [I assume that's the case?]). This results in unnecessary HTTP requests which costs us time.
Potential Solution
Adding a local cache system that caches every project name that was saved to lgtm.com will allow us to reduce the number of HTTP requests sent to lgtm.com.
Issues With This Issue
I actually don't know how much of this will be a problem. I don't have enough experience working with these scripts. But at the most basic level, this makes some sense to me.
What's the Problem
As these scripts are used more and more by a researcher, there will be duplicate projects saved to lgtm.com. lgtm.com can properly handle duplicate projects (returning an error stating the project is already saved [I assume that's the case?]). This results in unnecessary HTTP requests which costs us time.
Potential Solution
Adding a local cache system that caches every project name that was saved to lgtm.com will allow us to reduce the number of HTTP requests sent to lgtm.com.
Issues With This Issue
I actually don't know how much of this will be a problem. I don't have enough experience working with these scripts. But at the most basic level, this makes some sense to me.