Open charlottewright opened 1 year ago
This is related to the connection timing out after several minutes. breaking the list up into smaller files should be an effective workaround for now to save having to run them one at a time.
The current implementation stops searching if the connection breaks to save the API having to process abandoned queries, but I'll need to implement some way to save the results to be collected when the query has finished in genomehubs/genomehubs to make it possible to fetch large results from the CLI, API and UI.
@rjchallis do think this would be solved in part if the requests were made serially instead of concurrently?
Would make the progress bar issue (#22) easier to solve too.
I've been thinking about this one. We can make this much more efficient by combining multiple taxa into the same search. @rjchallis is there any constraint on number of taxa in URL requests apart from URL length (< 2048 chars)? Not sure how to do this yet, but we could check number of taxa on CLI/in a file and then make fewer requests but with more taxa per request.
I'll see how to implement this.
Hello!
When running
goat-cli taxon search
with a file as input I get an error for each query species which does not exist in goat. For species that do return a hit, the output is as expected. The error is:When running such species individually with
-t
flag, I get just the header output with no hits as expected so it seems the error is related to using a file with many queries.The exact command I run is:
goat-cli taxon search -f $i --assembly --country-list --karyotype
Thanks! :)
Charlotte