-
Check if there exist some RDF benchmarks, covering SPARQL queries and updates. Most current benchmark seem to focus on loading performance for a defined dataset dumps like dbpedia, but we want to figu…
-
Hello,
I am wondering learning results of `Ant-v2`'s dataset, like ant-random-v0/2 ant-medium-v0/v2. I think it is not listed in d4RL original paper but it's supplemented in github later.
Do y…
-
I made Benchmark AutoML libs, and TPOT showed very poor results, even worse than the usual CatBoost with standard parameters!
https://github.com/Alex-Lekov/AutoML-Benchmark/
I run the benchmark in d…
-
Is it possible to get a larger data set, say 2TB or 5TB? Testing on a 200GB data set that is easily compressible down to 50GB with modern compression algorithms might exclude disk IO from the equation…
-
## What's this paper about?
- Introduces 6 time series competitions held by Kaggle.
- Background: Real-life business forecasting tasks on Kaggle platform has been largely ignored by the academic …
-
- Measure HDF5 Sequential Performance for different chunk sizes and file sizes.
- Measure average read time and total read time for a fixed file and different chunk sizes
- Measure performance on Ar…
-
https://medium.com/radiant-earth-insights/creating-the-planets-digital-ecosystem-77a28fb76442
https://towardsdatascience.com/whats-growing-there-a5618a2e6933
https://www.mdpi.com/journal/remotesensi…
-
In LM Evaluation Harness, we work to match the "original" / "default" methods used to evaluate datasets. This includes using whatever answer extraction / post-processing is done by the original code i…
-
Hi,
It would be nice to have a 3rd column for data() output indicating whether the dataset can be used for regression or classification problems.
-
Hi all,
I met Tim (betatim) today who told me about this very interesting project, and I would like to contribute.
I have worked a bit on optimisation before, and have benchmarked some libraries lik…