-
I see the code in utils.py
idx_train = range(140)
idx_val = range(200, 500)
idx_test = range(500, 1500)
I cannot understand it, because the ratio of they may be 6:2:2 if the dataset is …
-
### Describe the bug
Does PaDiM model require validation and testing?
notebooks/500_use_cases/501_dobot/501a_training_a_model_with_cubes_from_a_robotic_arm.ipynb
when I see the above example co…
-
Hi Dr. Ge,
I am working on the PRScsx integrating target data (AFR, n=2000) and base data from EUR and AFR). I am not sure I fully understand how to split the target dataset into validation and tes…
-
### Describe the bug
We build a dataset which contains several hdf5 files and write a script using `h5py` to generate the dataset. The hdf5 files are large and the processed dataset cache takes more …
-
Hi!
I would like to follow your excellent work, but found out that there are only the split files for 5 datasets provided in the download link. Can you please provide the split files for the other …
-
**Why in the following prompt of LLama 3-8b tutorial notebook (https://colab.research.google.com/drive/135ced7oHytdxu3N2DNe1Z0kqjyYIkDXp?usp=sharing):**
_alpaca_prompt = """Below is an instruction …
-
**Description**
I believe the `scripts/generate-benchmark-lean4.ipynb` is buggy. I evaluated the performance of the ReProver premise retriever on a dataset I generated from Mathlib4. I should get a…
-
I uploaded this train code
import os
os.environ["TF_CPP_MIN_LOG_LEVEL"] = "2"
import numpy as np
import cv2
from glob import glob
from sklearn.model_selection import train_test_split
import t…
-
```
datasets = ['hotpotqa', '2wikimqa', 'musique', 'narrativeqa', 'qasper', 'multifieldqa_en', 'gov_report', 'qmsum', 'trec', 'samsum', 'triviaqa', 'passage_count', 'passage_retrieval_en', 'multi_new…
-
Hello,thanks for your excellent work! I would like to know if you use the scaffold-based splitting approach for all the 6 datasets of MoleculeNet? Or some datasets are randomly split?