-
Hi
I'm trying to impute summary-stats data. I could run properly the harmonization pipeline, but it give me an error when I run the imputation script on chr1. I'm using Python/3.8.2-GCCcore-9.3.0 a…
-
Parquet file import problem
The test file is displayed in the Parquet file opening tools without any problems
But in Zui, in the Format section, even though the Parquet file is selected, the imp…
-
### Describe the enhancement requested
Hello,
I may have misunderstood or misread the documentation but I didn't find it. I would love to have a field/column description just like we have in dat…
-
As mentioned in https://github.com/reichlab/variant-nowcast-hub/pull/116#issuecomment-2427130801:
> If we coerce to `pd.NA` (or `None`), the corresponding dtype that the validation tool receives is…
-
Columns schemas with `...encoding=DELTA_BINARY_PACKED, valuerepetitiontype=OPTIONAL"` cannot be read by other parquet tools if they contain all nulls. Tested with pyarrow and parquet-mr (java). This g…
bkief updated
2 years ago
-
Hello,
Copying the issue from the previous repo (https://github.com/segmentio/parquet-go/issues/483), just tested against 0.19 and the issue is still present.
---
I'm seeing a weird situation which…
-
After I execute
```
cargo run --release -- generate --benchmark tpcds \
--scale 1000 \
--partitions 48 \
--generator-path /path/to/DSGen-software-code-3.2.0rc1/tools \
--output /tmp/tpc…
-
#### Inspirational work
- [allo-montreal](https://github.com/jstonge/allo-montreal/tree/main/src) is a small project in observable framework where the backend is in python, and the charting is in ja…
-
Snappy compression is not working when writing to parquet for integer type data.
E.g. compare file sizes for:
```r
write_parquet(data.frame(x = 1:1e6), "snappy.parquet", compression = "snappy"…
-
I have a parquet file with two columns (int64 and double) and 9 million rows. The parquet tools (parquet, parquet-reader, parquet-schema...) read it perfectly. (I have many files, actually, but they a…