-
### Backend
VL (Velox)
### Bug description
```
import testImplicits._
Seq((1, 2)).toDF("Товары", "овары").write.mode("overwrite").parquet("tmp/t1")
spark.read.parquet("tm…
-
**Describe the problem you faced**
A flink write hudi job, we have hdfs jitter, cause flink task to fail over, and see this error
**To Reproduce**
Steps to reproduce the behavior:
*have ch…
-
### What happens?
I'm trying to use in DuckDB Pandas object that has string column filled with null values.
Unfortunately DuckDB is doing implicit conversion of this column to type INTEGER.
###…
-
### API
Other
### Description
I have this scenario where I have a CSV/Parquet file with just two columns:
```
product price
Laptop 1100.0
Mouse 150.0
Headphones 250.0
```
In my DDL, I want…
-
EDIT: SEE NEW COMMENTS.
When we read a parquet file that was written with to_parquet (ibis) and contains a geometry column, it reads it back as binary. This is a bug on our end, as this doesn't ha…
-
(from helpdesk)
I just noticed that the [FAQ for download formats](https://www.gbif.org/faq?question=download-formats#:~:text=Occurrence%20data%20can%20be%20downloaded%20in%20the%20following,distin…
-
Ensemble.from_parquet has evolved to do a lot of TAPE-specific things, from setting up the column mapper to setting the index on the chosen id column. As a result, there's some friction between the fu…
-
**Describe the bug**
If the `ParquetMetadataReader` tries to read metadata written by `ParquetMetaDataWriter` without first loading the page indexes, you get an error like "missing required field Co…
-
Verify that the Parquet docs are up-to-date. We've had quite a few updates to Parquet, so we should review the existing Parquet docs for accuracy.
Verify that the doc for dedicated columns is presen…
-
** Rational **
Large datasets are now often stored in the Parquet format (https://parquet.apache.org). In principle, other packages can read efficiently data in the parquet format.
**Related**
…
wfthi updated
3 weeks ago