-
Hey,
Currently we're trying to validate data stored in S3 Parquets, that are exposed via Glue/Athena. Currently, we can download the parquets, load it into duckdb, then use the DuckDB connector to do…
-
### Apache Iceberg version
1.5.2
### Query engine
Spark
### Please describe the bug 🐞
When I used spark to continuously rewrite the data file, I found a direct memory leak. The reason is that if …
-
## Prework
* [x] I understand and agree to [help guide](https://books.ropensci.org/targets/help.html).
* [x] I understand and agree to [contributing guide](https://github.com/ropensci/tarchetypes/…
-
### Checks
- [X] I have checked that this issue has not already been reported.
- [X] I have confirmed this bug exists on the [latest version](https://pypi.org/project/polars/) of Polars.
### Reprodu…
-
### Search before asking
- [X] I searched in the [issues](https://github.com/apache/paimon/issues) and found nothing similar.
### Paimon version
0.9
### Compute Engine
hive: 2.1-cdh-6.3-1
### M…
-
Totally fine if the answer to this question is "Because it is what we use", but I wonder if the capabilities of SPZ could be covered by standing upon something like Parquet or [Apache Feather](https:/…
-
### Description
If we pass the schema to scan_parquet/read_parquet it's easily possible to get errors such as this one:
```
polars.exceptions.SchemaError: dtypes differ for column col-d790df1a-b5af…
-
Hi,
When tried to write a data.frame that has Posixct columns where some values are `NA`, `write_parquet` throws the following error:
```
> write_parquet(df, "my.parquet")
Error in write_parqu…
-
We completely broke COPY support without us noticing. That problem is being fixed by #264, but it might happen again if we don't add some tests. For the 0.1.0 release we can manage with manual testing…
-
The first statement skips hidden objects correctly, but the 2nd statement doesn't.
```sql
ALTER TABLE region EXECUTE add_files('file:///tmp/add_files/.hidden', 'parquet');
ALTER TABLE region EXECUTE …