-
I did a quick test reading `DeltaLake` data in a notebook on a Coiled cluster from s3, with `dd.read_parquet` vs `ddt.read_deltalake`.
Cluster: https://cloud.coiled.io/clusters/245026/information?a…
-
delta-rs was included as an optional dependency in [aws-sdk-pandas](https://github.com/aws/aws-sdk-pandas), but that means it's still not included in the pre-built layer, so it's still hard to use del…
-
Hi,
When attempting to test (through polars), I was immediately met with:
```
deltalake.table.DeltaTableProtocolError: The table's minimum reader version is 2but deltalake only supports up to v…
-
Delta Lake can improve the efficiency of row deletions with deletion vectors, which is an "has this row been deleted" bitmap that elides rewriting Parquet files whenever a row is deleted. Our Delta La…
-
In Trino 444 with Alluxio cache enabled, when dropping and then recreating a delta table, I occasionally get the following error:
```
io.trino.spi.TrinoException: Error getting snapshot for my_schem…
-
The test failed in https://github.com/trinodb/trino/actions/runs/5745174053/job/15573014949?pr=17477
Unfortunately, there was no helpful error message.
```
Error: io.trino.plugin.deltalake.TestDe…
ebyhr updated
2 months ago
-
Arrow has adopted the (IMO) much better binary and string view types. Supporting these would mean we could move data zero-copy to delta-rs. Currently it fails:
```python
df = pl.DataFrame({"a": ["…
-
I've stored a bunch of data partitioned by date, and written it to delta using the deltalake package like so:
```python
for df in dfs:
write_deltalake("mytable", df, partition_by="date")
```…
-
I want to create a clickhouse table with DeltaLake engine and I use this query:
`create table deltatable Engine=DeltaLake('http://10.19.129.78:9090/source/enrichedweblimited', aws_access_key_id, aw…
-
In dask-deltatable, when calling dd.read_parquet, perhaps we can reuse the metadata already preserved in delta json, instead of collecting it from parquet files all over again.
Here:
https://git…