-
**Is your feature request related to a problem? Please describe.**
`cudf::size_type` is currently an `int32_t`, which limits column size to two billion elements (`MAX_INT`). Moreover, it limits child…
-
Version `"com.crealytics:spark-excel_2.12:0.14.0"`
```
val identifiers = Seq("s3a://bucket/wb1.xlsx", "s3a://bucket/wb1.xlsx")
spark.read
.options(options)
.format("co…
-
Remove or make it work and testable?
-
Does TransmorgrifAI support PySpark or is will this be supported in the future or is there a workaround for usage in PySpark?
Thanks in advance!
-
# The problem
Hello,
I'm working on a Change Data Capture and my goal is to replicate data from a parquet into a Delta table by making the required inserts, updates and deletes. I followed the t…
-
Looks like lates pytest shows some warning
```console
============================================================================= warnings summary =================================================…
-
I want to filter inside for loop like
```
sc
-
When using dplyr's `summarise()` with `.group = "drop"` parameter a new column called `.groups` is created as shown below.
```r
spark_version
-
- Are there any major differences other than `processor` being able to take >1 input?
---
My attempt at finding the differences from the docs:
`processor` is on the driver side and `transform…
-
Because of some stability issue in SnappyData 1.1.0 we thought of upgrading to recently released version i.e., SnappyData 1.2.0.
We have been using Snappydata for more than a year in production.
P…