-
Hi, I heard concurrent write to snowflake using same jdbc connection is [threadsafe](https://community.snowflake.com/s/question/0D50Z000076A6HeSAK/is-jdbc-driverconnection-thread-safe), as it ought to…
-
The main [README](https://github.com/sparklyr/sparklyr/blob/main/README.Rmd) mentions using `DBI::dbGetQuery()` to execute SQL. this seems very slow in comparison to use `sparklyr::sqf_sql()`. Could t…
-
I am observing a very weird issue when I try to read data from postgres databases via RStudio using Sparklyr/Spark with spark_read_jdbc.
The first spark_read_jdbc all works as expected and returns a …
-
Problem:
I'm able to run the code examples (with 4-row datasets) of the README file in https://github.com/catboost/catboost/tree/master/catboost/spark/catboost4j-spark
However, when I try to repla…
-
Hi. Sparklyr 1.6 is so slow compare with sparklyr 1.5.2 over the same table (cached) with tthe same resources.
Using sparklyr over aws emr cluster
```r
variable7 %
mutate("{variable7}_ca…
-
## Willingness to contribute
The MLflow Community encourages new feature contributions. Would you or another member of your organization be willing to contribute an implementation of this feature (ei…
-
**Describe the bug**
The year portion of the datetime string (e.g., "2012" in "2012-07-03") is being incorrectly encoded (e.g., "+43969-07-03") when dataframes are written as parquet files during p…
-
**Environment:**
1. Framework: PyTorch
2. Framework version: 1.6.0
3. Horovod version: 0.20.0
4. MPI version: 4.0.2 (Open MPI)
5. CUDA version: Not applicable, not GPU is used.
6. NCCL version: …
-
Write to R dataset directly from sparklyr without loading the dataframe to local memory.
spark_write_rds()
-
# Finding a standard dataset format for machine learning | OpenML Blog
Exploring new dataset format options for OpenML.org
[https://openml.github.io/blog/openml/data/2020/03/23/Finding-a-standard-da…