-
### Search before asking
- [X] I had searched in the [issues](https://github.com/apache/seatunnel/issues?q=is%3Aissue+label%3A%22bug%22) and found no similar issues.
### What happened
I encountere…
-
Hello,
I am trying to run spark-cep job. I have created the jar and trying to submit it using spark-submit and below query.
SELECT COUNT(DISTINCT t.duid) FROM stream_test OVER (WINDOW '300' SECONDS…
-
I have a question regarding the command for spark-submit:
```
spark-submit \
--packages org.apache.spark:spark-sql-kafka-0-10_2.12:3.1.2 \
stream_all_events.py
```
What is the meanin…
-
```
library(sparklyr)
library(dplyr)
spark_disconnect(sc)
setwd(paste0(Sys.getenv("SPARK_HOME"),"/"))
sc % stream_write_text(path = "file3")
```
Running the above code I want to consume a kafka…
DSoot updated
3 years ago
-
Hello.
I noticed that the `divolteStream` uses `createStream` method of `KafkaUtils` to establish a stream, however since Spark 1.3 it is recomended to use [DirectStream API](http://spark.apache.org/d…
-
I'm getting this error:
```
>>> directKafkaStream = KafkaUtils.createDirectStream(ssc, ["help_center.activity.events"], {"metadata.broker.list": "kafka.service.consul:9092"})
_______________________…
dasch updated
7 years ago
-
Background:
For now, the spark job is set to run on top of Kubernetes. Python reads about 35 images locally, each image get processed using Tesseract using Spark, the result is then collected in a py…
-
> Refer to roadmap [2023](https://github.com/StarRocks/starrocks/issues/16445) [2022](https://github.com/StarRocks/starrocks/issues/1244)
# Shared-data & StarOS
- Align with all functionalities…
-
## Issue Description
The existing data ingestion architecture, which relies on MQTT for IOT communication, KairosDB for Time Series storage, and a set of Rails functions orchestrated with Sidekiq a…
-
die compose Dateien sollen leichter über Umgebungsvariablen konfiguriert werden können.
Hier müssen wir unterscheiden zwischen Umgebungsvariablen die in der compose Datei ersetzt werden und solche …