exasol / spark-connector

A connector for Apache Spark to access Exasol
Apache License 2.0
12 stars 7 forks source link

Socket timeout when trying to read data #223

Closed flatballer closed 5 months ago

flatballer commented 7 months ago

Hi team, I can connect to our Exasol from Databricks and get the schema of the table I am trying to read, but whenever I try to read actual records from the table, I get this error: "org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 0.0 failed 4 times, most recent failure: Lost task 0.3 in stage 0.0 (TID 3) (172.16.12.159 executor 0): com.exasol.jdbc.ConnectFailed: java.net.SocketTimeoutException: connect timed out" my query is a simple "SELECT * FROM table".

Is this a bug or user error?

Shmuma commented 7 months ago

Hi!

This problem might be related to this issue: https://github.com/exasol/spark-connector/issues/218

Could you please check your DB configuration? If multiple datanodes are used, make sure that firewall is open between databricks (master and executors) and DB cluster.

flatballer commented 6 months ago

Hi! I tried the code with max_nodes=1. Same result. I also seem to have network connectivity:

8563/tcp  open     unknown
20000/tcp filtered dnp
20001/tcp filtered microsan
20002/tcp filtered commtact-http
Shmuma commented 6 months ago

Yes, looks good to me. Thanks for the reply!

Could you please create a support ticket? We have a special process to handle customer issues and support team might have better ideas what might be the cause.

flatballer commented 6 months ago

Thanks. I'm trying my luck there. I'll post the solution here if they find one.