Open liujiayi771 opened 1 day ago
cc @zhouyuan.
@liujiayi771
thanks for reporting, it looks like unload the libhhdfs.so is not working properly in your testing env.
would it be convenient to also paste the detail log in /root/hs_err_pid1258650.log
?
thanks, -yuan
@zhouyuan I have added the error stack in the description.
@liujiayi771 is the libhdfs.so
from a vanilla HDFS project or it's been customized?
Based on the stack, it looks like HDFS is trying to invoke the close
method
https://github.com/apache/hadoop/blob/branch-3.0/hadoop-hdfs-project/hadoop-hdfs-native-client/src/main/native/libhdfs/jni_helper.c#L123
but can not find the right symbol(?) then call the clean up function
https://github.com/apache/hadoop/blob/branch-3.0/hadoop-hdfs-project/hadoop-hdfs-native-client/src/main/native/libhdfs/jni_helper.c#L299
thanks, -yuan
@zhouyuan We have our customized HDFS. But the hdfs.c
and jni_helper.c
is same as branch-3.0 in hadoop repo. We have never modified the libhdfs code. But I will test it with the vanilla HDFS. Can you reproduce this issue?
I checked the code for FileSystem
in our code, and the close()
method, which is a basic interface, definitely hasn't been modified. It's strange that JNI couldn't find this method.
Hi @liujiayi771 , I tried locally but seems not able to trigger it. I think we may need to add more guards in Velox filesystem close() CC @JkSelf for her comments
Thanks, -yuan
@liujiayi771 Can you help to test adding this command before running your application? export CLASSPATH=$HADOOP_HOME/bin/hdfs classpath --glob
@JkSelf I have tested it, and it still results in a core dump. I will investigate this issue further in the next few days.
Backend
VL (Velox)
Bug description
After executing the SQL, if I exit the spark-sql command line using Ctrl+C or quit command, a core dump occurs. https://github.com/apache/incubator-gluten/pull/6172
Spark version
Spark-3.4.x
Spark configurations
No response
System information
No response
Relevant logs