Closed KelvinChi closed 1 week ago
把reader/writer中的hadoop/hive相关包全替换为3.3.3版本解决了该问题 xxxx@xxxx:libs ll hadoop- -rw-r--r-- 1 smcv hdfs 63348 Jul 11 14:45 hadoop-aliyun-3.3.3.jar -rw-r--r-- 1 smcv hdfs 25101 Jul 11 14:45 hadoop-annotations-3.3.3.jar -rw-r--r-- 1 smcv hdfs 104435 Jul 11 14:45 hadoop-auth-3.3.3.jar -rw-r--r-- 1 smcv hdfs 30524177 Jul 11 14:48 hadoop-client-runtime-3.3.3.jar -rw-r--r-- 1 smcv hdfs 4470534 Jul 11 14:45 hadoop-common-3.3.3.jar -rw-r--r-- 1 smcv hdfs 5500884 Jul 11 14:45 hadoop-hdfs-client-3.3.3.jar -rw-r--r-- 1 smcv hdfs 1636325 Jul 11 14:45 hadoop-mapreduce-client-core-3.3.3.jar -rw-r--r-- 1 smcv hdfs 3649778 Jul 11 14:45 hadoop-yarn-api-3.3.3.jar -rw-r--r-- 1 smcv hdfs 2965761 Jul 11 14:45 hadoop-yarn-common-3.3.3.jar -rw-r--r-- 1 smcv hdfs 258471 Jul 11 14:45 hadoop-yarn-server-applicationhistoryservice-3.3.3.jar -rw-r--r-- 1 smcv hdfs 1439997 Jul 11 14:45 hadoop-yarn-server-common-3.3.3.jar -rw-r--r-- 1 smcv hdfs 2492913 Jul 11 14:45 hadoop-yarn-server-resourcemanager-3.3.3.jar -rw-r--r-- 1 smcv hdfs 56807 Jul 11 14:45 hadoop-yarn-server-web-proxy-3.3.3.jar You have mail in /var/spool/mail/smcv xxxx@xxxx:libs ll hive- -rw-r--r-- 1 smcv hdfs 44704 Jul 11 14:45 hive-cli-2.3.9.jar -rw-r--r-- 1 smcv hdfs 436169 Jul 11 14:45 hive-common-2.3.9.jar -rw-r--r-- 1 smcv hdfs 45423312 Jul 11 14:45 hive-exec-2.3.9.jar -rw-r--r-- 1 smcv hdfs 265922 Jul 11 14:45 hive-hcatalog-core-2.3.9.jar -rw-r--r-- 1 smcv hdfs 116364 Jul 11 14:45 hive-jdbc-2.3.9.jar -rw-r--r-- 1 smcv hdfs 8195966 Jul 11 14:45 hive-metastore-2.3.9.jar -rw-r--r-- 1 smcv hdfs 916630 Jul 11 14:45 hive-serde-2.3.9.jar -rw-r--r-- 1 smcv hdfs 527783 Jul 11 14:45 hive-service-2.3.9.jar -rw-r--r-- 1 smcv hdfs 1549366 Jul 11 14:45 hive-service-rpc-2.3.9.jar -rw-r--r-- 1 smcv hdfs 53902 Jul 11 14:45 hive-shims-0.23-2.3.9.jar -rw-r--r-- 1 smcv hdfs 8786 Jul 11 14:45 hive-shims-2.3.9.jar -rw-r--r-- 1 smcv hdfs 119936 Jul 11 14:45 hive-shims-common-2.3.9.jar -rw-r--r-- 1 smcv hdfs 12923 Jul 11 14:45 hive-shims-scheduler-2.3.9.jar
cd "${HADOOP_HOME}"/share/hadoop/ cp client/hadoop-client-runtime-3.3.3.jar /data/datax_20240711/plugin/reader/hdfsreader/libs/ cp hdfs/lib/woodstox-core-5.3.0.jar /data/datax_20240711/plugin/reader/hdfsreader/libs/ cp hdfs/lib/stax2-api-4.2.1.jar /data/datax_20240711/plugin/reader/hdfsreader/libs/ cp hdfs/lib/commons-configuration2-2.1.1.jar /data/datax_20240711/plugin/reader/hdfsreader/libs/ cp hdfs/lib/re2j-1.1.jar /data/datax_20240711/plugin/reader/hdfsreader/libs/ cp client/hadoop-client-runtime-3.3.3.jar /data/datax_20240711/plugin/writer/hdfswriter/libs/ cp hdfs/lib/woodstox-core-5.3.0.jar /data/datax_20240711/plugin/writer/hdfswriter/libs/ cp hdfs/lib/stax2-api-4.2.1.jar /data/datax_20240711/plugin/writer/hdfswriter/libs/ cp hdfs/lib/commons-configuration2-2.1.1.jar /data/datax_20240711/plugin/writer/hdfswriter/libs/ cp hdfs/lib/re2j-1.1.jar /data/datax_20240711/plugin/writer/hdfswriter/libs/
DataX 3.0 读取HDFS EC地址异常
读取
当源地址通过hdfs ec -setPolicy -path "${sourcePath}" -policy RS-6-3-1024k转化后,源地址是一张orc hive表的数据路径。使用datax读取该地址并写入目标库,会出现以下异常:
写入
源数据库为MySQL,通过hdfswriter写入到经hdfs ec -setPolicy -path "${targetPath}" -policy RS-6-3-1024k修改过的路径后,出现异常:
DataX Json配置
读取
写入