Closed nvliyuan closed 1 month ago
Thanks @nvliyuan ! The java API has nothing to do with the XGBoost prediction. The java API only scana and analyzes the eventlog to generate raw_metrics and stage information.
The python API spark_rapids qualification
internally does the following:
rapids_4_spark_qualification_output
rapids_4_spark_qualification_output
xgboost_predictions
directory.Can you elaborate more on what the problem is? and what is not being in sync here?
Hi @amahussein, thanks for the quick reply, please ignore this FEA, I assume I already found the root cause of the issue, the customer write a udf to run qualx tool, but it hangs while running spark_rapids qualification
process, it is because they should update https://github.com/NVIDIA/spark-rapids-tools/blob/14a4213d54d3035b974e6598a8418c01090755c0/user_tools/src/spark_rapids_pytools/resources/onprem-configs.json#L4 to their customized repo uri...
The java api qualification output files cannot be used as the qualx tool prediction input since missing some files(features.csv...), it would be nice to keep the java API and python API in sync.
output:
output: