Open torvalds-dev-testbot[bot] opened 10 months ago
torvalds.dev is analyzing the ticket
The issue seems to be related to the Apache Hudi framework, which is used for managing storage of large analytical datasets over DFS (Distributed File Systems). The logs provided are from a Hudi job that is performing a merge operation on a dataset.
Here's a brief summary of the logs:
The relevant files to this issue are:
IOUtils.java
: This file contains the IOUtils
class which has methods for calculating the maximum memory allowed for merge and compaction tasks. It seems to be handling the memory allocation for the merge operation in the logs.
TestSparkIOUtils.java
: This file contains the TestSparkIOUtils
class which tests the methods in the IOUtils
class. It might be useful for understanding how the memory allocation methods are expected to work.
Suggested actions:
Check the memory allocation for the executor running the task. The logs show that the maximum memory per partition merge is set to a large value. If the executor does not have enough memory, it could cause issues.
Verify the payload size estimates. The logs show that the estimated payload size is updated shortly after the task starts. If these estimates are incorrect, it could lead to problems with memory allocation.
Monitor the producer and consumer threads. The logs show that these threads start and finish buffering and consuming records multiple times during the task. If these threads are not working correctly, it could affect the merge operation.
Review the merge operation. The logs show that the task is merging new data into an existing file. If there are issues with the data or the merge operation itself, it could cause the task to fail or perform poorly.
Check the creation of the marker file. The logs show that a marker file is created for the merge operation. If there are issues with the creation or use of this marker file, it could affect the merge operation.