Closed github-actions[bot] closed 6 months ago
linux-2xlarge-d29d1507
Affected builds
Action item: @claraberendsen - finish deploying the fix to ci.ros2.org
ERROR: [ssh-agent] Could not register BouncyCastle on the remote agent.
Affected builds
Previous job that failed seems like it could not clean itself properly. See reference
Action item: @claraberendsen Rotate the agent so it comes back clean. :white_check_mark:
nightly_win_deb
sros2 fatal exceptionThis seems to be the same issue that was happening last week see. Requires further investigation, it seems to be failing since May 16th.
packaging_linux
untracked files of rtiAffected builds
packaging_linux
#3445Not sure where these files came from, I'll try to run a packaging job on the same machine and see if the files are still present.
First failure: https://ci.ros2.org/view/nightly/job/nightly_win_deb/3095/ Last failure: https://ci.ros2.org/view/nightly/job/nightly_win_deb/3099/
Error
04:05:34 test\sros2\commands\security\verbs\test_create_permission.py . [ 65%]
04:05:34 Windows fatal exception: access violation
04:05:34
04:05:34 Current thread 0x00002c94 (most recent call first):
04:05:34 File "C:\ci\ws\src\ros2\sros2\sros2\sros2\verb\generate_artifacts.py", line 44 in main
Possible cause: https://github.com/ros2/sros2/commit/fb22661623426fadcc60804e10a964664330d98e
Affected builds
Seems like an issue when porting the coverage reports to the jenkins controller.
02:44:17 ERROR: ERROR: Failure to paint /home/jenkins-agent/workspace/nightly_linux_jazzy_coverage/ws/src/ros2/launch/launch/launch/event_handlers/on_include_launch_description.py to /var/lib/jenkins/jobs/nightly_linux_jazzy_coverage/cobertura
02:44:17 java.io.IOException: Failed to deserialize response to UserRequest:hudson.FilePath$WritePipe@2e47ea50: java.lang.SecurityException: Agent may not access a file path. See the system log for more details about the error ID '9da8cfdd-551a-403b-9459-8bee33963dd6' and https://www.jenkins.io/redirect/security-144 for more information.
Seems like an issue when porting the coverage reports to the jenkins controller.
The issues logged have been present for years, quite likely, based on recent reports. Although those errors affect the visualization of the source files in coverage reports, they don't actually break the build.
Th last time this was being investigated, the exit status of colcon-test was non-zero and that was the "actual" hard failure.
Th last time this was being investigated, the exit status of colcon-test was non-zero and that was the "actual" hard failure.
Yeah, we need to land https://github.com/ros2/ci/pull/771, which will switch this from a red build to a yellow one.
Yeah, we need to land https://github.com/ros2/ci/pull/771, which will switch this from a red build to a yellow one.
Thanks, we will wait for that to land then. Sorry for the noise, it's been a while since I've done buildfarm work :laughing:
linux-2xlarge-98771619
Affected builds
Action item: @claraberendsen - finish deploying the fix to ci.ros2.org. I was going to deploy today but need to patch buildfarm-deployment to deploy without mage.
Affected builds
Same issue as yesterday opening issue and reporting to the team. Issue opened https://github.com/ros2/sros2/issues/316
Same as yesterday, waiting on PR to merge.
Seems like an issue with dependencies not being imported correctly. Will open an issue. Seems to be happening for Rolling and Jazzy.
Same issue reported to the team here https://github.com/gazebosim/gz-rendering/issues/1004
rosbag2_compression.SequentialCompressionWriterTest.open_succeeds_twice
:female_detective:Affected jobs
Failing tests projectroot.test_sequential_compression_writer rosbag2_compression.SequentialCompressionWriterTest.open_succeeds_twice
Reference build: https://ci.ros2.org/job/nightly_linux-aarch64_repeated/2762 First build error: https://ci.ros2.org/job/nightly_linux-aarch64_repeated/2749/ This is happening consistent in nightlies. Possible cause for introduction: https://github.com/ros2/rosbag2/pull/1599 Error:
/home/jenkins-agent/workspace/nightly_linux-aarch64_repeated/ws/src/ros2/rosbag2/rosbag2_compression/test/rosbag2_compression/test_sequential_compression_writer.cpp:242
Expected: writer_->open(storage_options, {serialization_format_, serialization_format_}) doesn't throw an exception.
Actual: it throws std::runtime_error with description "Bag directory already exists (/tmp/path_not_empty), can't overwrite existing bag
@claraberendsen I've prepared a fix for the aforementioned CI regression. Please see https://github.com/ros2/rosbag2/pull/1667 for more information.
Reference build: https://build.ros.org/job/doc_rosindex/2003/console (started failing again since build 1998)
Log output:
00:00:14.591 # BEGIN SECTION: Run Docker - rosindex
00:00:14.591 + export PODMAN_USERNS=keep-id
00:00:14.591 + docker run --rm --net=host --cidfile=/home/jenkins-agent/workspace/doc_rosindex/docker_rosindex.cid -v /home/jenkins-agent/workspace/doc_rosindex/repositories/rosindex:/tmp/doc_repository -v /home/jenkins-agent/workspace/doc_rosindex/upload_repository:/tmp/upload_repository -e REPO=/tmp/doc_repository -e SITE=/tmp/upload_repository rosindex
00:00:14.603 docker: container ID file found, make sure the other container isn't running or delete /home/jenkins-agent/workspace/doc_rosindex/docker_rosindex.cid.
00:00:14.603 See 'docker run --help'.
00:00:14.613 Build step 'Execute shell' marked build as failure
00:00:14.621 Build did not succeed and the project is configured to only push after a successful build, so no pushing will occur.
00:00:14.631 Finished: FAILURE
It's being solved in: https://github.com/ros-infrastructure/ros_buildfarm/pull/1047
demo_nodes_cpp
| Test regression | Investigating :detective:Reported in: https://github.com/ros2/demos/issues/693
There is a parent test regression with different variations on ConnextDDS: demo_nodes_cpp.TestExecutablesTutorial.test_processes_output
Failing test regressions:
Flakiness report (projectroot.test_tutorial_parameter_events_async__rmw_connextdds):
job_name | last_fail | first_fail | build_count | failure_count | failure_percentage |
---|---|---|---|---|---|
nightly_linux_debug | 2024-05-22 | 2024-05-09 | 15 | 7 | 46.67 |
nightly_linux_repeated | 2024-05-19 | 2024-05-08 | 16 | 2 | 12.5 |
nightly_win_deb | 2024-05-15 | 2024-05-07 | 9 | 5 | 55.56 |
Flakiness report (projectroot.test_tutorial_parameter_events__rmw_connextdds):
job_name | last_fail | first_fail | build_count | failure_count | failure_percentage |
---|---|---|---|---|---|
nightly_win_rep | 2024-05-22 | 2024-05-07 | 13 | 13 | 100.0 |
nightly_win_rel | 2024-05-22 | 2024-05-22 | 14 | 1 | 7.14 |
nightly_linux-rhel_repeated | 2024-05-21 | 2024-05-12 | 16 | 3 | 18.75 |
nightly_win_deb | 2024-05-12 | 2024-05-09 | 9 | 2 | 22.22 |
I don't see any specific change that points to a reason why it started failing more in normal jobs, and not only repeated ones (package history).
Some linux nightlies failed due to a "No Space Left" error in linux-2xlarge-f8f1b088 machine:
I'm cleaning up the space in the agent
Reference build:https://build.ros2.org/view/Jci/job/Jci__nightly-release_ubuntu_noble_amd64/15/
@clalancette these warnings are in different packages. Where should I report this?
Previous log #51