Open mslavescu opened 6 years ago
Would be super if we could also get full dataset from different SDC prototypes (including from Udacity, Waterloo, MIT and other universities) driving in similar conditions to see if current sensor technology is enough (I assume it is) to detect a moving target (person with a bike) at 50m while driving at 60km/h, then notify the safety driver that an obstacle is in range (FCW), automatically switch to long beam range to allow long range cameras to see better and slow down while approaching the target to increase the chances to stop or avoid impact.
We should also try to see how recent ADAS system behave in those conditions, in Tesla, Volkswagen, etc.
If anyone has access to this kind of cars (SDC or advanced ADAS) please share your investigations and tests.
This is the minimum that should have happened in the accident scenario, would be good to see how cars (as many as possible) with this feature activated would have behaved:
Forward Collision Warning (FCW) Systems http://brainonboard.ca/safety_features/active_safety_features_fcw.php
"However, if an obstacle like a pedestrian or vehicle approaches quickly from the side, for example, the FCW system will have less time to assess the situation and therefore a driver relying exclusively on this safety feature to provide a warning of a collision will have less time to react."
Project page: http://www.pges.fr/2rt3d/
LIDAR scans: http://www.pges.fr/2rt3d/downloads.php
Videos: https://m.youtube.com/channel/UCQgt3-fe79kJhrR-tuK0M5Q
Lets see how visible is a person (or similarly shaped object) at around 50m and beyond.
To make it easier for everyone to understand the accident scenario, would be good to produce videos like these from real and simulated LIDAR scan:
Virtual scan of a pedestrian with a bicycle https://youtu.be/zauRVtAcXUc
The blog entry to this video is here: http://www.blensor.org/blog_entry_20180323.html We can do a more sophisticated simulation if necessary. This one was just to show that the person can be easily picked up from the LIDAR data
See the comments on this video, looks like Volvo XC90 has the best illumination between the 3 cars tested:
Volvo XC90, MB GLC & Audi Q7, night drive
We'll need to find if Uber car had active/dynamic beam mode on, that should have improved a lot the illumination range.
@mgschwan would be possible to convert some of the scans from 2rt3d dataset from above, to view them in your tool?
Also would be good to have similar videos with simulations at 25m and 50m for easy comparison.
Please add a bounding box with dimensions around the person+bike so we get a better feel of the scale and compare easier the point cloud.
@mslavescu the data is generated by our sensor simulation software. It can export to PCD point cloud data and be imported into any software that can read PCD files.
The visualization was done with Blender itself, no specialized tool was used. The colored pointcloud was generated with pcl_viewer from the pointcloud library.
I can do a scan at 25m and 50m, no problem, but I was wondering if there is some more information about the scene to make it a bit closer to reality?
This shows how bad is the video from Uber dash cam, here it is compared with another video in the same road spot, similar night conditions: https://twitter.com/emilyabell/status/977064719419879424
@mgschwan here is a picture with the Uber crash location, I added the Google StreetView link to the issue description:
https://www.ntsb.gov/news/press-releases/Pages/NR20180320.aspx
Developments on Tuesday include:
Meeting with representatives of Uber, the National Highway Traffic Safety Administration and the Tempe Police Department
Beginning examination of the accident vehicle, a Volvo XC90
Beginning examination of the accident site
Viewing a copy of a video that captured the crash from an aftermarket dash camera mounted in the test vehicle
Gathering information about the technology on the test vehicle
Collecting information about the pedestrian and the safety driver
Beginning collection of any and all electronic data stored on the test vehicle or transmitted to Uber
http://www-personal.acfr.usyd.edu.au/a.quadros/objects4.html
This object dataset was collected from the robot Shrimp (left), which has a Velodyne LIDAR. It spins around at 20Hz to produce a dynamic 3D point cloud of the scene. The dataset consists of 631 object instances segmented from Velodyne scans of the Sydney CBD. A few instances are shown below.
Performs real-time visualization and processing of live captured 3D LiDAR data from Velodyne's HDL sensors (HDL-64E, HDL-32E, VLP-32, VLP-16, Puck, Puck Lite, Puck HiRes). An introduction to features of VeloView is available as a video
https://www.paraview.org/Wiki/VeloView
Code and datasets:
Velodyne sample datasets (including from HDL 64E)
http://robots.engin.umich.edu/SoftwareData/Ford
The vehicle is outfitted with a professional (Applanix POS LV) and consumer (Xsens MTI-G) Inertial Measuring Unit (IMU), a Velodyne HDL 64E 3D-lidar scanner, two push-broom forward looking Riegl lidars, and a Point Grey Ladybug3 omnidirectional camera system.
Current data sets were collected during November-December 2009, in the future we plan to host more data sets corresponding to some other times of the year.
Paper: http://robots.engin.umich.edu/uploads/SoftwareData/Ford/ijrr2011.pdf
Videos:
http://robots.engin.umich.edu/nclt/ http://robots.engin.umich.edu/SoftwareData/NCLT
Sensors:
Velodyne HDL-32E LIDAR Ladybug3 Omnidirectional Vision Microstrain GX3 IMU KVH Fiber Optic Gyro Wheel Odometry Standard and RTK GPS
Paper: http://robots.engin.umich.edu/nclt/nclt.pdf
Video: 4D Mapping with Segway Robot https://youtube.com/watch?v=rD7VvAIbeag
Video with 2 cameras: Stereoscopic Reference for UBER accident location - Northbound on Mill approaching Curry https://youtu.be/8p0lpe-puOI
Here is a review of the camera used: https://www.cnet.com/products/drift-stealth-2/review/
It has a 1/3" 3 Mpix Aptina CMOS sensor with 2.0V/Lux-sec Low Light Sensitivity https://driftinnovation.com/products/drift-stealth2
I now made a more extensive simulation 50m approach with some surrounding terrain
To model this accident, you would want to consider the following:
a) Velodyne range starts at a little over 100m depending on reflectivity. At that point there is some curve to the road but the left lanes should be visible. b) There is a slight grade up to the road, it just went under an underpass. c) Car travels at 17 m/s according to reports d) At that speed, stopping distance is just under 25m. Thus the target must be decently sensed (though it need not be classified) by then. Generally, this involves seeing in in 3-4 frames of the Velodyne, which sweeps at 10hz. This adds perhaps 8m of distance e) In particular, you want enough frames to not just detect the obstacle but track its movement and velocity so you can model a vector of its future trajectory. f) An unsophisticated system which does not have a vector may only get concerned when it sees the obstacle in its own lane. The dashcam video shows the bicycle already in the lane when it appears from the darkness at 1.4 seconds prior to impact. What is lane width? Would guess 3.5m? For pedestrian at 3mph this suggests enters lane 2 to 2.5 seconds prior to impact. g) For radar, it is much harder to get the horizontal velocity vector, so you want a clear signal the target will incur into your lane in a simple system. However, she crossed 3.5 full lanes which suggests there should have been an ability to get this vector from radar as well.
Thanks @bradtem and @mgschwan !
If they used 64E I think the full scan rate is 20Hz.
We can use simulated data for now, hopefully they will release the raw data at some point.
Beside breaking in this situation I think by turning a bit to the left would have completely avoided the impact, which happened on the right side of the car, while the pedestrian was moving towards right side.
A human driver may have reacted differently, may have turned right instead, but software should be able to the calculate the optimum path considering all the surrounding objects (static or mobile).
We'll need to find an accurate way to simulate different scenarios.
CARLA simulator may be an option, not sure how good is their car model and physics engine though.
If you know better options (open source if possible) please list them here.
I think the 20 Hz is only available in the latest revision of the HDL-64e, please correct me if I am wrong.
HDL-64E S2 from 2010, has 5-15Hz scan rate: http://velodynelidar.com/lidar/products/brochure/HDL-64E%20S2%20datasheet_2010_lowres.pdf
HDL-64E S3 from 2013, has 5-20Hz scan rate: https://www.autonomoustuff.com/wp-content/uploads/2016/08/hdl-64E.pdf
Not sure which one was mounted on Uber car, and at what scan rate were they computing.
@mgschwan could you please do a video at 15Hz and 20Hz also.
I did a simulatio at 20Hz which reduces the angular resolution from the default 0.1728 degrees (at 10Hz) to 0.3456 degrees.
At a distance of 50 meters this makes a difference, at 25 meters it is still quite noticeable but definitely enough points to detect it. At 12.5 meters there are enough points to descern the bike from the person.
50 Meters - 20Hz
50 Meters - 10Hz
25 Meters - 20Hz
25 Meters - 10Hz
12.5 Meters - 20Hz
12.5 Meters - 10 Hz
Thanks @mgschwan!
How does BlenSor compare with CARLA LIDAR sensor, which seems it is supports at least HDL-32E (since yesterday): https://github.com/carla-simulator/carla/blob/master/CHANGELOG.md#carla-080 https://carla.readthedocs.io/en/latest/cameras_and_sensors/
This is HDL 64E S3 manual: http://velodynelidar.com/lidar/products/manual/HDL-64E%20S3%20manual.pdf
And here is the list of angular resolutions at different scan rates supported by HDL 64E S3 (at least since 2013):
An interesting discussion about angular resolution of LIDAR:
LIDAR also has limitations on angular resolution just as a function of how the sensor works. It's entirely possible that the size of the person/bike on LIDAR was just too small until it was too late to stop. https://news.ycombinator.com/item?id=16645627
Comparison between Velodyne Models - VLP-16, HDL-32, HDL-64: http://velodynelidar.com/docs/guides/LiDAR%20Comparison%20chart_Rev-A_2_Web.pdf
VLS-128 seems to be much better but it was just released (last year), I couldn't find the manual or specs, but these articles are representative:
Velodyne’s latest LIDAR lets driverless cars handle high-speed situations Discerning a butterfly from a broken tire at 70mph https://www.theverge.com/2017/11/29/16705674/velodyne-lidar-128-autonomous-vehicles-driverless-cars
128 Lasers on the Car Go Round and Round: David Hall on Velodyne’s New Sensor http://velodynelidar.com/blog/128-lasers-car-go-round-round-david-hall-velodynes-new-sensor/
A proper recreation also should involve the slight upward grade of the road, which may affect the line density on the target. Scans at 25 meters are interesting but not too valuable, in that this is the stopping distance at 40mph, and so an obstacle must be clearly visible well before that. In fact, it is generally desired to begin breaking even earlier to avoid the need for a hard full brake. This makes 50m a good distance to consider. At 10hz, you get one sweep every 1.7m, and you want a few sweeps for solid perception. As such, early detection by 35m is good and 50m is even better.
This is a nice video with DARPA 2007 challenge data:
Visualization of LIDAR data https://youtu.be/nXlqv_k4P8Q
@mgschwan how can we generate a 360° video with LIDAR cloud points (colored by distance) from car level in BlenSor? Even live with possible interaction to be viewed in VR systems (Oculus, Vive or Google Cardboard like). While the car moves from 50m to 5m towards the person.
@bradtem to summarise a few points:
This last point it makes very important to have adjustable beams and use them on long range and reduce the range/brightness and lower the angle while getting closer to an opposite car, in such a way that the combined beam would be enough.
Same applies to animals, especially when very silent electric cars will be deployed. Or the wind covers the car noise.
We need also elevated long range beams to see stop signs or other traffic signs.
Yes, you have 3 seconds to impact from 50m. However, generally impact is not your goal! To avoid impact, you must hard brake before 25m. To avoid hard braking it would be nice to start much earlier, like 40m. Which is fine, you get 6 sweeps at 10hz from 50m to 40m, and then you actuate the brakes.
However, for the pure emergency situation, you can consider starting at 35.5m, get 4 sweeps on your way to 25m, and then hard brake on dry pavement. So with a 400ms perception/analysis time, you really want to perceive before 36m.
@bradterm Do we have any info about the upward slope. I cant get that info from google maps. But if the slope remains constant up to the pedestrian it would hardly change the output as the car would already follow the slope.
@bradtem I agree, for full stop before the impact the car needs to act (start to break) at around 25m. Also by breaking at 25m we give the person more time to get out of the way, we should see how far the person will get in that fraction of second. Would be good to simulate different braking + steering (left or right) scenarios, and compare the impact (direct or lateral) force.
@mgschwan I found this topo map, may help a bit: https://www.topozone.com/arizona/maricopa-az/park/papago-park-tempe/ https://www.topozone.com/map-print/?lat=33.4422671&lon=-111.9420887&title=Papago%20Park-Tempe%20Topo%20Map%20in%20Maricopa%20County%20Arizona Here is the satellite view, the accident was a bit north east to Marque Theatre sign:
The Tempe weather on March 18, accident happened at 10PM: https://www.timeanddate.com/weather/usa/tempe/historic?month=3&year=2017#
There was no wind around 10PM, temp 15C, humidity 25%, visibility 16KM: Tempe Weather History for March 18, 2018 https://www.timeanddate.com/weather/usa/tempe/historic?hd=20180318
Here is a scan from a distance of 34 meters 10Hz scanning speed Angular resolution: 0.1728 degrees Speed of car: 17m/s
I can't do an animation in VR, but Sketchfab has a VR mode. So if you have a headset you can step into the scan
@mgschwan thanks! That is super cool! I'll try it tonight.
Looks good, and confirms what has been said by Velodyne itself and everybody else. No way this LIDAR doesn't readily spot the pedestrian. Now we have to wait for more data to see what failed.
A very nice video with object detection:
Motion-based Detection and Tracking in 3D LiDAR Scans https://youtu.be/cyufiAyTLE0
I'll try to see how we can run this kind of algorithms in BlenSor and CARLA Simulator.
Beside of understanding and reproducing the Uber accident scenario (more accurate when we get the car onboard sensors data), this exercise would help us build tests and simulate different (edge case) scenarios for OSSDC Initiative.
This look quite promising as well
https://www.youtube.com/watch?v=0Q1_nYwLhD0
But I think 3 year old papers are already out of date in this fast moving field.
If you got working implementations of those algorithms, getting simulated data from BlenSor into them probably works through a series of PCD or PLY files. Alternatively I think I can create a ROS Bag file to replay them for algorithms that work directly as a ROS node
@mgschwan how hard would be to write a ROS LIDAR/camera publisher in BlenSor?
That way we could easily create ROS bags and use it in live scenarios also.
Take a look at CARLA (https://github.com/carla-simulator/carla) and TrueVision (https://www.truevision.ai/) simulators, both based on Unity, if we could build similar features in a fully open source simulator, based on Blender (with BlenSor as base) would be super!
Here is a more recent work with neural nets based detection:
Vote3Deep: Fast Object Detection in 3D Point Clouds Using Efficient Convolutional Neural Networks https://youtu.be/WUOSmAfeXIw
We will try to integrate something like this in Colaboratory, like we did with other CNN based image detection/segmentation methods:
Try live: SSD object detection, Mask R-CNN object detection and instance segmentation, SfMLearner depth and ego motion estimation, directly from your browser! https://medium.com/@mslavescu/try-live-ssd-object-detection-mask-r-cnn-object-detection-and-instance-segmentation-sfmlearner-df62bdc97d52
Thanks @bradtem for this article:
How does a robocar see a pedestrian and how might Uber have gone wrong? http://ideas.4brad.com/how-does-robocar-see-pedestrian-and-how-might-uber-have-gone-wrong
Here is the link to a bag file with the 20Hz simulation.
http://www.blensor.org/pedestrian_approach.html
Can anyone with a running obstacle detection algorithm check if the bag file works. I created it inside ros docker container, so I can't run a GUI to check if the data makes sense
A great update by @bradtem: https://www.forbes.com/sites/bradtempleton/2019/11/06/new-ntsb-reports-on-uber-fatality-reveal-major-errors-by-uber/
At 2.6 seconds out, the classifier thought she was a bicyclist. She was indeed walking a bicycle. Again, with no history, her path was unknown. Oddly, after another LIDAR frame she was classed as moving along the lane to the left of the Uber car. This may be because the system doesn’t expect bicycles to be going sideways in the middle of a traffic lane, which would be another error. Either way, it isn’t until 1.5 seconds out that the system (switching to Unknown again) realizes she is coming into the Uber’s lane. Correctly, it plots to swerve around her.
The fatal moment comes at 1.2 seconds out. She is reclassified as a bicyclist and in the path of the vehicle. The swerving plan no longer can help. It’s time to do an emergency braking. It’s really time.
Some extra details here: https://www.azcentral.com/story/money/business/tech/2019/11/06/new-details-emerge-fatal-tempe-self-driving-uber-crash-pedestrian-elaine-herzberg/2508011001/
The new report offers the following timeline:
At 5.6 seconds before impact, the car’s radar detected Herzberg, classified her as a car and estimated her speed. At 5.2 seconds the lidar system detected her and classified her as an unknown object with a static path. Unknown objects were not considered obstacles unless they were in the path of the car, which she wasn't at this time. At 4.2 seconds before impact the lidar again detected her but reclassified her as a vehicle, predicting the “vehicle” would continue to travel in the same direction the Uber was heading because it was in the left travel lane. In the next seconds the lidar reclassified her several times as either “vehicle” or “other.” At 2.6 seconds before impact, lidar classified the woman as a bicycle. The new classification restarted the tracking history, and the car predicted her path as "static." At 2.5 seconds it again classified her as a bike that, like the vehicle classification earlier, would continue to travel in the same direction as the Uber. At 1.5 seconds before impact, the lidar reclassified her again to an unknown object, but now being partially in the Uber’s travel lane. The Uber planned to steer right to avoid the object. At 1.2 seconds the lidar classified her as a bike moving into the path of the Uber, and determined the prior plan to steer around her as no longer possible.
At this point, the Uber went into its 1-second “action suppression” mode.
It came out of the suppression mode 0.2 seconds before impact. The car could not safely brake and swerve to avoid impact, so it began a “controlled slowdown” and gave the operator an auditory alert. The Uber was traveling 39 mph at this instant.
The vehicle operator hit the brakes 0.7 seconds after the impact.
Collect datasets from similar situations and analyse them to help everyone understand what the car should have seen and what could have been done to avoid the accident.
Help us obtain the dataset from Uber SDC involved in the accident, at least 3 min before and 1 min after the impact (this is a reply to the police tweet with the video from the accident): https://twitter.com/GTARobotics/status/976628350331518976
A few initial pointers to accident info:
The Google Maps StreetView link where the accident happened:
642 North Mill Avenue, Tempe, Arizona, USA https://goo.gl/maps/wTDdCvSzc522
Brad Templeton analysis of the accident: https://twitter.com/GTARobotics/status/976726328488710150 https://twitter.com/bradtem/status/978013912359555072
Experts Break Down the Self-Driving Uber Crash https://twitter.com/GTARobotics/status/978025535807934470?s=09
Experts view on the fact that LIDAR should have detected the person from far away: https://twitter.com/GTARobotics/status/977764787328356352
This is the moment when we decide that human lives matter more than cars https://www.curbed.com/transportation/2018/3/20/17142090/uber-fatal-crash-driverless-pedestrian-safety
Uber self-driving system should have spotted woman, experts say http://www.cbc.ca/beta/news/world/uber-self-driving-accident-video-1.4587439
IIHS shows the Volvo XC90 with a range just under 250 feet (76 meters) with "low beams" on! https://twitter.com/GTARobotics/status/977995274122682368
Help us get current companies that test SDC to provide datasets from their own cars in similar situations as the accident: https://twitter.com/GTARobotics/status/977773180344512512
Lets also capture current SDC sensors configurations/specs in: https://github.com/OSSDC/OSSDC-Hacking-Book/wiki
Join the discussions on OSSDC Slack at http://ossdc.org