Closed Xuechun-Hu closed 2 years ago
Hi, Thank you for the reporting! Can you share the code and data for this experiment?
Hi @neka-nat
''' import cupoch as cph
cph.utility.set_verbosity_level(cph.utility.Debug)
pcl = cph.io.read_point_cloud( "data/pointcloud/block_0.ply") pcl_name = 'block_0' cph_labels = pcl.cluster_dbscan(eps=1, min_points=100, print_progress=True).cpu()
'''
I'm trying to cluster a (15636915, 2) dataset, which looks like this [[-230.19247 43.109245] [-225.90079 41.327675] [-222.1525 33.68174 ] ... [-224.55055 22.347908] [-224.50694 22.773636] [-223.6284 24.611273]]
Thank you for sharing the code! Can you also share "block_0.ply" pointcloud data?
https://drive.google.com/drive/folders/1FXQ9zT3uqEblHsw5xaC-w_KyUaczEAI4?usp=sharing there are two pointcloud files ,block_0 is the relatively smaller one
↑Can you give me access to Google Drive?
Test using a small point cloud with 15million points, and it takes extremely long to process. The gpu is nvidia a600 with 48gb memory. Cuda version is 11.2. Is cuml used in the library? It suffers the batch shrinking problem
[2021-10-12 12:00:22.331] [debug] Read geometry::PointCloud: 15636915 vertices.
[2021-10-12 12:00:22.340] [debug] [RemoveNoneFinitePoints] 15636915 nan points have been removed. (15636915, 3) (15636915, 3) block_0 has 15636915 points | ID | GPU | MEM |
| 0 | 18% | 2% | [2021-10-12 12:00:22.798] [debug] Precompute Neighbours Precompute Neighbours[================> ] 40%