HarshayuGirase / Human-Path-Prediction

State-of-the-art methods for human trajectory forecasting. Contains code for papers published at ECCV 2020 and ICCV 2021.
MIT License
341 stars 81 forks source link

ETH original dataset #50

Open 12num opened 2 years ago

12num commented 2 years ago

Thank you very much for your work, I have a request I hope you can agree to Could you share with me the data of your ETH processing script before processing, thank you very much!

mulplue commented 1 year ago

hello, I have the same question. I can't get the same *.pickle data from original data(sgan provided) using image2world function, do you have some progress?

12num commented 1 year ago

你好,我有同样的问题。我无法使用 image2world 函数从原始数据(提供 sgan)中获取相同的 *.pickle 数据,您有什么进展吗?

Hello, I'm sorry to tell you that I haven't made any progress on this issue, so I turned to other model explorations. If you have any questions, please contact me by email.

mulplue commented 1 year ago

Hello, I'm sorry to tell you that I haven't made any progress on this issue, so I turned to other model explorations. If you have any questions, please contact me by email.

Hello, now I know how to make the world-pixel transformation using the homography matrix, but I still don't know how Ynet filter the data, so I turned to other model explorations, too. Here's the world2image transformation(from ETH official guidance), I hope it can help someone who are concerned about this issue:

def world2image(traj_w, H_inv):    
    # Converts points from Euclidean to homogeneous space, by (x, y) \u2192 (x, y, 1)
    traj_homog = np.hstack((traj_w, np.ones((traj_w.shape[0], 1)))).T  
    # to camera frame
    traj_cam = np.matmul(H_inv, traj_homog)  
    # to pixel coords
    traj_uvz = np.transpose(traj_cam/traj_cam[2]) 
    return traj_uvz[:, :2]
Chenzhou727 commented 10 months ago

Hello, I'm sorry to tell you that I haven't made any progress on this issue, so I turned to other model explorations. If you have any questions, please contact me by email.

Hello, now I know how to make the world-pixel transformation using the homography matrix, but I still don't know how Ynet filter the data, so I turned to other model explorations, too. Here's the world2image transformation(from ETH official guidance), I hope it can help someone who are concerned about this issue:

def world2image(traj_w, H_inv):    
    # Converts points from Euclidean to homogeneous space, by (x, y) \u2192 (x, y, 1)
    traj_homog = np.hstack((traj_w, np.ones((traj_w.shape[0], 1)))).T  
    # to camera frame
    traj_cam = np.matmul(H_inv, traj_homog)  
    # to pixel coords
    traj_uvz = np.transpose(traj_cam/traj_cam[2]) 
    return traj_uvz[:, :2]

I would like to ask if the UCY dataset has the same homography matrix for converting coordinates between world coordinates and pixel coordinates. Similar work, Y-net, NSP-SFM, etc., all use map information in pixel space. The final indicators (ADE/FDE) are in world coordinates. How do they achieve conversion on UCY? Also I noticed that the original dataset of UCY seems to be in pixel coordinates, most of the existing work uses world coordinates, how is this converted, thank you very much