charlesq34 / pointnet

PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation
Other
4.64k stars 1.44k forks source link

problem with Area1 hallway6 #284

Open youssef962 opened 2 years ago

youssef962 commented 2 years ago

I'm preparing the raw s3dis for training, after I converted the data to .npy files I want to convert it to hdf5 so I must run the 'gen_indoor3d_h5.py'. the first trial it showed me error with line 125 in 'indoor3d_util' as we adding cannot be done so I removed the + and put a comma instead it showed me error as I'm returning three values and I should return only two and I don't know which to remove the range(N) or the list(sample). when removing any of them it showed me another error of the dimensions of the array, any help please! 1st trial:

    else:
        sample = np.random.choice(N, num_sample-N)
        dup_data = data[sample, ...]
        return np.concatenate([data, dup_data], 0), range(N)+list(sample)

the error:

return np.concatenate([data, dup_data], 0), range(N)+list(sample)
TypeError: unsupported operand type(s) for +: 'range' and 'list'

2nd trial:

    else:
        sample = np.random.choice(N, num_sample-N)
        dup_data = data[sample, ...]
        return np.concatenate([data, dup_data], 0), range(N), list(sample)

the error:

new_data, sample_indices = sample_data(data, num_sample)
ValueError: too many values to unpack (expected 2)

3rd trial:

    else:
        sample = np.random.choice(N, num_sample-N)
        dup_data = data[sample, ...]
        return np.concatenate([data, dup_data], 0), list(sample)

the error


File "<__array_function__ internals>", line 5, in concatenateValueError: all the input array dimensions for the concatenation axis must match exactly, but along dimension 1, the array at index 0 has size 4096 and the array at index 5 has size 3778```