Open W-Nabe opened 2 months ago
Thank you for noticing our work! We are sorry for the late reply.
Thank you for your response!
Regarding point 1, you mentioned that mp4 files can't be used directly and that "you need to first use a monocular depth estimation algorithm to predict the depth of each frame." Could you suggest specific tools that would be good for generating this?
For point 2, you mentioned "decreasing the number of synchronized tracking points." Could you provide more details on the specific steps to take and what code to write to achieve this?
As for point 3, I haven't been able to run SpaTracker on my GPU or Colab yet, so I haven't even generated anything. I'm not sure about the conversion process, so if I encounter any issues, I'll ask again.
I apologize, but my knowledge in this area is really limited, and for points 1 and 2, I can't envision the specific code or tools to use. Any advice you could provide would be greatly appreciated.
--track_point_num 256
in script/train_odyssey.sh
to --track_point_num x
, where x
is a value that does not exceed the video memory. Thank you for your reply!
I haven't been able to try it yet due to time constraints, but I'm planning to run it on WSL2 in Windows 11.
I apologize for taking up your time, but I'd like to confirm the overall execution procedure as it's still a bit unclear to me.
I fed the repository contents, questions, answers, and materials into Google AI Studio and asked for the execution procedure. The AI output the following procedure. Is it correct?
If necessary, modify --track_point_num 256 in script/train_odyssey.sh to --track_point_num x
.
data/demo
directory and place the video and depth maps there.
data/demo/my_video/
├── rgb.mp4
└── deps.npz
data/dataset.py
and change the data_root
variable in the WWOdyssey
class to include data/demo/my_video
.
data_root = 'data/demo/my_video' # Modified part
odyssey_root = data_root + 'LSFOdyssey'
driving_root = data_root + 'LSFDriving'
run_demo.py
and change the call to the validate_odyssey
function as follows:
run_test.validate_odyssey(model, split='demo', seq_name='my_video') # Modified part
run_test.py
and add the following code at the beginning of the validate_odyssey
function:
if seq_name: # Added part
val_set = WWOdyssey(seq_len=-1, track_point_num=-1, split=split, seq_name=seq_name)
else:
val_set = WWOdyssey(seq_len=-1, track_point_num=-1, split=split)
run_demo.py
.
python run_demo.py
track.npz
file will be generated in the data/demo/my_video
folder. This file contains the 3D coordinates of the tracked points.Also, how can I rewrite the code to specify the points to be tracked?
I'm interested in SceneTracker and have several questions about its usage:
I apologize if any of my questions are off-base, as I'm not very familiar with this field. I would greatly appreciate your responses when you have the time.