Open Elycyx opened 1 day ago
@Elycyx Maybe you can set the https://github.com/wzcai99/Pixel-Navigator/blob/f7077a4cfc3fc0d320f73beafcda75375dc1ddb0/config_utils.py#L32 as False and try to run the code again.
@Elycyx Maybe you can set the
as False and try to run the code again.
Thank you for your reply! When I set draw_goal_aabbs=False
, the errors above disappear, but I got a new error:
Traceback (most recent call last):
File "/mnt/sda1/Pixel-Navigator/objnav_benchmark.py", line 77, in <module>
goal_image,goal_mask,debug_image,goal_rotate,goal_flag = nav_planner.make_plan(episode_images[-12:])
File "/mnt/sda1/Pixel-Navigator/gpt4v_planner.py", line 44, in make_plan
direction,goal_flag = self.query_gpt4v(pano_images)
File "/mnt/sda1/Pixel-Navigator/gpt4v_planner.py", line 92, in query_gpt4v
self.gptv_trajectory.append("GPT-4V Answer:\n%s"%raw_answer)
UnboundLocalError: local variable 'raw_answer' referenced before assignment
It seems that the agent cannot successfully infer the answer. If I add raw_answer = None
in line 81, the file can be run. But due to the lack of the right answer from inference, the agent cannot move in the right direction.
And the log is like this:
0%|▊ | 1/200 [04:08<13:42:57, 248.13s/it]No semantic annotations found
huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
To disable this warning, you can either:
- Avoid using `tokenizers` before the fork if possible
- Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
IMAGEIO FFMPEG_WRITER WARNING: input image is not divisible by macro_block_size=16, resizing from (1189, 1024) to (1200, 1024) to ensure video compatibility with most codecs and players. To prevent resizing, make your input image divisible by the macro_block_size or set the macro_block_size to 1 (risking incompatibility).
huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
To disable this warning, you can either:
- Avoid using `tokenizers` before the fork if possible
- Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
[rawvideo @ 0x5cdbbcf034c0] Stream #0: not enough frames to estimate rate; consider increasing probesize
[swscaler @ 0x5cdbbcf1cfc0] Warning: data is not aligned! This can lead to a speed loss
It means that the gpt4v doesn't work. Did you prepare your openai_key in the environment variables? You should check the file https://github.com/wzcai99/Pixel-Navigator/blob/master/llm_utils/gpt_request.py and try a unit test about the API calling functions.
It means that the gpt4v doesn't work. Did you prepare your openai_key in the environment variables? You should check the file https://github.com/wzcai99/Pixel-Navigator/blob/master/llm_utils/gpt_request.py and try a unit test about the API calling functions.
Thanks a lot! I find that the code uses AzureOpenAI API while I used the standard OpenAI API. After changing the way of calling API, the code can be run, and the agent can move towards the right direction.
However, I still face the problem:
[rawvideo @ 0x5df61bd784c0] Stream #0: not enough frames to estimate rate; consider increasing probesize
[swscaler @ 0x5df61bd91fc0] Warning: data is not aligned! This can lead to a speed loss
Those are some warnings from writing data into mp4 using imageio. But it doesn't matter.
Those are some warnings from writing data into mp4 using imageio. But it doesn't matter.
OK, that makes sense. Thank you for your time in helping me solve these problems, you really did an excellent job!
Thank you for your interest!
Hello! When I tried to evaluate the ObjNav Benchmark, I found the following problems: When I run
Some errors emerged.
I have already executed the
evaluate_policy.py
successfully, so I believe most of my environment setup is correct. But it is noticed that when running theevaluate_policy.py
, some errors likehappened sometimes, though they did not have obvious impacts on the results.
Does anyone know how to solve it?