Open chch9907 opened 10 months ago
Thank you for sharing your insights and potential applications of KubeEdge in the field of Offline Visual Navigation.
The idea of leveraging KubeEdge to curate datasets from diverse edge devices aligns with our SIG's goals and could be a valuable addition to the realm of robotic navigation.
We appreciate your dedication and look forward to the progress of your work. I fully support your application for it to be a subproject of KubeEdge. Feel free to contact me if you need any assistance or clarification.
Community members, please share your thoughts and feedback on this issue~
/PTAL @kubeedge/sig-robotics-repo-admins
/assign @Poorunga
Sounds great, would you like to share it at the next sig meeting?
Currently, existing robotic navigation methods have been expert in simultaneous mapping and localization (SLAM) and path planning not only in the large-scale environment but also dense-crowd scenarios, among which the AI-powered methods intend to improve the algorithm efficiency and take advantage of the semantic information. However, there still remain some challanges to be solved:
Emergent scene representation for embodied AI: Scene (Map) representation have long been within occupancy grid map and signed distance field but both require rich 3D geometric features and unable to explain the non-observed view of the scene, neural radiance field or 3D gaussian splatting turn out to be the promising and effective techniques that can handle the above issues, benifiting the downstream embodied interaction or navigation. But training such neural map requires training dataset, and if the scenario is very large the computation cost can be drastically high. Fortunately, KubeEdge is skilled at doing these. That is, it can curate the dataset from different edge devices, meanwhile it feeds back the edge devices with abundant computation resource.
Goal searching in unknown environment: Given pre-built map, existing methods are mature to handle the navigation tasks even involving the high dynamic obstacle avoidance and lifelong SLAM, whereas goal-oriented navigation in unknown real-world environment is still an open task that need to be explored. Existing work mainly consider utilizing the common sense on the topology relation between objects or rooms in household scenarios. However, in real-world cases, the object arrangement can be dynamically changed, and even the scenario can be large and space without any objects (e.g., outdoor scenario). If utilizing the cloud capability of KubeEdge to help infer the large language model (LLM), the edge device can resort to the LLM-based policy to adaptively understand the unknown world.
Facilitating the deployment of learning-based methods: The sim-to-real gap is a long-standing issue that hinders deploying the learning-based methods to real-world application scenarios. This comes from the fact that existing open-source simulation data mainly concern the indoor cases and lack the outdoor data. Although scene reconstruction can build the world model of the deploying environment into simulation, it is often troubled by the texture defect and fidelity gap. Currently some effort are dedicated to directly learning from the offline data, i.e., the experience to learn the geometric attributes of the real world. Thanks to the edge-cloud communication capability of KubeEdge, I also contribute a work on the outdoor visual navigation based on offline reinforcement learning, which solely uses the visual sensor and train on self-collected offline dataset. It also handle the localization lost issue during navigation by a novel self-correcting method that generates future states in the latent space and estimates the novelty as the guidance to search the familiar places. This work is also submitted to 2024 ICRA. Once accepted, I will apply for it to be a subproject of KubeEdge. As a preprint version, one can check the progress at https://github.com/KubeEdge4Robotics/ScaleNav.
The above are my superficial opinions about the potential advantages and application of KubeEdge. I am eager to listen to the comments from the community. Thanks!