Skip to content

A unified, agentic system for general-purpose robots, enabling multi-modal perception, mapping and localization, and autonomous mobility and manipulation, with intelligent interaction with users.

License

Notifications You must be signed in to change notification settings

HorizonRobotics/HoloAgent

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

HoloAgent

A unified, agentic system for general-purpose robots, enabling multi-modal perception, mapping and localization, and autonomous mobility and manipulation, with intelligent interaction with users.

🤖 FSR-VLN

Projcet 📄 arXiv 中文介绍

FSR-VLN is a core component of the HoloAgent framework and will be open-sourced soon. It provides natural language guided navigation and intelligent interaction for general-purpose robots, and is built on core agent components such as mapping and localization, multimodal perception, decision-making and planning, and memory management. At its core, FSR-VLN is a vision–language navigation system that integrates a Hierarchical Multi-modal Scene Graph (HMSG) for coarse-to-fine environment representation with Fast-to-Slow Navigation Reasoning (FSR), leveraging VLM-driven refinement to enable efficient, real-time, long-range spatial reasoning.

Overall Framework

Checklist

  • Release the code of FSR-VLN.

🏗 Pipeline

1. Semantic Mapping and Retrieval Pipeline

  • Task: Implement the semantic mapping and retrieval system based on the instructions in fsr_vln/README.md.
  • Steps:
    1. Download the necessary pre-trained model checkpoints.
    2. Download and configure the required datasets.
    3. Set up the environment and dependencies as specified.
    4. Run the complete pipeline to verify its functionality for semantic mapping and visual place retrieval.

2. Navigation Agent Setup and Execution

  • Task: Set up and test the navigation agent according to nav_agent/README.md.
  • Steps:
    1. Install all required dependencies for the navigation environment.
    2. Configure the necessary parameters and environment settings.
    3. Execute the navigation agent to ensure it runs successfully and performs its intended tasks.

📚 Publications & Citation

If you find our project useful, please consider citing it:

@misc{zhou2025fsrvlnfastslowreasoning,
      title={FSR-VLN: Fast and Slow Reasoning for Vision-Language Navigation with Hierarchical Multi-modal Scene Graph}, 
      author={Xiaolin Zhou and Tingyang Xiao and Liu Liu and Yucheng Wang and Maiyue Chen and Xinrui Meng and Xinjie Wang and Wei Feng and Wei Sui and Zhizhong Su},
      year={2025},
      eprint={2509.13733},
      archivePrefix={arXiv},
      primaryClass={cs.RO},
      url={https://arxiv.org/abs/2509.13733}, 
}

⚖️ License

This project is licensed under the Apache License 2.0. See the LICENSE file for details.

About

A unified, agentic system for general-purpose robots, enabling multi-modal perception, mapping and localization, and autonomous mobility and manipulation, with intelligent interaction with users.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 2

  •  
  •