MIV-XJTU/JanusVLN
[ICLR2026] Official implementation for "JanusVLN: Decoupling Semantics and Spatiality with Dual Implicit Memory for Vision-Language Navigation"
This project helps create AI agents that can navigate complex indoor environments based on natural language instructions. You provide the AI with a written command, like "Go past the kitchen and turn left into the living room," and a 3D map of the space. The AI then plans and executes a path through the virtual environment. This is for researchers and developers working on embodied AI, robotics, and virtual reality applications.
508 stars.
Use this if you are developing or evaluating AI models that need to understand spatial relationships and follow human-like directions within simulated 3D spaces.
Not ideal if you need a pre-built navigation system for real-world robots or applications outside of research on vision-language navigation.
Stars
508
Forks
35
Language
Python
License
—
Category
Last pushed
Jan 26, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/MIV-XJTU/JanusVLN"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
KimMeen/Time-LLM
[ICLR 2024] Official implementation of " 🦙 Time-LLM: Time Series Forecasting by Reprogramming...
om-ai-lab/VLM-R1
Solve Visual Understanding with Reinforced VLMs
bytedance/SALMONN
SALMONN family: A suite of advanced multi-modal LLMs
NVlabs/OmniVinci
OmniVinci is an omni-modal LLM for joint understanding of vision, audio, and language.
fixie-ai/ultravox
A fast multimodal LLM for real-time voice