Skip to content

Commit a838ab8

Browse files
committed
[Fix] Pre-commit issues
1 parent 0489a7a commit a838ab8

11 files changed

Lines changed: 250 additions & 138 deletions

File tree

README.md

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -218,19 +218,19 @@ For example, raising issues, fixing bugs in the framework, and adapting or addin
218218

219219
## 🚀 Community Deployment & Best Practices
220220

221-
We are excited to see InternNav being deployed and extended by the community across different robots and real-world scenarios.
221+
We are excited to see InternNav being deployed and extended by the community across different robots and real-world scenarios.
222222
Below are selected community-driven deployment guides and solution write-ups, which may serve as practical references for advanced users.
223223

224-
- **IROS Challenge Nav Track: Champion Solution (2025)**
225-
A complete system-level solution and design analysis for Vision-and-Language Navigation in Physical Environments.
224+
- **IROS Challenge Nav Track: Champion Solution (2025)**
225+
A complete system-level solution and design analysis for Vision-and-Language Navigation in Physical Environments.
226226
🔗 https://zhuanlan.zhihu.com/p/1969046543286907790
227227

228-
- **Go2 Series Deployment Tutorial (ShanghaiTech University)**
229-
Step-by-step edge deployment guide for InternNav-based perception and navigation.
228+
- **Go2 Series Deployment Tutorial (ShanghaiTech University)**
229+
Step-by-step edge deployment guide for InternNav-based perception and navigation.
230230
🔗 https://github.com/cmjang/InternNav-deploy
231231

232-
- **G1 Series Deployment Tutorial (Wuhan University)**
233-
Detailed educational materials on vision-language navigation deployment.
232+
- **G1 Series Deployment Tutorial (Wuhan University)**
233+
Detailed educational materials on vision-language navigation deployment.
234234
🔗 [*Chapter 5: Vision-Language Navigation (Part II)*](https://mp.weixin.qq.com/s/p3cJzbRvecMajiTh9mXoAw)
235235

236236
## 🔗 Citation

internnav/agent/dialog_agent.py

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -16,13 +16,13 @@
1616
from internnav.configs.agent import AgentCfg
1717

1818
try:
19+
from depth_camera_filtering import filter_depth
20+
from habitat.tasks.nav.shortest_path_follower import ShortestPathFollower
1921
from transformers import (
2022
AutoProcessor,
2123
AutoTokenizer,
2224
Qwen2_5_VLForConditionalGeneration,
2325
)
24-
from depth_camera_filtering import filter_depth
25-
from habitat.tasks.nav.shortest_path_follower import ShortestPathFollower
2626
except Exception as e:
2727
print(f"Warning: ({e}), Habitat Evaluation is not loaded in this runtime. Ignore this if not using Habitat.")
2828

@@ -47,12 +47,12 @@ def split_and_clean(text):
4747
@Agent.register('dialog')
4848
class DialogAgent(Agent):
4949
"""Vision-language navigation agent that can either move or ask an oracle via dialog. The agent builds a multimodal
50-
chat prompt from current/historical RGB observations (and optional look-down frames), runs a Qwen2.5-VL model to
51-
produce either an action sequence, a pixel waypoint, or a dialog query, then converts the model output into
50+
chat prompt from current/historical RGB observations (and optional look-down frames), runs a Qwen2.5-VL model to
51+
produce either an action sequence, a pixel waypoint, or a dialog query, then converts the model output into
5252
simulator actions and (optionally) a world-space navigation goal.
5353
5454
Args:
55-
agent_config (AgentCfg): AgentCfg containing model_settings (e.g., task name, sensor config, model path, mode,
55+
agent_config (AgentCfg): AgentCfg containing model_settings (e.g., task name, sensor config, model path, mode,
5656
resizing, dialog flags, and generation parameters) and runtime info such as local_rank.
5757
"""
5858

@@ -440,7 +440,7 @@ def pixel_to_gps(self, pixel, depth, intrinsic, tf_camera_to_episodic):
440440
pixel (Tuple[int, int] | List[int] | np.ndarray): pixel coordinate in (v, u) indexing as used here.
441441
depth (np.ndarray): depth image of shape (H, W) in meters, where depth[v, u] returns the metric depth.
442442
intrinsic (np.ndarray): camera intrinsic matrix.
443-
tf_camera_to_episodic (np.ndarray): homogeneous transform of shape (4, 4) mapping camera-frame points to
443+
tf_camera_to_episodic (np.ndarray): homogeneous transform of shape (4, 4) mapping camera-frame points to
444444
the episodic frame.
445445
446446
Returns:

internnav/dataset/internvla_n1_lerobot_dataset.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -17,8 +17,8 @@
1717
from torchcodec.decoders import VideoDecoder
1818
from transformers.image_utils import to_numpy_array
1919

20-
from .vlln_lerobot_dataset import VLLNDataset
2120
from .rope2d import get_rope_index_2, get_rope_index_25
21+
from .vlln_lerobot_dataset import VLLNDataset
2222

2323
# Define placeholders for dataset paths
2424
CAMBRIAN_737K = {
@@ -1330,6 +1330,7 @@ def __call__(self, instances: Sequence[Dict]) -> Dict[str, torch.Tensor]:
13301330

13311331
return batch
13321332

1333+
13331334
class CombinedDataset(Dataset):
13341335
"""
13351336
Combine multiple datasets into a single dataset interface.
@@ -1338,6 +1339,7 @@ class CombinedDataset(Dataset):
13381339
It concatenates samples from all provided datasets and optionally shuffles
13391340
the global index mapping (without changing the underlying datasets).
13401341
"""
1342+
13411343
def __init__(self, datasets, shuffle=False):
13421344
super(CombinedDataset, self).__init__()
13431345
self.datasets = datasets

0 commit comments

Comments
 (0)