Skip to content

Commit db4e06d

Browse files
committed
save
1 parent b387073 commit db4e06d

2 files changed

Lines changed: 3 additions & 3 deletions

File tree

content/authors/admin/_index.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@ organizations:
1414
- name: Stanford University
1515
url: https://www.stanford.edu/
1616

17-
bio: "Autonomous systems (RL, perception, decision-making) and LLM-powered scientific agents for multi-omics workflows."
17+
bio: "Graduate student in Aeronautics & Astronautics at Stanford, working on robot learning and autonomous systems, with a focus on sim-to-real transfer and learned control policies."
1818

1919
interests:
2020
- Autonomous Robotics

content/project/Language-steered-drones/index.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ tags:
1212
- Autonomous Systems
1313
---
1414

15-
## Simulated Trajectory
15+
## Drone navigating to a leaf blower
1616
<iframe
1717
width="100%"
1818
height="420"
@@ -23,7 +23,7 @@ tags:
2323

2424
Developed a vision-language navigation (VLN) policy for autonomous drone flight in photorealistic 3D Gaussian Splatting environments. Given a natural language instruction like "go to the green leafblower," the drone autonomously identifies and navigates to the target — collision-free.
2525

26-
The video shows the drone's onboard view: RGB (left) and semantic similarity field (right) for the query "green and pink leafblower." The system first encodes the language instruction via CLIP embeddings, localizes the target using CLIPSeg semantic segmentation, and generates real-time control commands to navigate through a cluttered indoor environment while avoiding obstacles.
26+
The video shows the drone's onboard view: RGB (left) and semantic similarity field (right) for the query "green and pink leafblower." In the right view, red indicates high similarity with the query and blue indicates low similarity — the drone navigates towards the high-similarity region while avoiding obstacles. The system first encodes the language instruction via CLIP embeddings, localizes the target using CLIPSeg semantic segmentation, and generates real-time control commands to navigate through a cluttered indoor environment.
2727

2828
The control policy is a lightweight neural network (SqueezeNet Commander MLP) trained via Behavioral Cloning from an ACADOS-based MPC expert. A key contribution is the design and implementation of a full DAgger (Dataset Aggregation) pipeline — including mixed-policy rollouts, expert annotation filtering, iterative retraining with best-model checkpointing, and automated benchmarking — to systematically correct for compounding errors under distribution shift. A second key contribution is the introduction of explicit geometric features — bearing and elevation — extracted from the CLIPSeg heatmap centroid, providing the policy with a direct spatial signal for goal-directed control. This replaces the previous approach where target localization had to be implicitly learned from visual embeddings alone.
2929

0 commit comments

Comments
 (0)