Skip to content
/ IL_ViT Public

This is a demo to test ViT-based foundation models for image-navigation tasks, using immitation learning with Habitat Sim.

License

Notifications You must be signed in to change notification settings

DoeLoz/IL_ViT

Repository files navigation

IL_ViT

This is a demo to test ViT-based foundation models for image-navigation tasks, using immitation learning with Habitat Sim.

Environment Configuration

Ensure that Habitat-sim and Habitat-lab are installed properly. For datasets, please refer to GibsonEnv and PointNav (Task Dataset)

Data Generation

python collect_IL_data.py --ep-per-env 50 --num-procs 4 --split train --data-dir /your-path-to-save-data

This will generate data for immitation learning. Settings can be customed in collect_IL_data.py.

Training

python train_bc.py --config configs/vgm.yaml --stop --gpu 0

By default we use a simplified OVRL model with ViT and compression layer. In comparison to a single ViT or ResNet, rewrite policy = part。

Evaluation

Load your trained model and run:

python validation.py

If the agent reaches a distance of less than 1 meter to the goal, the current task episode is marked as a success. Continuous stagnation or exceeding the time limit results in a failure. The current and average SR, SPL, DTS are displayed for review. To mitigate the extreme impact of data sparsity on imitation learning, the output of the stop action 0 in the simulator is replaced with the forward action 1. After all, this is just a demo to test the foundation model.

About

This is a demo to test ViT-based foundation models for image-navigation tasks, using immitation learning with Habitat Sim.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages