Publication
I am currently exploring video generative model as control policy for robotics.
|
|
Context-Aware Replanning with Pre-Explored Semantic Map for Object Navigation
Po-Chen Ko*, Hung-Ting Su*, Ching-Yuan Chen*, Jia-Fong Yeh, Min Sun, Winston H. Hsu
CoRL 2024
arxiv /
website /
paper
We improved VLM-based semantic maps for Object Navigaiton by leveraging uncertainty information from the map to aid candidate selection.
|
|
Learning to Act from Actionless Video through Dense Correspondences
Po-Chen Ko, Jiayuan Mao, Yilun Du, Shao-Hua Sun, Joshua B. Tenenbaum
ICLR 2024 (Spotlight)
arxiv /
video /
website /
code /
paper
We developed a robot policy using images, which can be trained without action annotations. It’s trained on RGB videos, effective for table-top tasks and navigation. Our framework also allows rapid modeling with just 4 GPUs in a day.
|
Awards
|
NTUEE Undergraduate Innovation Award
1st. prize
|
|