Miao Liu (刘淼)

I am a Reserach Scientist at META Reality Labs GenAI. I finished my Ph.D. in Robotics at Georgia Tech, advised by Prof. James Rehg. I also work closely with Prof.Yin Li from University of WisconsinMadison. I'm fortunate to work with Prof. Siyu Tang and Prof. Michael Black during my visit at ETH Zurich and Max Planck Institute. I enjoyed a wonderful time interning at Facebook Reality Lab, where I collaborated with Dr. Chao Li, Dr. Lingni Ma, Dr. Kiran Somasundaram, and Prof. Kristen Grauman on egocentric action recognition and localization. Before I joined Georgia Tech, I earned my Master degree from Carnegie Mellon University and Bachelor degree from Beihang University.

My previous research focuses on understanding human attention and human actions from First-Person Vision perspective. Since I joined Meta GenAI, I have been working on multimodal large language model and diffusion model.





News

Oct. 2024: Our LEGO paper has been nominated as one of the 15 award candidates at ECCV 2024.

Feb. 2024: Two corresponding-author papers accepted to ECCV 2024 (1 Poster, 1 Oral).

Feb. 2024: Three papers accepted to CVPR 2024 (1 Poster, 1 Highlight, 1 Oral).

Nov. 2023: One paper accepted to IEEE TPAMI.

Nov. 2023: One paper accepted to IJCV.

June. 2023: One paper accepted to ACL 2023 as Findings.

Nov. 2022: Our paper on Egocentric Gaze Estimation won the Best Student Paper Prize for BMVC 2022!

Sep. 2022: One paper accepted to BMVC 2022 for spotlight presentation!

Aug. 2022: I started my new journey at META Reality Labs.

Jul. 2022: Two papers accepted at ECCV 2022.

Jun. 2022: I succesfully denfended my thesis!

Apr. 2022: Techinical talk at META AI Research.

Mar. 2022: Techinical talk at Amazon.

Mar. 2022: Our Ego4D paper was accepted to CVPR2022 for oral presentation, Best Paper Finalist .

Feb. 2022: Techinical talk at Apple.

Oct. 2021: Our Ego4D project has launched! Check out the arXiv paper.

Oct. 2021: One paper accepted to 3DV 2021.

Jul. 2021: I passed my thesis proposal.

Jan. 2021: One paper accepted to IEEE TPAMI.

Oct. 2020: Techinical talk "Towards an In-Depth Understanding of Egocentric Actions" at Facebook Reality Lab.

Aug. 2020: Invited talk at ECCV 2020 Workshop on Egocentric Perception, Interaction and Computing (EPIC)

Aug. 2020: One paper accepted to BMVC 2020 for oral presentation!

Jul. 2020: One paper accepted to IMWUT (UbiComp 2020).

Jul. 2020: One Paper accepted to ECCV 2020 for oral presentation!

Jun. 2020: Invited talk at CVPR 2020 Workshop on Egocentric Perception, Interaction and Computing (EPIC)

Jun. 2020: I won 2nd place in the EPIC-KITCHENS Challenge 2020 for Action Recognition in Unseen Environments.

Jun. 2020: Started my internship at Facebook Reality Lab with Dr. Chao Li, and Dr. Kiran Somasundaram during Jun. 2020 - Dec. 2020.

Jan. 2020: Started my internship with Prof. Siyu Tang at ETH Zurich during Jan. 2020 - Apr. 2020.

Sep. 2019: Started my internship with Prof. Siyu Tang and Prof. Michael Black at Max Planck Institute during Sep. 2019 - Jan. 2020.

Publication

Google Scholar

Check out my Google Scholar page for my latest publications.

Bolin Lai, Xiaoliang Dai, Lawrence Chen, Guan Pang, James M. Rehg, Miao Liu. LEGO: Learning EGOcentric Action Frame Generation via Visual Instruction Tuning, accepted by European Conference on Computer Vision (ECCV) 2024 (Oral, Best Paper Award Candidate 15/8585). [arXiv]

Bolin Lai, Fiona Ryan, Wenqi Jia, Miao Liu†, James M. Rehg†. Listen to Look into the Future: Audio-Visual Egocentric Gaze Anticipation, accepted by European Conference on Computer Vision (ECCV) 2024 . [arXiv] †: Co-corresponding Author

Yunhao Ge*, Yihe Tang*, Jiashu Xu*, Cem Gokmen*, Chengshu Li, Wensi Ai, Benjamin Jose Martinez, Arman Aydin, Mona Anvari, Ayush K Chakravarthy, Hong-Xing Yu, Josiah Wong, Sanjana Srivastava, Sharon Lee, Shengxin Zha, Laurent Itti, Yunzhu Li, Roberto Martín-Martín, Miao Liu, Pengchuan Zhang, Ruohan Zhang, Li Fei-Fei, Jiajun Wu. BEHAVIOR Vision Suite: Customizable Dataset Generation via Simulation, accepted by Computer Vision and Pattern Recognition Conference (CVPR) 2024 (Spotlight). [arXiv] *: Euqual Contribution

With Kristen Grauman, et al. Ego-Exo4D: Understanding Skilled Human Activity from First- and Third-Person Perspectives, accepted by Computer Vision and Pattern Recognition Conference (CVPR) 2024 (Oral). [arXiv]

Wenqi Jia, Miao Liu, Hao Jiang, Ishwarya Ananthabhotla, James Rehg, Vamsi Krishna Ithapu, Ruohan Gao. The Audio-Visual Conversational Graph: From an Egocentric-Exocentric Perspective, accepted by Computer Vision and Pattern Recognition Conference (CVPR) 2024. [arXiv]

Bolin Lai, Miao Liu+, Fiona Ryan, James M. Rehg. In the eye of transformer: Global–local correlation for egocentric gaze estimation and beyond, accepted by International Journal of Computer Vision (IJCV). †: Student Mentorx [arXiv]

Bolin Lai*, Hongxin Zhang*, Miao Liu*, Aryan Pariani*, Fiona Ryan, Wenqi Jia, Shirley Anugrah Hayati, James M. Rehg, Diyi Yang. In the Eye of Transformer: Global-Local Correlation for Egocentric Gaze Estimation, accepted by the Association for Computational Linguistics (ACL) 2023 (Findings). [arXiv] *: Euqual Contribution

Bolin Lai, Miao Liu†, Fiona Ryan, James M. Rehg. In the Eye of Transformer: Global-Local Correlation for Egocentric Gaze Estimation, accepted by British Machine Vision Conference (BMVC) 2022 †: Student Mentor, Co-corresponding Author (Spotlight, Best Student Paper Prize). [arXiv]

Wenqi Jia*, Miao Liu*, James M. Rehg. Generative Adversarial Network for Future Hand Segmentation from Egocentric Video, accepted by European Conference on Computer Vision (ECCV) 2022. [arXiv] *: Euqual Contribution

Miao Liu, Lingni Ma, Kiran Somasundaram, Yin Li, Kristen Grauman, James M. Rehg, Chao Li. Egocentric Activity Recognition and Localization on a 3D Map, accepted by European Conference on Computer Vision (ECCV) 2022 .[arXiv]

With Kristen Grauman, et al. Ego4D: Around the World in 3,000 Hours of Egocentric Video, accepted by Computer Vision and Pattern Recognition Conference (CVPR) 2022 (Oral, best paper finalist, 33/8161) [arXiv] Key driver for Social Benchmark and Forecasting Benchmark

Miao Liu, Dexin Yang, Yan Zhang, Zhaopeng Cui, James M. Rehg, and Siyu Tang. 4D Human Body Capture from Egocentric Video via 3D Scene Grounding. accpeted by Interantional Conference on 3D Vision. [arXiv] [project page]

Yin Li, Miao Liu, and James M. Rehg. In the Eye of the Beholder: Gaze and Actions in First Person Video, accepted by IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI) 2021. [arXiv]

Miao Liu, Xin Chen, Yun Zhang, Yin Li and James M. Rehg. Attention Distillation for Learning Video Representations, accepted by British Machine Vision Conference (BMVC) 2020 (Oral, acceptance rate 5.0%). [pdf] [project page]

Yun Zhang*, Shibo Zhang*, Miao Liu, Elyse Daly, Samuel Battalio, Santosh Kumar, Bonnie Spring, James M. Rehg, Dr Nabil Alshurafa. SyncWISE: Window Induced Shift Estimation for Synchronization of Video and Accelerometry from Wearable Sensors, accepted by Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. (IMUWT/UbiComp) 2020 (* denotes equal contribution) [pdf]

Miao Liu, Siyu Tang, Yin Li and James M. Rehg. Forecasting Human Object Interaction: Joint Prediction of Motor Attention and Actions in First Person Vision, accepted by European Conference on Computer Vision (ECCV) 2020 (Oral, acceptence rate 2.0%). [pdf] [project page]

Yin Li, Miao Liu, and James M. Rehg. In the Eye of Beholder: Joint Learning of Gaze and Actions in First Person Video, accepted by European Conference on Computer Vision (ECCV) 2018. [pdf]

Xingbang Yang, Tianmiao Wang, Jianhong Liang, Guocai Yao, and Miao Liu. Survey on the novel hybrid aquatic-aerial amphibious aircraft: Aquatic unmanned aerial vehicle (AquaUAV), accepted by Progress in Aerospace Sciences 74 (2015): 131-151.

Guocai Yao, Jianhong Liang, Tianmiao Wang, Xingbang Yang, Miao Liu, and Yicheng Zhang. Submersible unmanned flying boat: Design and experiment, accepted by IEEE International Conference on Robotics and Biomimetics (ROBIO) 2014.

Tianmiao Wang, Yicheng Zhang, Chaolei Wang, Jianhong Liang, Han Gao, Miao Liu, Qinpu Guan, and Anqi Sun. Indoor visual navigation system based on paired-landmark for small UAVs, accepted by IEEE International Conference on Robotics and Biomimetics (ROBIO) 2014.





Teaching

Teaching Assitant for CS 7643 Deep Learning, Georgia Tech, Spring 2019

Teaching Assitant for CS 4476 Intro to Computer Vision, Georgia Tech, Summer 2019





Talks

Oct. 2020 Techinical talk "Towards an In-Depth Understanding of Egocentric Actions" at Facebook Reality Lab.

Aug. 2020 Invited talk at ECCV2020 Workshop on Egocentric Perception, Interaction and Computing (EPIC)[Video]

Jun. 2020 Invited talk at CVPR2020 Workshop on Egocentric Perception, Interaction and Computing (EPIC)[Video]





Contact

I'm more than happy to discuss any potential research oppotunities.

Georgia Tech Email

Gmail