Bo Yang

I am an Assistant Professor (2020.11-) in the Department of Computing at The Hong Kong Polytechnic University. I completed my D.Phil degree (2016.10-2020.09) in the Department of Computer Science at University of Oxford, supervised by Profs. Niki Trigoni and Andrew Markham. Prior to Oxford, I obtained an M.Phil degree from The University of Hong Kong and a B.Eng degree from Beijing University of Posts and Telecommunications.

In my D.Phil study, I interned at the Augumented Reality team of Amazon (Palo Alto, CA). In my M.Phil study, I interned at Hong Kong Applied Science and Technology Research Institute. In my undergraduate study, I was an exchange student at Universitat Politècnica de València (Valencia, Spain).

Email / vLAR Group / Google Scholar / Github

Research

I lead the Visual Learning and Reasoning (vLAR) Group, focusing on the fundamental research problems in machine learning, computer vision, and robotics. Our research goal is to build intelligent systems which endow machines to recover, understand, and eventually interact with the real 3D world. This includes accurate and efficient recognition, segmentation and reconstruction of all individual objects within large-scale 3D scenes.


Openings (Sep 2024 - ):

Several fully funded PostDoc/PhD/RA positions for 3D Vision/Robot Learning are available now. Email me with your CV and transcripts!

(All emails/CVs are carefully read and evaluated. Only matched candidates will be responded.)

News & Activities

[2024.05.23] Our extension of OGC is accepted by TPAMI 2024.

[2024.05.02] Our paper OSN is accepted by ICML 2024.

[2024.01.29] Our paper DynCatch is accepted by ICRA 2024.

[2024.01.09] Our paper unsupervised object segmentation is accepted by IJCV 2024.

[2023.09.22] Our papers RayDF and NVFi are accepted by NeurIPS 2023.

[2023.03.31] We are going to organize The 3rd Challenge on Point Cloud Understanding at ICCV 2023.

[2023.02.28] Our paper GrowSP is accepted by CVPR 2023.

[2023.01.22] Our paper DM-NeRF is accepted by ICLR 2023.

[2023.01.18] Our paper DecoupSkill is accepted by ICRA 2023.

[2022.09.03] Our papers OGC and UnsupObjSeg are accepted by NeurIPS 2022.

[2022.07.03] Our paper SQN is accepted by ECCV 2022.

[2022.05.26] Our extension of SpinNet is accepted by TPAMI.

[2022.05.04] We are going to organize The 2nd Challenge on Point Cloud Understanding at ECCV 2022.

[2022.04.19] Our paper RangeUDF is on arXiv.

[2021.11.11] Our extension of SensatUrban is accepted by IJCV.

[2021.07.23] Our paper GRF for neural rendering is accepted by ICCV 2021.

[2021.05.15] Our extension of RandLA-Net is accepted by TPAMI.

[2021.04.09] We are going to organize The 1st Challenge on Point Cloud Understanding at ICCV 2021.

[2021.03.01] Our papers SensatUrban and SpinNet are accepted by CVPR 2021.

[2021.02.28] Our paper RadarLoc is accepted by ICRA 2021.

[2020.11.28] We organize a tutorial about 3D point cloud learning at 3DV 2020.

[2020.02.27] Our paper RandLA-Net for 3D semantic segmentation is accepted by CVPR 2020.

[2019.09.03] Our paper 3D-BoNet is accepted as a spotlight at NeurIPS 2019.

[2019.08.16] Our paper AttSets is accepted by IJCV.

Five Selected Recent Publications (Full list at vLAR research page)
PontTuset

OSN: Infinite Representations of Dynamic 3D Scenes from Monocular Videos
Z. Song, J. Li, B. Yang
International Conference on Machine Learning (ICML), 2024
arXiv / Code

We present the first framework to represent dynamic 3D scenes in infinitely many ways from a monocular RGB video.

PontTuset

RayDF: Neural Ray-surface Distance Fields with Multi-view Consistency
Z. Liu, B. Yang*, Y. Luximon, A. Kumar, J. Li
Advances in Neural Information Processing Systems (NeurIPS), 2023
arXiv / Project Page / Code
(* indicates corresponding author)

We propose a novel ray-based 3D shape representation, achieving a 1000x faster speed in rendering.

PontTuset

NVFi: Neural Velocity Fields for 3D Physics Learning from Dynamic Videos
J. Li, Z. Song, B. Yang
Advances in Neural Information Processing Systems (NeurIPS), 2023
arXiv / Code

We present a novel framework to simultaneously learn the geometry, appearance, and physical velocity of 3D scenes.

PontTuset

GrowSP: Unsupervised Semantic Segmentation of 3D Point Clouds
Z. Zhang, B. Yang*, B. Wang, B. Li
IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2023
arXiv / Code
(* indicates corresponding author)

We propose the first unsupervised 3D semantic segmentation method, learning from growing superpoints in point clouds.

PontTuset

OGC: Unsupervised 3D Object Segmentation from Rigid Dynamics of Point Clouds
Z. Song, B. Yang
Advances in Neural Information Processing Systems (NeurIPS), 2022
arXiv / Video/ Code

We introduce the first unsupervised 3D object segmentation method on point clouds.


DPhil (PhD) Thesis
PontTuset

Learning to Reconstruct and Segment 3D Objects

B. Yang
Oxford Research Archive (PDF), 2020; (News: 机器之心报道)

Thesis committee (Transfer/Confirmation/Viva):
Alessandro Abate, Andrew Davison, Pawan Kumar, Andrew Zisserman.

This thesis aims to understand scenes and the objects within them by learning general and robust representations using deep neural networks, trained on large-scale real-world 3D data. In particular, the thesis makes three core contributions from object-level 3D shape estimation from single or multiple views to scene-level semantic understanding.

Talks & Services

[2023.05] Invited talk about Unsupervised 3D Semantic and Instance Segmentation at VALSE webinar.

[2022.12] Invited talk about Unsupervised 2D/3D Object Segmentation at TechBeat forum.

[2022.06] Invited talk about 3D Scene Reconstruction, Decomposition and Manipulation at Xiamen University.

[2021.10] Invited talk about 3D Representation Learning at GAMES Webinar.

[2021.04] Invited talk about Beyond Supervised Learning for 3D Representations at a CSIG workshop.

[2020.10] Invited talk about 3D Scene Understanding at Wonderland AI Summit. Check out the trailer.

[2020.09] Invited talk about 3D Point Cloud Segmentation at MFI 2020.

[2020.03] Invited talk about our RandLA-Net and 3D-BoNet at Shenlan. Here are the Video and Slides.

[2018 -] Regularly reviewing for top-tier conferences/journals in machine learning, computer vision, and robotics.

Teaching

Fall, 2023: Machine Learning and Data Analytics (The Hong Kong Polytechnic University).

Spring, 2023: Creative Digital Media Design (The Hong Kong Polytechnic University).

Spring&Fall, 2021&2022: Machine Learning and Data Analytics (The Hong Kong Polytechnic University).

Hilary, 2019: Knowledge Representation & Reasoning (University of Oxford).

Michaelmas, 2018&2017: Machine Learning (University of Oxford).

Spring, 2014: C++ Programming (The University of Hong Kong).

Mentoring (Full list at vLAR member page)

Qingyong Hu (Oct 2018 - ): Department of Computer Science at University of Oxford.

Alexander Trevithick (Oct 2019 - Mar 2021): Now PhD at UCSD.

Jianan Wang (May - Dec 2018): Now with Google DeepMind.

Zihang Lai (Oct 2017 - Mar 2018): Now PhD at CMU.

About Me

In my free time, I like playing tennis on lawns, clays, and hard surfaces. I also like to fly drones for landscape photography. Here's a video over the historic Oxford [Youtube, 腾讯视频], and another video for the scenic Lake District [Youtube]. Remember to turn up the volume for the background music.


Last update: 2024.09. Thanks.