I am a fourth-year PhD student in the Department of Computer Science at Boston University (BU), advised by Prof. Eshed Ohn-Bar. My research spans various areas, focusing on machine intelligence and computer vision, especially in their application to autonomous driving systems. Before joining Prof. Eshed Ohn-Bar's group, I worked on prize-collecting Steiner tree (PCST) project with Prof. Alina Ene (2020-2022).
Prior to BU, I worked with Prof. Dingzhu Du as a research assistant, focusing on submodular function maximization problem (2019-2020). I received my master's degree in the Department of Mathematics at East China Normal University (ECNU), under the supervision of Prof. Changhong Lu, with a focus on Combinatorics and Graph Theory (2017-2020).
We present a highly scalable self-training framework for incrementally adapting vision-based end-to-end autonomous driving policies in a semi-supervised manner, i.e., over a continual stream of incoming video data. To facilitate large-scale model training (e.g., open web or unlabeled data), we do not assume access to ground-truth labels and instead estimate pseudo-label policy targets for each video. Our framework comprises three key components: knowledge distillation, a sample purification module, and an exploration and knowledge retention mechanism. Trained as a complete never-ending learning system, we demonstrate state-of-the-art performance on training from domain-changing data as well as millions of images from the open web.
We propose XVO, a semi-supervised learning method for training generalized monocular Visual Odometry (VO) models with robust off-the-shelf operation across diverse datasets and settings. We empirically demonstrate the benefits of semi-supervised training for learning a general-purpose direct VO regression network, alongside the benefits of multi-modal supervision to facilitate generalized representations for the VO task. Combined with the proposed semi-supervised step, XVO demonstrates off-the-shelf knowledge transfer across diverse conditions on KITTI, nuScenes, and Argoverse without fine-tuning.
We consider the problem of maximizing monotone submodular function over the bounded integer lattice with a cardinality constraint. We propose a random greedy \( (1-\frac{1}{e}) \)-approximation algorithm and a deterministic \( \frac{1}{e} \)-approximation algorithm. Both algorithms work in value oracle model. In the random greedy algorithm, we assume the monotone submodular function satisfies diminishing return property, which is not an equivalent definition of submodularity on integer lattice. Additionally, our random greedy algorithm makes \( \mathcal{O}((|\mathbb{E}|+1)\cdot k) \) value oracle queries and deterministic algorithm makes \( \mathcal{O}(|\mathbb{E}|\cdot B\cdot k^3) \) value oracle queries.