Weiyang Liu

 

Google Scholar   Github   Twitter

 

University of Cambridge
Max Planck Institute for Intelligent Systems

About Me

I conduct research at Cambridge and MPI Tübingen with Adrian Weller and Bernhard Schölkopf. Previously, I spent wonderful years at Georgia Tech. I have also spent time at Google Brain, Nvidia, and MERL.

I work on principled modeling of inductive bias in machine learning. My research seeks to understand how inductive bias determines generalization, and to develop "light-yet-sweet" generalizable models: (i) light: conceptually simple in methodology and easy to implement in practice, (ii) sweet: having clear intuitions and non-trivial theoretical guarantees.

Over the years, I always find myself fascinated by geometric invariance, symmetry, structures (graph, causality) and how they can benefit generalization as a guiding principle. More recently, I become very passionate about foundation models (how to simulate human-level intelligence) and 3D/4D generative modeling (how to recreate and simulate the physical world).

I always believe in two principles in my research: (i) insight must precede application, and (ii) everything should be made as simple as possible, but not simpler. I try to follow certain research values.

I am on the academic job market this upcoming year. Feel free to reach out if there is a good fit!

    - Focus on creating novel ideas, not publishing papers
    - Follow curiosity and passion, not trends
    - Ideas are not owned, but come with debts to those who came before
    - Ideas become stronger when shared, discussed and criticized
    - Life is surprisingly short, so solve problems that interest and excite you most
    - It is good to be quick, but it is more important to be deep
    - Think like an amateur, do as an expert
    - This is not only about how to do research, but also how to live your life

Mentoring

I take great pleasure to (co-)mentor a few talented and highly motivated students. Mentoring and working with junior students is truely a privilege, and I always learn from and get inspired by them. I am fortunate to work with (alphabetical order):

   - Zhen Liu (PhD student at University of Montreal)
   - Zeju Qiu (Master student at Technical University of Munich)
   - Longhui Yu (Master student at Peking University)

Recent Highlight

GraphDreamer: Compositional 3D Scene Synthesis from Scene Graphs
Gege Gao, Weiyang Liu*, Anpei Chen, Andreas Geiger, Bernhard Schölkopf

Preprint 2023

arXiv | code | project | bib

  @article{gao2023graphdreamer,
      author = {Gao, Gege and Liu, Weiyang and Chen, Anpei and Geiger, Andreas and Sch{\"o}lkopf, Bernhard},
      title = {GraphDreamer: Compositional 3D Scene Synthesis from Scene Graphs},
      journal = {arXiv preprint arXiv:2312.00093},
      year = {2023}}

Parameter-Efficient Orthogonal Finetuning via Butterfly Factorization
Weiyang Liu*, Zeju Qiu*, Yao Feng**, Yuliang Xiu**, Yuxuan Xue**, Longhui Yu**, Haiwen Feng, Zhen Liu, Juyeon Heo, Songyou Peng, Yandong Wen, Michael J. Black, Adrian Weller, Bernhard Schölkopf

Preprint 2023

arXiv | code | project | bib

  @article{liu2023boft,
      author = {Liu, Weiyang and Qiu, Zeju and Feng, Yao and Xiu, Yuliang and Xue, Yuxuan and Yu, Longhui and Feng, Haiwen and Liu, Zhen 
        and Heo, Juyeon and Peng, Songyou and Wen, Yandong and Black, Michael J. and Weller, Adrian and Sch{\"o}lkopf, Bernhard},
      title = {Parameter-Efficient Orthogonal Finetuning via Butterfly Factorization},
      journal = {arXiv preprint arXiv:2311.06243},
      year = {2023}}

Ghost on the Shell: An Expressive Representation of General 3D Shapes
Zhen Liu, Yao Feng, Yuliang Xiu, Weiyang Liu*, Liam Paull, Michael J. Black, Bernhard Schölkopf

Preprint 2023

arXiv | code | project | bib

  @article{Liu2023gshell,
      title={Ghost on the Shell: An Expressive Representation of General 3D Shapes},
      author={Liu, Zhen and Feng, Yao and Xiu, Yuliang and Liu, Weiyang 
        and Paull, Liam and Black, Michael J and Sch{\"o}lkopf, Bernhard},
      journal={arXiv preprint arXiv:2310.15168},
      year={2023}}

MetaMath: Bootstrap Your Own Mathematical Questions for Large Language Models
Longhui Yu, Weisen Jiang, Han Shi, Jincheng Yu, Zhengying Liu, Yu Zhang, James Kwok, Zhenguo Li, Adrian Weller, Weiyang Liu

Preprint 2023

arXiv | code | project | huggingface | bib

  @article{Yu2023MetaMath,
      title={MetaMath: Bootstrap Your Own Mathematical Questions for Large Language Models},
      author={Yu, Longhui and Jiang, Weisen and Shi, Han and Yu, Jincheng and Liu, Zhengying 
        and Zhang, Yu and Kwok, James T and Li, Zhenguo and Weller, Adrian and Liu, Weiyang},
      journal={arXiv preprint arXiv:2309.12284},
      year={2023}}

Controlling Text-to-Image Diffusion by Orthogonal Finetuning
Zeju Qiu*, Weiyang Liu*, Haiwen Feng, Yuxuan Xue, Yao Feng, Zhen Liu, Dan Zhang, Adrian Weller, Bernhard Schölkopf

NeurIPS 2023

arXiv | code | project | bib

  @InProceedings{Qiu2023OFT,
      title={Controlling Text-to-Image Diffusion by Orthogonal Finetuning},
      author={Qiu, Zeju and Liu, Weiyang and Feng, Haiwen and Xue, Yuxuan and Feng, Yao 
        and Liu, Zhen and Zhang, Dan and Weller, Adrian and Sch{\"o}lkopf, Bernhard},
      booktitle={NeurIPS},
      year={2023}}

Publications

Last updated on 27th October 2023.
This site has been visisted Free Web Counter times in total.