Lulu Tang
I am a researcher at Beijing Academy of Artificial Intelligence (BAAI).
I completed my postdoctoral research at BAAI and Tsinghua University, under the supervision of Prof. Tiejun Huang and Prof. Jiwen Lu. Prior to that, I obtained my Ph.D. degree from University of Macau, advised by Prof. Zhi-xin Yang and frequently collaborating with Prof. Kui Jia.
My research interests lie in the area of 3D computer vision, 3D generation models, and Vision-Language foundation models.
Email  | 
GitHub  | 
Google Scholar
|
|
Selected Publications
(*: Equal Contribution, ♣: corresponding author)
|
|
You See it, You Got it: Learning 3D Creation on Pose-Free Videos at Scale
Baorui Ma*,
Huachen Gao*,
Haoge Deng*,
Zhengxiong Luo,
Tiejun Huang,
Lulu Tang♣,
Xinlong Wang♣,
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2025,(Highlight)
[arxiv]
[Project page]
[Code]
[Dataset]
[Post]
See3D is a scalable visual-conditional MVD model for open-world 3D creation, which can be trained on web-scale video collections without camera pose annotations.
|
|
PiCo: Jailbreaking Multimodal Large Language Models via Pictorial Code Contextualization
Aofan Liu,
Lulu Tang♣,
Ting Pan,
Yuguo Yin,
Bin Wang,
Ao Yang,
IEEE International Conference on Multimedia & Expo (ICME), 2025
[arXiv]
PiCo is a jailbreaking framework that bypasses advanced MLLM defenses using token-level typographic attacks to evade input filters and embeds malicious intent in programming instructions to avoid runtime monitoring.
|
|
TAP: Tokenize Anything via Prompting
Ting Pan*,
Lulu Tang*,
Xinlong Wang♣,
Shiguang Shan,
European Conference on Computer Vision (ECCV), 2024
[arXiv]
[Code]
[Demo]
TAP is a unified and promptable model capable of simultaneously segmenting, recognizing, and captioning arbitrary regions, with flexible visual prompts (point, box and sketch).
|
|
Canonical correlation analysis regularization: An effective deep multiview learning baseline for RGB-D object recognition
Lulu Tang*,
Zhi-Xin Yang♣,
Kui Jia♣
IEEE Transactions on Cognitive and Developmental Systems, 2019
[Paper]
|
|