ZHOU Shuchang (周舒畅)
Ph D,
shuchang [dot] zhou [at] gmail.com
I graduated from Tsinghua University in 2004 and obtained my PhD from the Chinese Academy of Sciences.
(2025) Step-audio: Unified understanding and generation in intelligent speech interaction
(ECCV'24) Chat-edit-3d: Interactive 3d scene editing via text prompts
(ICCAD'23) Sole: Hardware-software co-design of softmax and layernorm for efficient transformer inference
(AAAI'23) One is all: Bridging the gap between neural radiance fields architectures with progressive volume distillation
(2023) Occdepth: A depth-aware method for 3d semantic scene completion
(ICCV'23) Occ$^2$Net: Robust Image Matching Based on 3D Occupancy Estimation for Occluded Regions
(CVPR'23 highlight) Xiaotao Hu, et al. "A Dynamic Multi-Scale Voxel Flow Network for Video Prediction." ( arxiv, github)
(CVPR'23 highlight) Shengchao Zhou, et al. "UniDistill: A Universal Cross-Modality Knowledge Distillation Framework for 3D Object Detection in Bird’s-Eye View." ( arxiv, github)
(CVPR'23) Yun-Hao Cao, et al. "Three Guidelines You Should Know for Universally Slimmable Self-Supervised Learning." ( arxiv, github)
(AAAI oral) Shuangkang Fang, et al. "One is All: Bridging the Gap Between Neural Radiance Fields Architectures with Progressive Volume Distillation." ( arxiv, github)
1st place in the listening head generation track and 2nd place in the talking head generation track, ACM Multimedia ViCo 2022 Conversational Head Generation Challenge, report
"Megvii-hzwer" won the 2nd place in NIPS'17 Learning to Run Challenge , a competition on teaching a skeleton to run as fast as possible. We proposed the Actor-Critic Ensemble (ACE) method ( PDF , Github).
"Megvii" won the 1st place in all tracks in NIST TRAIT '16 , a competition on Text Recognition in the wild (OCR).