Obserwuj
Di Huang
Di Huang
Zweryfikowany adres z uni.sydney.edu.au - Strona główna
Tytuł
Cytowane przez
Cytowane przez
Rok
Motiongpt: Finetuned llms are general-purpose motion generators
Y Zhang, D Huang, B Liu, S Tang, Y Lu, L Chen, L Bai, Q Chu, N Yu, ...
Proceedings of the AAAI Conference on Artificial Intelligence 38 (7), 7368-7376, 2024
842024
Onepose++: Keypoint-free one-shot object pose estimation without CAD models
X He, J Sun, Y Wang, D Huang, H Bao, X Zhou
Advances in Neural Information Processing Systems 35, 35103-35115, 2022
812022
Seeing is not always believing: A Quantitative Study on Human Perception of AI-Generated Images
Z Lu, D Huang, L Bai, X Liu, J Qu, W Ouyang
Thirty-seventh Conference on Neural Information Processing Systems Datasets …, 2023
76*2023
Ponder: Point cloud pre-training via neural rendering
D Huang, S Peng, T He, H Yang, X Zhou, W Ouyang
Proceedings of the IEEE/CVF International Conference on Computer Vision …, 2023
352023
Ponderv2: Pave the way for 3d foundataion model with a universal pre-training paradigm
H Zhu, H Yang, X Wu, D Huang, S Zhang, X He, T He, H Zhao, C Shen, ...
arXiv preprint arXiv:2310.08586, 2023
342023
Unipad: A universal pre-training paradigm for autonomous driving
H Yang, S Zhang, D Huang, X Wu, H Zhu, T He, S Tang, H Zhao, Q Qiu, ...
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern …, 2024
332024
Fit: Flexible vision transformer for diffusion model
Z Lu, Z Wang, D Huang, C Wu, X Liu, W Ouyang, L Bai
arXiv preprint arXiv:2402.12376, 2024
322024
Reconstructing hand-held objects from monocular video
D Huang, X Ji, X He, J Sun, T He, Q Shuai, W Ouyang, X Zhou
SIGGRAPH Asia 2022 Conference Papers, 1-9, 2022
242022
Gvgen: Text-to-3d generation with volumetric representation
X He, J Chen, S Peng, D Huang, Y Li, X Huang, C Yuan, W Ouyang, T He
European Conference on Computer Vision, 463-479, 2024
142024
MeshAnything: Artist-Created Mesh Generation with Autoregressive Transformers
Y Chen, T He, D Huang, W Ye, S Chen, J Tang, X Chen, Z Cai, L Yang, ...
arXiv preprint arXiv:2406.10163, 2024
132024
Agent3d-zero: An agent for zero-shot 3d understanding
S Zhang, D Huang, J Deng, S Tang, W Ouyang, T He, Y Zhang
European Conference on Computer Vision, 186-202, 2024
92024
Nerf-det++: Incorporating semantic cues and perspective-aware depth supervision for indoor multi-view 3d detection
C Huang, Y Hou, W Ye, D Huang, X Huang, B Lin, D Cai, W Ouyang
arXiv preprint arXiv:2402.14464, 2024
82024
Point Cloud Matters: Rethinking the Impact of Different Observation Spaces on Robot Learning
H Zhu, Y Wang, D Huang, W Ye, W Ouyang, T He
arXiv preprint arXiv:2402.02500, 2024
82024
Fitv2: Scalable and improved flexible vision transformer for diffusion model
ZD Wang, Z Lu, D Huang, C Zhou, W Ouyang
arXiv preprint arXiv:2410.13925, 2024
52024
MotionGPT-2: A General-Purpose Motion-Language Model for Motion Generation and Understanding
Y Wang, D Huang, Y Zhang, W Ouyang, J Jiao, X Feng, Y Zhou, P Wan, ...
arXiv preprint arXiv:2410.21747, 2024
42024
Gigags: Scaling up planar-based 3d gaussians for large scene surface reconstruction
J Chen, W Ye, Y Wang, D Chen, D Huang, W Ouyang, G Zhang, Y Qiao, ...
arXiv preprint arXiv:2409.06685, 2024
42024
GenAgent: Build Collaborative AI Systems with Automated Workflow Generation--Case Studies on ComfyUI
X Xue, Z Lu, D Huang, W Ouyang, L Bai
arXiv preprint arXiv:2409.01392, 2024
42024
Nd-sdf: Learning normal deflection fields for high-fidelity indoor reconstruction
Z Tang, W Ye, Y Wang, D Huang, H Bao, T He, G Zhang
arXiv preprint arXiv:2408.12598, 2024
42024
Neurodin: A two-stage framework for high-fidelity neural surface reconstruction
Y Wang, D Huang, W Ye, G Zhang, W Ouyang, T He
arXiv preprint arXiv:2408.10178, 2024
42024
Depth any video with scalable synthetic data
H Yang, D Huang, W Yin, C Shen, H Liu, X He, B Lin, W Ouyang, T He
arXiv preprint arXiv:2410.10815, 2024
32024
Nie można teraz wykonać tej operacji. Spróbuj ponownie później.
Prace 1–20