Withdraw
Loading…
Dynamic 3D Gaussian tracking for graph-based neural dynamics modeling
Zhang, Mingtong
Loading…
Permalink
https://hdl.handle.net/2142/125606
Description
- Title
- Dynamic 3D Gaussian tracking for graph-based neural dynamics modeling
- Author(s)
- Zhang, Mingtong
- Issue Date
- 2024-07-15
- Director of Research (if dissertation) or Advisor (if thesis)
- Li, Yunzhu
- Department of Study
- Siebel Computing &DataScience
- Discipline
- Computer Science
- Degree Granting Institution
- University of Illinois at Urbana-Champaign
- Degree Name
- M.S.
- Degree Level
- Thesis
- Keyword(s)
- Dynamics Model
- 3D Gaussian Splatting
- Action-Conditioned Video Prediction
- Model-Based Planning
- Abstract
- Videos of robots interacting with objects encode rich information about the objects' dynamics. However, existing video prediction approaches typically do not explicitly account for the 3D information from videos, such as robot actions and objects' 3D states, limiting their use in real-world robotic applications. In this work, we introduce a comprehensive framework to learn object dynamics directly from multi-view RGB videos by explicitly considering the robot's action trajectories and their effects on scene dynamics. Our approach utilizes the 3D Gaussian representation of 3D Gaussian Splatting (3DGS) to train a particle-based dynamics model using Graph Neural Networks (GNNs). This model operates on sparse control particles downsampled from the densely tracked 3D Gaussian reconstructions, ensuring that the critical dynamics of the scene are captured efficiently and accurately. By learning the neural dynamics model on offline robot interaction data, our method can predict object motions under varying initial configurations and unseen robot actions, providing a robust tool for real-world applications. The 3D transformations of Gaussians can be interpolated from the motions of control particles, enabling the rendering of predicted future object states and achieving action-conditioned video prediction. This capability allows for the generation of realistic and accurate future scenarios based on the robot's actions, facilitating advanced planning and decision-making processes. Furthermore, the dynamics model can be integrated into model-based planning frameworks for object manipulation tasks. By leveraging the learned dynamics, robots can plan and execute complex manipulation tasks with higher precision and reliability. This integration is particularly beneficial for tasks involving deformable materials, where accurate prediction of object behavior is crucial. We conduct extensive experiments on various kinds of deformable materials, including ropes, clothes, and stuffed animals, demonstrating our framework's ability to model complex shapes and dynamics. Our results show significant improvements in prediction accuracy and visual fidelity compared to existing methods, highlighting the effectiveness of incorporating 3D information and explicit action conditioning in dynamics modeling.
- Graduation Semester
- 2024-08
- Type of Resource
- Thesis
- Handle URL
- https://hdl.handle.net/2142/125606
- Copyright and License Information
- Copyright 2024 Mingtong Zhang
Owning Collections
Graduate Dissertations and Theses at Illinois PRIMARY
Graduate Theses and Dissertations at IllinoisManage Files
Loading…
Edit Collection Membership
Loading…
Edit Metadata
Loading…
Edit Properties
Loading…
Embargoes
Loading…