Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
awesome-3d-diffusion
A collection of papers on diffusion models for 3D generation.
https://github.com/cwchenwang/awesome-3d-diffusion
Last synced: about 24 hours ago
JSON representation
-
2D Diffusion with Pretraining
-
Text-to-3D Object Generation
- ScaleDreamer: Scalable Text-to-3D Synthesis with Asynchronous Score Distillation
- DreamFusion: Text-to-3D using 2D Diffusion
- Magic3D: High-Resolution Text-to-3D Content Creation
- Score jacobian chaining: Lifting pretrained 2d diffusion models for 3d generation
- Fantasia3D: Disentangling Geometry and Appearance for High-quality Text-to-3D Content Creation
- Let 2D Diffusion Model Know 3D-Consistency for Robust Text-to-3D Generation
- DITTO-NeRF: Diffusion-based Iterative Text To Omni-directional 3D Model
- TextMesh: Generation of Realistic 3D Meshes From Text Prompts
- Text-driven Visual Synthesis with Latent Diffusion Prior
- Re-imagine the Negative Prompt Algorithm: Transform 2D Diffusion into 3D, alleviate Janus problem and Beyond
- HiFA: High-fidelity Text-to-3D with Advanced Diffusion Guidance
- ATT3D: Amortized Text-to-3D Object Synthesis
- PanoGen: Text-Conditioned Panoramic Environment Generation for Vision-and-Language Navigation
- ProlificDreamer: High-Fidelity and Diverse Text-to-3D Generation with Variational Score Distillation
- DreamTime: An Improved Optimization Strategy for Text-to-3D Content Creation
- EfficientDreamer: High-Fidelity and Robust 3D Creation via Orthogonal-view Diffusion Prior
- DreamCraft3D: Hierarchical 3D Generation with Bootstrapped Diffusion Prior
- Text-to-3D with Classifier Score Distillation
- Instant3D: Instant Text-to-3D Generation
- HyperFields: Towards Zero-Shot Generation of NeRFs from Text
- DreamSpace: Dreaming Your Room Space with Text-Driven Panoramic Texture Propagation
- Enhancing High-Resolution 3D Generation through Pixel-wise Gradient Clipping
- GaussianDreamer: Fast Generation from Text to 3D Gaussian Splatting with Point Cloud Priors
- LucidDreamer: Domain-free Generation of 3D Gaussian Splatting Scenes
- CG3D: Compositional Generation for Text-to-3D via Gaussian Splatting
- LucidDreamer: Towards High-Fidelity Text-to-3D Generation via Interval Score Matching
- StableDreamer: Taming Noisy Score Distillation Sampling for Text-to-3D
- DreamComposer: Controllable 3D Object Generation via Multi-View Conditions
- GraphDreamer: Compositional 3D Scene Synthesis from Scene Graphs
- X-Dreamer: Creating High-quality 3D Content by Bridging the Domain Gap Between Text-to-2D and Text-to-3D Generation
- HyperDreamer: Hyper-Realistic 3D Content Generation and Editing from a Single Image
- RichDreamer: A Generalizable Normal-Depth Diffusion Model for Detail Richness in Text-to-3D
- Text-to-3D Generation with Bidirectional Diffusion using both 2D and 3D priors
- Sherpa3D: Boosting High-Fidelity Text-to-3D Generation via Coarse 3D Prior
- DreamPropeller: Supercharge Text-to-3D Generation with Parallel Sampling
- UniDream: Unifying Diffusion Priors for Relightable Text-to-3D Generation
- Stable Score Distillation for High-Quality 3D Generation
- Text-Image Conditioned Diffusion for Consistent Text-to-3D Generation
- Flow Score Distillation for Diverse Text-to-3D
- Dreamer XL: Towards High-Resolution Text-to-3D Generation via Trajectory Score Matching
- DreamControl: Control-Based Text-to-3D Generation with 3D Self-Prior
- PlacidDreamer: Advancing Harmony in Text-to-3D Generation
- Vista3D: Unravel the 3D Darkside of a Single Image
- SceneDreamer360: Text-Driven 3D-Consistent Scene Generation with Panoramic Gaussian Splatting
- Atlas3D: Physically Constrained Self-Supporting Text-to-3D for Simulation and Fabrication
- SweetDreamer: Aligning Geometric Priors in 2D Diffusion for Consistent Text-to-3D
- Connecting Consistency Distillation to Score Distillation for Text-to-3D Generation
- DreamCouple: Exploring High Quality Text-to-3D Generation Via Rectified Flow
- PlacidDreamer: Advancing Harmony in Text-to-3D Generation
- JointDreamer: Ensuring Geometry Consistency and Text Congruence in Text-to-3D Generation via Joint Score Distillation
- DreamDissector: Learning Disentangled Text-to-3D Generation from 2D Diffusion Priors
- DreamReward: Text-to-3D Generation with Human Preference
- DreamFlow: High-Quality Text-to-3D Generation by Approximating Probability Flow
- LATTE3D: Large-scale Amortized Text-To-Enhanced3D Synthesis
- DreamPolisher: Towards High-Quality Text-to-3D Generation via Geometric Diffusion
- VP3D: Unleashing 2D Visual Prompt for Text-to-3D Generation
- Retrieval-Augmented Score Distillation for Text-to-3D Generation
- DIRECT-3D: Learning Direct Text-to-3D Generation on Massive Noisy 3D Data
- Text-guided Controllable Mesh Refinement for Interactive 3D Modeling
- HexaGen3D: StableDiffusion is just one step away from Fast and Diverse Text-to-3D Generation
- BrightDreamer: Generic 3D Gaussian Generative Framework for Fast Text-to-3D Synthesis
- MicroDreamer: Zero-shot 3D Generation in ∼20 Seconds by Score-based Iterative Reconstruction
- SketchDream: Sketch-based Text-to-3D Generation and Editing
- Hash3D: Training-free Acceleration for 3D Generation
-
Human and Animal
- ScoreHMR: Score-Guided Diffusion for 3D Human Recovery
- Generative Proxemics: A Prior for 3D Social Interaction from Images
- DreamFace: Progressive Generation of Animatable 3D Faces under Text Guidance
- AvatarCraft: Transforming Text into Neural Human Avatars with Parameterized Shape and Pose Control
- DreamAvatar: Text-and-Shape Guided 3D Human Avatar Generation via Diffusion Models
- DreamWaltz: Make a Scene with Complex 3D Animatable Avatars
- ZeroAvatar: Zero-shot 3D Avatar Generation from a Single Image
- AvatarBooth: High-Quality and Customizable 3D Human Avatar Generation
- Farm3D: Learning Articulated 3D Animals by Distilling 2D Diffusion
- Anything 3D: Towards Single-view Anything Reconstruction in the Wild
- ARTIC3D: Learning Robust Articulated 3D Shapes from Noisy Web Image Collections
- TADA! Text to Animatable Digital Avatars
- Diffusion-Guided Reconstruction of Everyday Hand-Object Interaction Clips
- Text-Guided Generation and Editing of Compositional 3D Avatars
- HumanGaussian: Text-Driven 3D Human Generation with Gaussian Splatting
- AvatarStudio: High-fidelity and Animatable 3D Avatar Creation from Text
- Disentangled Clothed Avatar Generation from Text Descriptions
- SEEAvatar: Photorealistic Text-to-3D Avatar Generation with Constrained Geometry and Appearance
- GAvatar: Animatable 3D Gaussian Avatars with Implicit Mesh Learning
- Make-A-Character: High Quality Text-to-3D Character Generation within Minutes
- Morphable Diffusion: 3D-Consistent Diffusion for Single-image Avatar Creation
- Farm3D: Learning Articulated 3D Animals by Distilling 2D Diffusion
-
Multi-view Diffusion
- MVDiffusion: Enabling Holistic Multi-view Image Generation with Correspondence-Aware Diffusion
- Consistent123: Improve Consistency for One Image to 3D Object Synthesis
- Wonder3D: Single Image to 3D using Cross-Domain Diffusion
- Zero123++: a Single Image to Consistent Multi-view Diffusion Base Model
- TOSS:High-quality Text-guided Novel View Synthesis from a Single Image
- Text-Guided Texturing by Synchronized Multi-View Diffusion
- Direct2.5: Diverse Text-to-3D Generation via Multi-view 2.5D Diffusion
- ViVid-1-to-3: Novel View Synthesis with Video Diffusion Models
- MVDiffusion++: A Dense High-resolution Multi-view Diffusion Model for Single or Sparse-view 3D Object Reconstruction
- CRM: Single Image to 3D Textured Mesh with Convolutional Reconstruction Model
- MVDream: Multi-view Diffusion for 3D Generation
- IM-3D: Iterative Multiview Diffusion and Reconstruction for High-Quality 3D Generation - Kyriazi et al., Arxiv 2024
- Cycle3D: High-quality and Consistent Image-to-3D Generation via Generation-Reconstruction Cycle
- Envision3D: One Image to 3D with Anchor Views Interpolation
- CAT3D: Create Anything in 3D with Multi-View Diffusion Models
- CraftsMan: High-fidelity Mesh Generation with 3D Native Generation and Interactive Geometry Refiner
- Ouroboros3D: Image-to-3D Generation via 3D-aware Recursive Diffusion
- VFusion3D: Learning Scalable 3D Generative Models from Video Diffusion Models
- FDGaussian: Fast Gaussian Splatting from Single Image via Geometric-aware Diffusion Model
- Isotropic3D: Image-to-3D Generation Based on a Single CLIP Embedding
- SV3D: Novel Multi-view Synthesis and 3D Generation from a Single Image using Latent Video Diffusion
- Generic 3D Diffusion Adapter Using Controlled Multi-View Editing
- EscherNet: A Generative Model for Scalable View Synthesis
- LGM: Large Multi-View Gaussian Model for High-Resolution 3D Content Creation
- SPAD : Spatially Aware Multiview Diffusers
- V3D: Video Diffusion Models are Effective 3D Generators
- Make-Your-3D: Fast and Consistent Subject-Driven 3D Content Generation
- Controllable Text-to-3D Generation via Surface-Aligned Gaussian Splatting
- EpiDiff: Enhancing Multi-View Synthesis via Localized Epipolar-Constrained Diffusion
- Grounded Compositional and Diverse Text-to-3D with Pretrained Multi-View Diffusion Model
- Multi-view Image Prompted Multi-view Diffusion for Improved 3D Generation
- MVDiff: Scalable and Flexible Multi-View Diffusion for 3D Object Reconstruction from Single-View
- MVD-Fusion: Single-view 3D via Depth-consistent Multi-view Generation
- Magic-Boost: Boost 3D Generation with Mutli-View Conditioned Diffusion
- InstantMesh: Efficient 3D Mesh Generation from a Single Image with Sparse-view Large Reconstruction Models
-
Image-to-3D
- NeuralLift-360: Lifting An In-the-wild 2D Photo to A 3D Object with $360^{\deg}$ Views
- NeRDi: Single-View NeRF Synthesis with Language-Guided Diffusion as General Image Priors
- Latent-NeRF for Shape-Guided Generation of 3D Shapes and Textures
- RealFusion: 360{\deg} Reconstruction of Any Object from a Single Image - Kyriazi et al., Arxiv 2023
- Make-It-3D: High-Fidelity 3D Creation from A Single Image with Diffusion Prior
- Zero-1-to-3: Zero-shot One Image to 3D Object
- DreamBooth3D: Subject-Driven Text-to-3D Generation
- DreamSparse: Escaping from Plato's Cave with 2D Frozen Diffusion Model Given Sparse Views
- One-2-3-45: Any Single Image to 3D Mesh in 45 Seconds without Per-Shape Optimization
- Magic123: One Image to High-Quality 3D Object Generation Using Both 2D and 3D Diffusion Priors
- Viewpoint Textual Inversion: Unleashing Novel View Synthesis with Pretrained 2D Diffusion Models
- Customize-It-3D: High-Quality 3D Creation from A Single Image Using Subject-Specific Knowledge Prior
- IPDreamer: Appearance-Controllable 3D Object Generation with Image Prompts
- HarmonyView: Harmonizing Consistency and Diversity in One-Image-to-3D
- AGG: Amortized Generative 3D Gaussians for Single Image to 3D
- Part123: Part-aware 3D Reconstruction from a Single-view Image
- GECO: Generation Image-to-3D within a Second
- Fourier123: One Image to High-Quality 3D Object Generation with Hybrid Fourier Score Distillation
- Unique3D: High-Quality and Efficient 3D Mesh Generation from a Single Image
- Repaint123: Fast and High-quality One Image to 3D Generation with Progressive Controllable 2D Repainting
- 360◦ Reconstruction From a Single Image Using Space Carved Outpainting
-
3D Editing
- SKED: Sketch-guided Text-based 3D Editing
- Instruct-NeRF2NeRF: Editing 3D Scenes with Instructions
- Instruct 3D-to-3D: Text Instruction Guided 3D-to-3D conversion
- Edit-DiffNeRF: Editing 3D Neural Radiance Fields using 2D Diffusion Model
- Control4D: Dynamic Portrait Editing by Learning 4D GAN from 2D Diffusion-based Editor
- RePaint-NeRF: NeRF Editting via Semantic Masks and Diffusion Models
- DreamEditor: Text-Driven 3D Scene Editing with Neural Fields
- Language-driven Object Fusion into Neural Radiance Fields with Pose-Conditioned Dataset Updates
- ProteusNeRF: Fast Lightweight NeRF Editing using 3D-Aware Image Context
- ED-NeRF: Efficient Text-Guided Editing of 3D Scene using Latent Space NeRF
- 3D Paintbrush: Local Stylization of 3D Shapes with Cascaded Score Distillation
- GaussianEditor: Swift and Controllable 3D Editing with Gaussian Splatting
- Inpaint3D: 3D Scene Content Generation using 2D Inpainting Diffusion
- NeRFiller: Completing Scenes via Generative 3D Inpainting
- SHAP-EDITOR: Instruction-guided Latent 3D Editing in Seconds
- LatentEditor: Text Driven Local Editing of 3D Scenes
- Free-Editor: Zero-shot Text-driven 3D Scene Editing
- SIGNeRF: Scene Integrated Generation for Neural Radiance Fields
- Efficient-NeRF2NeRF: Streamlining Text-Driven 3D Editing with Multiview Correspondence-Enhanced Diffusion Models
- ReplaceAnything3D: Text-Guided 3D Scene Editing with Compositional Neural Radiance Fields
- GaussCtrl: Multi-View Consistent Text-Driven 3D Gaussian Splatting Editing
- View-Consistent 3D Editing with Gaussian Splatting
- Interactive3D: Create What You Want by Interactive 3D Generation
- DGE: Direct Gaussian 3D Editing by Consistent Multi-view Editing
- DATENeRF: Depth-Aware Text-based Editing of NeRFs
-
Texturing
- Text2Tex: Text-driven Texture Synthesis via Diffusion Models
- EucliDreamer: Fast and High-Quality Texturing for 3D Models with Stable Diffusion Depth
- Paint3D: Paint Anything 3D with Lighting-Less Texture Diffusion Models
- TexGen: Text-Guided 3D Texture Generation with Multi-view Sampling and Resampling
- DreamMat: High-quality PBR Material Generation with Geometry- and Light-aware Diffusion Models
- TextureDreamer: Image-guided Texture Synthesis through Geometry-aware Diffusion
- TEXTure: Text-Guided Texturing of 3D Shapes
- MaPa: Text-driven Photorealistic Material Painting for 3D Shapes
-
Compositional or Scene Generation
- A Quantitative Evaluation of Score Distillation Sampling Based Text-to-3D
- VividDreamer: Towards High-Fidelity and Efficient Text-to-3D Generation
- HoloDreamer: Holistic 3D Panoramic World Generation from Text Descriptions
- COMOGen: A Controllable Text-to-3D Multi-object Generation Framework
- Urban Architect: Steerable 3D Urban Scene Generation with Layout Prior
- DreamScape: 3D Scene Creation via Gaussian Splatting joint Correlation Modeling
- Director3D: Real-world Camera Trajectory and 3D Scene Generation from Text
- Sketch2Scene: Automatic Generation of Interactive 3D Game Scenes from User's Casual Sketches
- Scene123: One Prompt to 3D Scene Generation via Video-Assisted and Consistency-Enhanced MAE
- SceneWiz3D: Towards Text-guided 3D Scene Composition
- REPARO: Compositional 3D Assets Generation with Differentiable 3D Layout Alignment
- VividDream: Generating 3D Scene with Ambient Dynamics
- Text2Street: Controllable Text-to-image Generation for Street Views
- GALA3D: Towards Text-to-3D Complex Scene Generation via Layout-guided Generative Gaussian Splatting
- Sketch2NeRF: Multi-view Sketch-guided Text-to-3D Generation
- Invisible Stitch: Generating Smooth 3D Scenes with Depth Inpainting
- Text2Room: Extracting Textured 3D Meshes from 2D Text-to-Image Models
- SceneScape: Text-Driven Consistent Scene Generation
- Compositional 3D Scene Generation using Locally Conditioned Diffusion
- Set-the-Scene: Global-Local Training for Generating Controllable NeRF Scenes - Bar et al., Arxiv 2023
- CompoNeRF: Text-guided Multi-object Compositional NeRF with Editable 3D Scene Layout
- Text2NeRF: Text-Driven 3D Scene Generation with Neural Radiance Fields
- Ctrl-Room: Controllable Text-to-3D Room Meshes Generation with Layout Constraints
- ShowRoom3D: Text to High-Quality 3D Room Generation Using 3D Priors
- DreamScene: 3D Gaussian-based Text-to-3D Scene Generation via Formation Pattern Sampling
- DreamScene360: Unconstrained Text-to-3D Scene Generation with Panoramic Gaussian Splatting
- RealmDreamer: Text-Driven 3D Scene Generation with Inpainting and Depth Diffusion
-
-
Diffusion in 3D Space
-
Point Cloud, Meshs, Volumes
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- DiffFacto: Controllable Part-Based 3D Point Cloud Generation with Cross Diffusion
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- Point-E: A System for Generating 3D Point Clouds from Complex Prompts
- PolyDiff: Generating 3D Polygonal Meshes with Diffusion Models
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- Diffusion Probabilistic Models for 3D Point Cloud Generation
- 3d shape generation and completion through point-voxel diffusion
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- Sketch and Text Guided Diffusion Model for Colored Point Cloud Generation
- DiT-3D: Exploring Plain Diffusion Transformers for 3D Shape Generation
- MeshGPT: Generating Triangle Meshes with Decoder-Only Transformers
- ShapeGPT: 3D Shape Generation with A Unified Multi-modal Language Model
- VolumeDiffusion: Flexible Text-to-3D Generation with Efficient Volumetric Encoder
- SPiC·E: Structural Priors in 3D Diffusion Models using Cross-Entity Attention
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- Controllable Mesh Generation Through Sparse Latent Point Diffusion Models
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- Diffusion Probabilistic Models for 3D Point Cloud Generation
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
-
Triplane
- Rodin: A Generative Model for Sculpting 3D Digital Avatars Using Diffusion
- 3D Neural Field Generation using Triplane Diffusion
- DiffusionSDF: Conditional Generative Modeling of Signed Distance Functions
- Rodin: A Generative Model for Sculpting 3D Digital Avatars Using Diffusion
- 3DGen: Triplane Latent Diffusion for Textured Mesh Generation
- Single-Stage Diffusion NeRF: A Unified Approach to 3D Generation and Reconstruction
- Learning Controllable 3D Diffusion Models from Single-view Images
- Dual3D: Efficient and Consistent Text-to-3D Generation with Dual-mode Multi-view Latent Diffusion
- Single-Stage Diffusion NeRF: A Unified Approach to 3D Generation and Reconstruction
- Direct3D: Scalable Image-to-3D Generation via 3D Latent Diffusion Transformer
- Compress3D: a Compressed Latent Space for 3D Generation from a Single Image
-
Explicit Representation
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- 3d shape generation and completion through point-voxel diffusion
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- Controllable Mesh Generation Through Sparse Latent Point Diffusion Models
- Point-E: A System for Generating 3D Point Clouds from Complex Prompts
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
- A Diffusion-ReFinement Model for Sketch-to-Point Modeling
-
Implicit Representation
- Learning A Diffusion Prior For Nerfs
- Tetrahedral Diffusion Models for 3D Shape Generation
- MeshDiffusion: Score-based Generative 3D Mesh Modeling
- Neural Wavelet-domain Diffusion for 3D Shape Generation
- Neural Wavelet-domain Diffusion for 3D Shape Generation, Inversion, and Manipulation
- DiffRF: Rendering-Guided 3D Radiance Field Diffusion
- Locally Attentional SDF Diffusion for Controllable 3D Shape Generation
- HyperDiffusion: Generating Implicit Neural Fields with Weight-Space Diffusion
- DiffComplete: Diffusion-based Generative 3D Shape Completion
- DiffRoom: Diffusion-based High-Quality 3D Room Reconstruction and Generation
- Surf-D: High-Quality Surface Generation for Arbitrary Topologies using Diffusion Models
- Neural Wavelet-domain Diffusion for 3D Shape Generation
- DiffRoom: Diffusion-based High-Quality 3D Room Reconstruction and Generation
- HyperDiffusion: Generating Implicit Neural Fields with Weight-Space Diffusion
-
Latent Representation
- GAUDI: A Neural Architect for Immersive 3D Scene Generation
- LION: Latent Point Diffusion Models for 3D Shape Generation
- Diffusion-SDF: Text-to-Shape via Voxelized Diffusion
- 3D-LDM: Neural Implicit 3D Shape Generation with Latent Diffusion Models
- 3DShape2VecSet: A 3D Shape Representation for Neural Fields and Generative Diffusion Models
- Shap-E: Generating Conditional 3D Implicit Functions
- StyleAvatar3D: Leveraging Image-Text Diffusion Models for High-Fidelity 3D Avatar Generation
- AutoDecoding Latent 3D Diffusion Models
- LN3Diff: Scalable Latent Neural Fields Diffusion for Speedy 3D Generation
- XCube: Large-Scale 3D Generative Modeling using Sparse Voxel Hierarchies
- GAUDI: A Neural Architect for Immersive 3D Scene Generation
- AutoDecoding Latent 3D Diffusion Models
- Diffusion-SDF: Text-to-Shape via Voxelized Diffusion
- 3D-LDM: Neural Implicit 3D Shape Generation with Latent Diffusion Models
-
3D Gaussians
-
-
Diffusion for Motion
-
Human Motion
- Single Motion Diffusion
- Human Motion Diffusion as a Generative Prior
- DNO: Optimizing Diffusion Noise Can Serve As Universal Motion Priors
- MDM: Human Motion Diffusion Model
- EMDM: Efficient Motion Diffusion Model for Fast, High-Quality Human Motion Generation
- Executing your Commands via Motion Diffusion in Latent Space
- MotionDiffuse: Text-Driven Human Motion Generation with Diffusion Model
- OmniControl: Control Any Joint at Any Time for Human Motion Generation
- MotionMix: Weakly-Supervised Diffusion for Controllable Motion Generation
- RoHM: Robust Human Motion Reconstruction via Diffusion
- TLControl: Trajectory and Language Control for Human Motion Synthesis
- DiffusionPhase: Motion Diffusion in Frequency Domain
- InterControl: Generate Human Motion Interactions by Controlling Every Joint
- AAMDM: Accelerated Auto-regressive Motion Diffusion Model
- ReMoS: Reactive 3D Motion Synthesis for Two-Person Interactions
- HOI-Diff: Text-Driven Synthesis of 3D Human-Object Interactions using Diffusion Models
- Controllable Motion Diffusion Model
- MAS: Multi-view Ancestral Sampling for 3D motion generation using 2D diffusion
- CG-HOI: Contact-Guided 3D Human-Object Interaction Generation
- A Unified Framework for Multimodal, Multi-Part Human Motion Synthesis
- Controllable Motion Synthesis and Reconstruction with Autoregressive Diffusion Models
- Guided Motion Diffusion for Controllable Human Motion Synthesis
- Hierarchical Generation of Human-Object Interactions with Diffusion Probabilistic Models
- PhysDiff: Physics-Guided Human Motion Diffusion Model
- Priority-Centric Human Motion Generation in Discrete Latent Space
- ReMoDiffuse: Retrieval-Augmented Motion Diffusion Model
- Flame: Free-form language-based motion synthesis & editing
- Object Motion Guided Human Motion Synthesis
- Controllable Group Choreography using Contrastive Diffusion
- Listen, denoise, action! Audio-driven motion synthesis with diffusion models
- GestureDiffuCLIP: Gesture Diffusion Model with CLIP Latents
- Taming Diffusion Models for Audio-Driven Co-Speech Gesture Generation
- MoFusion: A Framework for Denoising-Diffusion-based Motion Synthesis
- RoHM: Robust Human Motion Reconstruction via Diffusion
- CG-HOI: Contact-Guided 3D Human-Object Interaction Generation
-
-
2D Diffusion without Pretraining
-
3D Objects
- Novel View Synthesis with Diffusion Models
- Generative Novel View Synthesis with 3D-Aware Diffusion Models
- NerfDiff: Single-image View Synthesis with NeRF-guided Distillation from 3D-aware Diffusion
- 3DDesigner: Towards Photorealistic 3D Object Generation and Editing with Text-guided Diffusion Models
- SparseFusSparseFusion: Distilling View-conditioned Diffusion for 3D Reconstruction
- HoloDiffusion: Training a 3D Diffusion Model using 2D Images
- Renderdiffusion: Image Diffusion for 3D Reconstruction, Inpainting and Generation
- Diffusion with Forward Models: Solving Stochastic Inverse Problems Without Direct Supervision
- 3D-aware Image Generation using 2D Diffusion Models
- Viewset Viewset Diffusion: (0-)Image-Conditioned 3D Generative Models from 2D Data
- HOLOFUSION: Towards Photo-realistic 3D Generative Modeling
- ZeroNVS: Zero-Shot 360-Degree View Synthesis from a Single Real Image
- Instant3D: Fast Text-to-3D with Sparse-view Generation and Large Reconstruction Model
- DMV3D: Denoising Multi-View Diffusion using 3D Large Reconstruction Model
- LRM: Large Reconstruction Model for Single Image to 3D
- WildFusion: Learning 3D-Aware Latent Diffusion Models in View Space
- ViewFusion: Learning Composable Diffusion Models for Novel View Synthesis
-
3D Scenes
- Consistent View Synthesis with Pose-Guided Diffusion Models
- Long-Term Photometric Consistent Novel View Synthesis with Diffusion Models
- DiffDreamer: Towards Consistent Unsupervised Single-view Scene Extrapolation with Conditional Diffusion Models
- SemCity: Semantic Scene Generation with Triplane Diffusion
-
-
Novel Representations
-
Latent Representation
-
Programming Languages
Categories
Sub Categories
Text-to-3D Object Generation
64
Point Cloud, Meshs, Volumes
44
Human Motion
35
Multi-view Diffusion
35
Compositional or Scene Generation
27
3D Editing
25
Explicit Representation
22
Human and Animal
22
Image-to-3D
21
3D Objects
17
Latent Representation
15
Implicit Representation
14
Triplane
11
Texturing
8
3D Scenes
4
3D Gaussians
2