Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
Awesome-Text-to-3D
A growing curation of Text-to-3D, Diffusion-to-3D works.
https://github.com/yyeboah/Awesome-Text-to-3D
- Zero-Shot Text-Guided Object Generation with Dream Fields - L6) | [site](https://ajayj.com/dreamfields) | [code](https://github.com/google-research/google-research/tree/master/dreamfields)
- CLIP-Forge: Towards Zero-Shot Text-to-Shape Generation - L13) | [site]() | [code](https://github.com/AutodeskAILab/Clip-Forge)
- PureCLIPNERF: Understanding Pure CLIP Guidance for Voxel Grid NeRF Models - Hung Lee et al., Arxiv 2022 | [citation](./references/citations.bib#L29-L34) | [site](https://hanhung.github.io/PureCLIPNeRF/) | [code](https://github.com/hanhung/PureCLIPNeRF)
- SDFusion: Multimodal 3D Shape Completion, Reconstruction, and Generation - Chi Cheng et al., CVPR 2023 | [citation](./references/citations.bib#L43-L48) | [site](https://yccyenchicheng.github.io/SDFusion/) | [code](https://github.com/yccyenchicheng/SDFusion)
- DreamFusion: Text-to-3D using 2D Diffusion - L62) | [site](https://dreamfusion3d.github.io/) | [code]()
- Dream3D: Zero-Shot Text-to-3D Synthesis Using 3D Shape Prior and Text-to-Image Diffusion Models - L69) | [site](https://bluestyle97.github.io/dream3d/) | [code]()
- Novel View Synthesis with Diffusion Models - L83) | [site](https://3d-diffusion.github.io/) | [code]()
- NeuralLift-360: Lifting An In-the-wild 2D Photo to A 3D Object with 360° Views - L90) | [site](https://vita-group.github.io/NeuralLift-360/) | [code](https://github.com/VITA-Group/NeuralLift-360)
- Point-E: A System for Generating 3D Point Clouds from Complex Prompts - L97) | [site]() | [code](https://github.com/openai/point-e)
- Latent-NeRF for Shape-Guided Generation of 3D Shapes and Textures - L104) | [site]() | [code](https://github.com/eladrich/latent-nerf)
- Magic3D: High-Resolution Text-to-3D Content Creation - Hsuan Linet et al., CVPR 2023 | [citation](./references/citations.bib#L106-L111) | [site](https://research.nvidia.com/labs/dir/magic3d/) | [code]()
- RealFusion: 360° Reconstruction of Any Object from a Single Image - Kyriazi et al., CVPR 2023 | [citation](./references/citations.bib#L113-L118) | [site](https://lukemelas.github.io/realfusion/) | [code](https://github.com/lukemelas/realfusion)
- Monocular Depth Estimation using Diffusion Models - L125) | [site](https://depth-gen.github.io/) | [code]()
- SparseFusion: Distilling View-conditioned Diffusion for 3D Reconstruction - L132) | [site](https://sparsefusion.github.io/) | [code](https://github.com/zhizdev/sparsefusion)
- NerfDiff: Single-image View Synthesis with NeRF-guided Distillation from 3D-aware Diffusion - L139) | [site](https://jiataogu.me/nerfdiff/) | [code]()
- Score Jacobian Chaining: Lifting Pretrained 2D Diffusion Models for 3D Generation - L146) | [site](https://pals.ttic.edu/p/score-jacobian-chaining) | [code](https://github.com/pals-ttic/sjc/)
- High-fidelity 3D Face Generation from Natural Language Descriptions - L153) | [site](https://mhwu2017.github.io/) | [code](https://github.com/zhuhao-nju/describe3d)
- TEXTure: Text-Guided Texturing of 3D Shapes - L160) | [site](https://texturepaper.github.io/TEXTurePaper/) | [code](https://github.com/TEXTurePaper/TEXTurePaper)
- NeRDi: Single-View NeRF Synthesis with Language-Guided Diffusion as General Image Priors - L167) | [site]() | [code]()
- DiffusioNeRF: Regularizing Neural Radiance Fields with Denoising Diffusion Models - L174) | [site]() | [code](https://github.com/nianticlabs/diffusionerf)
- 3DQD: Generalized Deep 3D Shape Prior via Part-Discretized Diffusion Process - L545) | [site]() | [code](https://github.com/colorful-liyu/3DQD)
- DATID-3D: Diversity-Preserved Domain Adaptation Using Text-to-Image Diffusion for 3D Generative Model - L181) | [site](https://gwang-kim.github.io/datid_3d/) | [code](https://github.com/gwang-kim/DATID-3D)
- Novel View Synthesis with Diffusion Models - L188) | [site]() | [code]()
- ProlificDreamer: High-Fidelity and Diverse Text-to-3D Generation with Variational Score Distillation - L195) | [site]() | [code]()
- 3D-aware Image Generation using 2D Diffusion Models - L209) | [site]() | [code]()
- Make-It-3D: High-Fidelity 3D Creation from A Single Image with Diffusion Prior - L216) | [site]() | [code]()
- GECCO: Geometrically-Conditioned Point Diffusion Models - L699) | [site]() | [code]()
- Re-imagine the Negative Prompt Algorithm: Transform 2D Diffusion into 3D, alleviate Janus problem and Beyond - L223) | [site]() | [code]()
- Generative Novel View Synthesis with 3D-Aware Diffusion Models - L237) | [site]() | [code]()
- Text2NeRF: Text-Driven 3D Scene Generation with Neural Radiance Fields - L244) | [site]() | [code]()
- Magic123: One Image to High-Quality 3D Object Generation Using Both 2D and 3D Diffusion Priors - L251) | [site]() | [code]()
- DreamBooth3D: Subject-Driven Text-to-3D Generation - L258) | [site]() | [code]()
- Zero-1-to-3: Zero-shot One Image to 3D Object - L265) | [site]() | [code]()
- ATT3D: Amortized Text-to-3D Object Synthesis - L293) | [site]() | [code]()
- Conditional 3D Shape Generation based on Shape-Image-Text Aligned Latent Representation - L300) | [site]() | [code]()
- Diffusion-SDF: Conditional Generative Modeling of Signed Distance Functions - L307) | [site]() | [code]()
- HiFA: High-fidelity Text-to-3D with Advanced Diffusion Guidance - L314) | [site]() | [code]()
- LERF: Language Embedded Radiance Fields - L321) | [site]() | [code]()
- 3DFuse: Let 2D Diffusion Model Know 3D-Consistency for Robust Text-to-3D Generation - L335) | [site]() | [code]()
- MVDiffusion: Enabling Holistic Multi-view Image Generation with Correspondence-Aware Diffusion - L342) | [site]() | [code]()
- One-2-3-45: Any Single Image to 3D Mesh in 45 Seconds without Per-Shape Optimization - L349) | [site]() | [code]()
- TextMesh: Generation of Realistic 3D Meshes From Text Prompts - L356) | [site]() | [code]()
- Prompt-Free Diffusion: Taking "Text" out of Text-to-Image Diffusion Models - L363) | [site]() | [code]()
- SceneScape: Text-Driven Consistent Scene Generation - L370) | [site]() | [code]()
- CLIP-Mesh: Generating textured meshes from text using pretrained image-text models - L384) | [site]() | [code]()
- Text2Room: Extracting Textured 3D Meshes from 2D Text-to-Image Models - L391) | [site]() | [code]()
- Single-Stage Diffusion NeRF: A Unified Approach to 3D Generation and Reconstruction - L398) | [site]() | [code]()
- PODIA-3D: Domain Adaptation of 3D Generative Model Across Large Domain Gap Using Pose-Preserved Text-to-Image Diffusion - L566) | [site]() | [code]()
- Shap-E: Generating Conditional 3D Implicit Functions - L405) | [site]() | [code]()
- Sketch-A-Shape: Zero-Shot Sketch-to-3D Shape Generation - L412) | [site]() | [code]()
- 3D VADER - AutoDecoding Latent 3D Diffusion Models - L433) | [site]() | [code]()
- DreamSparse: Escaping from Plato's Cave with 2D Frozen Diffusion Model Given Sparse Views - L447) | [site]() | [code]()
- Cap3D: Scalable 3D Captioning with Pretrained Models - L482) | [site]() | [code]()
- InstructP2P: Learning to Edit 3D Point Clouds with Text Instructions - L489) | [site]() | [code]()
- 3D-LLM: Injecting the 3D World into Large Language Models - L503) | [site]() | [code]()
- Points-to-3D: Bridging the Gap between Sparse Points and Shape-Controllable Text-to-3D Generation - L510) | [site]() | [code]()
- RGB-D-Fusion: Image Conditioned Depth Diffusion of Humanoid Subjects - L517) | [site]() | [code]()
- IT3D: Improved Text-to-3D Generation with Explicit View Synthesis - L608) | [site]() | [code]()
- MVDream: Multi-view Diffusion for 3D Generation - L629) | [site]() | [code]()
- PointLLM: Empowering Large Language Models to Understand Point Clouds - L636) | [site]() | [code]()
- SyncDreamer: Generating Multiview-consistent Images from a Single-view Image - L650) | [site]() | [code]()
- Large-Vocabulary 3D Diffusion Model with Transformer - L678) | [site]() | [code]()
- Progressive Text-to-3D Generation for Automatic 3D Prototyping - L685) | [site]() | [code]()
- DreamGaussian: Generative Gaussian Splatting for Efficient 3D Content Creation - L692) | [site]() | [code]()
- SweetDreamer: Aligning Geometric Priors in 2D Diffusion for Consistent Text-to-3D - L706) | [site]() | [code]()
- Consistent123: One Image to Highly Consistent 3D Asset Using Case-Aware Diffusion Priors - L720) | [site]() | [code]()
- GaussianDreamer: Fast Generation from Text to 3D Gaussian Splatting with Point Cloud Priors - L727) | [site]() | [code]()
- Text-to-3D using Gaussian Splatting - L734) | [site]() | [code]()
- Zero123++: a Single Image to Consistent Multi-view Diffusion Base Model - L755) | [site]() | [code]()
- DreamCraft3D: Hierarchical 3D Generation with Bootstrapped Diffusion Prior - L762) | [site]() | [code]()
- HyperFields: Towards Zero-Shot Generation of NeRFs from Text - L769) | [site]() | [code]()
- Enhancing High-Resolution 3D Generation through Pixel-wise Gradient Clipping - L776) | [site]() | [code]()
- Text-to-3D with classifier score distillation - L783) | [site]() | [code]()
- Noise-Free Score Distillation - L790) | [site]() | [code]()
- LRM: Large Reconstruction Model for Single Image to 3D - L811) | [site]() | [code]()
- One-2-3-45++: Fast Single Image to 3D Objects with Consistent Multi-View Generation and 3D Diffusion - L818) | [site]() | [code]()
- LucidDreamer: Towards High-Fidelity Text-to-3D Generation via Interval Score Matching - L825) | [site]() | [code]()
- MetaDreamer: Efficient Text-to-3D Creation With Disentangling Geometry and Texture - L832) | [site]() | [code]()
- Adversarial Diffusion Distillation - L854) | [site]() | [code]()
- MeshGPT: Generating Triangle Meshes with Decoder-Only Transformers - L868) | [site]() | [code]()
- DreamPropeller: Supercharge Text-to-3D Generation with Parallel Sampling - L875) | [site]() | [code]()
- X-Dreamer: Creating High-quality 3D Content by Bridging the Domain Gap Between Text-to-2D and Text-to-3D Generation - L889) | [site]() | [code]()
- StableDreamer: Taming Noisy Score Distillation Sampling for Text-to-3D - L903) | [site]() | [code]()
- CAD: Photorealistic 3D Generation via Adversarial Distillation - L910) | [site]() | [code]()
- RichDreamer: A Generalizable Normal-Depth Diffusion Model for Detail Richness in Text-to-3D - L917) | [site]() | [code]()
- Inpaint3D: 3D Scene Content Generation using 2D Inpainting Diffusion - L924) | [site]() | [code]()
- Text-to-3D Generation with Bidirectional Diffusion using both 2D and 3D priors - L931) | [site]() | [code]()
- Text2Immersion: Generative Immersive Scene with 3D Gaussians - L945) | [site]() | [code]()
- Stable Score Distillation for High-Quality 3D Generation - L980) | [site]() | [code]()
- Hyper-VolTran: Fast and Generalizable One-Shot Image to 3D Object Structure via HyperNetworks - L987) | [site]() | [code]()
- HarmonyView: Harmonizing Consistency and Diversity in One-Image-to-3D - L994) | [site]() | [code]()
- SteinDreamer: Variance Reduction for Text-to-3D Score Distillation via Stein Identity - L1022) | [site]() | [code]()
- AGG: Amortized Generative 3D Gaussians for Single Image to 3D - L1036) | [site]() | [code]()
- Topology-Aware Latent Diffusion for 3D Shape Generation - L1050) | [site]() | [code]()
- AToM: Amortized Text-to-Mesh using 2D Diffusion - L1085)
- LGM: Large Multi-View Gaussian Model for High-Resolution 3D Content Creation - L1092) | [site]() | [code]()
- IM-3D: : Iterative Multiview Diffusion and Reconstruction for High-Quality 3D Generation - Kyriazi et al., Arxiv 2024 | [citation](./references/citations.bib#L1108-L1113) | [site]() | [code]()
- L3GO: Language Agents with Chain-of-3D-Thoughts for Generating Unconventional Objects - L1120) | [site]() | [code]()
- MVD2: Efficient Multiview 3D Reconstruction for Multiview Diffusion - Yang Zheng et al., Arxiv 2024 | [citation](./references/citations.bib#L1122-L1127) | [site]() | [code]()
- Pushing Auto-regressive Models for 3D Shape Generation at Capacity and Scalability - L1134) | [site]() | [code]()
- SceneWiz3D: Towards Text-guided 3D Scene Composition - L1141) | [site]() | [code]()
- TripoSR: Fast 3D Object Reconstruction from a Single Image - L1155) | [site]() | [code]()
- V3D: Video Diffusion Models are Effective 3D Generators - L1169) | [site]() | [code]()
- CRM: Single Image to 3D Textured Mesh with Convolutional Reconstruction Model - L1176) | [site]() | [code]()
- Make-Your-3D: Fast and Consistent Subject-Driven 3D Content Generation - L1183) | [site]() | [code]()
- Isotropic3D: Image-to-3D Generation Based on a Single CLIP Embedding - L1190) | [site]() | [code]()
- SV3D: Novel Multi-view Synthesis and 3D Generation from a Single Image using Latent Video Diffusion - L1197) | [site]() | [code]()
- Generic 3D Diffusion Adapter Using Controlled Multi-View Editing - L1204) | [site]() | [code]()
- GVGEN: Text-to-3D Generation with Volumetric Representation - L1211) | [site]() | [code]()
- BrightDreamer: Generic 3D Gaussian Generative Framework for Fast Text-to-3D Synthesis - L1218) | [site]() | [code]()
- LATTE3D: Large-scale Amortized Text-To-Enhanced3D Synthesis - L1239) | [site]() | [code]()
- Make-Your-3D: Fast and Consistent Subject-Driven 3D Content Generation - L1246) | [site]() | [code]()
- GRM: Large Gaussian Reconstruction Model for Efficient 3D Reconstruction and Generation - L1253) | [site]() | [code]()
- VP3D: Unleashing 2D Visual Prompt for Text-to-3D Generation - L1260) | [site]() | [code]()
- DreamPolisher: Towards High-Quality Text-to-3D Generation via Geometric Diffusion - L1267) | [site]() | [code]()
- PointInfinity: Resolution-Invariant Point Diffusion Models - L1274) | [site](https://zixuanh.com/projects/pointinfinity.html) | [code]()
- The More You See in 2D, the More You Perceive in 3D - L1295) | [site](https://sap3d.github.io/) | [code]()
- Hash3D: Training-free Acceleration for 3D Generation - L1302) | [site](https://adamdad.github.io/hash3D/) | [code](https://github.com/Adamdad/hash3D)
- RealmDreamer: Text-Driven 3D Scene Generation with Inpainting and Depth Diffusion - L1309) | [site](https://realmdreamer.github.io/) | [code]()
- TC4D: Trajectory-Conditioned Text-to-4D Generation - L1316) | [site](https://sherwinbahmani.github.io/tc4d/) | [code]()
- Zero-shot Point Cloud Completion Via 2D Priors - L1323) | [site]() | [code]()
- InstantMesh: Efficient 3D Mesh Generation from a Single Image with Sparse-view Large Reconstruction Models - L1330) | [site]() | [code](https://github.com/TencentARC/InstantMesh)
- Zero-shot Point Cloud Completion Via 2D Priors - L1337) | [site]() | [code]()
- CLIP-GS: CLIP-Informed Gaussian Splatting for Real-time and View-consistent 3D Semantic Understanding - L1365) | [site]() | [code]()
- CAT3D: Create Anything in 3D with Multi-View Diffusion Models - L1372) | [site](https://cat3d.github.io) | [code]()
- CLIP-NeRF: Text-and-Image Driven Manipulation of Neural Radiance Fields - L20) | [site](https://cassiepython.github.io/clipnerf/) | [code](https://github.com/cassiePython/CLIPNeRF)
- CG-NeRF: Conditional Generative Neural Radiance Fields - L27) | [site]() | [code]()
- TANGO: Text-driven Photorealistic and Robust 3D Stylization via Lighting Decomposition - L41) | [site](https://cyw-3d.github.io/tango/) | [code](https://github.com/Gorilla-Lab-SCUT/tango)
- 3DDesigner: Towards Photorealistic 3D Object Generation and Editing with Text-guided Diffusion Models - L55) | [site](https://3ddesigner-diffusion.github.io/) | [code]()
- NeRF-Art: Text-Driven Neural Radiance Fields Stylization - L76) | [site](https://cassiepython.github.io/nerfart/) | [code](https://github.com/cassiePython/NeRF-Art)
- Instruct-NeRF2NeRF: Editing 3D Scenes with Instructions - L328) | [site](https://instruct-nerf2nerf.github.io/) | [code](https://github.com/ayaanzhaque/instruct-nerf2nerf)
- Local 3D Editing via 3D Distillation of CLIP Knowledge - L377) | [site]() | [code]()
- RePaint-NeRF: NeRF Editing via Semantic Masks and Diffusion Models - L419) | [site](https://starstesla.github.io/repaintnerf/) | [code](https://github.com/StarsTesla/RePaint-NeRF)
- Text2Tex: Text-driven Texture Synthesis via Diffusion Models - L426) | [site](https://daveredrum.github.io/Text2Tex/) | [code](https://github.com/daveredrum/Text2Tex)
- Control4D: Dynamic Portrait Editing by Learning 4D GAN from 2D Diffusion-based Editor - L440) | [site](https://fantasia3d.github.io/) | [code](https://github.com/Gorilla-Lab-SCUT/Fantasia3D)
- Fantasia3D: Disentangling Geometry and Appearance for High-quality Text-to-3D Content Creation - L454) | [site]() | [code]()
- Set-the-Scene: Global-Local Training for Generating Controllable NeRF Scenes - Bar et al., Arxiv 2023 | [citation](./references/citations.bib#L463-L468) | [site](https://control4darxiv.github.io/) | [code]()
- MATLABER: Material-Aware Text-to-3D via LAtent BRDF auto-EncodeR - L601) | [site](https://sheldontsui.github.io/projects/Matlaber) | [code](https://github.com/SheldonTsui/Matlaber)
- SATR: Zero-Shot Semantic Segmentation of 3D Shapes - L615) | [site](https://samir55.github.io/SATR/) | [code](https://github.com/Samir55/SATR)
- Texture Generation on 3D Meshes with Point-UV Diffusion - L643) | [site](https://cvmi-lab.github.io/Point-UV-Diffusion/) | [code](https://github.com/CVMI-Lab/Point-UV-Diffusion)
- Progressive3D: Progressively Local Editing for Text-to-3D Content Creation with Complex Semantic Prompts - L741) | [site](https://cxh0519.github.io/projects/Progressive3D/) | [code](https://github.com/cxh0519/Progressive3D)
- 3D-GPT: Procedural 3D Modeling with Large Language Models - L748) | [site](https://chuny1.github.io/3DGPT/3dgpt.html) | [code]()
- CustomNet: Zero-shot Object Customization with Variable-Viewpoints in Text-to-Image Diffusion Models - L797) | [site]() | [code]()
- Decorate3D: Text-Driven High-Quality Texture Generation for Mesh Decoration in the Wild - L840) | [site](https://decorate3d.github.io/Decorate3D/) | [code](https://github.com/Decorate3D/Decorate3D)
- HyperDreamer: Hyper-Realistic 3D Content Generation and Editing from a Single Image - L882) | [site]() | [code]()
- InseRF: Text-Driven Generative Object Insertion in Neural 3D Scenes - L1029) | [site](https://mohamad-shahbazi.github.io/inserf/) | [code]()
- ReplaceAnything3D:Text-Guided 3D Scene Editing with Compositional Neural Radiance Fields - L1057) | [site](https://replaceanything3d.github.io/) | [code]()
- Sketch2NeRF: Multi-view Sketch-guided Text-to-3D Generation - L1078) | [site]() | [code]()
- BoostDream: Efficient Refining for High-Quality Text-to-3D Generation from Multi-View Diffusion - L1064) | [site]() | [code]()
- 2L3: Lifting Imperfect Generated 2D Images into Accurate 3D - L1071) | [site]() | [code]()
- GALA3D: Towards Text-to-3D Complex Scene Generation via Layout-guided Generative Gaussian Splatting - L1106) | [site](https://gala3d.github.io/) | [code](https://github.com/VDIGPKU/GALA3D)
- Disentangled 3D Scene Generation with Layout Learning - L1148) | [site](https://dave.ml/layoutlearning/) | [code]()
- MagicClay: Sculpting Meshes With Generative Neural Fields - L1162) | [site]() | [code]()
- TexDreamer: Towards Zero-Shot High-Fidelity 3D Human Texture Generation - L1225) | [site](https://ggxxii.github.io/texdreamer/) | [code]()
- InTeX: Interactive Text-to-texture Synthesis via Unified Depth-aware Inpainting - L1232) | [site](https://me.kiui.moe/intex/) | [code](https://github.com/ashawkey/InTeX)
- SC4D: Sparse-Controlled Video-to-4D Generation and Motion Transfer - L1288) | [site](https://sc4d.github.io/) | [code]()
- TELA: Text to Layer-wise 3D Clothed Human Generation - L1344) | [site](http://jtdong.com/tela_layer/) | [code]()
- Interactive3D: Create What You Want by Interactive 3D Generation - L1351) | [site](https://interactive-3d.github.io) | [code](https://github.com/interactive-3d/interactive3d)
- TIP-Editor: An Accurate 3D Editor Following Both Text-Prompts And Image-Prompts - L1358) | [site](https://zjy526223908.github.io/TIP-Editor) | [code](https://github.com/zjy526223908/TIP-Editor)
- Coin3D: Controllable and Interactive 3D Assets Generation with Proxy-Guided Conditioning - L1379) | [site]() | [code]()
- Rodin: A Generative Model for Sculpting 3D Digital Avatars Using Diffusion - L202) | [site](https://3d-avatar-diffusion.microsoft.com/) | [code]()
- DINAR: Diffusion Inpainting of Neural Textures for One-Shot Human Avatars - L671) | [site]() | [code]()
- ZeroAvatar: Zero-shot 3D Avatar Generation from a Single Image - L272)
- AvatarCraft: Transforming Text into Neural Human Avatars with Parameterized Shape and Pose Control - L279) | [site](https://avatar-craft.github.io/) | [code](https://github.com/songrise/avatarcraft)
- Chupa: Carving 3D Clothed Humans from Skinned Shape Priors using 2D Diffusion Probabilistic Models - L573) | [site](https://snuvclab.github.io/chupa/) | [code](https://github.com/snuvclab/chupa)
- DreamFace: Progressive Generation of Animatable 3D Faces under Text Guidance - L461) | [site](https://sites.google.com/view/dreamface) | [code](https://huggingface.co/spaces/DEEMOSTECH/3D-Avatar-Generator)
- HeadSculpt: Crafting 3D Head Avatars with Text - L475) | [site](https://brandonhan.uk/HeadSculpt/) | [code]()
- DreamHuman: Animatable 3D Avatars from Text - L559) | [site](https://dream-human.github.io/) | [code]()
- FaceCLIPNeRF: Text-driven 3D Face Manipulation using Deformable Neural Radiance Fields - L496) | [site]() | [code]()
- AvatarVerse: High-quality & Stable 3D Avatar Creation from Text and Pose - L552) | [site](https://avatarverse3d.github.io/) | [code](https://github.com/bytedance/AvatarVerse)
- TeCH: Text-guided Reconstruction of Lifelike Clothed Humans - L580) | [site](https://huangyangyi.github.io/TeCH/) | [code](https://github.com/huangyangyi/TeCH)
- HumanLiff: Layer-wise 3D Human Generation with Diffusion Model - L587) | [site](https://skhu101.github.io/HumanLiff/) | [code](https://github.com/skhu101/HumanLiff)
- TADA! Text to Animatable Digital Avatars - L594) | [site](https://tada.is.tue.mpg.de/) | [code](https://github.com/TingtingLiao/TADA)
- One-shot Implicit Animatable Avatars with Model-based Priors - L622) | [site](https://huangyangyi.github.io/ELICIT/) | [code](https://github.com/huangyangyi/ELICIT)
- Text2Control3D: Controllable 3D Avatar Generation in Neural Radiance Fields using Geometry-Guided Text-to-Image Diffusion Model - L657) | [site]() | [code]()
- Text-Guided Generation and Editing of Compositional 3D Avatars - L664) | [site](https://yfeng95.github.io/teca/) | [code]()
- HumanNorm: Learning Normal Diffusion Model for High-quality and Realistic 3D Human Generation - L713) | [site](https://humannorm.github.io/) | [code](https://github.com/xhuangcv/humannorm)
- HumanGaussian: Text-Driven 3D Human Generation with Gaussian Splatting - L861) | [site](https://alvinliu0.github.io/projects/HumanGaussian) | [code](https://github.com/alvinliu0/HumanGaussian)
- Text-Guided 3D Face Synthesis: From Generation to Editing - L896) | [site](https://faceg2e.github.io/) | [code]()
- SEEAvatar: Photorealistic Text-to-3D Avatar Generation with Constrained Geometry and Appearance - L938) | [site](https://yoxu515.github.io/SEEAvatar/) | [code](https://github.com/yoxu515/SEEAvatar)
- GAvatar: Animatable 3D Gaussian Avatars with Implicit Mesh Learning - L973) | [site](https://nvlabs.github.io/GAvatar/) | [code]()
- Make-A-Character: High Quality Text-to-3D Character Generation within Minutes - L1008) | [site](https://human3daigc.github.io/MACH/) | [code](https://github.com/Human3DAIGC/Make-A-Character)
- En3D: An Enhanced Generative Model for Sculpting 3D Humans from 2D Synthetic Data - L1015) | [site](https://menyifang.github.io/projects/En3D/index.html) | [code](https://github.com/menyifang/En3D)
- HeadStudio: Text to Animatable Head Avatars with 3D Gaussian Splatting - L1099) | [site](https://zhenglinzhou.github.io/HeadStudio-ProjectPage/) | [code](https://github.com/ZhenglinZhou/HeadStudio/)
- InstructHumans: Editing Animatable 3D Human Textures with Instructions - L1281) | [site](https://jyzhu.top/instruct-humans/) | [code](https://github.com/viridityzhu/InstructHumans)
- X-Oscar: A Progressive Framework for High-quality Text-guided 3D Animatable Avatar Generation - L1386) | [site](https://xmu-xiaoma666.github.io/Projects/X-Oscar/) | [code](https://github.com/LinZhekai/X-Oscar)
- Text-To-4D Dynamic Scene Generation - L230) | [site](https://make-a-video3d.github.io/) | [code]()
- TextDeformer: Geometry Manipulation using Text Guidance - L286) | [site]() | [code]()
- Consistent4D: Consistent 360 Degree Dynamic Object Generation from Monocular Video - L809) | [site](https://consistent4d.github.io/) | [code](https://github.com/yanqinJiang/Consistent4D)
- 4D-fy:Text-to-4D Generation Using Hybrid Score Distillation Sampling - L847) | [site](https://sherwinbahmani.github.io/4dfy/) | [code](https://github.com/sherwinbahmani/4dfy)
- Objaverse: A Universe of Annotated 3D Objects - L524)
- Objaverse-XL: A Universe of 10M+ 3D Objects - L531)
- Describe3D: High-Fidelity 3D Face Generation from Natural Language Descriptions - L538)
- Repaint123: Fast and High-quality One Image to 3D Generation with Progressive Controllable 2D Repainting - L952)
- Customize-It-3D: High-Quality 3D Creation from A Single Image Using Subject-Specific Knowledge Prior - L959)
- Paint-it: Text-to-Texture Synthesis via Deep Convolutional Texture Map Optimization and Physically-Based Rendering - L966)
- SceneVerse: Scaling 3D Vision-Language Learning for Grounded Scene Understanding - L1043)
- Scalable 3D Captioning with Pretrained Models - L1393)
- threestudio: A unified framework for 3D content generation - Chen Guo et al., Github 2023
- Nerfstudio: A Modular Framework for Neural Radiance Field Development
- Mirage3D: Open-Source Implementations of 3D Diffusion Models Optimized for GLB Output
- AI 3D Generation, explained