Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
ai4artists
A list of AI Art courses, tools, libraries, people, and places.
https://github.com/jonathandinu/ai4artists
- Practical Deep Learning for Coders (fast.ai)
- Deep Learning (NYU)
- Introduction to Deep Learning (CMU)
- Deep Learning for Computer Vision (UMich)
- Deep Learning for Computer Vision (Stanford CS231n)
- Natural Language Processing with Deep Learning (Stanford CS224n)
- Deep Generative Models (Stanford)
- Deep Unsupervised Learning (UC Berkeley)
- Differentiable Inference and Generative Models (Toronto)
- Learning-Based Image Synthesis (CMU)
- Learning Discrete Latent Structure (Toronto)
- From Deep Learning Foundations to Stable Diffusion (fast.ai)
- Deep Learning for Art, Aesthetics, and Creativity (MIT)
- Machine Learning for the Web (ITP/NYU)
- Art and Machine Learning (CMU)
- New Media Installation: Art that Learns (CMU)
- Media course
- Code course
- The AI that creates any picture you want, explained (Vox)
- I Created a Neural Network and Tried Teaching it to Recognize Doodles (Sebastian Lague)
- Neural Network Series (3Blue1Brown)
- Beginner's Guide to Machine Learning in JavaScript (Coding Train)
- Two Minute Papers
- Dive into Deep Learning (Zhang, Lipton, Li, and Smola)
- Deep Learning (Goodfellow, Bengio, and Courville)
- Computer Vision: Algorithms and Applications (Szeliski)
- Procedural Content Generation in Games (Shaker, Togelius, and Nelson)
- Generative Design (Benedikt GroΓ)
- VQGAN-CLIP: Open Domain Image Generation and Editing with Natural Language Guidance (Crowson and Biderman)
- Tutorial on Deep Generative Models (IJCAI-ECAI 2018)
- Tutorial on GANs (CVPR 2018)
- Lil'Log (Lilian Weng)
- Distill [on hiatus
- Making Generative Art with Simple Mathematics
- Book of Shaders: Generative Designs
- Mike Bostock: Visualizing Algorithms
- Generative Examples in Processing
- Generative Music
- SDEdit: Guided Image Synthesis and Editing with Stochastic Differential Equations
- GLIDE: Towards Photorealistic Image Generation and Editing with Text-Guided Diffusion Models
- High-Resolution Image Synthesis with Latent Diffusion Models
- Prompt-to-Prompt Image Editing with Cross-Attention Control
- An Image is Worth One Word: Personalizing Text-to-Image Generation using Textual Inversion - to-prompt but instead takes an input image and a text description. Kinda like Style Transfer... but with Stable diffusion.
- DreamBooth: Fine Tuning Text-to-Image Diffusion Models for Subject-Driven Generation
- Novel View Synthesis with Diffusion Models
- AudioGen: Textually Guided Audio Generation
- Make-A-Video: Text-to-Video Generation without Text-Video Data
- Imagic: Text-Based Real Image Editing with Diffusion Models
- MDM: Human Motion Diffusion Model
- Soft Diffusion: Score Matching for General Corruptions
- Multi-Concept Customization of Text-to-Image Diffusion
- eDiff-I: Text-to-Image Diffusion Models with an Ensemble of Expert Denoisers
- Elucidating the Design Space of Diffusion-Based Generative Models (EDM)
- Tackling the Generative Learning Trilemma with Denoising Diffusion GANs
- Imagen Video: High Definition Video Generation with Diffusion Models
- Structure-from-Motion Revisited
- DeepSDF: Learning Continuous Signed Distance Functions for Shape Representation
- Deferred Neural Rendering: Image Synthesis using Neural Textures
- Neural Volumes: Learning Dynamic Renderable Volumes from Images
- NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis
- Neural Radiance Fields for Unconstrained Photo Collections
- Nerfies: Deformable Neural Radiance Fields - the-wild photos and videos (like from a cellphone)
- Mip-NeRF: A Multiscale Representation for Anti-Aliasing Neural Radiance Fields
- Depth-supervised NeRF: Fewer Views and Faster Training for Free
- Instant Neural Graphics Primitives with a Multiresolution Hash Encoding
- Understanding Pure CLIP Guidance for Voxel Grid NeRF Models - to-3D using CLIP
- NeRF-SLAM: Real-Time Dense Monocular SLAM with Neural Radiance Fields
- nerf2nerf: Pairwise Registration of Neural Radiance Fields
- The One Where They Reconstructed 3D Humans and Environments in TV Shows
- ClimateNeRF: Physically-based Neural Rendering for Extreme Climate Synthesis
- Realistic one-shot mesh-based head avatars
- Neural Point Catacaustics for Novel-View Synthesis of Reflections
- 3D Moments from Near-Duplicate Photos
- NeRDi: Single-View NeRF Synthesis with Language-Guided Diffusion as General Image Priors
- DreamFusion: Text-to-3D using 2D Diffusion (Google)
- ULIP: Learning Unified Representation of Language, Image and Point Cloud for 3D Understanding (Salesforce)
- Extracting Triangular 3D Models, Materials, and Lighting From Images (NVIDIA)
- GET3D: A Generative Model of High Quality 3D Textured Shapes Learned from Images (NVIDIA)
- 3D Neural Field Generation using Triplane Diffusion
- π MagicPony: Learning Articulated 3D Animals in the Wild
- ObjectStitch: Generative Object Compositing (Adobe)
- LADIS: Language Disentanglement for 3D Shape Editing (Snap)
- Rodin: A Generative Model for Sculpting 3D Digital Avatars Using Diffusion (Microsoft)
- SDFusion: Multimodal 3D Shape Completion, Reconstruction, and Generation (Snap)
- DiffRF: Rendering-guided 3D Radiance Field Diffusion (Meta)
- Novel View Synthesis with Diffusion Models (Google)
- Magic3D: High-Resolution Text-to-3D Content Creation (NVIDIA)
- Sampling Generative Networks
- Neural Discrete Representation Learning (VQVAE)
- Progressive Growing of GANs for Improved Quality, Stability, and Variation
- A Style-Based Generator Architecture for Generative Adversarial Networks (StyleGAN)
- Analyzing and Improving the Image Quality of StyleGAN (StyleGAN2)
- Training Generative Adversarial Networks with Limited Data (StyleGAN2-ADA)
- Alias-Free Generative Adversarial Networks (StyleGAN3)
- Generating Diverse High-Fidelity Images with VQ-VAE-2
- Taming Transformers for High-Resolution Image Synthesis (VQGAN)
- Diffusion Models Beat GANs on Image Synthesis
- StyleNAT: Giving Each Head a New Perspective
- StyleGAN-XL: Scaling StyleGAN to Large Diverse Datasets
- Image-to-Image Translation with Conditional Adversarial Nets (pix2pix)
- Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks (CycleGAN)
- High-Resolution Image Synthesis and Semantic Manipulation with Conditional GANs (pix2pixHD)
- Semantic Editing of Scenes by Adding, Manipulating or Erasing Objects (SESAME)
- Semantic Image Synthesis with Spatially-Adaptive Normalization (SPADE)
- You Only Need Adversarial Supervision for Semantic Image Synthesis (OASIS)
- Encoding in Style: a StyleGAN Encoder for Image-to-Image Translation
- Multimodal Conditional Image Synthesis with Product-of-Experts GANs
- Palette: Image-to-Image Diffusion Models
- Sketch-Guided Text-to-Image Diffusion Models
- HRDA: Context-Aware High-Resolution Domain-Adaptive Semantic Segmentation
- PiPa: Pixel- and Patch-wise Self-supervised Learning for Domain Adaptative Semantic Segmentation
- MIC: Masked Image Consistency for Context-Enhanced Domain Adaptation
- Pretraining is All You Need for Image-to-Image Translation (PITI)
- Generative Visual Manipulation on the Natural Image Manifold (iGAN)
- In-Domain GAN Inversion for Real Image Editing
- Image2StyleGAN: How to Embed Images Into the StyleGAN Latent Space?
- Designing an Encoder for StyleGAN Image Manipulation
- Pivotal Tuning for Latent-based Editing of Real Images
- HyperStyle: StyleGAN Inversion with HyperNetworks for Real Image Editing
- StyleCLIP: Text-Driven Manipulation of StyleGAN Imagery
- High-Fidelity GAN Inversion for Image Attribute Editing
- Swapping Autoencoder for Deep Image Manipulation
- Sketch Your Own GAN
- Rewriting Geometric Rules of a GAN
- Anycost GANs for Interactive Image Synthesis and Editing
- Third Timeβs the Charm? Image and Video Editing with StyleGAN3
- Discovering Interpretable GAN Controls (GANspace)
- Interpreting the Latent Space of GANs for Semantic Face Editing
- GAN Dissection: Visualizing and Understanding Generative Adversarial Networks
- Unsupervised Extraction of StyleGAN Edit Directions (CLIP2StyleGAN)
- Seeing What a GAN Cannot Generate
- Deep Image Matting
- Background Matting: The World is Your Green Screen
- Robust Video Matting
- Semantic Image Matting
- Privacy-Preserving Portrait Matting
- Deep Automatic Natural Image Matting
- MatteFormer
- MODNet: Real-Time Trimap-Free Portrait Matting via Objective Decomposition
- Robust Human Matting via Semantic Guidance
- NVIDIA Imaginaire
- NVIDIA Omniverse
- mmgeneration
- Modelverse - Based Search for Deep Generative Models
- PaddleGAN
- Tensorflow.js
- ml5.js
- MediaPipe
- Magenta
- Wekinator
- ofxAddons
- PyTorch
- Keras
- Tensorflow
- π€ Transformers
- π€ Diffusers
- JAX
- dlib
- Darknet
- FFCV: an Optimized Data Pipeline for Accelerating ML Training
- ONNX Runtime
- DeepSpeed (training, inference, compression)
- TensorRT
- Tensorflow Lite
- TorchScript
- TorchServe
- AITemplate
- Stable Diffusion
- Imagen
- DALLE 2
- VQGAN+CLIP
- Parti
- Muse: Text-To-Image Generation via Masked Generative Transformers - to-image models used masked image modeling w/ transformers
- Dream Studio
- Stable Diffusion Web UI - diffusion-webui/wiki/Features) to make common workflows easy.
- AI render (Blender)
- Dream Textures (Blender)
- lexica.art - SD Prompt Search.
- koi (Krita)
- Alpaca (Photoshop)
- Christian Cantrell's Plugin (Photoshop)
- Stable Diffusion Studio
- DeepSpeed-MII - latency and high-throughput inference for a variety (20,000+) models/tasks, including SD.
- COLMAP
- nerfstudio
- NVlabs/instant-ngp
- NerfAcc
- Processing (Java)
- openFrameworks (C++)
- Cinder (C++)
- nannou (Rust)
- vvvv
- TouchDesigner
- Max/MSP/Jitter
- Pure Data
- Sonic Pi
- SuperCollider
- Overtone
- Tone.js
- ChucK
- sjfricke/awesome-webgl
- three.js
- regl
- stack.gl
- spacy
- natural
- Tracery
- Learning Deep Generative Models (Salakhutdinov 2015)
- LAION Datasets - text pairs datasets (notably used to train the open source [Stable Diffusion](https://stability.ai) models)
- Unsplash Images
- Open Images - level labels, object bounding boxes, object segmentation masks, visual relationships, and localized narratives:
- Mozilla Common Voice
- Labeled Faces in the Wild (LFW)
- CelebA
- LFWA+
- CelebAMask-HQ
- CelebA-Spoof
- UTKFace
- SSHQ
- Brutus Light Field
- Artbreeder
- Midjourney
- DALLE 2 (OpenAI)
- Runway - AI powered video editor.
- Facet AI - AI powered image editor.
- Adobe Sensei - AI powered features for the Creative Cloud suite.
- NVIDIA AI Demos
- ClipDrop
- Memo Akten
- Neural Bricolage (helena sarin)
- Sofia Crespo
- Lauren McCarthy
- Philipp Schmitt
- Anna Ridler
- Tom White
- Ivona Tau
- Trevor Paglen
- Sasha Stiles
- Mario Klingemann
- Tega Brain
- Mimi Onuoha
- Allison Parrish
- Caroline Sinders
- Robbie Barrat
- Kyle McDonald
- Golan Levin
- STUDIO for Creative Inquiry
- ITP @ NYU
- Gray Area Foundation for the Arts
- Stability AI (Eleuther, LAION, et al.)
- Goldsmiths @ University of London
- UCLA Design Media Arts
- Berkeley Center for New Media
- Google Artists and Machine Intelligence
- Google Creative Lab
- The Lab at the Google Cultural Institute
- Tokyo
- Machine Learning for Art
- Tools and Resources for AI Art (pharmapsychotic) - Big list of Google Colab notebooks for generative text-to-image techniques as well as general tools and resources.
- Awesome Generative Deep Art - A curated list of Generative Deep Art / Generative AI projects, tools, artworks, and models
Programming Languages
Keywords
deep-learning
16
generative-adversarial-network
9
pytorch
9
computer-vision
9
gan
7
image-generation
7
machine-learning
6
stylegan
5
computer-graphics
5
stylegan-encoder
4
pix2pix
4
image-editing
4
neural-network
3
inference
3
stable-diffusion
3
ai
3
image-manipulation
3
text-to-image
2
artificial-intelligence
2
jax
2
signed-distance-functions
2
3d-reconstruction
2
cuda
2
function-approximation
2
nerf
2
real-time
2
real-time-rendering
2
realtime
2
cyclegan
2
gans
2
torch
2
hci
2
gan-inversion
2
text2image
2
diffusion
2
image2image
2
semantic-segmentation
1
multi-resolution
1
high-resolution
1
attention
1
psp-model
1
psp-framework
1
avatar
1
pixel2style2pixel
1
image-translation
1
cvpr2021
1
neural-rendering
1
dcgan
1
image-to-image-translation
1
diffusion-models
1