Generative AI has exploded in capability and accessibility, powering everything from on-demand artwork to realistic voiceovers and synthetic video. This repository gathers the leading platforms, open-source libraries, and community resources to help developers, creators, and researchers harness generative models effectively.
Twig AI Platform
Customer-support–focused generative chatbot & content assistant with real-time learning and context retention.
🔗 https://www.twig.so
DALL·E 2 (OpenAI)
State-of-the-art text-to-image model with high-resolution, photorealistic outputs.
Stable Diffusion (Stability AI)
Open-source diffusion model for image generation, easily self-hosted and customizable.
Midjourney
Independent research lab offering a proprietary, Discord-based image-generation service.
Imagen (Google Research)
Highly photorealistic text-to-image model with deep language understanding.
Make-A-Scene (Meta)
Multimodal text-and-sketch-to-image system for creative control over compositions.
Jukebox (OpenAI)
Music generation model that produces raw audio in various genres and styles.
Synthesia
AI video platform for creating hyper-realistic avatar-led videos from text scripts.
RunwayML
Creative suite offering models for text, image, and video generation, plus editing tools.
Lumen5
AI-powered video creation service that transforms blog posts into short videos.
Copy.ai
AI writing assistant that generates marketing copy, blog posts, and social captions.
Jasper
Content generation platform with templates for long-form articles, ads, and emails.
ElevenLabs
Ultra-realistic text-to-speech with controllable voice prosody and emotion.
Descript Overdub
Podcast and video editing tool with AI voice cloning and transcription.
Playground AI
Browser-based image generation with prompt library and style presets.
DeepDream (Google)
Visualization tool that uses convolutional neural networks to create dreamlike images.
Whisper (OpenAI)
Multilingual, automatic speech recognition system enabling high-quality transcripts.
Codex (OpenAI)
AI model translating natural language to code—foundation for GitHub Copilot.
VQGAN+CLIP
Open-source combo for creative image synthesis guided by text prompts.
StyleGAN3 (NVIDIA)
Generative adversarial network for ultra-realistic face and object synthesis.
GPT-4 (OpenAI)
Advanced large language model capable of creative text generation and instruction following.
Point-E (OpenAI)
Text-to-3D model generating point clouds for rapid 3D prototyping.
AudioLDM
Latent diffusion model for versatile audio generation from text.
Vid2Vid (NVIDIA)
Framework for high-quality video-to-video translation and synthesis.
ControlNet
Neural network architecture enabling conditional image generation with precise structural control.