Generative AI Models – What Are They?

Introduction

The conversations around artificial intelligence (AI) in media and entertainment can be conflicting because there are varying perspectives on how it fits into the narrative of creating and consuming content. Depending on who you talk to, their experiences or personal philosophies will shape how they interpret AI’s role in content creation pipelines.

Generative AI Models Article Featured Image with an Artificial Brain in the Background

At the heart of these discussions are generative AI models, trained on large datasets to generate specific types of content, such as images, videos, music, voice, and 3D assets. These models are improving rapidly, with new versions and updates released frequently, which can make it difficult to keep pace with the latest advancements.

However, no single model truly ‘works’ for all aspects of the content pipeline. Different applications, hardware, interfaces, and workflows can affect inference, which is the operation of a trained model using inputs such as prompts, images, or other control variables to predict and generate content as an output. In turn, inference impacts processing time, output quality, and computational requirements, meaning the most suitable model and platform may vary depending on your needs.

While there are many avenues for discussion around generative AI, it’s worth mentioning that this post is not an attempt to identify the ‘best’ model or tool for content creation. That question is inherently subjective, and I believe your own perspectives will help you determine whether a given model or tool’s output meets your standards. What this post will cover, at a high level, is the different types of generative AI models and where you can access them.

Generative AI Explained

Artificial intelligence is a broad term that refers to many different technologies. In content creation, it can describe various AI models, software, tools, and features – making it difficult to distinguish what each actually does or how they differ from one another.

To make sense of these differences, it can be helpful to categorize AI based on how they are offered and used:

  • AI as Products – Artificial intelligence solutions offered commercially for access or deployment.
  • AI as Tools – Products functioning within a creative pipeline or workflow to generate content.
  • AI as Features – Capabilities embedded in software that support or enhance parts of the creative process.

As a product, generative AI can be commercially distributed through web-based platforms, APIs, downloadable models, or managed hosting solutions. For organizations deploying generative AI solutions, your infrastructure influences how effectively systems scale to meet demand, maintain responsiveness under variable workloads, and integrate with existing applications and workflows.

Within the content creation pipeline, generative AI tools enable creators to produce various types of content. These tools are powered by AI models, each trained to generate specific types of outputs, such as images, video, music, sound effects, or 3D assets. As a result, content creators have the opportunity to experiment, iterate, and explore alternative workflows that might have been difficult, time-consuming, or costly using traditional production. However, not all models within a given category produce the same type of content. Developers train models on datasets that shape the content they generate, and these datasets can influence stylistic outcomes, consistency, quality, and performance. There is no single model that works ‘best’ for every task, so it’s important to research and test models to find those that most align with your needs.

Aside from generative AI tools, some mainstream software applications include features that leverage other types of AI or machine learning (ML) to automate tedious or repetitive tasks that support the creative process. Classification AI labels or categorizes content and encompasses activities like analyzing and tagging media files, object detection, and masking. Other ML-assisted functions include enhancing audio, speech-to-text transcriptions, stabilizing footage, and motion tracking.

Different Types of Generative AI Models

As mentioned in the intro, the generative AI landscape is evolving rapidly, with new models and updates constantly emerging. To provide a clearer view of the current tools shaping the field, I have included four tables below that break down many of the popular generative AI base models available for different types of content.

It’s worth noting that some models may be missing, as my research was limited to what could be reasonably reviewed at the time of writing, and the total number of models available is too large to comprehensively cover for this type of post. The tables include only base models and exclude LoRAs or fine-tuned variants. Covering topics such as data security, copyright, and licensing for commercial or personal use is also outside the scope of this overview. Please note that these tables have multiple pages of entries and can be sorted or filtered if you want to narrow the list down.

Image Models

The table below highlights generative AI models for creating 2D images.

wdt_ID wdt_created_by wdt_created_at wdt_last_edited_by wdt_last_edited_at Model Family / Base Model Developer / Organization Developer Country Access Model Variants
1 Peter Feb 2026 03:54 PM Peter Feb 2026 03:54 PM Adobe Firefly Adobe United States Cloud Firefly Model 3, Firefly Model 4, Firefly Model 5
2 Peter Feb 2026 03:54 PM Peter Feb 2026 03:54 PM Ovis Alibaba China Local, Cloud
3 Peter Feb 2026 03:54 PM Peter Feb 2026 03:54 PM Qwen Image Alibaba China Local, Cloud Qwen Image, Qwen Image 2512, Qwen Image Edit 2511
4 Peter Feb 2026 03:54 PM Peter Feb 2026 03:54 PM Z-Image Alibaba China Local, Cloud Z-Image Standard, Z-Image Turbo
5 Peter Feb 2026 03:54 PM Peter Feb 2026 03:54 PM Flux 1 Series Black Forest Labs Germany Local, Cloud Flux 1.0, Flux 1.1, Flux 1.1 Pro, Flux 1.1 Ultra, Flux 1.1 Kontext, Flux 1 Schnell
6 Peter Feb 2026 03:54 PM Peter Feb 2026 03:54 PM Flux 2 Series Black Forest Labs Germany Local, Cloud Klein 4B, Klein 4B Fast, Klein 9B, Kontext Max, Kontext Pro, Flex, Max, Pro
7 Peter Feb 2026 03:54 PM Peter Feb 2026 03:54 PM Bria Bria AI Israel / United States Cloud Bria 4.0
8 Peter Feb 2026 03:54 PM Peter Feb 2026 03:54 PM Cosmos‑Predict2 NVIDIA United States Local Cosmos-Predict2-2B-Text2Image, Cosmos-Predict2-14B-Text2Image
9 Peter Feb 2026 03:54 PM Peter Feb 2026 03:54 PM DALL·E OpenAI United States Cloud DALL·E 2, DALL·E 3
10 Peter Feb 2026 03:54 PM Peter Feb 2026 03:54 PM GPT Image OpenAI United States Cloud GPT Image 1, GPT Image 1.5
Model Family / Base Model Developer / Organization Developer Country Access Model Variants

Video Models

The table below highlights generative AI models designed for video generation.

wdt_ID wdt_created_by wdt_created_at wdt_last_edited_by wdt_last_edited_at Model Family / Base Model Developer / Organization Developer Country Access Model Variants
1 Peter Feb 2026 03:58 PM Peter Feb 2026 03:58 PM Adobe Firefly Video Adobe United States Cloud Firefly Video Model 1, Firefly Video Model 2
2 Peter Feb 2026 03:58 PM Peter Feb 2026 03:58 PM Wan Video Alibaba China Cloud Wan 2.1 Standard, Wan 2.2 Standard, Wan 2.5 Standard, Wan 2.6 Standard, Wan 2.2 Fast, Wan 2.5 Fast, Wan 2.6 Fast
3 Peter Feb 2026 03:58 PM Peter Feb 2026 03:58 PM SwitchLight 3.0 Beeble Labs United States Local & Cloud
4 Peter Feb 2026 03:58 PM Peter Feb 2026 03:58 PM SHARP Apple United States Local --
5 Peter Feb 2026 03:58 PM Peter Feb 2026 03:58 PM Omni Human Beijing Academy of Artificial Intelligence China Cloud OmniHuman 1.5
6 Peter Feb 2026 03:58 PM Peter Feb 2026 03:58 PM Seedance Video ByteDance China Cloud Seedance 1.5 Pro, Seedance Pro, Seedance Pro Fast, Seedance Lite, Seedance 2.0
7 Peter Feb 2026 03:58 PM Peter Feb 2026 03:58 PM Veo Video Google (DeepMind) United States Cloud Veo 2, Veo 3.0, Veo 3.0 Fast, Veo 3.1, Veo 3.1 Fast
8 Peter Feb 2026 03:58 PM Peter Feb 2026 03:58 PM Higgsfield Video Higgsfield AI United States Cloud Higgsfield Lite, Higgsfield Standard, Higgsfield Turbo
9 Peter Feb 2026 03:58 PM Peter Feb 2026 03:58 PM Kandinsky 5.0 Video KandinskyLab Russia Local, Cloud Kandinsky 5.0 Lite, Kandinsky 5.0 Pro
10 Peter Feb 2026 03:58 PM Peter Feb 2026 03:58 PM Kling Video Kuaishou Technology China Cloud Kling 2.1, Kling 2.1 Master, Kling 2.5 Turbo, Kling 2.6, Kling 2.6 Motion Control, Kling 3.0, Kling Avatars 2.0, Kling O1 Video, Kling O1 Video Edit
Model Family / Base Model Developer / Organization Developer Country Access Model Variants

Audio Models

The table below highlights generative AI models pertaining to music, voice, lip-sync, and sound effects.

wdt_ID wdt_created_by wdt_created_at wdt_last_edited_by wdt_last_edited_at Model Family / Base Model Developer / Organization Developer Country Access Model Variants Content Type
1 Peter Feb 2026 04:02 PM Peter Feb 2026 04:02 PM Ace Step ACE Studio China Local, Cloud Ace Step 1.5 Music
2 Peter Feb 2026 04:02 PM Peter Feb 2026 04:02 PM Adobe Podcast / Firefly Sound Adobe United States Cloud Enhance Speech, Firefly Sound Effects Voice, SFX
3 Peter Feb 2026 04:02 PM Peter Feb 2026 04:02 PM Bark Suno Labs United States Local, Cloud Bark Voice
4 Peter Feb 2026 04:02 PM Peter Feb 2026 04:02 PM Cartesia Voice Cartesia United States Cloud Cartesia Voice Voice
5 Peter Feb 2026 04:02 PM Peter Feb 2026 04:02 PM Eleven Music ElevenLabs United States Cloud Eleven Music Music
6 Peter Feb 2026 04:02 PM Peter Feb 2026 04:02 PM Eleven TTS (Text-to-Speech) ElevenLabs United States Cloud Eleven v3, Eleven Multilingual v2, Eleven Flash v2.5, Eleven Turbo v2.5 Voice
7 Peter Feb 2026 04:02 PM Peter Feb 2026 04:02 PM Eleven TTV (Text-to-Voice) ElevenLabs United States Cloud Eleven TTV v3 Voice
8 Peter Feb 2026 04:02 PM Peter Feb 2026 04:02 PM Scribe ElevenLabs United States Cloud Scribe v2, Scribe v2 Realtime Voice
9 Peter Feb 2026 04:02 PM Peter Feb 2026 04:02 PM Higgsfield Voice Higgsfield AI United States Cloud Higgsfield Voice Standard, Higgsfield Voice Pro Voice
10 Peter Feb 2026 04:02 PM Peter Feb 2026 04:02 PM AudioGen Google Research United States Cloud AudioGen 1, AudioGen 2 SFX
Model Family / Base Model Developer / Organization Developer Country Access Model Variants Content Type

3D Models

The table below highlights generative AI models for various types of 3D content, including 2D→3D outputs, environments, meshes, textures, and point cloud generation.

wdt_ID wdt_created_by wdt_created_at wdt_last_edited_by wdt_last_edited_at Model Family / Base Model Developer / Organization Developer Country Access Model Variants Content Type
1 Peter Feb 2026 04:02 PM Peter Feb 2026 04:02 PM Adobe Firefly (Substance 3D Integration) Adobe United States Local & Cloud -- Materials, Textures
2 Peter Feb 2026 04:02 PM Peter Feb 2026 04:02 PM SHARP Apple United States Local -- 2D → 3D Point Cloud (Gaussian splat)
3 Peter Feb 2026 04:02 PM Peter Feb 2026 04:02 PM Rodin 3D AI3D Labs United States Cloud Rodin Gen-1 2D→3D Mesh, Animation
4 Peter Feb 2026 04:02 PM Peter Feb 2026 04:02 PM Tripo AI Alibaba China Cloud Tripo 2.0 2D→3D Mesh, Textured Mesh
5 Peter Feb 2026 04:02 PM Peter Feb 2026 04:02 PM Hunyuan 3D Tencent China Cloud Hunyuan 3D 2.5, Hunyuan 3D 3.0 2D→3D Mesh, Textures
6 Peter Feb 2026 04:02 PM Peter Feb 2026 04:02 PM Meshy AI Meshy LLC United States Cloud Meshy 4 2D→3D Mesh, Textures
7 Peter Feb 2026 04:02 PM Peter Feb 2026 04:02 PM GET3D NVIDIA United States Local GET3D 2D→3D Mesh, Textured Mesh
8 Peter Feb 2026 04:02 PM Peter Feb 2026 04:02 PM Point-E OpenAI United States Local Point-E Point Cloud → Mesh
9 Peter Feb 2026 04:02 PM Peter Feb 2026 04:02 PM Sloyd Sloyd Norway Cloud Sloyd 2.0 Text‑to‑3D, Image‑to‑3D, Parametric Templates, Export (STL/OBJ/GLB)
10 Peter Feb 2026 04:02 PM Peter Feb 2026 04:02 PM Spline AI 3D Generation Spline United States Cloud Spline AI 3D 2D→3D Mesh, Scene Layout
Model Family / Base Model Developer / Organization Developer Country Access Model Variants Content Type

How to Access and Use Generative AI Models

Generative AI models can be accessed through software applications hosted locally on your system or through cloud and web-based platforms. This includes new tools explicitly built for generating content, as well as traditional applications integrating AI features into their software. That said, not all platforms offer the same models or features. For example, Adobe Photoshop runs inferences in their cloud servers from prompts and tools such as generative fill – but it only offers a limited set of models, whereas other applications or platforms may provide similar features with access to a larger number of models.

Adobe Photoshop UI showing the Generative Fill tool and available Generative AI models that can be used for editing.

Generative Fill tool in Adobe Photoshop with generative AI model options

To perform inference on your own hardware, you can download AI models from repositories such as Huggingface or GitHub and use tools like Comfy UI or TouchDesigner for generating content locally. Running models on your own system allows you to work offline and maintain full control over your data, creating a self-contained workflow. However, not all models can run on local hardware, and larger or more complex models may only be accessible through cloud-based platforms or other hosted services.

Cloud- and web-based platforms such as Adobe Firefly, Artlist, Freepik, Higgsfield, Krea, and OpenArt provide access to multiple generative AI models. All processing is done on the platform’s servers, so content creators don’t need to invest in high-performance systems; they only need a web browser and an internet connection to run inferences. These platforms also differentiate themselves by offering tools and features that are exclusive to their subscribers. For example, MidJourney makes its models available only through its website and requires a subscription. Other web-based platforms like ComfyUI Cloud, Flora, SOTA, and Weavy offer a different type of interface, such as node-based workflows that allow users to combine models, adjust parameters, and manage outputs in a structured way, which can lead to more predictable and consistent results when generating content.

ComfyUI interface with a node-based workflow utilizing the Flux.2 Klein 4B model for generating images.

Node-based workflow in ComfyUI using Flux.2 Klein 4B model for image generation

Where Can Generative AI Models Be Useful in Content Creation?

There isn’t a single “right” way to use generative AI in content creation. Some models run locally on consumer or workstation-class hardware, giving more control over your workflow. Others are accessed through cloud- or web-based platforms, where computations are handled remotely, so creators don’t need a high-end workstation to generate content.

From my perspective, regardless of the models or applications used, generative AI should be considered as an alternative method or supplementary tool that bridges gaps across different stages of the content creation pipeline. Rather than replacing people, I hope that AI models can let artists and creators do more and avoid tedious, repetitive tasks. For example, within pre-production, they can help you generate mood boards, refine storyboards, or mock up placeholders for conceptual ideas. In production, they can be used as an alternative method to traditional practices for music, graphics, photography, or video productions. In post-production, generative AI tools can assist with tasks such as refining visual effects, completing or reconstructing missing frames, enhancing audio clarity, adjusting colors or lighting, removing unwanted elements, and preparing assets for final assembly.

For content creators, it’s important to understand both the capabilities of generative AI models and the resources required to select the right hardware and platform for your needs. It is equally important to understand how models are trained and the terms under which outputs can be used. Not all models produce content that is safe for commercial use, so reviewing licensing and copyright terms before generating or distributing outputs, and understanding which models are trained on “clean” data, will help protect your projects from legal risks and support ethical and responsible use of generative AI.

Nonetheless, whether you are looking to deploy, host, or generate content using AI models, hardware will be the core component that enables you to offer or produce great content. If you are a creator looking to experiment with AI or incorporate it into your workflow, we also have recommended systems for Generative AI to help you get started. For those looking to develop or deploy AI for a small team, you can check out this list of recommended systems as a starting point. Organizations and teams looking to host, train, and scale the development of their models will eventually need server-class solutions.


Whatever your workflow, if you are looking for a new computer, the Puget Systems workstations on our solutions page are tailored to excel in various software packages from content creation to engineering and scientific computing. If you prefer to take a more hands-on approach, our custom configuration page can help you to configure a system that matches your exact needs. Or, if you would like more guidance in configuring a workstation that aligns with your unique workflow, our knowledgeable technology consultants are here to lend their expertise.


Tower Computer Icon in Puget Systems Colors

Looking for a Content Creation workstation?

We build computers tailor-made for your workflow. 

Talking Head Icon in Puget Systems Colors

Don’t know where to start?
We can help!

Get in touch with one of our technical consultants today.