Models

  • Our catalog features the most popular open-source AI models from developers worldwide, including large language models (LLMs), multimodal, and diffusion models. Try any model in one place — we’ve made it easy for you.
  • To explore and test a model, you can query it through our public endpoint. For production use, fine-tuning, or custom weights, we recommend renting a virtual or a dedicated GPU server.

Qwen2.5-32B

Qwen2.5-32B is a 32B parameter model with a 128K context window, offering top-tier performance for complex enterprise and research tasks. It is ideal for legal, scientific, and large-scale content analysis

19.09.2024

FLUX.1-schnell

FLUX.1 [schnell] is a 12 billion parameter rectified flow transformer capable of generating images from text descriptions. 

01.08.2024

FLUX.1-dev

FLUX.1 [dev] is a 12 billion parameter rectified flow transformer capable of generating images from text descriptions. 

01.08.2024

Qwen2-57B-A14B

Qwen2-57B-A14B is a multilingual MoE model with 57 billion parameters, optimized for complex text generation tasks in question-answering systems, analytics, and programming, with high resource efficiency and computational effectiveness.

27.07.2024

Qwen2-72B

Qwen2-72B is the flagship model of the second series, featuring 72 billion parameters and a context window of 128K tokens, delivering performance on par with leading proprietary models. The model is suitable for the most demanding and accuracy-critical applications.

24.07.2024

Qwen2-7B

Qwen2-7B is a 7-billion-parameter model that delivers high performance and accuracy. It efficiently runs on mid-range GPUs and serves as a foundation for developing specialized solutions across various domains.

24.07.2024

Qwen2-1.5B

Qwen2-1.5B is a lightweight, balanced model with 1.5 billion parameters, designed for basic tasks on local machines and small servers. It delivers solid performance in text generation, summarization, and translation while maintaining moderate resource requirements.

24.07.2024

Qwen2-0.5B

Qwen2-0.5B is an ultra-compact model with 0.5 billion parameters and a 32K context window, optimized for deployment on mobile devices and IoT systems. It is suitable for building simple applications and text autocompletion systems.

24.07.2024

Llama-3.1-8B

An incredibly popular multilingual model in the community, trained on 15 trillion tokens, with 8 billion parameters and a context window of 128K. The model is adapted to solve a wide range of tasks, supports function calling, and is ideally suited for building intelligent dialogue systems, software assistants, and agent applications.

23.07.2024

Phi-3.5-mini

Phi-3.5-mini is a compact and highly efficient language model capable of running on mobile and edge devices, delivering generation quality comparable to that of larger models. Thanks to optimized training on high-quality data and multilingual support, it is ideal for chatbots, educational applications, and tasks with limited computational resources.

23.04.2024

Playground v2.5 – 1024px Aesthetic Model

This is diffusion text-to-image model designed to generate high-aesthetic images of 1024x1024 pixels, including portraits and landscapes. It is the successor to Playground v2 and demonstrates superiority over models like SDXL, PixArt-α, DALL-E 3, and Midjourney 5.2 in user studies focused on aesthetic quality.

16.02.2024

Dreamshaper XL v2 Turbo

This is an image generation model based on text prompts (text-to-image), built on the architecture of Stable Diffusion XL (SDXL). It represents a fine-tuned version of the base model stabilityai/stable-diffusion-xl-base-1.0.

07.02.2024

Stable Diffusion XL-Turbo

SDXL-Turbo is a distilled version of SDXL 1.0, which allows sampling large-scale foundational image diffusion models in 1 to 4 steps at high image quality. 

27.11.2023

Kandinsky 3.0

Kandinsky-3 is a diffusion model for text-to-image generation, developed based on previous versions of the Kandinsky2-x family. It has been improved through expanded data volume, including information related to Russian culture, enabling the generation of images reflecting this theme. The model also demonstrates enhanced text understanding and improved visual quality due to larger text encoder and Diffusion U-Net model sizes.

21.11.2023

Stable Video Diffusion Image-to-Video (SVD-Image-to-Video)

This is a diffusion model developed by Stability AI for generating short video clips from a static image (image-to-video). The model creates videos up to 4 seconds long (25 frames at 576×1024 resolution), using the input image as a conditional frame.

20.11.2023

Stable Diffusion XL-refiner-1.0

Refiner Model specializes in the final stages of noise reduction and enhances the visual accuracy of images generated by the base model.

26.07.2023

Stable Diffusion XL-base-1.0

Stable Diffusion XL-base-1.0 is the base* text-to-image generation model, improved upon previous versions of Stable Diffusion models. It is designed to generate images at 1024x1024 pixels. It is also not recommended to choose an image size smaller than 512x512 pixels.

26.07.2023

Kandinsky 2 2 decoder

Kandinsky 2.2 is a free Russian neural network for image generation developed by Sber AI. It operates based on a diffusion model: initially adding noise to images it was trained on, then restoring them through a reverse diffusion process to create new unique images.

12.07.2023

Stable Diffusion 2.1

The Stable Diffusion v2-1 is a diffusion-based text-to-image generation model. It is fine-tuned from the Stable Diffusion v2 checkpoint. Generates high-resolution images (up to 768x768) and supports text-guided generation.

06.12.2022

Stable Diffusion x4 upscaler

Stable Diffusion x4 upscaler - is a diffusion model for 4x image resolution upscaling based on text prompts. It takes as input a low-resolution image and a text prompt, along with the noise_level parameter.

23.11.2022