Video Generation API is now live!

Models

Explore the active model market,from a local OpenRouter snapshot.

This page reads from a local JSON snapshot synced from OpenRouter, so the catalog stays fast, indexable, and stable. Use it to browse current model coverage by provider, modality, reasoning support, context window, and pricing metadata.

Reset

Results

Showing 48 of 683 matching models

Snapshot source: OpenRouter. Synced April 21, 2026 at 8:00 AM. Page 10 of 15.

This route is built from local JSON so the catalog stays stable for browsing and SEO. If you need a specific model on ImaRouter, treat this page as a discovery reference and then contact the team for availability.

Text

Unknown provider

WizardLM-2 7B

WizardLM-2 7B is the smaller variant of Microsoft AI's latest Wizard model. It is the fastest and achieves comparable performance with existing 10x larger opensource leading models It is a finetune of [Mistral 7B Instruct](/models/mistralai/mistral-7b-instruct), using the same technique as [WizardLM-2 8x22B](/models/microsoft/wizardlm-2-8x22b). To read more about the model release, [click here](https://wizardlm.github.io/WizardLM2/). #moe

Text

Context

32K

Group

Mistral

Pricing preview

No display pricing published in the current snapshot.

Slug

microsoft/wizardlm-2-7b

Text

NovitaAI

WizardLM-2 8x22B

WizardLM-2 8x22B is Microsoft AI's most advanced Wizard model. It demonstrates highly competitive performance compared to leading proprietary models, and it consistently outperforms all existing state-of-the-art opensource models. It is an instruct finetune of [Mixtral 8x22B](/models/mistralai/mixtral-8x22b). To read more about the model release, [click here](https://wizardlm.github.io/WizardLM2/). #moe

Text

Context

65.5K

Group

Mistral

Pricing preview

Input Price: $0.62 /M tokens

Output Price: $0.62 /M tokens

Slug

microsoft/wizardlm-2-8x22b

Text

Unknown provider

Zephyr 141B-A35B

Zephyr 141B-A35B is A Mixture of Experts (MoE) model with 141B total parameters and 35B active parameters. Fine-tuned on a mix of publicly available, synthetic datasets. It is an instruct finetune of [Mixtral 8x22B](/models/mistralai/mixtral-8x22b). #moe

Text

Context

65.5K

Group

Mistral

Pricing preview

No display pricing published in the current snapshot.

Slug

huggingfaceh4/zephyr-orpo-141b-a35b

Text

Unknown provider

Mistral: Mixtral 8x22B (base)

Mixtral 8x22B is a large-scale language model from Mistral AI. It consists of 8 experts, each 22 billion parameters, with each token using 2 experts at a time. It was released via [X](https://twitter.com/MistralAI/status/1777869263778291896). #moe

Text

Context

65.5K

Group

Mistral

Pricing preview

No display pricing published in the current snapshot.

Slug

mistralai/mixtral-8x22b

Text

Unknown provider

Google: Gemini 1.5 Pro

Google's latest multimodal model, supports image and video[0] in text or chat prompts. Optimized for language tasks including: - Code generation - Text generation - Text editing - Problem solving - Recommendations - Information extraction - Data extraction or generation - AI agents Usage of Gemini is subject to Google's [Gemini Terms of Use](https://ai.google.dev/terms). * [0]: Video input is not available through OpenRouter at this time.

TextImage

Context

2M

Group

Gemini

Pricing preview

No display pricing published in the current snapshot.

Slug

google/gemini-pro-1.5

Text

OpenAI

OpenAI: GPT-4 Turbo

The latest GPT-4 Turbo model with vision capabilities. Vision requests can now use JSON mode and function calling. Training data: up to December 2023.

TextImage

Context

128K

Group

GPT

Pricing preview

Input Price: $10 /M tokens

Output Price: $30 /M tokens

Slug

openai/gpt-4-turbo

Text

Unknown provider

Cohere: Command R+

Command R+ is a new, 104B-parameter LLM from Cohere. It's useful for roleplay, general consumer usecases, and Retrieval Augmented Generation (RAG). It offers multilingual support for ten key languages to facilitate global business operations. See benchmarks and the launch post [here](https://txt.cohere.com/command-r-plus-microsoft-azure/). Use of this model is subject to Cohere's [Usage Policy](https://docs.cohere.com/docs/usage-policy) and [SaaS Agreement](https://cohere.com/saas-agreement).

Text

Context

128K

Group

Cohere

Pricing preview

No display pricing published in the current snapshot.

Slug

cohere/command-r-plus

Text

Unknown provider

Cohere: Command R+ (04-2024)

Command R+ is a new, 104B-parameter LLM from Cohere. It's useful for roleplay, general consumer usecases, and Retrieval Augmented Generation (RAG). It offers multilingual support for ten key languages to facilitate global business operations. See benchmarks and the launch post [here](https://txt.cohere.com/command-r-plus-microsoft-azure/). Use of this model is subject to Cohere's [Usage Policy](https://docs.cohere.com/docs/usage-policy) and [SaaS Agreement](https://cohere.com/saas-agreement).

Text

Context

128K

Group

Cohere

Pricing preview

No display pricing published in the current snapshot.

Slug

cohere/command-r-plus-04-2024

Text

Unknown provider

Databricks: DBRX 132B Instruct

DBRX is a new open source large language model developed by Databricks. At 132B, it outperforms existing open source LLMs like Llama 2 70B and [Mixtral-8x7b](/models/mistralai/mixtral-8x7b) on standard industry benchmarks for language understanding, programming, math, and logic. It uses a fine-grained mixture-of-experts (MoE) architecture. 36B parameters are active on any input. It was pre-trained on 12T tokens of text and code data. Compared to other open MoE models like Mixtral-8x7B and Grok-1, DBRX is fine-grained, meaning it uses a larger number of smaller experts. See the launch announcement and benchmark results [here](https://www.databricks.com/blog/introducing-dbrx-new-state-art-open-llm). #moe

Text

Context

32.8K

Group

Other

Pricing preview

No display pricing published in the current snapshot.

Slug

databricks/dbrx-instruct

Text

Unknown provider

Midnight Rose 70B

A merge with a complex family tree, this model was crafted for roleplaying and storytelling. Midnight Rose is a successor to Rogue Rose and Aurora Nights and improves upon them both. It wants to produce lengthy output by default and is the best creative writing merge produced so far by sophosympatheia. Descending from earlier versions of Midnight Rose and [Wizard Tulu Dolphin 70B](https://huggingface.co/sophosympatheia/Wizard-Tulu-Dolphin-70B-v1.0), it inherits the best qualities of each.

Text

Context

4.1K

Group

Llama2

Pricing preview

No display pricing published in the current snapshot.

Slug

sophosympatheia/midnight-rose-70b

Text

Unknown provider

Cohere: Command

Command is an instruction-following conversational model that performs language tasks with high quality, more reliably and with a longer context than our base generative models. Use of this model is subject to Cohere's [Usage Policy](https://docs.cohere.com/docs/usage-policy) and [SaaS Agreement](https://cohere.com/saas-agreement).

Text

Context

4.1K

Group

Cohere

Pricing preview

No display pricing published in the current snapshot.

Slug

cohere/command

Text

Unknown provider

Cohere: Command R

Command-R is a 35B parameter model that performs conversational language tasks at a higher quality, more reliably, and with a longer context than previous models. It can be used for complex workflows like code generation, retrieval augmented generation (RAG), tool use, and agents. Read the launch post [here](https://txt.cohere.com/command-r/). Use of this model is subject to Cohere's [Usage Policy](https://docs.cohere.com/docs/usage-policy) and [SaaS Agreement](https://cohere.com/saas-agreement).

Text

Context

128K

Group

Cohere

Pricing preview

No display pricing published in the current snapshot.

Slug

cohere/command-r

Text

Amazon Bedrock

Anthropic: Claude 3 Haiku

Claude 3 Haiku is Anthropic's fastest and most compact model for near-instant responsiveness. Quick and accurate targeted performance. See the launch announcement and benchmark results [here](https://www.anthropic.com/news/claude-3-haiku) #multimodal

TextImage

Context

200K

Group

Claude

Pricing preview

Input Price: $0.25 /M tokens

Output Price: $1.25 /M tokens

Slug

anthropic/claude-3-haiku

Text

Unknown provider

Anthropic: Claude 3 Sonnet

Claude 3 Sonnet is an ideal balance of intelligence and speed for enterprise workloads. Maximum utility at a lower price, dependable, balanced for scaled deployments. See the launch announcement and benchmark results [here](https://www.anthropic.com/news/claude-3-family) #multimodal

TextImage

Context

200K

Group

Claude

Pricing preview

No display pricing published in the current snapshot.

Slug

anthropic/claude-3-sonnet

Text

Unknown provider

Anthropic: Claude 3 Opus

Claude 3 Opus is Anthropic's most powerful model for highly complex tasks. It boasts top-level performance, intelligence, fluency, and understanding. See the launch announcement and benchmark results [here](https://www.anthropic.com/news/claude-3-family) #multimodal

TextImage

Context

200K

Group

Claude

Pricing preview

No display pricing published in the current snapshot.

Slug

anthropic/claude-3-opus

Text

Unknown provider

Cohere: Command R (03-2024)

Command-R is a 35B parameter model that performs conversational language tasks at a higher quality, more reliably, and with a longer context than previous models. It can be used for complex workflows like code generation, retrieval augmented generation (RAG), tool use, and agents. Read the launch post [here](https://txt.cohere.com/command-r/). Use of this model is subject to Cohere's [Usage Policy](https://docs.cohere.com/docs/usage-policy) and [SaaS Agreement](https://cohere.com/saas-agreement).

Text

Context

128K

Group

Cohere

Pricing preview

No display pricing published in the current snapshot.

Slug

cohere/command-r-03-2024

Text

Mistral

Mistral Large

This is Mistral AI's flagship model, Mistral Large 2 (version `mistral-large-2407`). It's a proprietary weights-available model and excels at reasoning, code, JSON, chat, and more. Read the launch announcement [here](https://mistral.ai/news/mistral-large-2407/). It supports dozens of languages including French, German, Spanish, Italian, Portuguese, Arabic, Hindi, Russian, Chinese, Japanese, and Korean, along with 80+ coding languages including Python, Java, C, C++, JavaScript, and Bash. Its long context window allows precise information recall from large documents.

Text

Context

128K

Group

Mistral

Pricing preview

Input Price: $2 /M tokens

Output Price: $6 /M tokens

Slug

mistralai/mistral-large

Text

Unknown provider

Google: Gemma 7B

Gemma by Google is an advanced, open-source language model family, leveraging the latest in decoder-only, text-to-text technology. It offers English language capabilities across text generation tasks like question answering, summarization, and reasoning. The Gemma 7B variant is comparable in performance to leading open source models. Usage of Gemma is subject to Google's [Gemma Terms of Use](https://ai.google.dev/gemma/terms).

Text

Context

8.2K

Group

Gemini

Pricing preview

No display pricing published in the current snapshot.

Slug

google/gemma-7b-it

Text

Unknown provider

Nous: Hermes 2 Mistral 7B DPO

This is the flagship 7B Hermes model, a Direct Preference Optimization (DPO) of [Teknium/OpenHermes-2.5-Mistral-7B](/models/teknium/openhermes-2.5-mistral-7b). It shows improvement across the board on all benchmarks tested - AGIEval, BigBench Reasoning, GPT4All, and TruthfulQA. The model prior to DPO was trained on 1,000,000 instructions/chats of GPT-4 quality or better, primarily synthetic data as well as other high quality datasets.

Text

Context

8.2K

Group

Mistral

Pricing preview

No display pricing published in the current snapshot.

Slug

nousresearch/nous-hermes-2-mistral-7b-dpo

Text

Unknown provider

Meta: CodeLlama 70B Instruct

Code Llama is a family of large language models for code. This one is based on [Llama 2 70B](/models/meta-llama/llama-2-70b-chat) and provides zero-shot instruction-following ability for programming tasks.

Text

Context

2K

Group

Llama2

Pricing preview

No display pricing published in the current snapshot.

Slug

meta-llama/codellama-70b-instruct

Text

Unknown provider

RWKV v5: Eagle 7B

Eagle 7B is trained on 1.1 Trillion Tokens across 100+ world languages (70% English, 15% multilang, 15% code). - Built on the [RWKV-v5](/models?q=rwkv) architecture (a linear transformer with 10-100x+ lower inference cost) - Ranks as the world's greenest 7B model (per token) - Outperforms all 7B class models in multi-lingual benchmarks - Approaches Falcon (1.5T), LLaMA2 (2T), Mistral (>2T?) level of performance in English evals - Trade blows with MPT-7B (1T) in English evals - All while being an ["Attention-Free Transformer"](https://www.isattentionallyouneed.com/) Eagle 7B models are provided for free, by [Recursal.AI](https://recursal.ai), for the beta period till end of March 2024 Find out more [here](https://blog.rwkv.com/p/eagle-7b-soaring-past-transformers) [rnn](/models?q=rwkv)

Text

Context

10K

Group

RWKV

Pricing preview

No display pricing published in the current snapshot.

Slug

recursal/eagle-7b

Text

Azure

OpenAI: GPT-3.5 Turbo (older v0613)

GPT-3.5 Turbo is OpenAI's fastest model. It can understand and generate natural language or code, and is optimized for chat and traditional completion tasks. Training data up to Sep 2021.

Text

Context

4.1K

Group

GPT

Pricing preview

Input Price: $1 /M tokens

Output Price: $2 /M tokens

Slug

openai/gpt-3.5-turbo-0613

Text

OpenAI

OpenAI: GPT-4 Turbo Preview

The preview GPT-4 model with improved instruction following, JSON mode, reproducible outputs, parallel function calling, and more. Training data: up to Dec 2023. **Note:** heavily rate limited by OpenAI while in preview.

Text

Context

128K

Group

GPT

Pricing preview

Input Price: $10 /M tokens

Output Price: $30 /M tokens

Slug

openai/gpt-4-turbo-preview

Text

Unknown provider

Yi 34B 200K

The Yi series models are large language models trained from scratch by developers at [01.AI](https://01.ai/). This version was trained on a large context length, allowing ~200k words (1000 paragraphs) of combined input and output.

Text

Context

200K

Group

Yi

Pricing preview

No display pricing published in the current snapshot.

Slug

01-ai/yi-34b-200k

Text

Unknown provider

Nous: Hermes 2 Mixtral 8x7B DPO

Nous Hermes 2 Mixtral 8x7B DPO is the new flagship Nous Research model trained over the [Mixtral 8x7B MoE LLM](/models/mistralai/mixtral-8x7b). The model was trained on over 1,000,000 entries of primarily [GPT-4](/models/openai/gpt-4) generated data, as well as other high quality data from open datasets across the AI landscape, achieving state of the art performance on a variety of tasks. #moe

Text

Context

32.8K

Group

Mistral

Pricing preview

No display pricing published in the current snapshot.

Slug

nousresearch/nous-hermes-2-mixtral-8x7b-dpo

Text

Unknown provider

Nous: Hermes 2 Mixtral 8x7B SFT

Nous Hermes 2 Mixtral 8x7B SFT is the supervised finetune only version of [the Nous Research model](/models/nousresearch/nous-hermes-2-mixtral-8x7b-dpo) trained over the [Mixtral 8x7B MoE LLM](/models/mistralai/mixtral-8x7b). The model was trained on over 1,000,000 entries of primarily GPT-4 generated data, as well as other high quality data from open datasets across the AI landscape, achieving state of the art performance on a variety of tasks. #moe

Text

Context

32.8K

Group

Mistral

Pricing preview

No display pricing published in the current snapshot.

Slug

nousresearch/nous-hermes-2-mixtral-8x7b-sft

Text

Unknown provider

Mistral Tiny

Note: This model is being deprecated. Recommended replacement is the newer [Ministral 8B](/mistral/ministral-8b) This model is currently powered by Mistral-7B-v0.2, and incorporates a "better" fine-tuning than [Mistral 7B](/models/mistralai/mistral-7b-instruct-v0.1), inspired by community work. It's best used for large batch processing tasks where cost is a significant factor but reasoning capabilities are not crucial.

Text

Context

32K

Group

Mistral

Pricing preview

No display pricing published in the current snapshot.

Slug

mistralai/mistral-tiny

Text

Unknown provider

Mistral Medium

This is Mistral AI's closed-source, medium-sided model. It's powered by a closed-source prototype and excels at reasoning, code, JSON, chat, and more. In benchmarks, it compares with many of the flagship models of other companies.

Text

Context

32K

Group

Mistral

Pricing preview

No display pricing published in the current snapshot.

Slug

mistralai/mistral-medium

Text

Unknown provider

Mistral Small

With 22 billion parameters, Mistral Small v24.09 offers a convenient mid-point between (Mistral NeMo 12B)[/mistralai/mistral-nemo] and (Mistral Large 2)[/mistralai/mistral-large], providing a cost-effective solution that can be deployed across various platforms and environments. It has better reasoning, exhibits more capabilities, can produce and reason about code, and is multiligual, supporting English, French, German, Italian, and Spanish.

Text

Context

32K

Group

Mistral

Pricing preview

No display pricing published in the current snapshot.

Slug

mistralai/mistral-small

Text

Unknown provider

Bagel 34B v0.2

An experimental fine-tune of [Yi 34b 200k](/models/01-ai/yi-34b-200k) using [bagel](https://github.com/jondurbin/bagel). This is the version of the fine-tune before direct preference optimization (DPO) has been applied. DPO performs better on benchmarks, but this version is likely better for creative writing, roleplay, etc.

Text

Context

200K

Group

Yi

Pricing preview

No display pricing published in the current snapshot.

Slug

jondurbin/bagel-34b

Text

Unknown provider

Nous: Hermes 2 Yi 34B

Nous Hermes 2 Yi 34B was trained on 1,000,000 entries of primarily GPT-4 generated data, as well as other high quality data from open datasets across the AI landscape. Nous-Hermes 2 on Yi 34B outperforms all Nous-Hermes & Open-Hermes models of the past, achieving new heights in all benchmarks for a Nous Research LLM as well as surpassing many popular finetunes.

Text

Context

4.1K

Group

Yi

Pricing preview

No display pricing published in the current snapshot.

Slug

nousresearch/nous-hermes-yi-34b

Text

Unknown provider

Noromaid Mixtral 8x7B Instruct

This model was trained for 8h(v1) + 8h(v2) + 12h(v3) on customized modified datasets, focusing on RP, uncensoring, and a modified version of the Alpaca prompting (that was already used in LimaRP), which should be at the same conversational level as ChatLM or Llama2-Chat without adding any additional special tokens.

Text

Context

8K

Group

Mistral

Pricing preview

No display pricing published in the current snapshot.

Slug

neversleep/noromaid-mixtral-8x7b-instruct

Text

Unknown provider

Mistral: Mistral 7B Instruct v0.2

A high-performing, industry-standard 7.3B parameter model, with optimizations for speed and context length. An improved version of [Mistral 7B Instruct](/modelsmistralai/mistral-7b-instruct-v0.1), with the following changes: - 32k context window (vs 8k context in v0.1) - Rope-theta = 1e6 - No Sliding-Window Attention

Text

Context

32.8K

Group

Mistral

Pricing preview

No display pricing published in the current snapshot.

Slug

mistralai/mistral-7b-instruct-v0.2

Text

Unknown provider

Dolphin 2.6 Mixtral 8x7B 🐬

This is a 16k context fine-tune of [Mixtral-8x7b](/models/mistralai/mixtral-8x7b). It excels in coding tasks due to extensive training with coding data and is known for its obedience, although it lacks DPO tuning. The model is uncensored and is stripped of alignment and bias. It requires an external alignment layer for ethical use. Users are cautioned to use this highly compliant model responsibly, as detailed in a blog post about uncensored models at [erichartford.com/uncensored-models](https://erichartford.com/uncensored-models). #moe #uncensored

Text

Context

32.8K

Group

Mistral

Pricing preview

No display pricing published in the current snapshot.

Slug

cognitivecomputations/dolphin-mixtral-8x7b

Text

Unknown provider

RWKV v5 World 3B

[RWKV](https://wiki.rwkv.com) is an RNN (recurrent neural network) with transformer-level performance. It aims to combine the best of RNNs and transformers - great performance, fast inference, low VRAM, fast training, "infinite" context length, and free sentence embedding. RWKV-5 is trained on 100+ world languages (70% English, 15% multilang, 15% code). RWKV 3B models are provided for free, by Recursal.AI, for the beta period. More details [here](https://substack.recursal.ai/p/public-rwkv-3b-model-via-openrouter). #rnn

Text

Context

10K

Group

RWKV

Pricing preview

No display pricing published in the current snapshot.

Slug

rwkv/rwkv-5-world-3b

Text

DeepInfra

Mistral: Mixtral 8x7B Instruct

Mixtral 8x7B Instruct is a pretrained generative Sparse Mixture of Experts, by Mistral AI, for chat and instruction use. Incorporates 8 experts (feed-forward networks) for a total of 47 billion parameters. Instruct model fine-tuned by Mistral. #moe

Text

Context

32.8K

Group

Mistral

Pricing preview

Input Price: $0.54 /M tokens

Output Price: $0.54 /M tokens

Slug

mistralai/mixtral-8x7b-instruct

Text

Unknown provider

RWKV v5 3B AI Town

This is an [RWKV 3B model](/models/rwkv/rwkv-5-world-3b) finetuned specifically for the [AI Town](https://github.com/a16z-infra/ai-town) project. [RWKV](https://wiki.rwkv.com) is an RNN (recurrent neural network) with transformer-level performance. It aims to combine the best of RNNs and transformers - great performance, fast inference, low VRAM, fast training, "infinite" context length, and free sentence embedding. RWKV 3B models are provided for free, by Recursal.AI, for the beta period. More details [here](https://substack.recursal.ai/p/public-rwkv-3b-model-via-openrouter). #rnn

Text

Context

10K

Group

RWKV

Pricing preview

No display pricing published in the current snapshot.

Slug

recursal/rwkv-5-3b-ai-town

Text

Unknown provider

StripedHyena Nous 7B

This is the chat model variant of the [StripedHyena series](/models?q=stripedhyena) developed by Together in collaboration with Nous Research. StripedHyena uses a new architecture that competes with traditional Transformers, particularly in long-context data processing. It combines attention mechanisms with gated convolutions for improved speed, efficiency, and scaling. This model marks a significant advancement in AI architecture for sequence modeling tasks.

Text

Context

32.8K

Group

Mistral

Pricing preview

No display pricing published in the current snapshot.

Slug

togethercomputer/stripedhyena-nous-7b

Text

Unknown provider

StripedHyena Hessian 7B (base)

This is the base model variant of the [StripedHyena series](/models?q=stripedhyena), developed by Together. StripedHyena uses a new architecture that competes with traditional Transformers, particularly in long-context data processing. It combines attention mechanisms with gated convolutions for improved speed, efficiency, and scaling. This model marks an advancement in AI architecture for sequence modeling tasks.

Text

Context

32.8K

Group

Mistral

Pricing preview

No display pricing published in the current snapshot.

Slug

togethercomputer/stripedhyena-hessian-7b

Text

Unknown provider

Psyfighter v2 13B

The v2 of [Psyfighter](/models/jebcarter/psyfighter-13b) - a merged model created by the KoboldAI community members Jeb Carter and TwistedShadows, made possible thanks to the KoboldAI merge request service. The intent was to add medical data to supplement the model's fictional ability with more details on anatomy and mental states. This model should not be used for medical advice or therapy because of its high likelihood of pulling in fictional data. It's a merge between: - [KoboldAI/LLaMA2-13B-Tiefighter](https://huggingface.co/KoboldAI/LLaMA2-13B-Tiefighter) - [Doctor-Shotgun/cat-v1.0-13b](https://huggingface.co/Doctor-Shotgun/cat-v1.0-13b) - [Doctor-Shotgun/llama-2-13b-chat-limarp-v2-merged](https://huggingface.co/Doctor-Shotgun/llama-2-13b-chat-limarp-v2-merged). #merge

Text

Context

4.1K

Group

Llama2

Pricing preview

No display pricing published in the current snapshot.

Slug

koboldai/psyfighter-13b-2

Text

Unknown provider

MythoMist 7B

From the creator of [MythoMax](/models/gryphe/mythomax-l2-13b), merges a suite of models to reduce word anticipation, ministrations, and other undesirable words in ChatGPT roleplaying data. It combines [Neural Chat 7B](/models/intel/neural-chat-7b), Airoboros 7b, [Toppy M 7B](/models/undi95/toppy-m-7b), [Zepher 7b beta](/models/huggingfaceh4/zephyr-7b-beta), [Nous Capybara 34B](/models/nousresearch/nous-capybara-34b), [OpenHeremes 2.5](/models/teknium/openhermes-2.5-mistral-7b), and many others. #merge

Text

Context

32.8K

Group

Mistral

Pricing preview

No display pricing published in the current snapshot.

Slug

gryphe/mythomist-7b

Text

Unknown provider

Yi 34B Chat

The Yi series models are large language models trained from scratch by developers at [01.AI](https://01.ai/). This 34B parameter model has been instruct-tuned for chat.

Text

Context

4.1K

Group

Yi

Pricing preview

No display pricing published in the current snapshot.

Slug

01-ai/yi-34b-chat

Text

Unknown provider

Yi 6B (base)

The Yi series models are large language models trained from scratch by developers at [01.AI](https://01.ai/). This is the base 6B parameter model.

Text

Context

4.1K

Group

Yi

Pricing preview

No display pricing published in the current snapshot.

Slug

01-ai/yi-6b

Text

Unknown provider

Nous: Hermes 2 Vision 7B (alpha)

This vision-language model builds on innovations from the popular [OpenHermes-2.5](/models/teknium/openhermes-2.5-mistral-7b) model, by Teknium. It adds vision support, and is trained on a custom dataset enriched with function calling This project is led by [qnguyen3](https://twitter.com/stablequan) and [teknium](https://twitter.com/Teknium1). #multimodal

TextImage

Context

4.1K

Group

Mistral

Pricing preview

No display pricing published in the current snapshot.

Slug

nousresearch/nous-hermes-2-vision-7b

Text

Unknown provider

Yi 34B (base)

The Yi series models are large language models trained from scratch by developers at [01.AI](https://01.ai/). This is the base 34B parameter model.

Text

Context

4.1K

Group

Yi

Pricing preview

No display pricing published in the current snapshot.

Slug

01-ai/yi-34b

Text

Unknown provider

Nous: Capybara 7B

The Capybara series is a collection of datasets and models made by fine-tuning on data created by Nous, mostly in-house. V1.9 uses unalignment techniques for more consistent and dynamic control. It also leverages a significantly better foundation model, [Mistral 7B](/models/mistralai/mistral-7b-instruct-v0.1).

Text

Context

8.2K

Group

Mistral

Pricing preview

No display pricing published in the current snapshot.

Slug

nousresearch/nous-capybara-7b

Text

Unknown provider

Psyfighter 13B

A merge model based on [Llama-2-13B](/models/meta-llama/llama-2-13b-chat) and made possible thanks to the compute provided by the KoboldAI community. It's a merge between: - [KoboldAI/LLaMA2-13B-Tiefighter](https://huggingface.co/KoboldAI/LLaMA2-13B-Tiefighter) - [chaoyi-wu/MedLLaMA_13B](https://huggingface.co/chaoyi-wu/MedLLaMA_13B) - [Doctor-Shotgun/llama-2-13b-chat-limarp-v2-merged](https://huggingface.co/Doctor-Shotgun/llama-2-13b-chat-limarp-v2-merged). #merge

Text

Context

4.1K

Group

Llama2

Pricing preview

No display pricing published in the current snapshot.

Slug

jebcarter/psyfighter-13b

Text

Unknown provider

OpenChat 3.5 7B

OpenChat 7B is a library of open-source language models, fine-tuned with "C-RLFT (Conditioned Reinforcement Learning Fine-Tuning)" - a strategy inspired by offline reinforcement learning. It has been trained on mixed-quality data without preference labels. - For OpenChat fine-tuned on Mistral 7B, check out [OpenChat 7B](/models/openchat/openchat-7b). - For OpenChat fine-tuned on Llama 8B, check out [OpenChat 8B](/models/openchat/openchat-8b). #open-source

Text

Context

8.2K

Group

Mistral

Pricing preview

No display pricing published in the current snapshot.

Slug

openchat/openchat-7b

Page 10 of 15

Need a model request?

Use the market snapshot for discovery, then ask ImaRouter for rollout.

If a model matters for your product, send the slug, expected traffic, target region, and latency expectations. The team can confirm support status, onboarding priority, or a migration path to an equivalent route on ImaRouter.

Contact

support@imarouter.com

Best for model availability questions, onboarding priority, routing strategy, and enterprise rollout planning.

Models | ImaRouter