Seamlessly migrate from closed-source solutions with OpenAI-compatible APIs.
Chat-optimized LLM leveraging public datasets and 1M+ human annotations.
Try this modelChat-optimized LLM leveraging public datasets and 1M+ human annotations.
Try this modelAuto-regressive LLM with optimized transformers, SFT, and RLHF for alignment with helpfulness and safety preferences.
Try this modelAuto-regressive LLM with optimized transformers, SFT, and RLHF for alignment with helpfulness and safety preferences.
Try this modelAuto-regressive LLM with optimized transformers, SFT, and RLHF for alignment with helpfulness and safety preferences.
Try this modelAuto-regressive LLM with optimized transformers, SFT, and RLHF for alignment with helpfulness and safety preferences.
Try this modelExperimental merge of MythoLogic-L2 and Huginn using tensor intermingling for enhanced front and end tensor integration.
Try this modelExperimental merge of MythoLogic-L2 and Huginn using tensor intermingling for enhanced front and end tensor integration.
Try this modelInstruct Thai large language model with 8 billion parameters based on Llama3.1-8B.
Try this modelInstruct Thai large language model with 70 billion parameters, based on Llama3.1-70B.
Try this modelTransformer-based decoder-only LLM, pretrained on extensive data, offering improvements over the previous Qwen model.
Try this modelFlagship Nous Research MoE model trained on 1M+ GPT-4 and high-quality open dataset entries, excelling across diverse tasks.
Try this modelMultilingual LLM pre-trained and instruction-tuned, surpassing open and closed models on key benchmarks.
Try this modelImproved instruct fine-tuned version of Mistral-7B-Instruct-v0.1.
Try this modelInstruction-tuned 7.61B Qwen2.5 causal LLM with 131K context, RoPE, SwiGLU, RMSNorm, and advanced attention mechanisms.
Try this modelLightweight, SOTA open models from Google, leveraging research and tech behind the Gemini models.
Try this model2B instruct Gemma model by Google: lightweight, open, text-to-text LLM for QA, summarization, reasoning, and resource-efficient deployment.
Try this modelMultimodal LLM optimized for visual recognition, image reasoning, captioning, and answering image-related questions.
Try this modelMoE LLM trained from scratch and specialized in few-turn interactions for enhanced performance.
Try this modelLightweight Gemma 3 model with 128K context, vision-language input, and multilingual support for on-device AI.
Try this modelLightweight Gemma 3 model (1B) with 128K context, vision-language input, and multilingual support for on-device AI.
Try this modelMost lightweight Gemma 3 model, with 128K context, vision-language input, and multilingual support for on-device AI.
Try this modelSmall Qwen 1.5B distilled with reasoning capabilities from Deepseek R1. Beats GPT-4o on MATH-500 whilst being a fraction of the size.
Try this modelQwen 14B distilled with reasoning capabilities from Deepseek R1. Outperforms GPT-4o in math & matches o1-mini on coding.
Try this modelLlama 70B distilled with reasoning capabilities from Deepseek R1. Surpasses GPT-4o with 94.5% on MATH-500 & matches o1-mini on coding.
Try this modelMultilingual LLM pre-trained and instruction-tuned, surpassing open and closed models on key benchmarks.
Try this modelLightweight, SOTA open models from Google, leveraging research and tech behind the Gemini models.
Try this modelMultilingual LLM pre-trained and instruction-tuned, surpassing open and closed models on key benchmarks.
Try this model70B multilingual LLM, pretrained and instruction-tuned, excels in dialogue use cases, surpassing open and closed models.
Try this modelBest-in-class open-source LLM trained with IDA for alignment, reasoning, and self-reflective, agentic applications.
Try this modelCustom NVIDIA LLM optimized to enhance the helpfulness and relevance of generated responses to user queries.
Try this modelBest-in-class open-source LLM trained with IDA for alignment, reasoning, and self-reflective, agentic applications.
Try this model24B model rivaling GPT-4o mini, and larger models like Llama 3.3 70B. Ideal for chat use cases like customer support, translation and summarization.
Try this modelBest-in-class open-source LLM trained with IDA for alignment, reasoning, and self-reflective, agentic applications.
Try this modelBest-in-class open-source LLM trained with IDA for alignment, reasoning, and self-reflective, agentic applications.
Try this modelQwen series reasoning model excelling in complex tasks, outperforming conventional instruction-tuned models on hard problems.
Try this modelMixture-of-Experts model challenging top AI models at much lower cost. Updated on March 24th, 2025.
Try this modelBest-in-class open-source LLM trained with IDA for alignment, reasoning, and self-reflective, agentic applications.
Try this modelVision-language model with advanced visual reasoning, video understanding, structured outputs, and agentic capabilities.
Try this modelFree endpoint to try this 70B multilingual LLM optimized for dialogue, excelling in benchmarks and surpassing many chat models.
Try this modelLightweight model with vision-language input, multilingual support, visual reasoning, and top-tier performance per size.
Try this modelFree endpoint to experiment the power of reasoning models. This distilled model beats GPT-4o on math & matches o1-mini on coding.
Try this modelOpen-source reasoning model rivaling OpenAI-o1, excelling in math, code, reasoning, and cost efficiency.
Try this modelSOTA 109B model with 17B active params & large context, excelling at multi-document analysis, codebase reasoning, and personalized tasks.
Try this modelSOTA 128-expert MoE powerhouse for multilingual image/text understanding, creative writing, and enterprise-scale applications.
Try this model12 billion parameter rectified flow transformer capable of generating images from text descriptions.
Try this modelFastest available endpoint for the SOTA open-source image generation model by Black Forest Labs.
Try this modelAdapter for FLUX.1 models enabling image variation, refining input images, and integrating into advanced restyling workflows.
Try this model12 billion parameter rectified flow transformer capable of generating an image based on a text description while following the structure of a given input image.
Try this model12 billion parameter rectified flow transformer capable of generating an image based on a text description while following the structure of a given input image.
Try this model12 billion parameter rectified flow transformer capable of generating images from text descriptions.
Try this modelAdvanced image generation model with FLUX.1-dev architecture, offering high-quality outputs for artistic and commercial use.
Try this modelFree endpoint for the SOTA open-source image generation model by Black Forest Labs.
Try this modelMultimodal LLM optimized for visual recognition, image reasoning, captioning, and answering image-related questions.
Try this modelMultimodal LLM optimized for visual recognition, image reasoning, captioning, and answering image-related questions.
Try this modelLightweight Gemma 3 model with 128K context, vision-language input, and multilingual support for on-device AI.
Try this modelLightweight Gemma 3 model (1B) with 128K context, vision-language input, and multilingual support for on-device AI.
Try this modelMost lightweight Gemma 3 model, with 128K context, vision-language input, and multilingual support for on-device AI.
Try this modelOSS vision model merging advanced vision with instruction-tuned language understanding for visual reasoning.
Try this modelFree endpoint to test this auto-regressive language model that uses an optimized transformer architecture.
Try this modelVision-language model with advanced visual reasoning, video understanding, structured outputs, and agentic capabilities.
Try this modelLightweight model with vision-language input, multilingual support, visual reasoning, and top-tier performance per size.
Try this modelSOTA 109B model with 17B active params & large context, excelling at multi-document analysis, codebase reasoning, and personalized tasks.
Try this modelSOTA 128-expert MoE powerhouse for multilingual image/text understanding, creative writing, and enterprise-scale applications.
Try this modelLow-latency, ultra-realistic voice model, served in partnership with Cartesia.
Try this modelLLM trained on 2T tokens with double Llama 1's context length, available in 7B, 13B, and 70B parameter sizes.
Try this model8B Llama 3.1 model fine-tuned for content safety, moderating prompts and responses in 8 languages with MLCommons alignment.
Try this model11B Llama 3.2 model fine-tuned for content safety, detecting harmful multimodal prompts and text in image reasoning use cases.
Try this model8B Llama 3-based safeguard model for classifying LLM inputs and outputs, detecting unsafe content and policy violations.
Try this model7B Llama 2-based safeguard model for classifying LLM inputs and outputs, detecting unsafe content and policy violations.
Try this model7.3B model surpassing Llama 2 13B, nearing CodeLlama 7B on code, with GQA for speed and SWA for efficient long-sequence handling.
Try this modelLightweight Gemma 3 model with 128K context, vision-language input, and multilingual support for on-device AI.
Try this modelLightweight Gemma 3 model (1B) with 128K context, vision-language input, and multilingual support for on-device AI.
Try this modelMost lightweight Gemma 3 model, with 128K context, vision-language input, and multilingual support for on-device AI.
Try this modelSOTA code LLM with advanced code generation, reasoning, fixing, and support for up to 128K tokens.
Try this modelBest-in-class open-source LLM trained with IDA for alignment, reasoning, and self-reflective, agentic applications.
Try this modelBest-in-class open-source LLM trained with IDA for alignment, reasoning, and self-reflective, agentic applications.
Try this modelBest-in-class open-source LLM trained with IDA for alignment, reasoning, and self-reflective, agentic applications.
Try this modelBest-in-class open-source LLM trained with IDA for alignment, reasoning, and self-reflective, agentic applications.
Try this modelBest-in-class open-source LLM trained with IDA for alignment, reasoning, and self-reflective, agentic applications.
Try this modelLightweight model with vision-language input, multilingual support, visual reasoning, and top-tier performance per size.
Try this model80M checkpoint of M2-BERT, pretrained with sequence length 32768, and it has been fine-tuned for long-context retrieval.
Try this modelThis model maps any text to a low-dimensional dense vector using FlagEmbedding.
Try this model80M checkpoint of M2-BERT, pretrained with sequence length 8192, and it has been fine-tuned for long-context retrieval.
Try this model80M checkpoint of M2-BERT, pretrained with sequence length 2048, and it has been fine-tuned for long-context retrieval.
Try this modelUniversal English sentence embedding model by WhereIsAI with 1024-dim embeddings and 512 context length support.
Try this modelBAAI v1.5 large maps text to dense vectors for retrieval, classification, clustering, semantic search, and LLM databases.
Try this modelSalesforce Research's proprietary fine-tuned rerank model with 8K context, outperforming Cohere Rerank for superior document retrieval.
Try this model