Aitlas — Map the AI Universe

      inference

      Total 16 articles sites
      Sorting
      releaseupdateViewsLike
      KoboldCpp

      KoboldCpp

      Single-binary llama.cpp fork with a lightweight built-in UI; fast local inference for RP with long context options.
      0370
      Roleplay Frontends (Local/Self-Hosted UI)# adult# GGUF# inference
      RunPod Serverless Endpoints

      RunPod Serverless Endpoints

      Always-on, pre-warmed GPU endpoints for low-latency model inference at scale.
      0320
      Inference/Hosting & APIs# endpoints# GPU# inference
      Modal Inference

      Modal Inference

      Serverless GPU inference with sub-second cold starts and Python-first workflows.
      0310
      Inference/Hosting & APIs# GPU# inference# Modal
      Baseten

      Baseten

      Production inference platform—dedicated deployments, autoscaling, and GPU options.
      0320
      Inference/Hosting & APIs# autoscaling# Baseten# dedicated
      SambaNova Cloud

      SambaNova Cloud

      RDU-accelerated inference platform with OpenAI-compatible API keys for top open models.
      0440
      Inference/Hosting & APIs# API# inference# LLM
      Cerebras Inference

      Cerebras Inference

      Wafer-scale engine cloud with OpenAI-style APIs for ultra-fast open-model inference.
      0450
      Inference/Hosting & APIs# API# Cerebras# inference
      Together AI Inference

      Together AI Inference

      Fast serverless APIs and dedicated endpoints for 200+ open models.
      0460
      Inference/Hosting & APIs# API# dedicated# inference
      GroqCloud

      GroqCloud

      Ultra-low latency LPU-powered inference for text, speech, and vision models.
      0480
      Inference/Hosting & APIs# API# Groq# inference
      NVIDIA NIM

      NVIDIA NIM

      Prebuilt, optimized inference microservices for leading models on any NVIDIA-accelerated stack.
      0310
      Inference/Hosting & APIs# GPU# inference# microservices
      Azure AI Foundry Models / OpenAI

      Azure AI Foundry Models / OpenAI

      Catalog of OpenAI and open models with enterprise governance and Azure AI Inference APIs.
      0330
      Inference/Hosting & APIs# Azure# enterprise# governance
      OpenVINO Open Model Zoo

      OpenVINO Open Model Zoo

      Optimized Intel OpenVINO reference models and demos for high-performance inference.
      0310
      Model Hubs# inference# Intel# model zoo
      Together AI Model Library

      Together AI Model Library

      Fast inference and fine-tuning for 200+ open models with a unified developer experience.
      0300
      Model Hubs# fine-tuning# inference# model library
      NVIDIA NGC Models

      NVIDIA NGC Models

      Optimized model catalog for NVIDIA GPUs—LLMs, vision, speech—with containers and inference recipes.
      0300
      Model Hubs# containers# GPU# inference
      GitHub Models

      GitHub Models

      A model catalog and API integrated with GitHub—evaluate, compare, and run many vendor models via one interface.
      0460
      Model Hubs# API# evaluation# GitHub
      Hugging Face Hub

      Hugging Face Hub

      The largest open model repo to browse, download, and deploy LLMs, vision, audio, and multimodal models with rich metadata and tooling.
      0330
      Model Hubs# audio# datasets# inference
      加载更多
      Aitlas — Map the AI Universe
      AITLAS is a curated AI tools directory for creators and developers. With a minimal, distraction-free UI and a clear 3-level taxonomy, you can scan categories fast and jump straight to what matters. Each entry includes a concise overview, key tags, and official links, while deep collections keep growing across roleplay & worldbuilding, research writing, automation, translation, and generative media. Broad coverage, frequent updates—so you find the right tool, faster.
      aitlas.orgAitlas — Map the AI Universe
      aitlas.org
      © 2025 AITLAS.org All rights reserved.  
      feedback
      Let's build a civilized community together! Your feedback is important!
      sites
      sitesposts