Completions

Conversational Chat Completions

👍

New! Tools for chat completions now available. Click here for more details.

Why Use Standard Chat Completions?

Standard chat completions—like OpenAI’s chat/completions endpoint—are quickly becoming the industry default for building conversational AI. Whether you're developing a chatbot, virtual assistant, or AI-powered automation, here's why sticking with the standard just makes sense.

🥳

The Gloo AI Completions API allows you to build with many of the leading open-source models and ensure that it has Human Flourishing alignment and guardrails for any use.


Familiar Developer Experience

If you've used tools like OpenAI, Anthropic, or other LLM providers, you've likely already encountered the chat format: a sequence of messages between system, user, and assistant.

[
  { "role": "system", "content": "You are a helpful assistant." },
  { "role": "user", "content": "How do I reset my password?" }
]

This simple message-based interface has become the standard across APIs—making it easier to build, test, and scale without needing to relearn a custom format for every provider.


Plug-and-Play with Tools and SDKs

Standard chat completions are supported out-of-the-box by:

  • Popular SDKs (openai, langchain, llamaindex, etc.)
  • Prompt engineering tools
  • Debugging/observability dashboards
  • Prompt versioning platforms
  • Orchestration frameworks

That means less boilerplate and more productivity.


Easily Portable Across Models

Using a standard chat format lets you switch between:

  • OpenAI's GPT models
  • Anthropic's Claude
  • Google Gemini
  • Open-source chat models like Mistral, LLaMA, or Mixtral

...without rewriting your app logic.

Standardization lets you benchmark models, compare outputs, and even build model-agnostic fallbacks or ensembles.


Better Alignment with Human Intent

The chat format encourages natural interaction flow, including:

  • Clarification (multi-turn)
  • Function/tool calling
  • Role-based prompting (e.g. system as context setter)

It also makes features like memory, tool use, and retrieval-augmented generation (RAG) feel native, rather than bolted on.


Built-In Support for Advanced Features

Using chat completions gives you access to:

  • Function/tool calling
  • JSON mode
  • Temperature/top_p control
  • Streaming support
  • Multi-turn context windows

These features are essential for reliable, context-aware, and real-time applications.


Supported by the Ecosystem

From hosting providers to vector databases, the whole AI tooling ecosystem now expects the standard chat format. It’s the easiest way to stay compatible and future-proof your codebase.


TL;DR: Why Use Standard Chat Completions?

BenefitDescription
FamiliarCommon API shape across many providers
PluggableCompatible with SDKs and tools
PortableEasy to swap or compare models
PowerfulSupports advanced features like tool calling
Ecosystem-ReadyPlays nicely with the AI stack
Human FlourishingBuild with AI and Human Flourishing capabilities at it's core with the world's leading open source models.

Ready to build?

The chat format is your foundation for production-ready AI apps—backed by best practices, trusted by industry, and ready to scale.

Quickstart Completions API


What’s Next

Learn more about our Chat (Threads) API.