1
0
mirror of synced 2025-12-21 10:57:10 -05:00
Files
docs/content/rest/models/inference.md
2025-05-29 14:30:23 +00:00

1.0 KiB

title, shortTitle, intro, versions, topics, autogenerated, allowTitleToDifferFromFilename
title shortTitle intro versions topics autogenerated allowTitleToDifferFromFilename
REST API endpoints for models inference Inference Use the REST API to submit a chat completion request to a specified model, with or without organizational attribution.
fpt
*
API
rest true

About {% data variables.product.prodname_github_models %} inference

You can use the REST API to run inference requests using the {% data variables.product.prodname_github_models %} platform. The API requires the models: read scope when using a {% data variables.product.pat_v2 %} or when authenticating using a {% data variables.product.prodname_github_app %}.

The API supports:

  • Accessing top models from OpenAI, DeepSeek, Microsoft, Llama, and more.
  • Running chat-based inference requests with full control over sampling and response parameters.
  • Streaming or non-streaming completions.
  • Organizational attribution and usage tracking.