Co-authored-by: Copilot <175728472+Copilot@users.noreply.github.com> Co-authored-by: Ben Ahmady <32935794+subatoi@users.noreply.github.com>
1.0 KiB
1.0 KiB
title, shortTitle, intro, versions, topics, autogenerated, allowTitleToDifferFromFilename
| title | shortTitle | intro | versions | topics | autogenerated | allowTitleToDifferFromFilename | |||
|---|---|---|---|---|---|---|---|---|---|
| REST API endpoints for models inference | Inference | Use the REST API to submit a chat completion request to a specified model, with or without organizational attribution. |
|
|
rest | true |
About {% data variables.product.prodname_github_models %} inference
You can use the REST API to run inference requests using the {% data variables.product.prodname_github_models %} platform. The API requires the models: read scope when using a {% data variables.product.pat_v2 %} or when authenticating using a {% data variables.product.prodname_github_app %}.
The API supports:
- Accessing top models from OpenAI, DeepSeek, Microsoft, Llama, and more.
- Running chat-based inference requests with full control over sampling and response parameters.
- Streaming or non-streaming completions.
- Organizational attribution and usage tracking.