Skip to content

DeepInfra

DeepInfra inference provider using the OpenAI-compatible Chat Completions API.

Setup

bash
go get github.com/zendev-sh/goai@latest
go
import "github.com/zendev-sh/goai/provider/deepinfra"

Set the DEEPINFRA_API_KEY environment variable, or pass WithAPIKey() directly.

Models

  • meta-llama/Llama-3.3-70B-Instruct
  • mistralai/Mixtral-8x7B-Instruct-v0.1
  • Qwen/Qwen2.5-72B-Instruct

Model IDs use the org/model prefix format.

Tested Models

Unit tested (mock HTTP server, 2026-03-15): meta-llama/Llama-3.3-70B-Instruct

Usage

go
model := deepinfra.Chat("meta-llama/Llama-3.3-70B-Instruct")

result, err := goai.GenerateText(ctx, model, goai.WithPrompt("Hello"))
if err != nil {
    log.Fatal(err)
}
fmt.Println(result.Text)

Options

OptionTypeDescription
WithAPIKey(key)stringSet a static API key
WithTokenSource(ts)provider.TokenSourceSet a dynamic token source
WithBaseURL(url)stringOverride the default https://api.deepinfra.com/v1/openai endpoint
WithHeaders(h)map[string]stringSet additional HTTP headers
WithHTTPClient(c)*http.ClientSet a custom *http.Client

Notes

  • Environment variable DEEPINFRA_BASE_URL can override the default endpoint.

Released under the MIT License.