Last updated:
0 purchases
openai dart
OpenAI Dart Client #
Unofficial Dart client for OpenAI API.
Features #
Generated from the official OpenAI OpenAPI specification
Fully type-safe, documented and tested
All platforms supported
Authentication with organization support
Custom base URL, headers and query params support (e.g. HTTP proxies)
Custom HTTP client support (e.g. SOCKS5 proxies or advanced use cases)
Partial Azure OpenAI API support
It can be used to consume OpenAI-compatible APIs like TogetherAI, Anyscale, OpenRouter, One API, Groq, Llamafile, GPT4All, FastChat, etc.
Supported endpoints:
Chat (with tools and streaming support)
Completions (legacy)
Embeddings
Fine-tuning
Batch
Images
Models
Moderations
Assistants v2 (with tools and streaming support) beta
Threads
Messages
Runs
Run Steps
Vector Stores
Vector Store Files
Vector Store File Batches
Table of contents #
Usage
Authentication
Organization (optional)
Chat
Completions (legacy)
Embeddings
Fine-tuning
Batch
Images
Models
Moderations
Assistants (beta)
Threads (beta)
Messages (beta)
Runs (beta)
Vector Stores (beta)
Vector Store Files (beta)
Vector Store File Batches (beta)
Advance Usage
Azure OpenAI Service
OpenAI-compatible APIs
Default HTTP client
Custom HTTP client
Using a proxy
HTTP proxy
SOCKS5 proxy
Acknowledgements
License
Usage #
Authentication #
The OpenAI API uses API keys for authentication. Visit your API Keys page to retrieve the API key you'll use in your requests.
Remember that your API key is a secret!
Do not share it with others or expose it in any client-side code (browsers, apps). Production requests must be routed through your own backend server where your API key can be securely loaded from an environment variable or key management service.
final openaiApiKey = Platform.environment['OPENAI_API_KEY'];
final client = OpenAIClient(apiKey: openaiApiKey);
copied to clipboard
Organization (optional)
For users who belong to multiple organizations, you can specify which organization is used for an API request. Usage from these API requests will count against the specified organization's subscription quota.
final client = OpenAIClient(
apiKey: openaiApiKey,
organization: 'org-dtDDtkEGoFccn5xaP5W1p3Rr',
);
copied to clipboard
Chat #
Given a list of messages comprising a conversation, the model will return a response.
Related guide: Chat Completions
Create chat completion:
final res = await client.createChatCompletion(
request: CreateChatCompletionRequest(
model: ChatCompletionModel.modelId('gpt-4'),
messages: [
ChatCompletionMessage.system(
content: 'You are a helpful assistant.',
),
ChatCompletionMessage.user(
content: ChatCompletionUserMessageContent.string('Hello!'),
),
],
temperature: 0,
),
);
print(res.choices.first.message.content);
// Hello! How can I assist you today?
copied to clipboard
ChatCompletionModel is a sealed class that offers two ways to specify the model:
ChatCompletionModel.modelId('model-id'): the model ID as string (e.g. 'gpt-4' or your fine-tuned model ID).
ChatCompletionModel.model(ChatCompletionModels.gpt4): a value from ChatCompletionModels enum which lists all of the available models.
ChatCompletionMessage is a sealed class that supports the following message types:
ChatCompletionMessage.system(): a system message.
ChatCompletionMessage.user(): a user message.
ChatCompletionMessage.assistant(): an assistant message.
ChatCompletionMessage.tool(): a tool message.
ChatCompletionMessage.function(): a function message.
ChatCompletionMessage.user() takes a ChatCompletionUserMessageContent object that supports the following content types:
ChatCompletionUserMessageContent.string('content'): string content.
ChatCompletionUserMessageContent.parts([...]): multi-modal content (check the 'Multi-modal prompt' section below).
ChatCompletionMessageContentPart.text('content'): text content.
ChatCompletionMessageContentPart.image(imageUrl: ...): image content.
Stream chat completion:
final stream = client.createChatCompletionStream(
request: CreateChatCompletionRequest(
model: ChatCompletionModel.modelId('gpt-4-turbo'),
messages: [
ChatCompletionMessage.system(
content:
'You are a helpful assistant that replies only with numbers '
'in order without any spaces or commas',
),
ChatCompletionMessage.user(
content: ChatCompletionUserMessageContent.string(
'List the numbers from 1 to 9',
),
),
],
),
);
await for (final res in stream) {
print(res.choices.first.delta.content);
}
// 123
// 456
// 789
copied to clipboard
Multi-modal prompt: (docs)
final res = await client.createChatCompletion(
request: CreateChatCompletionRequest(
model: ChatCompletionModel.model(
ChatCompletionModels.gpt4VisionPreview,
),
messages: [
ChatCompletionMessage.system(
content: 'You are a helpful assistant.',
),
ChatCompletionMessage.user(
content: ChatCompletionUserMessageContent.parts(
[
ChatCompletionMessageContentPart.text(
text: 'What fruit is this?',
),
ChatCompletionMessageContentPart.image(
imageUrl: ChatCompletionMessageImageUrl(
url: 'https://upload.wikimedia.org/wikipedia/commons/9/92/95apple.jpeg',
),
),
],
),
),
],
),
);
print(res.choices.first.message.content);
// The fruit in the image is an apple.
copied to clipboard
JSON mode: (docs)
final res = await client.createChatCompletion(
request: CreateChatCompletionRequest(
model: ChatCompletionModel.model(
ChatCompletionModels.gpt41106Preview,
),
messages: [
ChatCompletionMessage.system(
content:
'You are a helpful assistant. That extracts names from text '
'and returns them in a JSON array.',
),
ChatCompletionMessage.user(
content: ChatCompletionUserMessageContent.string(
'John, Mary, and Peter.',
),
),
],
temperature: 0,
responseFormat: ChatCompletionResponseFormat(
type: ChatCompletionResponseFormatType.jsonObject,
),
),
);
// { "names": ["John", "Mary", "Peter"] }
copied to clipboard
Structured output: (docs)
final res = await client.createChatCompletion(
request: CreateChatCompletionRequest(
model: ChatCompletionModel.model(
ChatCompletionModels.gpt4oMini,
),
messages: [
ChatCompletionMessage.system(
content:
'You are a helpful assistant. That extracts names from text.',
),
ChatCompletionMessage.user(
content: ChatCompletionUserMessageContent.string(
'John, Mary, and Peter.',
),
),
],
temperature: 0,
responseFormat: ResponseFormat.jsonSchema(
jsonSchema: JsonSchemaObject(
name: 'Names',
description: 'A list of names',
strict: true,
schema: {
'type': 'object',
'properties': {
'names': {
'type': 'array',
'items': {
'type': 'string',
},
},
},
'additionalProperties': false,
'required': ['names'],
},
),
),
),
);
// {"names":["John","Mary","Peter"]}
copied to clipboard
Tools: (docs)
const function = FunctionObject(
name: 'get_current_weather',
description: 'Get the current weather in a given location',
parameters: {
'type': 'object',
'properties': {
'location': {
'type': 'string',
'description': 'The city and state, e.g. San Francisco, CA',
},
'unit': {
'type': 'string',
'description': 'The unit of temperature to return',
'enum': ['celsius', 'fahrenheit'],
},
},
'required': ['location'],
},
);
const tool = ChatCompletionTool(
type: ChatCompletionToolType.function,
function: function,
);
final res1 = await client.createChatCompletion(
request: CreateChatCompletionRequest(
model: const ChatCompletionModel.model(
ChatCompletionModels.gpt4oMini,
),
messages: [
ChatCompletionMessage.system(
content: 'You are a helpful assistant.',
),
ChatCompletionMessage.user(
content: ChatCompletionUserMessageContent.string(
'What’s the weather like in Boston right now?',
),
),
],
tools: [tool],
toolChoice: ChatCompletionToolChoiceOption.tool(
ChatCompletionNamedToolChoice(
type: ChatCompletionNamedToolChoiceType.function,
function: ChatCompletionFunctionCallOption(name: function.name),
),
),
),
);
final toolCall = res1.choices.first.message.toolCalls!.first;
final functionCall = toolCall.function;
final arguments = json.decode(functionCall.arguments) as Map<String, dynamic>;
final functionResult = getCurrentWeather(arguments['location'], arguments['unit']);
final res2 = await client.createChatCompletion(
request: CreateChatCompletionRequest(
model: ChatCompletionModel.modelId('gpt-4-turbo'),
messages: [
ChatCompletionMessage.system(
content: 'You are a helpful assistant.',
),
ChatCompletionMessage.user(
content: ChatCompletionUserMessageContent.string(
'What’s the weather like in Boston right now?',
),
),
ChatCompletionMessage.tool(
toolCallId: toolCall.id,
content: json.encode(functionResult),
),
],
tools: [tool],
),
);
final answer = res2.choices.first.message.content;
// The weather in Boston right now is sunny with a temperature of 22°C
copied to clipboard
Function calling: (deprecated in favor of tools)
const function = FunctionObject(
name: 'get_current_weather',
description: 'Get the current weather in a given location',
parameters: {
'type': 'object',
'properties': {
'location': {
'type': 'string',
'description': 'The city and state, e.g. San Francisco, CA',
},
'unit': {
'type': 'string',
'description': 'The unit of temperature to return',
'enum': ['celsius', 'fahrenheit'],
},
},
'required': ['location'],
},
);
final res1 = await client.createChatCompletion(
request: CreateChatCompletionRequest(
model: ChatCompletionModel.modelId('gpt-4o-mini'),
messages: [
ChatCompletionMessage.system(
content: 'You are a helpful assistant.',
),
ChatCompletionMessage.user(
content: ChatCompletionUserMessageContent.string(
'What’s the weather like in Boston right now?',
),
),
],
functions: [function],
),
);
final arguments = json.decode(
res1.choices.first.message.functionCall!.arguments,
) as Map<String, dynamic>;
final functionResult = getCurrentWeather(arguments['location'], arguments['unit']);
final res2 = await client.createChatCompletion(
request: CreateChatCompletionRequest(
model: ChatCompletionModel.modelId('gpt-4o-mini'),
messages: [
ChatCompletionMessage.system(
content: 'You are a helpful assistant.',
),
ChatCompletionMessage.user(
content: ChatCompletionUserMessageContent.string(
'What’s the weather like in Boston right now?',
),
),
ChatCompletionMessage.function(
name: function.name,
content: json.encode(functionResult),
),
],
functions: [function],
),
);
final answer = res2.choices.first.message.content;
// The weather in Boston right now is sunny with a temperature of 22°C
copied to clipboard
Completions (legacy) #
Given a prompt, the model will return one or more predicted completions, and can also return the probabilities of alternative tokens at each position.
Most developer should use our Chat Completions API to leverage our best and newest models.
Most models that support the legacy Completions endpoint will be shut off on January 4th, 2024.
Create completion:
final res = await client.createCompletion(
request: CreateCompletionRequest(
model: CompletionModel.modelId('gpt-3.5-turbo-instruct'),
prompt: CompletionPrompt.string('Say this is a test'),
maxTokens: 7,
temperature: 0,
),
);
print(res.choices.first.text);
// This is a test.
copied to clipboard
CompletionModel is a sealed class that offers two ways to specify the model:
CompletionModel.modelId('model-id'): the model ID as string (e.g. 'gpt-3.5-turbo-instruct' or your fine-tuned model ID).
CompletionModel.model(CompletionModels.gpt35TurboInstruct): a value from CompletionModels enum which lists all of the available models.
CompletionPrompt is a sealed class that offers four ways to specify the prompt:
CompletionPrompt.string('prompt'): the prompt as string.
CompletionPrompt.tokens([...]): the tokenized prompt.
CompletionPrompt.listString(['prompt']): batch of string prompts.
CompletionPrompt.listTokens([[...]]): batch of tokenized prompts.
Stream completion:
final stream = client.createCompletionStream(
request: CreateCompletionRequest(
model: 'gpt-3.5-turbo-instruct',
prompt: [
'Say this is a test',
],
maxTokens: 7,
temperature: 0,
),
);
await for (final res in stream) {
print(res.choices.first.text);
}
copied to clipboard
Embeddings #
Get a vector representation of a given input that can be easily consumed by machine learning models and algorithms.
Related guide: Embeddings
Create embedding:
final res = await client.createEmbedding(
request: CreateEmbeddingRequest(
model: EmbeddingModel.modelId('text-embedding-3-small'),
input: EmbeddingInput.string('The food was delicious and the waiter...'),
),
);
print(res.data.first.embeddingVector);
// [0.002253932, -0.009333183, 0.01574578, -0.007790351, -0.004711035, ...]
copied to clipboard
EmbeddingModel is a sealed class that offers two ways to specify the model:
EmbeddingModel.modelId('model-id'): the model ID as string.
EmbeddingModel.model(EmbeddingModels.textEmbedding3Small): a value from EmbeddingModels enum which lists all of the available models.
EmbeddingInput is a sealed class that offers four ways to specify the embedding input:
EmbeddingInput.string('input'): the input as string.
EmbeddingInput.tokens([...]): the tokenized input.
EmbeddingInput.listString(['input']): batch of string inputs.
EmbeddingInput.listTokens([[...]]): batch of tokenized inputs.
You can also request the embedding vector encoded as a base64 string:
final res = await client.createEmbedding(
request: CreateEmbeddingRequest(
model: EmbeddingModel.modelId('text-embedding-3-small'),
input: EmbeddingInput.string('The food was delicious and the waiter...'),
encodingFormat: EmbeddingEncodingFormat.base64,
),
);
print(res.data.first.embeddingVectorBase64);
// tLYTOzXqGLxL/YA8M0b/uwdfmrsdNXM8iJIfvEOOHL3IJeK7Ok3rv...
copied to clipboard
Fine-tuning #
Manage fine-tuning jobs to tailor a model to your specific training data.
Related guide: Fine-tune models
Create fine-tuning job:
const request = CreateFineTuningJobRequest(
model: FineTuningModel.modelId('gpt-4o-mini'),
trainingFile: 'file-abc123',
validationFile: 'file-abc123',
hyperparameters: FineTuningJobHyperparameters(
nEpochs: FineTuningNEpochs.mode(FineTuningNEpochsOptions.auto),
),
integrations: [],
seed: 999,
);
final res = await client.createFineTuningJob(request: request);
copied to clipboard
List fine-tuning jobs:
final res = await client.listPaginatedFineTuningJobs();
copied to clipboard
Retrieve fine-tuning job:
final res = await client.retrieveFineTuningJob(
fineTuningJobId: 'ft-AF1WoRqd3aJAHsqc9NY7iL8F',
);
copied to clipboard
List fine-tuning job checkpoints:
final res = await client.listFineTuningJobCheckpoints(
fineTuningJobId: 'ft-AF1WoRqd3a
);
copied to clipboard
Cancel fine-tuning job:
final res = await client.cancelFineTuningJob(
fineTuningJobId: 'ft-AF1WoRqd3aJAHsqc9NY7iL8F',
);
copied to clipboard
List fine-tuning events:
final res = await client.listFineTuningEvents(
fineTuningJobId: 'ft-AF1WoRqd3aJAHsqc9NY7iL8F',
);
copied to clipboard
Batch #
Create large batches of API requests to run asynchronously
Create batch:
const request = CreateBatchRequest(
inputFileId: 'file-abc123',
endpoint: BatchEndpoint.v1ChatCompletions,
completionWindow: BatchCompletionWindow.v24h,
);
final res = await client.createBatch(request: request);
copied to clipboard
Retrieve batch:
final res = await client.retrieveBatch(
batchId: 'batch_abc123',
);
copied to clipboard
Cancel batch:
final res = await client.cancelBatch(
batchId: 'batch_abc123',
);
copied to clipboard
List batches:
final res = await client.listBatches();
copied to clipboard
Images #
Given a prompt and/or an input image, the model will generate a new image.
Related guide: Image generation
Create image:
final res = await client.createImage(
request: CreateImageRequest(
model: CreateImageRequestModel.model(ImageModels.dallE3),
prompt: 'A cute baby sea otter',
quality: ImageQuality.hd,
size: ImageSize.v1024x1792,
style: ImageStyle.natural,
),
);
print(res.data.first.url);
// https://oaidalleapiprodscus.blob.core.windows.net/private/...
copied to clipboard
Models #
List and describe the various models available in the API. You can refer to the Models documentation to understand what models are available and the differences between them.
List models:
final res = await client.listModels();
print(res.data.first.id);
// text-search-babbage-doc-001
copied to clipboard
Retrieve model:
final res = await client.retrieveModel(model: 'gpt-4');
print(res.ownedBy);
// openai
copied to clipboard
Delete fine-tune model:
final res = await client.deleteModel(
model: 'ft:gpt-3.5-turbo-0613:langchain::7qTVM5AR',
);
print(res.deleted);
// true
copied to clipboard
Moderations #
Given an input text, outputs if the model classifies it as violating OpenAI's content policy.
Related guide: Moderations
Create moderation:
final res = await client.createModeration(
request: CreateModerationRequest(
model: ModerationModel.modelId('text-moderation-latest'),
input: ModerationInput.string('I want to kill them.'),
),
);
print(res.results.first.categories.violence);
// true
print(res.results.first.categoryScores.violence);
// 0.9925811290740967
copied to clipboard
ModerationModel is a sealed class that offers two ways to specify the model:
ModerationModel.modelId('model-id'): the model ID as string.
ModerationModel.model(ModerationModels.textModerationLatest): a value from ModerationModels enum which lists all of the available models.
ModerationInput is a sealed class that offers four ways to specify the embedding input:
ModerationInput.string('input'): the input as string.
ModerationInput.listString(['input']): batch of string inputs.
Assistants (beta) #
Build assistants that can call models and use tools to perform tasks.
Get started with the Assistants API
Create assistant:
final res = await client.createAssistant(
request: CreateAssistantRequest(
model: CreateAssistantRequestModel.string('gpt-4'),
name: 'Math Tutor',
description: 'Help students with math homework',
instructions: 'You are a personal math tutor. Write and run code to answer math questions.',
tools: [AssistantTools.codeInterpreter()],
),
);
copied to clipboard
List assistants:
final res = await client.listAssistants();
copied to clipboard
Retrieve assistant:
final res = await client.getAssistant(assistantId: assistantId);
copied to clipboard
Modify assistant:
final res = await client.modifyAssistant(
assistantId: assistantId,
request: ModifyAssistantRequest(name: 'New name'),
);
copied to clipboard
Delete assistant:
final res = await client.deleteAssistant(assistantId: assistantId);
copied to clipboard
Threads (beta) #
Create threads that assistants can interact with.
Related guide: Assistants
Create thread:
final res = await client.createThread(
request: CreateThreadRequest(),
);
copied to clipboard
Retrieve thread:
final res = await client.getThread(threadId: threadId);
copied to clipboard
Modify thread:
final res = await client.modifyThread(
threadId: threadId,
request: ModifyThreadRequest(metadata: {'new': 'metadata'}),
);
copied to clipboard
Delete thread:
final res = await client.deleteThread(threadId: threadId);
copied to clipboard
Messages (beta) #
Create messages within threads.
Related guide: Assistants
Create message:
final res = await client.createThreadMessage(
threadId: threadId,
request: CreateMessageRequest(
role: CreateMessageRequestRole.user,
content: CreateMessageRequestContent.text(
'I need to solve the equation `3x + 11 = 14`. Can you help me?',
),
),
);
copied to clipboard
If you need to send multi-modal content, you can use the CreateMessageRequestContent.parts([...]) method:
final res = await client.createThreadMessage(
threadId: threadId,
request: CreateMessageRequest(
role: CreateMessageRequestRole.user,
content: CreateMessageRequestContent.parts([
MessageContent.text(
text: MessageContentText(
value: 'Some text...',
),
),
MessageContent.imageFile(
imageFile: MessageContentImageFile(fileId: 'file-abc123'),
),
MessageContent.imageUrl(
imageUrl: MessageContentImageUrl(
url: 'https://example.com/image.jpg',
),
),
]),
),
);
copied to clipboard
List messages:
final res = await client.listThreadMessages(threadId: threadId);
copied to clipboard
Retrieve message:
final res = await client.getThreadMessage(threadId: threadId, messageId: messageId);
copied to clipboard
Modify message:
final res = await client.modifyThreadMessage(
threadId: threadId,
messageId: messageId,
request: ModifyMessageRequest(metadata: {'new': 'metadata'}),
);
copied to clipboard
Delete message:
final res await deleteThreadMessage(threadId: threadId, messageId: messageId);
copied to clipboard
Runs (beta) #
Represents an execution run on a thread.
Related guide: Assistants
Create run:
final res = await client.createThreadRun(
threadId: threadId,
request: CreateRunRequest(
assistantId: assistantId,
instructions: 'Please address the user as Jane Doe. The user has a premium account.',
),
);
copied to clipboard
Create run: (streaming)
final stream = client.createThreadRunStream(
threadId: threadId,
request: CreateRunRequest(
assistantId: assistantId,
instructions: 'Please address the user as Jane Doe. The user has a premium account.',
),
);
copied to clipboard
Create thread and run:
final res = await client.createThreadAndRun(
request: CreateThreadAndRunRequest(
assistantId: assistantId,
instructions: 'Please address the user as Jane Doe. The user has a premium account.',
),
);
copied to clipboard
Create thread and run: (streaming)
final res = await client.createThreadAndRunStream(
request: CreateThreadAndRunRequest(
assistantId: assistantId,
instructions: 'Please address the user as Jane Doe. The user has a premium account.',
),
);
copied to clipboard
List runs:
final res = await client.listThreadRuns(threadId: threadId);
copied to clipboard
List run steps:
final res = await client.listThreadRunSteps(threadId: threadId, runId: runId);
copied to clipboard
Retrieve run:
final res = await client.getThreadRun(threadId: threadId, runId: runId);
copied to clipboard
Retrieve run step:
final res = await client.getThreadRunStep(threadId: threadId, runId: runId, stepId: stepId);
copied to clipboard
Modify run:
final res = await client.modifyThreadRun(
threadId: threadId,
runId: runId,
request: ModifyRunRequest(metadata: {'new': 'metadata'}),
);
copied to clipboard
Submit tool outputs to run:
final res = await client.submitThreadToolOutputsToRun(
threadId: threadId,
runId: runId,
request: SubmitToolOutputsRunRequest(
toolOutputs: [
RunSubmitToolOutput(
toolCallId: 'call_abc123',
output: '28C',
),
],
),
);
copied to clipboard
Submit tool outputs to run (streaming):
final res = await client.submitThreadToolOutputsToRunStream(
threadId: threadId,
runId: runId,
request: SubmitToolOutputsRunRequest(
toolOutputs: [
RunSubmitToolOutput(
toolCallId: 'call_abc123',
output: '28C',
),
],
),
);
copied to clipboard
Cancel a run:
final res = await client.cancelThreadRun(threadId: threadId, runId: runId);
copied to clipboard
Vector Stores (beta) #
Vector stores are used to store files for use by the file_search tool.
Related guide: File Search
Create vector store:
final res = await client.createVectorStore(
request: CreateVectorStoreRequest(
name: 'Support FAQ',
),
);
copied to clipboard
List vector stores:
final res = await client.listVectorStores();
copied to clipboard
Retrieve vector store:
final res = await client.getVectorStore(vectorStoreId: vectorStoreId);
copied to clipboard
Modify vector store:
final res = await client.modifyVectorStore(
vectorStoreId: vectorStoreId,
request: UpdateVectorStoreRequest(
name: 'New name',
),
);
copied to clipboard
Delete vector store:
final res = await client.deleteVectorStore(vectorStoreId: vectorStoreId);
copied to clipboard
Vector Store Files (beta) #
Vector store files represent files inside a vector store.
Related guide: File Search
Create vector store file:
final res = await client.createVectorStoreFile(
vectorStoreId: vectorStoreId,
request: CreateVectorStoreFileRequest(
fileId: 'file-abc123',
),
);
copied to clipboard
List vector store files:
final res = await client.listVectorStoreFiles(vectorStoreId: vectorStoreId);
copied to clipboard
Retrieve vector store file:
final res = await client.getVectorStoreFile(
vectorStoreId: vectorStoreId,
fileId: fileId,
);
copied to clipboard
Delete vector store file:
final res = await client.deleteVectorStoreFile(
vectorStoreId: vectorStoreId,
fileId: fileId,
);
copied to clipboard
Vector Store File Batches (beta) #
Vector store file batches represent operations to add multiple files to a vector store.
Related guide: File Search
Create vector store file batch:
final res = await client.createVectorStoreFileBatch(
vectorStoreId: vectorStoreId,
request: CreateVectorStoreFileBatchRequest(
fileIds: ['file-abc123', 'file-abc456'],
),
);
copied to clipboard
Retrieve vector store file batch:
final res = await client.getVectorStoreFileBatch(
vectorStoreId: vectorStoreId,
batchId: batchId,
);
copied to clipboard
Cancel vector store file batch:
final res = await client.cancelVectorStoreFileBatch(
vectorStoreId: vectorStoreId,
batchId: batchId,
);
copied to clipboard
List vector store files in a batch:
final res = await client.listFilesInVectorStoreBatch(
vectorStoreId: vectorStoreId,
batchId: batchId,
);
copied to clipboard
Advance Usage #
Azure OpenAI Service #
OpenAI's models are also available as an Azure service.
Although the Azure OpenAI API is similar to the official OpenAI API, there are subtle differences between them. This client is intended to be used with the official OpenAI API, but most of the functionality should work with the Azure OpenAI API as well.
If you want to use this client with the Azure OpenAI API (at your own risk), you can do so by instantiating the client as follows:
final client = OpenAIClient(
baseUrl: 'https://YOUR_RESOURCE_NAME.openai.azure.com/openai/deployments/YOUR_DEPLOYMENT_NAME',
headers: { 'api-key': 'YOUR_API_KEY' },
queryParams: { 'api-version': 'API_VERSION' },
);
copied to clipboard
YOUR_RESOURCE_NAME: This value can be found in the Keys & Endpoint section when examining your resource from the Azure portal.
YOUR_DEPLOYMENT_NAME: This value will correspond to the custom name you chose for your deployment when you deployed a model. This value can be found under Resource Management > Deployments in the Azure portal.
YOUR_API_KEY: This value can be found in the Keys & Endpoint section when examining your resource from the Azure portal.
API_VERSION: The Azure OpenAI API version to use (e.g. 2023-05-15). Try to use the latest version available, it will probably be the closest to the official OpenAI API.
OpenAI-compatible APIs #
This client can be used to consume APIs that are compatible with the OpenAI API spec.
TogetherAI:
final client = OpenAIClient(
baseUrl: 'https://api.together.xyz/v1',
headers: { 'api-key': 'YOUR_TOGETHER_AI_API_KEY' },
);
copied to clipboard
Anyscale:
final client = OpenAIClient(
baseUrl: 'https://api.endpoints.anyscale.com/v1',
headers: { 'api-key': 'YOUR_ANYSCALE_API_KEY' },
);
copied to clipboard
OpenRouter:
final client = OpenAIClient(
baseUrl: 'https://openrouter.ai/api/v1',
headers: { 'api-key': 'YOUR_OPEN_ROUTER_API_KEY' },
);
copied to clipboard
Etc.
Default HTTP client #
By default, the client uses the following implementation of http.Client:
Non-web: IOClient
Web: FetchClient (to support streaming on web)
Custom HTTP client #
You can always provide your own implementation of http.Client for further customization:
final client = OpenAIClient(
apiKey: 'OPENAI_API_KEY',
client: MyHttpClient(),
);
copied to clipboard
Using a proxy #
HTTP proxy
You can use your own HTTP proxy by overriding the baseUrl and providing your required headers:
final client = OpenAIClient(
baseUrl: 'https://my-proxy.com',
headers: {
'x-my-proxy-header': 'value',
},
);
copied to clipboard
If you need further customization, you can always provide your own http.Client.
SOCKS5 proxy
To use a SOCKS5 proxy, you can use the socks5_proxy package:
final baseHttpClient = HttpClient();
SocksTCPClient.assignToHttpClient(baseHttpClient, [
ProxySettings(InternetAddress.loopbackIPv4, 1080),
]);
final httpClient = IOClient(baseClient);
final client = OpenAIClient(
client: httpClient,
);
copied to clipboard
Acknowledgements #
The generation of this client was made possible by the openapi_spec package.
License #
OpenAI Dart Client is licensed under the MIT License.
For personal and professional use. You cannot resell or redistribute these repositories in their original state.
There are no reviews.