diff --git a/.claude/skills/tryagi-openai/commands.md b/.claude/skills/tryagi-openai/commands.md index 7a54c9376..b7d33d1e0 100644 --- a/.claude/skills/tryagi-openai/commands.md +++ b/.claude/skills/tryagi-openai/commands.md @@ -150,7 +150,7 @@ Files are used to upload documents that can be used with features like Assistant | `list-files` | `GET /files` | Returns a list of files. | | `retrieve-file` | `GET /files/{file_id}` | Returns information about a specific file. | | `retrieve-file-content` | `GET /files/{file_id}/content` | Returns the contents of the specified file. | -| `upload-file` | `POST /files` | Upload a file that can be used across various endpoints. Individual files can be up to 512 MB, and each project can store up to 2.5 TB of files in total. There is no organization-wide storage limit. Uploads to this endpoint are rate-limited to 2,000 files per minute per organization. - The Assistants API supports files up to 2 million tokens and of specific file types. See the [Assistants Tools guide](/docs/assistants/tools) for details. - The Fine-tuning API only supports `.jsonl` files. The input also has certain required formats for fine-tuning [chat](/docs/api-reference/fine-tuning/chat-input) or [completions](/docs/api-reference/fine-tuning/completions-input) models. - The Batch API only supports `.jsonl` files up to 200 MB in size. The input also has a specific required [format](/docs/api-reference/batch/request-input). - For Retrieval or `file_search` ingestion, upload files here first. If you need to attach multiple uploaded files to the same vector store, use [`/vector_stores/{vector_store_id}/file_batches`](/docs/api-reference/vector-stores-file-batches/createBatch) instead of attaching them one by one. Please [contact us](https://help.openai.com/) if you need to increase these storage limits. | +| `upload-file` | `POST /files` | Upload a file that can be used across various endpoints. Individual files can be up to 512 MB, and each project can store up to 2.5 TB of files in total. There is no organization-wide storage limit. Uploads to this endpoint are rate-limited to 1,000 requests per minute per authenticated user. - The Assistants API supports files up to 2 million tokens and of specific file types. See the [Assistants Tools guide](/docs/assistants/tools) for details. - The Fine-tuning API only supports `.jsonl` files. The input also has certain required formats for fine-tuning [chat](/docs/api-reference/fine-tuning/chat-input) or [completions](/docs/api-reference/fine-tuning/completions-input) models. - The Batch API only supports `.jsonl` files up to 200 MB in size. The input also has a specific required [format](/docs/api-reference/batch/request-input). - For Retrieval or `file_search` ingestion, upload files here first. If you need to attach multiple uploaded files to the same vector store, use [`/vector_stores/{vector_store_id}/file_batches`](/docs/api-reference/vector-stores-file-batches/createBatch) instead of attaching them one by one. Vector store attachment has separate limits from file upload, including 2,000 attached files per minute per organization. Please [contact us](https://help.openai.com/) if you need to increase these storage limits. | ## `fine-tuning` diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.FilesClient.UploadFile.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.FilesClient.UploadFile.g.cs index c07f06b8e..3e92fb875 100644 --- a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.FilesClient.UploadFile.g.cs +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.FilesClient.UploadFile.g.cs @@ -45,7 +45,8 @@ partial void ProcessUploadFileResponseContent( /// Upload a file that can be used across various endpoints. Individual files
/// can be up to 512 MB, and each project can store up to 2.5 TB of files in
/// total. There is no organization-wide storage limit. Uploads to this
- /// endpoint are rate-limited to 2,000 files per minute per organization.
+ /// endpoint are rate-limited to 1,000 requests per minute per authenticated
+ /// user.
/// - The Assistants API supports files up to 2 million tokens and of specific
/// file types. See the [Assistants Tools guide](/docs/assistants/tools) for
/// details.
@@ -59,7 +60,9 @@ partial void ProcessUploadFileResponseContent( /// - For Retrieval or `file_search` ingestion, upload files here first. If
/// you need to attach multiple uploaded files to the same vector store, use
/// [`/vector_stores/{vector_store_id}/file_batches`](/docs/api-reference/vector-stores-file-batches/createBatch)
- /// instead of attaching them one by one.
+ /// instead of attaching them one by one. Vector store attachment has separate
+ /// limits from file upload, including 2,000 attached files per minute per
+ /// organization.
/// Please [contact us](https://help.openai.com/) if you need to increase these
/// storage limits. /// @@ -424,7 +427,8 @@ partial void ProcessUploadFileResponseContent( /// Upload a file that can be used across various endpoints. Individual files
/// can be up to 512 MB, and each project can store up to 2.5 TB of files in
/// total. There is no organization-wide storage limit. Uploads to this
- /// endpoint are rate-limited to 2,000 files per minute per organization.
+ /// endpoint are rate-limited to 1,000 requests per minute per authenticated
+ /// user.
/// - The Assistants API supports files up to 2 million tokens and of specific
/// file types. See the [Assistants Tools guide](/docs/assistants/tools) for
/// details.
@@ -438,7 +442,9 @@ partial void ProcessUploadFileResponseContent( /// - For Retrieval or `file_search` ingestion, upload files here first. If
/// you need to attach multiple uploaded files to the same vector store, use
/// [`/vector_stores/{vector_store_id}/file_batches`](/docs/api-reference/vector-stores-file-batches/createBatch)
- /// instead of attaching them one by one.
+ /// instead of attaching them one by one. Vector store attachment has separate
+ /// limits from file upload, including 2,000 attached files per minute per
+ /// organization.
/// Please [contact us](https://help.openai.com/) if you need to increase these
/// storage limits. /// diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.IFilesClient.UploadFile.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.IFilesClient.UploadFile.g.cs index 425e59940..5c66530e7 100644 --- a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.IFilesClient.UploadFile.g.cs +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.IFilesClient.UploadFile.g.cs @@ -8,7 +8,8 @@ public partial interface IFilesClient /// Upload a file that can be used across various endpoints. Individual files
/// can be up to 512 MB, and each project can store up to 2.5 TB of files in
/// total. There is no organization-wide storage limit. Uploads to this
- /// endpoint are rate-limited to 2,000 files per minute per organization.
+ /// endpoint are rate-limited to 1,000 requests per minute per authenticated
+ /// user.
/// - The Assistants API supports files up to 2 million tokens and of specific
/// file types. See the [Assistants Tools guide](/docs/assistants/tools) for
/// details.
@@ -22,7 +23,9 @@ public partial interface IFilesClient /// - For Retrieval or `file_search` ingestion, upload files here first. If
/// you need to attach multiple uploaded files to the same vector store, use
/// [`/vector_stores/{vector_store_id}/file_batches`](/docs/api-reference/vector-stores-file-batches/createBatch)
- /// instead of attaching them one by one.
+ /// instead of attaching them one by one. Vector store attachment has separate
+ /// limits from file upload, including 2,000 attached files per minute per
+ /// organization.
/// Please [contact us](https://help.openai.com/) if you need to increase these
/// storage limits. /// @@ -39,7 +42,8 @@ public partial interface IFilesClient /// Upload a file that can be used across various endpoints. Individual files
/// can be up to 512 MB, and each project can store up to 2.5 TB of files in
/// total. There is no organization-wide storage limit. Uploads to this
- /// endpoint are rate-limited to 2,000 files per minute per organization.
+ /// endpoint are rate-limited to 1,000 requests per minute per authenticated
+ /// user.
/// - The Assistants API supports files up to 2 million tokens and of specific
/// file types. See the [Assistants Tools guide](/docs/assistants/tools) for
/// details.
@@ -53,7 +57,9 @@ public partial interface IFilesClient /// - For Retrieval or `file_search` ingestion, upload files here first. If
/// you need to attach multiple uploaded files to the same vector store, use
/// [`/vector_stores/{vector_store_id}/file_batches`](/docs/api-reference/vector-stores-file-batches/createBatch)
- /// instead of attaching them one by one.
+ /// instead of attaching them one by one. Vector store attachment has separate
+ /// limits from file upload, including 2,000 attached files per minute per
+ /// organization.
/// Please [contact us](https://help.openai.com/) if you need to increase these
/// storage limits. /// diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonConverters.LocalShellCallOutputStatusEnum.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonConverters.FunctionShellCallOutputStatusEnum.g.cs similarity index 64% rename from src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonConverters.LocalShellCallOutputStatusEnum.g.cs rename to src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonConverters.FunctionShellCallOutputStatusEnum.g.cs index d709158df..e4afc70ca 100644 --- a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonConverters.LocalShellCallOutputStatusEnum.g.cs +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonConverters.FunctionShellCallOutputStatusEnum.g.cs @@ -3,10 +3,10 @@ namespace tryAGI.OpenAI.JsonConverters { /// - public sealed class LocalShellCallOutputStatusEnumJsonConverter : global::System.Text.Json.Serialization.JsonConverter + public sealed class FunctionShellCallOutputStatusEnumJsonConverter : global::System.Text.Json.Serialization.JsonConverter { /// - public override global::tryAGI.OpenAI.LocalShellCallOutputStatusEnum Read( + public override global::tryAGI.OpenAI.FunctionShellCallOutputStatusEnum Read( ref global::System.Text.Json.Utf8JsonReader reader, global::System.Type typeToConvert, global::System.Text.Json.JsonSerializerOptions options) @@ -18,7 +18,7 @@ public sealed class LocalShellCallOutputStatusEnumJsonConverter : global::System var stringValue = reader.GetString(); if (stringValue != null) { - return global::tryAGI.OpenAI.LocalShellCallOutputStatusEnumExtensions.ToEnum(stringValue) ?? default; + return global::tryAGI.OpenAI.FunctionShellCallOutputStatusEnumExtensions.ToEnum(stringValue) ?? default; } break; @@ -26,11 +26,11 @@ public sealed class LocalShellCallOutputStatusEnumJsonConverter : global::System case global::System.Text.Json.JsonTokenType.Number: { var numValue = reader.GetInt32(); - return (global::tryAGI.OpenAI.LocalShellCallOutputStatusEnum)numValue; + return (global::tryAGI.OpenAI.FunctionShellCallOutputStatusEnum)numValue; } case global::System.Text.Json.JsonTokenType.Null: { - return default(global::tryAGI.OpenAI.LocalShellCallOutputStatusEnum); + return default(global::tryAGI.OpenAI.FunctionShellCallOutputStatusEnum); } default: throw new global::System.ArgumentOutOfRangeException(nameof(reader)); @@ -42,12 +42,12 @@ public sealed class LocalShellCallOutputStatusEnumJsonConverter : global::System /// public override void Write( global::System.Text.Json.Utf8JsonWriter writer, - global::tryAGI.OpenAI.LocalShellCallOutputStatusEnum value, + global::tryAGI.OpenAI.FunctionShellCallOutputStatusEnum value, global::System.Text.Json.JsonSerializerOptions options) { writer = writer ?? throw new global::System.ArgumentNullException(nameof(writer)); - writer.WriteStringValue(global::tryAGI.OpenAI.LocalShellCallOutputStatusEnumExtensions.ToValueString(value)); + writer.WriteStringValue(global::tryAGI.OpenAI.FunctionShellCallOutputStatusEnumExtensions.ToValueString(value)); } } } diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonConverters.LocalShellCallOutputStatusEnumNullable.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonConverters.FunctionShellCallOutputStatusEnumNullable.g.cs similarity index 65% rename from src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonConverters.LocalShellCallOutputStatusEnumNullable.g.cs rename to src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonConverters.FunctionShellCallOutputStatusEnumNullable.g.cs index 117851de9..ce6264f3b 100644 --- a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonConverters.LocalShellCallOutputStatusEnumNullable.g.cs +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonConverters.FunctionShellCallOutputStatusEnumNullable.g.cs @@ -3,10 +3,10 @@ namespace tryAGI.OpenAI.JsonConverters { /// - public sealed class LocalShellCallOutputStatusEnumNullableJsonConverter : global::System.Text.Json.Serialization.JsonConverter + public sealed class FunctionShellCallOutputStatusEnumNullableJsonConverter : global::System.Text.Json.Serialization.JsonConverter { /// - public override global::tryAGI.OpenAI.LocalShellCallOutputStatusEnum? Read( + public override global::tryAGI.OpenAI.FunctionShellCallOutputStatusEnum? Read( ref global::System.Text.Json.Utf8JsonReader reader, global::System.Type typeToConvert, global::System.Text.Json.JsonSerializerOptions options) @@ -18,7 +18,7 @@ public sealed class LocalShellCallOutputStatusEnumNullableJsonConverter : global var stringValue = reader.GetString(); if (stringValue != null) { - return global::tryAGI.OpenAI.LocalShellCallOutputStatusEnumExtensions.ToEnum(stringValue); + return global::tryAGI.OpenAI.FunctionShellCallOutputStatusEnumExtensions.ToEnum(stringValue); } break; @@ -26,11 +26,11 @@ public sealed class LocalShellCallOutputStatusEnumNullableJsonConverter : global case global::System.Text.Json.JsonTokenType.Number: { var numValue = reader.GetInt32(); - return (global::tryAGI.OpenAI.LocalShellCallOutputStatusEnum)numValue; + return (global::tryAGI.OpenAI.FunctionShellCallOutputStatusEnum)numValue; } case global::System.Text.Json.JsonTokenType.Null: { - return default(global::tryAGI.OpenAI.LocalShellCallOutputStatusEnum?); + return default(global::tryAGI.OpenAI.FunctionShellCallOutputStatusEnum?); } default: throw new global::System.ArgumentOutOfRangeException(nameof(reader)); @@ -42,7 +42,7 @@ public sealed class LocalShellCallOutputStatusEnumNullableJsonConverter : global /// public override void Write( global::System.Text.Json.Utf8JsonWriter writer, - global::tryAGI.OpenAI.LocalShellCallOutputStatusEnum? value, + global::tryAGI.OpenAI.FunctionShellCallOutputStatusEnum? value, global::System.Text.Json.JsonSerializerOptions options) { writer = writer ?? throw new global::System.ArgumentNullException(nameof(writer)); @@ -53,7 +53,7 @@ public override void Write( } else { - writer.WriteStringValue(global::tryAGI.OpenAI.LocalShellCallOutputStatusEnumExtensions.ToValueString(value.Value)); + writer.WriteStringValue(global::tryAGI.OpenAI.FunctionShellCallOutputStatusEnumExtensions.ToValueString(value.Value)); } } } diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonConverters.LocalShellCallStatus.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonConverters.FunctionShellCallStatus.g.cs similarity index 66% rename from src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonConverters.LocalShellCallStatus.g.cs rename to src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonConverters.FunctionShellCallStatus.g.cs index 4b4f81be3..db36db379 100644 --- a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonConverters.LocalShellCallStatus.g.cs +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonConverters.FunctionShellCallStatus.g.cs @@ -3,10 +3,10 @@ namespace tryAGI.OpenAI.JsonConverters { /// - public sealed class LocalShellCallStatusJsonConverter : global::System.Text.Json.Serialization.JsonConverter + public sealed class FunctionShellCallStatusJsonConverter : global::System.Text.Json.Serialization.JsonConverter { /// - public override global::tryAGI.OpenAI.LocalShellCallStatus Read( + public override global::tryAGI.OpenAI.FunctionShellCallStatus Read( ref global::System.Text.Json.Utf8JsonReader reader, global::System.Type typeToConvert, global::System.Text.Json.JsonSerializerOptions options) @@ -18,7 +18,7 @@ public sealed class LocalShellCallStatusJsonConverter : global::System.Text.Json var stringValue = reader.GetString(); if (stringValue != null) { - return global::tryAGI.OpenAI.LocalShellCallStatusExtensions.ToEnum(stringValue) ?? default; + return global::tryAGI.OpenAI.FunctionShellCallStatusExtensions.ToEnum(stringValue) ?? default; } break; @@ -26,11 +26,11 @@ public sealed class LocalShellCallStatusJsonConverter : global::System.Text.Json case global::System.Text.Json.JsonTokenType.Number: { var numValue = reader.GetInt32(); - return (global::tryAGI.OpenAI.LocalShellCallStatus)numValue; + return (global::tryAGI.OpenAI.FunctionShellCallStatus)numValue; } case global::System.Text.Json.JsonTokenType.Null: { - return default(global::tryAGI.OpenAI.LocalShellCallStatus); + return default(global::tryAGI.OpenAI.FunctionShellCallStatus); } default: throw new global::System.ArgumentOutOfRangeException(nameof(reader)); @@ -42,12 +42,12 @@ public sealed class LocalShellCallStatusJsonConverter : global::System.Text.Json /// public override void Write( global::System.Text.Json.Utf8JsonWriter writer, - global::tryAGI.OpenAI.LocalShellCallStatus value, + global::tryAGI.OpenAI.FunctionShellCallStatus value, global::System.Text.Json.JsonSerializerOptions options) { writer = writer ?? throw new global::System.ArgumentNullException(nameof(writer)); - writer.WriteStringValue(global::tryAGI.OpenAI.LocalShellCallStatusExtensions.ToValueString(value)); + writer.WriteStringValue(global::tryAGI.OpenAI.FunctionShellCallStatusExtensions.ToValueString(value)); } } } diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonConverters.LocalShellCallStatusNullable.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonConverters.FunctionShellCallStatusNullable.g.cs similarity index 68% rename from src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonConverters.LocalShellCallStatusNullable.g.cs rename to src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonConverters.FunctionShellCallStatusNullable.g.cs index f76a781ee..7ddbd0403 100644 --- a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonConverters.LocalShellCallStatusNullable.g.cs +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonConverters.FunctionShellCallStatusNullable.g.cs @@ -3,10 +3,10 @@ namespace tryAGI.OpenAI.JsonConverters { /// - public sealed class LocalShellCallStatusNullableJsonConverter : global::System.Text.Json.Serialization.JsonConverter + public sealed class FunctionShellCallStatusNullableJsonConverter : global::System.Text.Json.Serialization.JsonConverter { /// - public override global::tryAGI.OpenAI.LocalShellCallStatus? Read( + public override global::tryAGI.OpenAI.FunctionShellCallStatus? Read( ref global::System.Text.Json.Utf8JsonReader reader, global::System.Type typeToConvert, global::System.Text.Json.JsonSerializerOptions options) @@ -18,7 +18,7 @@ public sealed class LocalShellCallStatusNullableJsonConverter : global::System.T var stringValue = reader.GetString(); if (stringValue != null) { - return global::tryAGI.OpenAI.LocalShellCallStatusExtensions.ToEnum(stringValue); + return global::tryAGI.OpenAI.FunctionShellCallStatusExtensions.ToEnum(stringValue); } break; @@ -26,11 +26,11 @@ public sealed class LocalShellCallStatusNullableJsonConverter : global::System.T case global::System.Text.Json.JsonTokenType.Number: { var numValue = reader.GetInt32(); - return (global::tryAGI.OpenAI.LocalShellCallStatus)numValue; + return (global::tryAGI.OpenAI.FunctionShellCallStatus)numValue; } case global::System.Text.Json.JsonTokenType.Null: { - return default(global::tryAGI.OpenAI.LocalShellCallStatus?); + return default(global::tryAGI.OpenAI.FunctionShellCallStatus?); } default: throw new global::System.ArgumentOutOfRangeException(nameof(reader)); @@ -42,7 +42,7 @@ public sealed class LocalShellCallStatusNullableJsonConverter : global::System.T /// public override void Write( global::System.Text.Json.Utf8JsonWriter writer, - global::tryAGI.OpenAI.LocalShellCallStatus? value, + global::tryAGI.OpenAI.FunctionShellCallStatus? value, global::System.Text.Json.JsonSerializerOptions options) { writer = writer ?? throw new global::System.ArgumentNullException(nameof(writer)); @@ -53,7 +53,7 @@ public override void Write( } else { - writer.WriteStringValue(global::tryAGI.OpenAI.LocalShellCallStatusExtensions.ToValueString(value.Value)); + writer.WriteStringValue(global::tryAGI.OpenAI.FunctionShellCallStatusExtensions.ToValueString(value.Value)); } } } diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonSerializerContext.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonSerializerContext.g.cs index 2969214dd..256f792b8 100644 --- a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonSerializerContext.g.cs +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonSerializerContext.g.cs @@ -3561,9 +3561,9 @@ namespace tryAGI.OpenAI typeof(global::tryAGI.OpenAI.JsonConverters.LocalShellExecActionTypeNullableJsonConverter), - typeof(global::tryAGI.OpenAI.JsonConverters.LocalShellCallStatusJsonConverter), + typeof(global::tryAGI.OpenAI.JsonConverters.FunctionShellCallStatusJsonConverter), - typeof(global::tryAGI.OpenAI.JsonConverters.LocalShellCallStatusNullableJsonConverter), + typeof(global::tryAGI.OpenAI.JsonConverters.FunctionShellCallStatusNullableJsonConverter), typeof(global::tryAGI.OpenAI.JsonConverters.LocalEnvironmentResourceTypeJsonConverter), @@ -3577,9 +3577,9 @@ namespace tryAGI.OpenAI typeof(global::tryAGI.OpenAI.JsonConverters.FunctionShellCallTypeNullableJsonConverter), - typeof(global::tryAGI.OpenAI.JsonConverters.LocalShellCallOutputStatusEnumJsonConverter), + typeof(global::tryAGI.OpenAI.JsonConverters.FunctionShellCallOutputStatusEnumJsonConverter), - typeof(global::tryAGI.OpenAI.JsonConverters.LocalShellCallOutputStatusEnumNullableJsonConverter), + typeof(global::tryAGI.OpenAI.JsonConverters.FunctionShellCallOutputStatusEnumNullableJsonConverter), typeof(global::tryAGI.OpenAI.JsonConverters.FunctionShellCallOutputTimeoutOutcomeTypeJsonConverter), @@ -7133,7 +7133,7 @@ namespace tryAGI.OpenAI [global::System.Text.Json.Serialization.JsonSerializable(typeof(global::tryAGI.OpenAI.CodeInterpreterOutputImageType), TypeInfoPropertyName = "CodeInterpreterOutputImageType2")] [global::System.Text.Json.Serialization.JsonSerializable(typeof(global::tryAGI.OpenAI.LocalShellExecActionType), TypeInfoPropertyName = "LocalShellExecActionType2")] [global::System.Text.Json.Serialization.JsonSerializable(typeof(global::tryAGI.OpenAI.FunctionShellAction))] - [global::System.Text.Json.Serialization.JsonSerializable(typeof(global::tryAGI.OpenAI.LocalShellCallStatus), TypeInfoPropertyName = "LocalShellCallStatus2")] + [global::System.Text.Json.Serialization.JsonSerializable(typeof(global::tryAGI.OpenAI.FunctionShellCallStatus), TypeInfoPropertyName = "FunctionShellCallStatus2")] [global::System.Text.Json.Serialization.JsonSerializable(typeof(global::tryAGI.OpenAI.LocalEnvironmentResource))] [global::System.Text.Json.Serialization.JsonSerializable(typeof(global::tryAGI.OpenAI.LocalEnvironmentResourceType), TypeInfoPropertyName = "LocalEnvironmentResourceType2")] [global::System.Text.Json.Serialization.JsonSerializable(typeof(global::tryAGI.OpenAI.ContainerReferenceResource))] @@ -7141,7 +7141,7 @@ namespace tryAGI.OpenAI [global::System.Text.Json.Serialization.JsonSerializable(typeof(global::tryAGI.OpenAI.FunctionShellCallType), TypeInfoPropertyName = "FunctionShellCallType2")] [global::System.Text.Json.Serialization.JsonSerializable(typeof(global::tryAGI.OpenAI.EnvironmentVariant12))] [global::System.Text.Json.Serialization.JsonSerializable(typeof(global::tryAGI.OpenAI.FunctionShellCallEnvironmentVariant1Discriminator))] - [global::System.Text.Json.Serialization.JsonSerializable(typeof(global::tryAGI.OpenAI.LocalShellCallOutputStatusEnum), TypeInfoPropertyName = "LocalShellCallOutputStatusEnum2")] + [global::System.Text.Json.Serialization.JsonSerializable(typeof(global::tryAGI.OpenAI.FunctionShellCallOutputStatusEnum), TypeInfoPropertyName = "FunctionShellCallOutputStatusEnum2")] [global::System.Text.Json.Serialization.JsonSerializable(typeof(global::tryAGI.OpenAI.FunctionShellCallOutputTimeoutOutcome))] [global::System.Text.Json.Serialization.JsonSerializable(typeof(global::tryAGI.OpenAI.FunctionShellCallOutputTimeoutOutcomeType), TypeInfoPropertyName = "FunctionShellCallOutputTimeoutOutcomeType2")] [global::System.Text.Json.Serialization.JsonSerializable(typeof(global::tryAGI.OpenAI.FunctionShellCallOutputExitOutcome))] diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonSerializerContextTypes.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonSerializerContextTypes.g.cs index 0fb396903..baef1218c 100644 --- a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonSerializerContextTypes.g.cs +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonSerializerContextTypes.g.cs @@ -9600,7 +9600,7 @@ public sealed partial class JsonSerializerContextTypes /// /// /// - public global::tryAGI.OpenAI.LocalShellCallStatus? Type2393 { get; set; } + public global::tryAGI.OpenAI.FunctionShellCallStatus? Type2393 { get; set; } /// /// /// @@ -9632,7 +9632,7 @@ public sealed partial class JsonSerializerContextTypes /// /// /// - public global::tryAGI.OpenAI.LocalShellCallOutputStatusEnum? Type2401 { get; set; } + public global::tryAGI.OpenAI.FunctionShellCallOutputStatusEnum? Type2401 { get; set; } /// /// /// diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.FunctionShellCall.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.FunctionShellCall.g.cs index d1c2de6dd..b323c46c7 100644 --- a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.FunctionShellCall.g.cs +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.FunctionShellCall.g.cs @@ -42,9 +42,9 @@ public sealed partial class FunctionShellCall /// The status of the shell call. One of `in_progress`, `completed`, or `incomplete`. /// [global::System.Text.Json.Serialization.JsonPropertyName("status")] - [global::System.Text.Json.Serialization.JsonConverter(typeof(global::tryAGI.OpenAI.JsonConverters.LocalShellCallStatusJsonConverter))] + [global::System.Text.Json.Serialization.JsonConverter(typeof(global::tryAGI.OpenAI.JsonConverters.FunctionShellCallStatusJsonConverter))] [global::System.Text.Json.Serialization.JsonRequired] - public required global::tryAGI.OpenAI.LocalShellCallStatus Status { get; set; } + public required global::tryAGI.OpenAI.FunctionShellCallStatus Status { get; set; } /// /// @@ -94,7 +94,7 @@ public FunctionShellCall( string id, string callId, global::tryAGI.OpenAI.FunctionShellAction action, - global::tryAGI.OpenAI.LocalShellCallStatus status, + global::tryAGI.OpenAI.FunctionShellCallStatus status, global::tryAGI.OpenAI.EnvironmentVariant12? environment, string? createdBy, global::tryAGI.OpenAI.FunctionShellCallType type = global::tryAGI.OpenAI.FunctionShellCallType.ShellCall) diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.FunctionShellCallOutput.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.FunctionShellCallOutput.g.cs index 4fc65021d..3e8e5db16 100644 --- a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.FunctionShellCallOutput.g.cs +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.FunctionShellCallOutput.g.cs @@ -35,9 +35,9 @@ public sealed partial class FunctionShellCallOutput /// The status of the shell call output. One of `in_progress`, `completed`, or `incomplete`. /// [global::System.Text.Json.Serialization.JsonPropertyName("status")] - [global::System.Text.Json.Serialization.JsonConverter(typeof(global::tryAGI.OpenAI.JsonConverters.LocalShellCallOutputStatusEnumJsonConverter))] + [global::System.Text.Json.Serialization.JsonConverter(typeof(global::tryAGI.OpenAI.JsonConverters.FunctionShellCallOutputStatusEnumJsonConverter))] [global::System.Text.Json.Serialization.JsonRequired] - public required global::tryAGI.OpenAI.LocalShellCallOutputStatusEnum Status { get; set; } + public required global::tryAGI.OpenAI.FunctionShellCallOutputStatusEnum Status { get; set; } /// /// An array of shell call output contents @@ -93,7 +93,7 @@ public sealed partial class FunctionShellCallOutput public FunctionShellCallOutput( string id, string callId, - global::tryAGI.OpenAI.LocalShellCallOutputStatusEnum status, + global::tryAGI.OpenAI.FunctionShellCallOutputStatusEnum status, global::System.Collections.Generic.IList output, int? maxOutputLength, string? createdBy, diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.LocalShellCallOutputStatusEnum.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.FunctionShellCallOutputStatusEnum.g.cs similarity index 55% rename from src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.LocalShellCallOutputStatusEnum.g.cs rename to src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.FunctionShellCallOutputStatusEnum.g.cs index ce45eb233..5e2464abe 100644 --- a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.LocalShellCallOutputStatusEnum.g.cs +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.FunctionShellCallOutputStatusEnum.g.cs @@ -6,7 +6,7 @@ namespace tryAGI.OpenAI /// /// /// - public enum LocalShellCallOutputStatusEnum + public enum FunctionShellCallOutputStatusEnum { /// /// @@ -25,31 +25,31 @@ public enum LocalShellCallOutputStatusEnum /// /// Enum extensions to do fast conversions without the reflection. /// - public static class LocalShellCallOutputStatusEnumExtensions + public static class FunctionShellCallOutputStatusEnumExtensions { /// /// Converts an enum to a string. /// - public static string ToValueString(this LocalShellCallOutputStatusEnum value) + public static string ToValueString(this FunctionShellCallOutputStatusEnum value) { return value switch { - LocalShellCallOutputStatusEnum.Completed => "completed", - LocalShellCallOutputStatusEnum.InProgress => "in_progress", - LocalShellCallOutputStatusEnum.Incomplete => "incomplete", + FunctionShellCallOutputStatusEnum.Completed => "completed", + FunctionShellCallOutputStatusEnum.InProgress => "in_progress", + FunctionShellCallOutputStatusEnum.Incomplete => "incomplete", _ => throw new global::System.ArgumentOutOfRangeException(nameof(value), value, null), }; } /// /// Converts an string to a enum. /// - public static LocalShellCallOutputStatusEnum? ToEnum(string value) + public static FunctionShellCallOutputStatusEnum? ToEnum(string value) { return value switch { - "completed" => LocalShellCallOutputStatusEnum.Completed, - "in_progress" => LocalShellCallOutputStatusEnum.InProgress, - "incomplete" => LocalShellCallOutputStatusEnum.Incomplete, + "completed" => FunctionShellCallOutputStatusEnum.Completed, + "in_progress" => FunctionShellCallOutputStatusEnum.InProgress, + "incomplete" => FunctionShellCallOutputStatusEnum.Incomplete, _ => null, }; } diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.LocalShellCallStatus.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.FunctionShellCallStatus.g.cs similarity index 58% rename from src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.LocalShellCallStatus.g.cs rename to src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.FunctionShellCallStatus.g.cs index dea3c8e71..695507aba 100644 --- a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.LocalShellCallStatus.g.cs +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.FunctionShellCallStatus.g.cs @@ -6,7 +6,7 @@ namespace tryAGI.OpenAI /// /// /// - public enum LocalShellCallStatus + public enum FunctionShellCallStatus { /// /// @@ -25,31 +25,31 @@ public enum LocalShellCallStatus /// /// Enum extensions to do fast conversions without the reflection. /// - public static class LocalShellCallStatusExtensions + public static class FunctionShellCallStatusExtensions { /// /// Converts an enum to a string. /// - public static string ToValueString(this LocalShellCallStatus value) + public static string ToValueString(this FunctionShellCallStatus value) { return value switch { - LocalShellCallStatus.Completed => "completed", - LocalShellCallStatus.InProgress => "in_progress", - LocalShellCallStatus.Incomplete => "incomplete", + FunctionShellCallStatus.Completed => "completed", + FunctionShellCallStatus.InProgress => "in_progress", + FunctionShellCallStatus.Incomplete => "incomplete", _ => throw new global::System.ArgumentOutOfRangeException(nameof(value), value, null), }; } /// /// Converts an string to a enum. /// - public static LocalShellCallStatus? ToEnum(string value) + public static FunctionShellCallStatus? ToEnum(string value) { return value switch { - "completed" => LocalShellCallStatus.Completed, - "in_progress" => LocalShellCallStatus.InProgress, - "incomplete" => LocalShellCallStatus.Incomplete, + "completed" => FunctionShellCallStatus.Completed, + "in_progress" => FunctionShellCallStatus.InProgress, + "incomplete" => FunctionShellCallStatus.Incomplete, _ => null, }; } diff --git a/src/libs/tryAGI.OpenAI/openapi.yaml b/src/libs/tryAGI.OpenAI/openapi.yaml index 4d9690e0e..763c04d83 100644 --- a/src/libs/tryAGI.OpenAI/openapi.yaml +++ b/src/libs/tryAGI.OpenAI/openapi.yaml @@ -9469,7 +9469,9 @@ paths: total. There is no organization-wide storage limit. Uploads to this - endpoint are rate-limited to 2,000 files per minute per organization. + endpoint are rate-limited to 1,000 requests per minute per authenticated + + user. - The Assistants API supports files up to 2 million tokens and of @@ -9487,7 +9489,9 @@ paths: - For Retrieval or `file_search` ingestion, upload files here first. If you need to attach multiple uploaded files to the same vector store, use [`/vector_stores/{vector_store_id}/file_batches`](/docs/api-reference/vector-stores-file-batches/createBatch) - instead of attaching them one by one. + instead of attaching them one by one. Vector store attachment has separate + limits from file upload, including 2,000 attached files per minute per + organization. Please [contact us](https://help.openai.com/) if you need to increase these @@ -9511,10 +9515,11 @@ paths: group: files description: > Uploads a file for later use across OpenAI APIs. Uploads to this - endpoint are rate-limited to 2,000 files per minute per organization. - For Retrieval or `file_search` ingestion, upload files here first. If - you need to attach multiple uploaded files to the same vector store, - use vector store file batches instead of attaching them one by one. + endpoint are rate-limited to 1,000 requests per minute per + authenticated user. For Retrieval or `file_search` ingestion, upload + files here first. If you need to attach multiple uploaded files to the + same vector store, use vector store file batches instead of attaching + them one by one. examples: request: curl: | @@ -69848,7 +69853,7 @@ components: - max_output_length title: Shell exec action description: Execute a shell command. - LocalShellCallStatus: + FunctionShellCallStatus: type: string enum: - in_progress @@ -69908,7 +69913,7 @@ components: The shell commands and limits that describe how to run the tool call. status: - $ref: '#/components/schemas/LocalShellCallStatus' + $ref: '#/components/schemas/FunctionShellCallStatus' description: >- The status of the shell call. One of `in_progress`, `completed`, or `incomplete`. @@ -69935,7 +69940,7 @@ components: description: >- A tool call that executes one or more shell commands in a managed environment. - LocalShellCallOutputStatusEnum: + FunctionShellCallOutputStatusEnum: type: string enum: - in_progress @@ -70019,7 +70024,7 @@ components: type: string description: The unique ID of the shell tool call generated by the model. status: - $ref: '#/components/schemas/LocalShellCallOutputStatusEnum' + $ref: '#/components/schemas/FunctionShellCallOutputStatusEnum' description: >- The status of the shell call output. One of `in_progress`, `completed`, or `incomplete`. @@ -71747,7 +71752,9 @@ components: - $ref: '#/components/schemas/CompactionBody' - $ref: '#/components/schemas/CodeInterpreterToolCall' - $ref: '#/components/schemas/LocalShellToolCall' + deprecated: true - $ref: '#/components/schemas/LocalShellToolCallOutput' + deprecated: true - $ref: '#/components/schemas/FunctionShellCall' - $ref: '#/components/schemas/FunctionShellCallOutput' - $ref: '#/components/schemas/ApplyPatchToolCall'