diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI..JsonSerializerContext.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI..JsonSerializerContext.g.cs index 9e7113e1..47d8750f 100644 --- a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI..JsonSerializerContext.g.cs +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI..JsonSerializerContext.g.cs @@ -2427,6 +2427,9 @@ namespace tryAGI.OpenAI typeof(global::tryAGI.OpenAI.JsonConverters.OneOfJsonConverter), typeof(global::tryAGI.OpenAI.JsonConverters.OneOfJsonConverter>), typeof(global::tryAGI.OpenAI.JsonConverters.OneOfJsonConverter), + typeof(global::tryAGI.OpenAI.JsonConverters.OneOfJsonConverter), + typeof(global::tryAGI.OpenAI.JsonConverters.OneOfJsonConverter), + typeof(global::tryAGI.OpenAI.JsonConverters.OneOfJsonConverter), typeof(global::tryAGI.OpenAI.JsonConverters.OneOfJsonConverter>), typeof(global::tryAGI.OpenAI.JsonConverters.OneOfJsonConverter>), typeof(global::tryAGI.OpenAI.JsonConverters.OneOfJsonConverter, byte[]>), diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.IVideosClient.CreateVideo.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.IVideosClient.CreateVideo.g.cs index df000285..5ec71394 100644 --- a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.IVideosClient.CreateVideo.g.cs +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.IVideosClient.CreateVideo.g.cs @@ -12,7 +12,7 @@ public partial interface IVideosClient /// global::System.Threading.Tasks.Task CreateVideoAsync( - global::tryAGI.OpenAI.CreateVideoBody request, + global::tryAGI.OpenAI.CreateVideoMultipartBody request, global::System.Threading.CancellationToken cancellationToken = default); /// @@ -24,15 +24,7 @@ public partial interface IVideosClient /// /// Text prompt that describes the video to generate. /// - /// - /// Optional multipart reference asset that guides generation. - /// - /// - /// Optional multipart reference asset that guides generation. - /// - /// - /// Optional JSON-safe image reference that guides generation. Provide exactly one of `image_url` or `file_id`. - /// + /// /// /// Clip duration in seconds (allowed values: 4, 8, 12). Defaults to 4 seconds. /// @@ -44,9 +36,7 @@ public partial interface IVideosClient global::System.Threading.Tasks.Task CreateVideoAsync( string prompt, global::tryAGI.OpenAI.VideoModel? model = default, - byte[]? inputReference = default, - string? inputReferencename = default, - global::tryAGI.OpenAI.ImageRefParam2? imageReference = default, + global::tryAGI.OpenAI.OneOf? inputReference = default, global::tryAGI.OpenAI.VideoSeconds? seconds = default, global::tryAGI.OpenAI.VideoSize? size = default, global::System.Threading.CancellationToken cancellationToken = default); diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.IVideosClient.CreateVideoCharacter.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.IVideosClient.CreateVideoCharacter.g.cs new file mode 100644 index 00000000..7bd58226 --- /dev/null +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.IVideosClient.CreateVideoCharacter.g.cs @@ -0,0 +1,38 @@ +#nullable enable + +namespace tryAGI.OpenAI +{ + public partial interface IVideosClient + { + /// + /// Create a character from an uploaded video. + /// + /// + /// The token to cancel the operation with + /// + global::System.Threading.Tasks.Task CreateVideoCharacterAsync( + + global::tryAGI.OpenAI.CreateVideoCharacterBody request, + global::System.Threading.CancellationToken cancellationToken = default); + + /// + /// Create a character from an uploaded video. + /// + /// + /// Video file used to create a character. + /// + /// + /// Video file used to create a character. + /// + /// + /// Display name for this API character. + /// + /// The token to cancel the operation with + /// + global::System.Threading.Tasks.Task CreateVideoCharacterAsync( + byte[] video, + string videoname, + string name, + global::System.Threading.CancellationToken cancellationToken = default); + } +} \ No newline at end of file diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.IVideosClient.CreateVideoEdit.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.IVideosClient.CreateVideoEdit.g.cs new file mode 100644 index 00000000..4a20d0fb --- /dev/null +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.IVideosClient.CreateVideoEdit.g.cs @@ -0,0 +1,32 @@ +#nullable enable + +namespace tryAGI.OpenAI +{ + public partial interface IVideosClient + { + /// + /// Create a new video generation job by editing a source video or existing generated video. + /// + /// + /// The token to cancel the operation with + /// + global::System.Threading.Tasks.Task CreateVideoEditAsync( + + global::tryAGI.OpenAI.CreateVideoEditMultipartBody request, + global::System.Threading.CancellationToken cancellationToken = default); + + /// + /// Create a new video generation job by editing a source video or existing generated video. + /// + /// + /// + /// Text prompt that describes how to edit the source video. + /// + /// The token to cancel the operation with + /// + global::System.Threading.Tasks.Task CreateVideoEditAsync( + global::tryAGI.OpenAI.OneOf video, + string prompt, + global::System.Threading.CancellationToken cancellationToken = default); + } +} \ No newline at end of file diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.IVideosClient.CreateVideoExtend.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.IVideosClient.CreateVideoExtend.g.cs new file mode 100644 index 00000000..1ed1d554 --- /dev/null +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.IVideosClient.CreateVideoExtend.g.cs @@ -0,0 +1,36 @@ +#nullable enable + +namespace tryAGI.OpenAI +{ + public partial interface IVideosClient + { + /// + /// Create an extension of a completed video. + /// + /// + /// The token to cancel the operation with + /// + global::System.Threading.Tasks.Task CreateVideoExtendAsync( + + global::tryAGI.OpenAI.CreateVideoExtendMultipartBody request, + global::System.Threading.CancellationToken cancellationToken = default); + + /// + /// Create an extension of a completed video. + /// + /// + /// + /// Updated text prompt that directs the extension generation. + /// + /// + /// Length of the newly generated extension segment in seconds (allowed values: 4, 8, 12, 16, 20). + /// + /// The token to cancel the operation with + /// + global::System.Threading.Tasks.Task CreateVideoExtendAsync( + global::tryAGI.OpenAI.OneOf video, + string prompt, + global::tryAGI.OpenAI.VideoSeconds seconds, + global::System.Threading.CancellationToken cancellationToken = default); + } +} \ No newline at end of file diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.IVideosClient.GetVideoCharacter.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.IVideosClient.GetVideoCharacter.g.cs new file mode 100644 index 00000000..1dc9ee13 --- /dev/null +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.IVideosClient.GetVideoCharacter.g.cs @@ -0,0 +1,19 @@ +#nullable enable + +namespace tryAGI.OpenAI +{ + public partial interface IVideosClient + { + /// + /// Fetch a character. + /// + /// + /// Example: char_123 + /// + /// The token to cancel the operation with + /// + global::System.Threading.Tasks.Task GetVideoCharacterAsync( + string characterId, + global::System.Threading.CancellationToken cancellationToken = default); + } +} \ No newline at end of file diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonSerializerContextTypes.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonSerializerContextTypes.g.cs index 6cf98ac1..66ddad5c 100644 --- a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonSerializerContextTypes.g.cs +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.JsonSerializerContextTypes.g.cs @@ -10220,967 +10220,1011 @@ public sealed partial class JsonSerializerContextTypes /// /// /// - public global::tryAGI.OpenAI.CreateVideoBody? Type2548 { get; set; } + public global::tryAGI.OpenAI.CreateVideoMultipartBody? Type2548 { get; set; } /// /// /// - public global::tryAGI.OpenAI.DeletedVideoResource? Type2549 { get; set; } + public global::tryAGI.OpenAI.OneOf? Type2549 { get; set; } /// /// /// - public global::tryAGI.OpenAI.DeletedVideoResourceObject? Type2550 { get; set; } + public global::tryAGI.OpenAI.CreateVideoJsonBody? Type2550 { get; set; } /// /// /// - public global::tryAGI.OpenAI.VideoContentVariant? Type2551 { get; set; } + public global::tryAGI.OpenAI.CreateVideoCharacterBody? Type2551 { get; set; } /// /// /// - public global::tryAGI.OpenAI.CreateVideoRemixBody? Type2552 { get; set; } + public global::tryAGI.OpenAI.VideoCharacterResource? Type2552 { get; set; } /// /// /// - public global::tryAGI.OpenAI.TruncationEnum? Type2553 { get; set; } + public global::tryAGI.OpenAI.VideoReferenceInputParam? Type2553 { get; set; } /// /// /// - public global::tryAGI.OpenAI.TokenCountsBody? Type2554 { get; set; } + public global::tryAGI.OpenAI.CreateVideoEditMultipartBody? Type2554 { get; set; } /// /// /// - public global::tryAGI.OpenAI.TokenCountsResource? Type2555 { get; set; } + public global::tryAGI.OpenAI.OneOf? Type2555 { get; set; } /// /// /// - public global::tryAGI.OpenAI.TokenCountsResourceObject? Type2556 { get; set; } + public global::tryAGI.OpenAI.CreateVideoEditJsonBody? Type2556 { get; set; } /// /// /// - public global::tryAGI.OpenAI.CompactResponseMethodPublicBody? Type2557 { get; set; } + public global::tryAGI.OpenAI.CreateVideoExtendMultipartBody? Type2557 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ItemField? Type2558 { get; set; } + public global::tryAGI.OpenAI.OneOf? Type2558 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ItemFieldDiscriminator? Type2559 { get; set; } + public global::tryAGI.OpenAI.CreateVideoExtendJsonBody? Type2559 { get; set; } /// /// /// - public global::tryAGI.OpenAI.CompactResource? Type2560 { get; set; } + public global::tryAGI.OpenAI.DeletedVideoResource? Type2560 { get; set; } /// /// /// - public global::tryAGI.OpenAI.CompactResourceObject? Type2561 { get; set; } + public global::tryAGI.OpenAI.DeletedVideoResourceObject? Type2561 { get; set; } /// /// /// - public global::System.Collections.Generic.IList? Type2562 { get; set; } + public global::tryAGI.OpenAI.VideoContentVariant? Type2562 { get; set; } /// /// /// - public global::tryAGI.OpenAI.SkillResource? Type2563 { get; set; } + public global::tryAGI.OpenAI.CreateVideoRemixBody? Type2563 { get; set; } /// /// /// - public global::tryAGI.OpenAI.SkillResourceObject? Type2564 { get; set; } + public global::tryAGI.OpenAI.TruncationEnum? Type2564 { get; set; } /// /// /// - public global::tryAGI.OpenAI.SkillListResource? Type2565 { get; set; } + public global::tryAGI.OpenAI.TokenCountsBody? Type2565 { get; set; } /// /// /// - public global::tryAGI.OpenAI.SkillListResourceObject? Type2566 { get; set; } + public global::tryAGI.OpenAI.TokenCountsResource? Type2566 { get; set; } /// /// /// - public global::System.Collections.Generic.IList? Type2567 { get; set; } + public global::tryAGI.OpenAI.TokenCountsResourceObject? Type2567 { get; set; } /// /// /// - public global::tryAGI.OpenAI.CreateSkillBody? Type2568 { get; set; } + public global::tryAGI.OpenAI.CompactResponseMethodPublicBody? Type2568 { get; set; } /// /// /// - public global::tryAGI.OpenAI.OneOf, byte[]>? Type2569 { get; set; } + public global::tryAGI.OpenAI.ItemField? Type2569 { get; set; } /// /// /// - public global::tryAGI.OpenAI.SetDefaultSkillVersionBody? Type2570 { get; set; } + public global::tryAGI.OpenAI.ItemFieldDiscriminator? Type2570 { get; set; } /// /// /// - public global::tryAGI.OpenAI.DeletedSkillResource? Type2571 { get; set; } + public global::tryAGI.OpenAI.CompactResource? Type2571 { get; set; } /// /// /// - public global::tryAGI.OpenAI.DeletedSkillResourceObject? Type2572 { get; set; } + public global::tryAGI.OpenAI.CompactResourceObject? Type2572 { get; set; } /// /// /// - public global::tryAGI.OpenAI.SkillVersionResource? Type2573 { get; set; } + public global::System.Collections.Generic.IList? Type2573 { get; set; } /// /// /// - public global::tryAGI.OpenAI.SkillVersionResourceObject? Type2574 { get; set; } + public global::tryAGI.OpenAI.SkillResource? Type2574 { get; set; } /// /// /// - public global::tryAGI.OpenAI.SkillVersionListResource? Type2575 { get; set; } + public global::tryAGI.OpenAI.SkillResourceObject? Type2575 { get; set; } /// /// /// - public global::tryAGI.OpenAI.SkillVersionListResourceObject? Type2576 { get; set; } + public global::tryAGI.OpenAI.SkillListResource? Type2576 { get; set; } /// /// /// - public global::System.Collections.Generic.IList? Type2577 { get; set; } + public global::tryAGI.OpenAI.SkillListResourceObject? Type2577 { get; set; } /// /// /// - public global::tryAGI.OpenAI.CreateSkillVersionBody? Type2578 { get; set; } + public global::System.Collections.Generic.IList? Type2578 { get; set; } /// /// /// - public global::tryAGI.OpenAI.DeletedSkillVersionResource? Type2579 { get; set; } + public global::tryAGI.OpenAI.CreateSkillBody? Type2579 { get; set; } /// /// /// - public global::tryAGI.OpenAI.DeletedSkillVersionResourceObject? Type2580 { get; set; } + public global::tryAGI.OpenAI.OneOf, byte[]>? Type2580 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ChatkitWorkflowTracing? Type2581 { get; set; } + public global::tryAGI.OpenAI.SetDefaultSkillVersionBody? Type2581 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ChatkitWorkflow? Type2582 { get; set; } + public global::tryAGI.OpenAI.DeletedSkillResource? Type2582 { get; set; } /// /// /// - public global::tryAGI.OpenAI.OneOf? Type2583 { get; set; } + public global::tryAGI.OpenAI.DeletedSkillResourceObject? Type2583 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ChatSessionRateLimits? Type2584 { get; set; } + public global::tryAGI.OpenAI.SkillVersionResource? Type2584 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ChatSessionStatus? Type2585 { get; set; } + public global::tryAGI.OpenAI.SkillVersionResourceObject? Type2585 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ChatSessionAutomaticThreadTitling? Type2586 { get; set; } + public global::tryAGI.OpenAI.SkillVersionListResource? Type2586 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ChatSessionFileUpload? Type2587 { get; set; } + public global::tryAGI.OpenAI.SkillVersionListResourceObject? Type2587 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ChatSessionHistory? Type2588 { get; set; } + public global::System.Collections.Generic.IList? Type2588 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ChatSessionChatkitConfiguration? Type2589 { get; set; } + public global::tryAGI.OpenAI.CreateSkillVersionBody? Type2589 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ChatSessionResource? Type2590 { get; set; } + public global::tryAGI.OpenAI.DeletedSkillVersionResource? Type2590 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ChatSessionResourceObject? Type2591 { get; set; } + public global::tryAGI.OpenAI.DeletedSkillVersionResourceObject? Type2591 { get; set; } /// /// /// - public global::tryAGI.OpenAI.WorkflowTracingParam? Type2592 { get; set; } + public global::tryAGI.OpenAI.ChatkitWorkflowTracing? Type2592 { get; set; } /// /// /// - public global::tryAGI.OpenAI.WorkflowParam? Type2593 { get; set; } + public global::tryAGI.OpenAI.ChatkitWorkflow? Type2593 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ExpiresAfterParam? Type2594 { get; set; } + public global::tryAGI.OpenAI.OneOf? Type2594 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ExpiresAfterParamAnchor? Type2595 { get; set; } + public global::tryAGI.OpenAI.ChatSessionRateLimits? Type2595 { get; set; } /// /// /// - public global::tryAGI.OpenAI.RateLimitsParam? Type2596 { get; set; } + public global::tryAGI.OpenAI.ChatSessionStatus? Type2596 { get; set; } /// /// /// - public global::tryAGI.OpenAI.AutomaticThreadTitlingParam? Type2597 { get; set; } + public global::tryAGI.OpenAI.ChatSessionAutomaticThreadTitling? Type2597 { get; set; } /// /// /// - public global::tryAGI.OpenAI.FileUploadParam? Type2598 { get; set; } + public global::tryAGI.OpenAI.ChatSessionFileUpload? Type2598 { get; set; } /// /// /// - public global::tryAGI.OpenAI.HistoryParam? Type2599 { get; set; } + public global::tryAGI.OpenAI.ChatSessionHistory? Type2599 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ChatkitConfigurationParam? Type2600 { get; set; } + public global::tryAGI.OpenAI.ChatSessionChatkitConfiguration? Type2600 { get; set; } /// /// /// - public global::tryAGI.OpenAI.CreateChatSessionBody? Type2601 { get; set; } + public global::tryAGI.OpenAI.ChatSessionResource? Type2601 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UserMessageInputText? Type2602 { get; set; } + public global::tryAGI.OpenAI.ChatSessionResourceObject? Type2602 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UserMessageInputTextType? Type2603 { get; set; } + public global::tryAGI.OpenAI.WorkflowTracingParam? Type2603 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UserMessageQuotedText? Type2604 { get; set; } + public global::tryAGI.OpenAI.WorkflowParam? Type2604 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UserMessageQuotedTextType? Type2605 { get; set; } + public global::tryAGI.OpenAI.ExpiresAfterParam? Type2605 { get; set; } /// /// /// - public global::tryAGI.OpenAI.AttachmentType? Type2606 { get; set; } + public global::tryAGI.OpenAI.ExpiresAfterParamAnchor? Type2606 { get; set; } /// /// /// - public global::tryAGI.OpenAI.Attachment? Type2607 { get; set; } + public global::tryAGI.OpenAI.RateLimitsParam? Type2607 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ToolChoice8? Type2608 { get; set; } + public global::tryAGI.OpenAI.AutomaticThreadTitlingParam? Type2608 { get; set; } /// /// /// - public global::tryAGI.OpenAI.InferenceOptions? Type2609 { get; set; } + public global::tryAGI.OpenAI.FileUploadParam? Type2609 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UserMessageItem? Type2610 { get; set; } + public global::tryAGI.OpenAI.HistoryParam? Type2610 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UserMessageItemObject? Type2611 { get; set; } + public global::tryAGI.OpenAI.ChatkitConfigurationParam? Type2611 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UserMessageItemType? Type2612 { get; set; } + public global::tryAGI.OpenAI.CreateChatSessionBody? Type2612 { get; set; } /// /// /// - public global::System.Collections.Generic.IList? Type2613 { get; set; } + public global::tryAGI.OpenAI.UserMessageInputText? Type2613 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ContentItem4? Type2614 { get; set; } + public global::tryAGI.OpenAI.UserMessageInputTextType? Type2614 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UserMessageItemContentItemDiscriminator? Type2615 { get; set; } + public global::tryAGI.OpenAI.UserMessageQuotedText? Type2615 { get; set; } /// /// /// - public global::System.Collections.Generic.IList? Type2616 { get; set; } + public global::tryAGI.OpenAI.UserMessageQuotedTextType? Type2616 { get; set; } /// /// /// - public global::tryAGI.OpenAI.FileAnnotationSource? Type2617 { get; set; } + public global::tryAGI.OpenAI.AttachmentType? Type2617 { get; set; } /// /// /// - public global::tryAGI.OpenAI.FileAnnotationSourceType? Type2618 { get; set; } + public global::tryAGI.OpenAI.Attachment? Type2618 { get; set; } /// /// /// - public global::tryAGI.OpenAI.FileAnnotation? Type2619 { get; set; } + public global::tryAGI.OpenAI.ToolChoice8? Type2619 { get; set; } /// /// /// - public global::tryAGI.OpenAI.FileAnnotationType? Type2620 { get; set; } + public global::tryAGI.OpenAI.InferenceOptions? Type2620 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UrlAnnotationSource? Type2621 { get; set; } + public global::tryAGI.OpenAI.UserMessageItem? Type2621 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UrlAnnotationSourceType? Type2622 { get; set; } + public global::tryAGI.OpenAI.UserMessageItemObject? Type2622 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UrlAnnotation? Type2623 { get; set; } + public global::tryAGI.OpenAI.UserMessageItemType? Type2623 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UrlAnnotationType? Type2624 { get; set; } + public global::System.Collections.Generic.IList? Type2624 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ResponseOutputText? Type2625 { get; set; } + public global::tryAGI.OpenAI.ContentItem4? Type2625 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ResponseOutputTextType? Type2626 { get; set; } + public global::tryAGI.OpenAI.UserMessageItemContentItemDiscriminator? Type2626 { get; set; } /// /// /// - public global::System.Collections.Generic.IList? Type2627 { get; set; } + public global::System.Collections.Generic.IList? Type2627 { get; set; } /// /// /// - public global::tryAGI.OpenAI.AnnotationsItem3? Type2628 { get; set; } + public global::tryAGI.OpenAI.FileAnnotationSource? Type2628 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ResponseOutputTextAnnotationDiscriminator? Type2629 { get; set; } + public global::tryAGI.OpenAI.FileAnnotationSourceType? Type2629 { get; set; } /// /// /// - public global::tryAGI.OpenAI.AssistantMessageItem? Type2630 { get; set; } + public global::tryAGI.OpenAI.FileAnnotation? Type2630 { get; set; } /// /// /// - public global::tryAGI.OpenAI.AssistantMessageItemObject? Type2631 { get; set; } + public global::tryAGI.OpenAI.FileAnnotationType? Type2631 { get; set; } /// /// /// - public global::tryAGI.OpenAI.AssistantMessageItemType? Type2632 { get; set; } + public global::tryAGI.OpenAI.UrlAnnotationSource? Type2632 { get; set; } /// /// /// - public global::System.Collections.Generic.IList? Type2633 { get; set; } + public global::tryAGI.OpenAI.UrlAnnotationSourceType? Type2633 { get; set; } /// /// /// - public global::tryAGI.OpenAI.WidgetMessageItem? Type2634 { get; set; } + public global::tryAGI.OpenAI.UrlAnnotation? Type2634 { get; set; } /// /// /// - public global::tryAGI.OpenAI.WidgetMessageItemObject? Type2635 { get; set; } + public global::tryAGI.OpenAI.UrlAnnotationType? Type2635 { get; set; } /// /// /// - public global::tryAGI.OpenAI.WidgetMessageItemType? Type2636 { get; set; } + public global::tryAGI.OpenAI.ResponseOutputText? Type2636 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ClientToolCallStatus? Type2637 { get; set; } + public global::tryAGI.OpenAI.ResponseOutputTextType? Type2637 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ClientToolCallItem? Type2638 { get; set; } + public global::System.Collections.Generic.IList? Type2638 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ClientToolCallItemObject? Type2639 { get; set; } + public global::tryAGI.OpenAI.AnnotationsItem3? Type2639 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ClientToolCallItemType? Type2640 { get; set; } + public global::tryAGI.OpenAI.ResponseOutputTextAnnotationDiscriminator? Type2640 { get; set; } /// /// /// - public global::tryAGI.OpenAI.TaskType? Type2641 { get; set; } + public global::tryAGI.OpenAI.AssistantMessageItem? Type2641 { get; set; } /// /// /// - public global::tryAGI.OpenAI.TaskItem? Type2642 { get; set; } + public global::tryAGI.OpenAI.AssistantMessageItemObject? Type2642 { get; set; } /// /// /// - public global::tryAGI.OpenAI.TaskItemObject? Type2643 { get; set; } + public global::tryAGI.OpenAI.AssistantMessageItemType? Type2643 { get; set; } /// /// /// - public global::tryAGI.OpenAI.TaskItemType? Type2644 { get; set; } + public global::System.Collections.Generic.IList? Type2644 { get; set; } /// /// /// - public global::tryAGI.OpenAI.TaskGroupTask? Type2645 { get; set; } + public global::tryAGI.OpenAI.WidgetMessageItem? Type2645 { get; set; } /// /// /// - public global::tryAGI.OpenAI.TaskGroupItem? Type2646 { get; set; } + public global::tryAGI.OpenAI.WidgetMessageItemObject? Type2646 { get; set; } /// /// /// - public global::tryAGI.OpenAI.TaskGroupItemObject? Type2647 { get; set; } + public global::tryAGI.OpenAI.WidgetMessageItemType? Type2647 { get; set; } /// /// /// - public global::tryAGI.OpenAI.TaskGroupItemType? Type2648 { get; set; } + public global::tryAGI.OpenAI.ClientToolCallStatus? Type2648 { get; set; } /// /// /// - public global::System.Collections.Generic.IList? Type2649 { get; set; } + public global::tryAGI.OpenAI.ClientToolCallItem? Type2649 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ThreadItem? Type2650 { get; set; } + public global::tryAGI.OpenAI.ClientToolCallItemObject? Type2650 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ThreadItemDiscriminator? Type2651 { get; set; } + public global::tryAGI.OpenAI.ClientToolCallItemType? Type2651 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ThreadItemListResource? Type2652 { get; set; } + public global::tryAGI.OpenAI.TaskType? Type2652 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ThreadItemListResourceObject? Type2653 { get; set; } + public global::tryAGI.OpenAI.TaskItem? Type2653 { get; set; } /// /// /// - public global::System.Collections.Generic.IList? Type2654 { get; set; } + public global::tryAGI.OpenAI.TaskItemObject? Type2654 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ActiveStatus? Type2655 { get; set; } + public global::tryAGI.OpenAI.TaskItemType? Type2655 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ActiveStatusType? Type2656 { get; set; } + public global::tryAGI.OpenAI.TaskGroupTask? Type2656 { get; set; } /// /// /// - public global::tryAGI.OpenAI.LockedStatus? Type2657 { get; set; } + public global::tryAGI.OpenAI.TaskGroupItem? Type2657 { get; set; } /// /// /// - public global::tryAGI.OpenAI.LockedStatusType? Type2658 { get; set; } + public global::tryAGI.OpenAI.TaskGroupItemObject? Type2658 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ClosedStatus? Type2659 { get; set; } + public global::tryAGI.OpenAI.TaskGroupItemType? Type2659 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ClosedStatusType? Type2660 { get; set; } + public global::System.Collections.Generic.IList? Type2660 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ThreadResource? Type2661 { get; set; } + public global::tryAGI.OpenAI.ThreadItem? Type2661 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ThreadResourceObject? Type2662 { get; set; } + public global::tryAGI.OpenAI.ThreadItemDiscriminator? Type2662 { get; set; } /// /// /// - public global::tryAGI.OpenAI.Status? Type2663 { get; set; } + public global::tryAGI.OpenAI.ThreadItemListResource? Type2663 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ThreadResourceStatusDiscriminator? Type2664 { get; set; } + public global::tryAGI.OpenAI.ThreadItemListResourceObject? Type2664 { get; set; } /// /// /// - public global::tryAGI.OpenAI.DeletedThreadResource? Type2665 { get; set; } + public global::System.Collections.Generic.IList? Type2665 { get; set; } /// /// /// - public global::tryAGI.OpenAI.DeletedThreadResourceObject? Type2666 { get; set; } + public global::tryAGI.OpenAI.ActiveStatus? Type2666 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ThreadListResource? Type2667 { get; set; } + public global::tryAGI.OpenAI.ActiveStatusType? Type2667 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ThreadListResourceObject? Type2668 { get; set; } + public global::tryAGI.OpenAI.LockedStatus? Type2668 { get; set; } /// /// /// - public global::System.Collections.Generic.IList? Type2669 { get; set; } + public global::tryAGI.OpenAI.LockedStatusType? Type2669 { get; set; } /// /// /// - public global::tryAGI.OpenAI.DragPoint? Type2670 { get; set; } + public global::tryAGI.OpenAI.ClosedStatus? Type2670 { get; set; } /// /// /// - public global::tryAGI.OpenAI.RealtimeServerEventType? Type2671 { get; set; } + public global::tryAGI.OpenAI.ClosedStatusType? Type2671 { get; set; } /// /// /// - public global::tryAGI.OpenAI.RealtimeServerEventBase? Type2672 { get; set; } + public global::tryAGI.OpenAI.ThreadResource? Type2672 { get; set; } /// /// /// - public global::tryAGI.OpenAI.RealtimeSessionUpdate? Type2673 { get; set; } + public global::tryAGI.OpenAI.ThreadResourceObject? Type2673 { get; set; } /// /// /// - public global::tryAGI.OpenAI.RealtimeSessionUpdateType? Type2674 { get; set; } + public global::tryAGI.OpenAI.Status? Type2674 { get; set; } /// /// /// - public global::tryAGI.OpenAI.RealtimeInputAudioBufferAppend? Type2675 { get; set; } + public global::tryAGI.OpenAI.ThreadResourceStatusDiscriminator? Type2675 { get; set; } /// /// /// - public global::tryAGI.OpenAI.RealtimeInputAudioBufferAppendType? Type2676 { get; set; } + public global::tryAGI.OpenAI.DeletedThreadResource? Type2676 { get; set; } /// /// /// - public global::tryAGI.OpenAI.RealtimeInputAudioBufferCommit? Type2677 { get; set; } + public global::tryAGI.OpenAI.DeletedThreadResourceObject? Type2677 { get; set; } /// /// /// - public global::tryAGI.OpenAI.RealtimeInputAudioBufferCommitType? Type2678 { get; set; } + public global::tryAGI.OpenAI.ThreadListResource? Type2678 { get; set; } /// /// /// - public global::tryAGI.OpenAI.RealtimeInputAudioBufferClear? Type2679 { get; set; } + public global::tryAGI.OpenAI.ThreadListResourceObject? Type2679 { get; set; } /// /// /// - public global::tryAGI.OpenAI.RealtimeInputAudioBufferClearType? Type2680 { get; set; } + public global::System.Collections.Generic.IList? Type2680 { get; set; } /// /// /// - public global::tryAGI.OpenAI.RealtimeConversationItemCreate? Type2681 { get; set; } + public global::tryAGI.OpenAI.DragPoint? Type2681 { get; set; } /// /// /// - public global::tryAGI.OpenAI.RealtimeConversationItemCreateType? Type2682 { get; set; } + public global::tryAGI.OpenAI.RealtimeServerEventType? Type2682 { get; set; } /// /// /// - public global::tryAGI.OpenAI.RealtimeConversationItemTruncate? Type2683 { get; set; } + public global::tryAGI.OpenAI.RealtimeServerEventBase? Type2683 { get; set; } /// /// /// - public global::tryAGI.OpenAI.RealtimeConversationItemTruncateType? Type2684 { get; set; } + public global::tryAGI.OpenAI.RealtimeSessionUpdate? Type2684 { get; set; } /// /// /// - public global::tryAGI.OpenAI.RealtimeConversationItemDelete? Type2685 { get; set; } + public global::tryAGI.OpenAI.RealtimeSessionUpdateType? Type2685 { get; set; } /// /// /// - public global::tryAGI.OpenAI.RealtimeConversationItemDeleteType? Type2686 { get; set; } + public global::tryAGI.OpenAI.RealtimeInputAudioBufferAppend? Type2686 { get; set; } /// /// /// - public global::tryAGI.OpenAI.RealtimeResponseCreate? Type2687 { get; set; } + public global::tryAGI.OpenAI.RealtimeInputAudioBufferAppendType? Type2687 { get; set; } /// /// /// - public global::tryAGI.OpenAI.RealtimeResponseCreateType? Type2688 { get; set; } + public global::tryAGI.OpenAI.RealtimeInputAudioBufferCommit? Type2688 { get; set; } /// /// /// - public global::tryAGI.OpenAI.RealtimeResponseCreateResponse? Type2689 { get; set; } + public global::tryAGI.OpenAI.RealtimeInputAudioBufferCommitType? Type2689 { get; set; } /// /// /// - public global::System.Collections.Generic.IList? Type2690 { get; set; } + public global::tryAGI.OpenAI.RealtimeInputAudioBufferClear? Type2690 { get; set; } /// /// /// - public global::tryAGI.OpenAI.RealtimeResponseCreateResponseModalitie? Type2691 { get; set; } + public global::tryAGI.OpenAI.RealtimeInputAudioBufferClearType? Type2691 { get; set; } /// /// /// - public global::tryAGI.OpenAI.RealtimeResponseCreateResponseVoice? Type2692 { get; set; } + public global::tryAGI.OpenAI.RealtimeConversationItemCreate? Type2692 { get; set; } /// /// /// - public global::System.Collections.Generic.IList? Type2693 { get; set; } + public global::tryAGI.OpenAI.RealtimeConversationItemCreateType? Type2693 { get; set; } /// /// /// - public global::tryAGI.OpenAI.RealtimeResponseCreateResponseTool? Type2694 { get; set; } + public global::tryAGI.OpenAI.RealtimeConversationItemTruncate? Type2694 { get; set; } /// /// /// - public global::tryAGI.OpenAI.OneOf? Type2695 { get; set; } + public global::tryAGI.OpenAI.RealtimeConversationItemTruncateType? Type2695 { get; set; } /// /// /// - public global::tryAGI.OpenAI.RealtimeResponseCreateResponseToolChoice? Type2696 { get; set; } + public global::tryAGI.OpenAI.RealtimeConversationItemDelete? Type2696 { get; set; } /// /// /// - public global::tryAGI.OpenAI.OneOf? Type2697 { get; set; } + public global::tryAGI.OpenAI.RealtimeConversationItemDeleteType? Type2697 { get; set; } /// /// /// - public global::tryAGI.OpenAI.RealtimeResponseCreateResponseMaxOutputTokens? Type2698 { get; set; } + public global::tryAGI.OpenAI.RealtimeResponseCreate? Type2698 { get; set; } /// /// /// - public global::tryAGI.OpenAI.RealtimeResponseCancel? Type2699 { get; set; } + public global::tryAGI.OpenAI.RealtimeResponseCreateType? Type2699 { get; set; } /// /// /// - public global::tryAGI.OpenAI.RealtimeResponseCancelType? Type2700 { get; set; } + public global::tryAGI.OpenAI.RealtimeResponseCreateResponse? Type2700 { get; set; } /// /// /// - public global::tryAGI.OpenAI.CreateBatchRequest? Type2701 { get; set; } + public global::System.Collections.Generic.IList? Type2701 { get; set; } /// /// /// - public global::tryAGI.OpenAI.CreateBatchRequestEndpoint? Type2702 { get; set; } + public global::tryAGI.OpenAI.RealtimeResponseCreateResponseModalitie? Type2702 { get; set; } /// /// /// - public global::tryAGI.OpenAI.CreateBatchRequestCompletionWindow? Type2703 { get; set; } + public global::tryAGI.OpenAI.RealtimeResponseCreateResponseVoice? Type2703 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UpdateChatCompletionRequest? Type2704 { get; set; } + public global::System.Collections.Generic.IList? Type2704 { get; set; } /// /// /// - public global::tryAGI.OpenAI.CreateConversationItemsRequest? Type2705 { get; set; } + public global::tryAGI.OpenAI.RealtimeResponseCreateResponseTool? Type2705 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UpdateEvalRequest? Type2706 { get; set; } + public global::tryAGI.OpenAI.OneOf? Type2706 { get; set; } /// /// /// - public global::tryAGI.OpenAI.AdminApiKeysCreateRequest? Type2707 { get; set; } + public global::tryAGI.OpenAI.RealtimeResponseCreateResponseToolChoice? Type2707 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListAssistantsOrder? Type2708 { get; set; } + public global::tryAGI.OpenAI.OneOf? Type2708 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListChatCompletionsOrder? Type2709 { get; set; } + public global::tryAGI.OpenAI.RealtimeResponseCreateResponseMaxOutputTokens? Type2709 { get; set; } /// /// /// - public global::tryAGI.OpenAI.GetChatCompletionMessagesOrder? Type2710 { get; set; } + public global::tryAGI.OpenAI.RealtimeResponseCancel? Type2710 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListContainersOrder? Type2711 { get; set; } + public global::tryAGI.OpenAI.RealtimeResponseCancelType? Type2711 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListContainerFilesOrder? Type2712 { get; set; } + public global::tryAGI.OpenAI.CreateBatchRequest? Type2712 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListConversationItemsOrder? Type2713 { get; set; } + public global::tryAGI.OpenAI.CreateBatchRequestEndpoint? Type2713 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListEvalsOrder? Type2714 { get; set; } + public global::tryAGI.OpenAI.CreateBatchRequestCompletionWindow? Type2714 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListEvalsOrderBy? Type2715 { get; set; } + public global::tryAGI.OpenAI.UpdateChatCompletionRequest? Type2715 { get; set; } /// /// /// - public global::tryAGI.OpenAI.GetEvalRunsOrder? Type2716 { get; set; } + public global::tryAGI.OpenAI.CreateConversationItemsRequest? Type2716 { get; set; } /// /// /// - public global::tryAGI.OpenAI.GetEvalRunsStatus? Type2717 { get; set; } + public global::tryAGI.OpenAI.UpdateEvalRequest? Type2717 { get; set; } /// /// /// - public global::tryAGI.OpenAI.GetEvalRunOutputItemsStatus? Type2718 { get; set; } + public global::tryAGI.OpenAI.AdminApiKeysCreateRequest? Type2718 { get; set; } /// /// /// - public global::tryAGI.OpenAI.GetEvalRunOutputItemsOrder? Type2719 { get; set; } + public global::tryAGI.OpenAI.ListAssistantsOrder? Type2719 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListFilesOrder? Type2720 { get; set; } + public global::tryAGI.OpenAI.ListChatCompletionsOrder? Type2720 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListFineTuningCheckpointPermissionsOrder? Type2721 { get; set; } + public global::tryAGI.OpenAI.GetChatCompletionMessagesOrder? Type2721 { get; set; } /// /// /// - public global::tryAGI.OpenAI.AdminApiKeysListOrder? Type2722 { get; set; } + public global::tryAGI.OpenAI.ListContainersOrder? Type2722 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListAuditLogsEffectiveAt? Type2723 { get; set; } + public global::tryAGI.OpenAI.ListContainerFilesOrder? Type2723 { get; set; } /// /// /// - public global::System.Collections.Generic.IList? Type2724 { get; set; } + public global::tryAGI.OpenAI.ListConversationItemsOrder? Type2724 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListOrganizationCertificatesOrder? Type2725 { get; set; } + public global::tryAGI.OpenAI.ListEvalsOrder? Type2725 { get; set; } /// /// /// - public global::System.Collections.Generic.IList? Type2726 { get; set; } + public global::tryAGI.OpenAI.ListEvalsOrderBy? Type2726 { get; set; } /// /// /// - public global::tryAGI.OpenAI.GetCertificateIncludeItem? Type2727 { get; set; } + public global::tryAGI.OpenAI.GetEvalRunsOrder? Type2727 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UsageCostsBucketWidth? Type2728 { get; set; } + public global::tryAGI.OpenAI.GetEvalRunsStatus? Type2728 { get; set; } /// /// /// - public global::System.Collections.Generic.IList? Type2729 { get; set; } + public global::tryAGI.OpenAI.GetEvalRunOutputItemsStatus? Type2729 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UsageCostsGroupByItem? Type2730 { get; set; } + public global::tryAGI.OpenAI.GetEvalRunOutputItemsOrder? Type2730 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListGroupsOrder? Type2731 { get; set; } + public global::tryAGI.OpenAI.ListFilesOrder? Type2731 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListGroupRoleAssignmentsOrder? Type2732 { get; set; } + public global::tryAGI.OpenAI.ListFineTuningCheckpointPermissionsOrder? Type2732 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListGroupUsersOrder? Type2733 { get; set; } + public global::tryAGI.OpenAI.AdminApiKeysListOrder? Type2733 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListProjectCertificatesOrder? Type2734 { get; set; } + public global::tryAGI.OpenAI.ListAuditLogsEffectiveAt? Type2734 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListProjectGroupsOrder? Type2735 { get; set; } + public global::System.Collections.Generic.IList? Type2735 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListRolesOrder? Type2736 { get; set; } + public global::tryAGI.OpenAI.ListOrganizationCertificatesOrder? Type2736 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UsageAudioSpeechesBucketWidth? Type2737 { get; set; } + public global::System.Collections.Generic.IList? Type2737 { get; set; } /// /// /// - public global::System.Collections.Generic.IList? Type2738 { get; set; } + public global::tryAGI.OpenAI.GetCertificateIncludeItem? Type2738 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UsageAudioSpeechesGroupByItem? Type2739 { get; set; } + public global::tryAGI.OpenAI.UsageCostsBucketWidth? Type2739 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UsageAudioTranscriptionsBucketWidth? Type2740 { get; set; } + public global::System.Collections.Generic.IList? Type2740 { get; set; } /// /// /// - public global::System.Collections.Generic.IList? Type2741 { get; set; } + public global::tryAGI.OpenAI.UsageCostsGroupByItem? Type2741 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UsageAudioTranscriptionsGroupByItem? Type2742 { get; set; } + public global::tryAGI.OpenAI.ListGroupsOrder? Type2742 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UsageCodeInterpreterSessionsBucketWidth? Type2743 { get; set; } + public global::tryAGI.OpenAI.ListGroupRoleAssignmentsOrder? Type2743 { get; set; } /// /// /// - public global::System.Collections.Generic.IList? Type2744 { get; set; } + public global::tryAGI.OpenAI.ListGroupUsersOrder? Type2744 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UsageCodeInterpreterSessionsGroupByItem? Type2745 { get; set; } + public global::tryAGI.OpenAI.ListProjectCertificatesOrder? Type2745 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UsageCompletionsBucketWidth? Type2746 { get; set; } + public global::tryAGI.OpenAI.ListProjectGroupsOrder? Type2746 { get; set; } /// /// /// - public global::System.Collections.Generic.IList? Type2747 { get; set; } + public global::tryAGI.OpenAI.ListRolesOrder? Type2747 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UsageCompletionsGroupByItem? Type2748 { get; set; } + public global::tryAGI.OpenAI.UsageAudioSpeechesBucketWidth? Type2748 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UsageEmbeddingsBucketWidth? Type2749 { get; set; } + public global::System.Collections.Generic.IList? Type2749 { get; set; } /// /// /// - public global::System.Collections.Generic.IList? Type2750 { get; set; } + public global::tryAGI.OpenAI.UsageAudioSpeechesGroupByItem? Type2750 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UsageEmbeddingsGroupByItem? Type2751 { get; set; } + public global::tryAGI.OpenAI.UsageAudioTranscriptionsBucketWidth? Type2751 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UsageImagesBucketWidth? Type2752 { get; set; } + public global::System.Collections.Generic.IList? Type2752 { get; set; } /// /// /// - public global::System.Collections.Generic.IList? Type2753 { get; set; } + public global::tryAGI.OpenAI.UsageAudioTranscriptionsGroupByItem? Type2753 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UsageImagesSource? Type2754 { get; set; } + public global::tryAGI.OpenAI.UsageCodeInterpreterSessionsBucketWidth? Type2754 { get; set; } /// /// /// - public global::System.Collections.Generic.IList? Type2755 { get; set; } + public global::System.Collections.Generic.IList? Type2755 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UsageImagesSize? Type2756 { get; set; } + public global::tryAGI.OpenAI.UsageCodeInterpreterSessionsGroupByItem? Type2756 { get; set; } /// /// /// - public global::System.Collections.Generic.IList? Type2757 { get; set; } + public global::tryAGI.OpenAI.UsageCompletionsBucketWidth? Type2757 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UsageImagesGroupByItem? Type2758 { get; set; } + public global::System.Collections.Generic.IList? Type2758 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UsageModerationsBucketWidth? Type2759 { get; set; } + public global::tryAGI.OpenAI.UsageCompletionsGroupByItem? Type2759 { get; set; } /// /// /// - public global::System.Collections.Generic.IList? Type2760 { get; set; } + public global::tryAGI.OpenAI.UsageEmbeddingsBucketWidth? Type2760 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UsageModerationsGroupByItem? Type2761 { get; set; } + public global::System.Collections.Generic.IList? Type2761 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UsageVectorStoresBucketWidth? Type2762 { get; set; } + public global::tryAGI.OpenAI.UsageEmbeddingsGroupByItem? Type2762 { get; set; } /// /// /// - public global::System.Collections.Generic.IList? Type2763 { get; set; } + public global::tryAGI.OpenAI.UsageImagesBucketWidth? Type2763 { get; set; } /// /// /// - public global::tryAGI.OpenAI.UsageVectorStoresGroupByItem? Type2764 { get; set; } + public global::System.Collections.Generic.IList? Type2764 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListUserRoleAssignmentsOrder? Type2765 { get; set; } + public global::tryAGI.OpenAI.UsageImagesSource? Type2765 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListProjectGroupRoleAssignmentsOrder? Type2766 { get; set; } + public global::System.Collections.Generic.IList? Type2766 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListProjectRolesOrder? Type2767 { get; set; } + public global::tryAGI.OpenAI.UsageImagesSize? Type2767 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListProjectUserRoleAssignmentsOrder? Type2768 { get; set; } + public global::System.Collections.Generic.IList? Type2768 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListInputItemsOrder? Type2769 { get; set; } + public global::tryAGI.OpenAI.UsageImagesGroupByItem? Type2769 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListMessagesOrder? Type2770 { get; set; } + public global::tryAGI.OpenAI.UsageModerationsBucketWidth? Type2770 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListRunsOrder? Type2771 { get; set; } + public global::System.Collections.Generic.IList? Type2771 { get; set; } /// /// /// - public global::System.Collections.Generic.IList? Type2772 { get; set; } + public global::tryAGI.OpenAI.UsageModerationsGroupByItem? Type2772 { get; set; } /// /// /// - public global::tryAGI.OpenAI.CreateRunIncludeItem? Type2773 { get; set; } + public global::tryAGI.OpenAI.UsageVectorStoresBucketWidth? Type2773 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListRunStepsOrder? Type2774 { get; set; } + public global::System.Collections.Generic.IList? Type2774 { get; set; } /// /// /// - public global::System.Collections.Generic.IList? Type2775 { get; set; } + public global::tryAGI.OpenAI.UsageVectorStoresGroupByItem? Type2775 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListRunStepsIncludeItem? Type2776 { get; set; } + public global::tryAGI.OpenAI.ListUserRoleAssignmentsOrder? Type2776 { get; set; } /// /// /// - public global::System.Collections.Generic.IList? Type2777 { get; set; } + public global::tryAGI.OpenAI.ListProjectGroupRoleAssignmentsOrder? Type2777 { get; set; } /// /// /// - public global::tryAGI.OpenAI.GetRunStepIncludeItem? Type2778 { get; set; } + public global::tryAGI.OpenAI.ListProjectRolesOrder? Type2778 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListVectorStoresOrder? Type2779 { get; set; } + public global::tryAGI.OpenAI.ListProjectUserRoleAssignmentsOrder? Type2779 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListFilesInVectorStoreBatchOrder? Type2780 { get; set; } + public global::tryAGI.OpenAI.ListInputItemsOrder? Type2780 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListFilesInVectorStoreBatchFilter? Type2781 { get; set; } + public global::tryAGI.OpenAI.ListMessagesOrder? Type2781 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListVectorStoreFilesOrder? Type2782 { get; set; } + public global::tryAGI.OpenAI.ListRunsOrder? Type2782 { get; set; } /// /// /// - public global::tryAGI.OpenAI.ListVectorStoreFilesFilter? Type2783 { get; set; } + public global::System.Collections.Generic.IList? Type2783 { get; set; } /// /// /// - public global::tryAGI.OpenAI.OneOf? Type2784 { get; set; } + public global::tryAGI.OpenAI.CreateRunIncludeItem? Type2784 { get; set; } /// /// /// - public global::tryAGI.OpenAI.OneOf? Type2785 { get; set; } + public global::tryAGI.OpenAI.ListRunStepsOrder? Type2785 { get; set; } /// /// /// - public global::tryAGI.OpenAI.DeleteEvalResponse? Type2786 { get; set; } + public global::System.Collections.Generic.IList? Type2786 { get; set; } /// /// /// - public global::tryAGI.OpenAI.DeleteEvalRunResponse? Type2787 { get; set; } + public global::tryAGI.OpenAI.ListRunStepsIncludeItem? Type2787 { get; set; } /// /// /// - public global::tryAGI.OpenAI.AdminApiKeysDeleteResponse? Type2788 { get; set; } + public global::System.Collections.Generic.IList? Type2788 { get; set; } + /// + /// + /// + public global::tryAGI.OpenAI.GetRunStepIncludeItem? Type2789 { get; set; } + /// + /// + /// + public global::tryAGI.OpenAI.ListVectorStoresOrder? Type2790 { get; set; } + /// + /// + /// + public global::tryAGI.OpenAI.ListFilesInVectorStoreBatchOrder? Type2791 { get; set; } + /// + /// + /// + public global::tryAGI.OpenAI.ListFilesInVectorStoreBatchFilter? Type2792 { get; set; } + /// + /// + /// + public global::tryAGI.OpenAI.ListVectorStoreFilesOrder? Type2793 { get; set; } + /// + /// + /// + public global::tryAGI.OpenAI.ListVectorStoreFilesFilter? Type2794 { get; set; } + /// + /// + /// + public global::tryAGI.OpenAI.OneOf? Type2795 { get; set; } + /// + /// + /// + public global::tryAGI.OpenAI.OneOf? Type2796 { get; set; } + /// + /// + /// + public global::tryAGI.OpenAI.DeleteEvalResponse? Type2797 { get; set; } + /// + /// + /// + public global::tryAGI.OpenAI.DeleteEvalRunResponse? Type2798 { get; set; } + /// + /// + /// + public global::tryAGI.OpenAI.AdminApiKeysDeleteResponse? Type2799 { get; set; } /// /// diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoCharacterBody.Json.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoCharacterBody.Json.g.cs new file mode 100644 index 00000000..7e29683f --- /dev/null +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoCharacterBody.Json.g.cs @@ -0,0 +1,92 @@ +#nullable enable + +namespace tryAGI.OpenAI +{ + public sealed partial class CreateVideoCharacterBody + { + /// + /// Serializes the current instance to a JSON string using the provided JsonSerializerContext. + /// + public string ToJson( + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return global::System.Text.Json.JsonSerializer.Serialize( + this, + this.GetType(), + jsonSerializerContext); + } + + /// + /// Serializes the current instance to a JSON string using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public string ToJson( + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.Serialize( + this, + jsonSerializerOptions); + } + + /// + /// Deserializes a JSON string using the provided JsonSerializerContext. + /// + public static global::tryAGI.OpenAI.CreateVideoCharacterBody? FromJson( + string json, + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return global::System.Text.Json.JsonSerializer.Deserialize( + json, + typeof(global::tryAGI.OpenAI.CreateVideoCharacterBody), + jsonSerializerContext) as global::tryAGI.OpenAI.CreateVideoCharacterBody; + } + + /// + /// Deserializes a JSON string using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public static global::tryAGI.OpenAI.CreateVideoCharacterBody? FromJson( + string json, + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.Deserialize( + json, + jsonSerializerOptions); + } + + /// + /// Deserializes a JSON stream using the provided JsonSerializerContext. + /// + public static async global::System.Threading.Tasks.ValueTask FromJsonStreamAsync( + global::System.IO.Stream jsonStream, + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return (await global::System.Text.Json.JsonSerializer.DeserializeAsync( + jsonStream, + typeof(global::tryAGI.OpenAI.CreateVideoCharacterBody), + jsonSerializerContext).ConfigureAwait(false)) as global::tryAGI.OpenAI.CreateVideoCharacterBody; + } + + /// + /// Deserializes a JSON stream using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public static global::System.Threading.Tasks.ValueTask FromJsonStreamAsync( + global::System.IO.Stream jsonStream, + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.DeserializeAsync( + jsonStream, + jsonSerializerOptions); + } + } +} diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoCharacterBody.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoCharacterBody.g.cs new file mode 100644 index 00000000..f14253f8 --- /dev/null +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoCharacterBody.g.cs @@ -0,0 +1,70 @@ + +#nullable enable + +namespace tryAGI.OpenAI +{ + /// + /// Parameters for creating a character from an uploaded video. + /// + public sealed partial class CreateVideoCharacterBody + { + /// + /// Video file used to create a character. + /// + [global::System.Text.Json.Serialization.JsonPropertyName("video")] + [global::System.Text.Json.Serialization.JsonRequired] + public required byte[] Video { get; set; } + + /// + /// Video file used to create a character. + /// + [global::System.Text.Json.Serialization.JsonPropertyName("videoname")] + [global::System.Text.Json.Serialization.JsonRequired] + public required string Videoname { get; set; } + + /// + /// Display name for this API character. + /// + [global::System.Text.Json.Serialization.JsonPropertyName("name")] + [global::System.Text.Json.Serialization.JsonRequired] + public required string Name { get; set; } + + /// + /// Additional properties that are not explicitly defined in the schema + /// + [global::System.Text.Json.Serialization.JsonExtensionData] + public global::System.Collections.Generic.IDictionary AdditionalProperties { get; set; } = new global::System.Collections.Generic.Dictionary(); + + /// + /// Initializes a new instance of the class. + /// + /// + /// Video file used to create a character. + /// + /// + /// Video file used to create a character. + /// + /// + /// Display name for this API character. + /// +#if NET7_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.SetsRequiredMembers] +#endif + public CreateVideoCharacterBody( + byte[] video, + string videoname, + string name) + { + this.Video = video ?? throw new global::System.ArgumentNullException(nameof(video)); + this.Videoname = videoname ?? throw new global::System.ArgumentNullException(nameof(videoname)); + this.Name = name ?? throw new global::System.ArgumentNullException(nameof(name)); + } + + /// + /// Initializes a new instance of the class. + /// + public CreateVideoCharacterBody() + { + } + } +} \ No newline at end of file diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoEditJsonBody.Json.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoEditJsonBody.Json.g.cs new file mode 100644 index 00000000..47fab4a6 --- /dev/null +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoEditJsonBody.Json.g.cs @@ -0,0 +1,92 @@ +#nullable enable + +namespace tryAGI.OpenAI +{ + public sealed partial class CreateVideoEditJsonBody + { + /// + /// Serializes the current instance to a JSON string using the provided JsonSerializerContext. + /// + public string ToJson( + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return global::System.Text.Json.JsonSerializer.Serialize( + this, + this.GetType(), + jsonSerializerContext); + } + + /// + /// Serializes the current instance to a JSON string using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public string ToJson( + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.Serialize( + this, + jsonSerializerOptions); + } + + /// + /// Deserializes a JSON string using the provided JsonSerializerContext. + /// + public static global::tryAGI.OpenAI.CreateVideoEditJsonBody? FromJson( + string json, + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return global::System.Text.Json.JsonSerializer.Deserialize( + json, + typeof(global::tryAGI.OpenAI.CreateVideoEditJsonBody), + jsonSerializerContext) as global::tryAGI.OpenAI.CreateVideoEditJsonBody; + } + + /// + /// Deserializes a JSON string using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public static global::tryAGI.OpenAI.CreateVideoEditJsonBody? FromJson( + string json, + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.Deserialize( + json, + jsonSerializerOptions); + } + + /// + /// Deserializes a JSON stream using the provided JsonSerializerContext. + /// + public static async global::System.Threading.Tasks.ValueTask FromJsonStreamAsync( + global::System.IO.Stream jsonStream, + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return (await global::System.Text.Json.JsonSerializer.DeserializeAsync( + jsonStream, + typeof(global::tryAGI.OpenAI.CreateVideoEditJsonBody), + jsonSerializerContext).ConfigureAwait(false)) as global::tryAGI.OpenAI.CreateVideoEditJsonBody; + } + + /// + /// Deserializes a JSON stream using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public static global::System.Threading.Tasks.ValueTask FromJsonStreamAsync( + global::System.IO.Stream jsonStream, + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.DeserializeAsync( + jsonStream, + jsonSerializerOptions); + } + } +} diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoEditJsonBody.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoEditJsonBody.g.cs new file mode 100644 index 00000000..17c55c9a --- /dev/null +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoEditJsonBody.g.cs @@ -0,0 +1,58 @@ + +#nullable enable + +namespace tryAGI.OpenAI +{ + /// + /// JSON parameters for editing an existing generated video. + /// + public sealed partial class CreateVideoEditJsonBody + { + /// + /// Reference to the completed video to edit. + /// + [global::System.Text.Json.Serialization.JsonPropertyName("video")] + [global::System.Text.Json.Serialization.JsonRequired] + public required global::tryAGI.OpenAI.VideoReferenceInputParam Video { get; set; } + + /// + /// Text prompt that describes how to edit the source video. + /// + [global::System.Text.Json.Serialization.JsonPropertyName("prompt")] + [global::System.Text.Json.Serialization.JsonRequired] + public required string Prompt { get; set; } + + /// + /// Additional properties that are not explicitly defined in the schema + /// + [global::System.Text.Json.Serialization.JsonExtensionData] + public global::System.Collections.Generic.IDictionary AdditionalProperties { get; set; } = new global::System.Collections.Generic.Dictionary(); + + /// + /// Initializes a new instance of the class. + /// + /// + /// Reference to the completed video to edit. + /// + /// + /// Text prompt that describes how to edit the source video. + /// +#if NET7_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.SetsRequiredMembers] +#endif + public CreateVideoEditJsonBody( + global::tryAGI.OpenAI.VideoReferenceInputParam video, + string prompt) + { + this.Video = video ?? throw new global::System.ArgumentNullException(nameof(video)); + this.Prompt = prompt ?? throw new global::System.ArgumentNullException(nameof(prompt)); + } + + /// + /// Initializes a new instance of the class. + /// + public CreateVideoEditJsonBody() + { + } + } +} \ No newline at end of file diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoEditMultipartBody.Json.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoEditMultipartBody.Json.g.cs new file mode 100644 index 00000000..99ccf728 --- /dev/null +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoEditMultipartBody.Json.g.cs @@ -0,0 +1,92 @@ +#nullable enable + +namespace tryAGI.OpenAI +{ + public sealed partial class CreateVideoEditMultipartBody + { + /// + /// Serializes the current instance to a JSON string using the provided JsonSerializerContext. + /// + public string ToJson( + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return global::System.Text.Json.JsonSerializer.Serialize( + this, + this.GetType(), + jsonSerializerContext); + } + + /// + /// Serializes the current instance to a JSON string using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public string ToJson( + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.Serialize( + this, + jsonSerializerOptions); + } + + /// + /// Deserializes a JSON string using the provided JsonSerializerContext. + /// + public static global::tryAGI.OpenAI.CreateVideoEditMultipartBody? FromJson( + string json, + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return global::System.Text.Json.JsonSerializer.Deserialize( + json, + typeof(global::tryAGI.OpenAI.CreateVideoEditMultipartBody), + jsonSerializerContext) as global::tryAGI.OpenAI.CreateVideoEditMultipartBody; + } + + /// + /// Deserializes a JSON string using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public static global::tryAGI.OpenAI.CreateVideoEditMultipartBody? FromJson( + string json, + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.Deserialize( + json, + jsonSerializerOptions); + } + + /// + /// Deserializes a JSON stream using the provided JsonSerializerContext. + /// + public static async global::System.Threading.Tasks.ValueTask FromJsonStreamAsync( + global::System.IO.Stream jsonStream, + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return (await global::System.Text.Json.JsonSerializer.DeserializeAsync( + jsonStream, + typeof(global::tryAGI.OpenAI.CreateVideoEditMultipartBody), + jsonSerializerContext).ConfigureAwait(false)) as global::tryAGI.OpenAI.CreateVideoEditMultipartBody; + } + + /// + /// Deserializes a JSON stream using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public static global::System.Threading.Tasks.ValueTask FromJsonStreamAsync( + global::System.IO.Stream jsonStream, + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.DeserializeAsync( + jsonStream, + jsonSerializerOptions); + } + } +} diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoEditMultipartBody.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoEditMultipartBody.g.cs new file mode 100644 index 00000000..d58a6745 --- /dev/null +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoEditMultipartBody.g.cs @@ -0,0 +1,59 @@ + +#pragma warning disable CS0618 // Type or member is obsolete + +#nullable enable + +namespace tryAGI.OpenAI +{ + /// + /// Parameters for editing an existing generated video. + /// + public sealed partial class CreateVideoEditMultipartBody + { + /// + /// + /// + [global::System.Text.Json.Serialization.JsonPropertyName("video")] + [global::System.Text.Json.Serialization.JsonConverter(typeof(global::tryAGI.OpenAI.JsonConverters.OneOfJsonConverter))] + [global::System.Text.Json.Serialization.JsonRequired] + public required global::tryAGI.OpenAI.OneOf Video { get; set; } + + /// + /// Text prompt that describes how to edit the source video. + /// + [global::System.Text.Json.Serialization.JsonPropertyName("prompt")] + [global::System.Text.Json.Serialization.JsonRequired] + public required string Prompt { get; set; } + + /// + /// Additional properties that are not explicitly defined in the schema + /// + [global::System.Text.Json.Serialization.JsonExtensionData] + public global::System.Collections.Generic.IDictionary AdditionalProperties { get; set; } = new global::System.Collections.Generic.Dictionary(); + + /// + /// Initializes a new instance of the class. + /// + /// + /// + /// Text prompt that describes how to edit the source video. + /// +#if NET7_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.SetsRequiredMembers] +#endif + public CreateVideoEditMultipartBody( + global::tryAGI.OpenAI.OneOf video, + string prompt) + { + this.Video = video; + this.Prompt = prompt ?? throw new global::System.ArgumentNullException(nameof(prompt)); + } + + /// + /// Initializes a new instance of the class. + /// + public CreateVideoEditMultipartBody() + { + } + } +} \ No newline at end of file diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoExtendJsonBody.Json.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoExtendJsonBody.Json.g.cs new file mode 100644 index 00000000..299e8f63 --- /dev/null +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoExtendJsonBody.Json.g.cs @@ -0,0 +1,92 @@ +#nullable enable + +namespace tryAGI.OpenAI +{ + public sealed partial class CreateVideoExtendJsonBody + { + /// + /// Serializes the current instance to a JSON string using the provided JsonSerializerContext. + /// + public string ToJson( + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return global::System.Text.Json.JsonSerializer.Serialize( + this, + this.GetType(), + jsonSerializerContext); + } + + /// + /// Serializes the current instance to a JSON string using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public string ToJson( + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.Serialize( + this, + jsonSerializerOptions); + } + + /// + /// Deserializes a JSON string using the provided JsonSerializerContext. + /// + public static global::tryAGI.OpenAI.CreateVideoExtendJsonBody? FromJson( + string json, + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return global::System.Text.Json.JsonSerializer.Deserialize( + json, + typeof(global::tryAGI.OpenAI.CreateVideoExtendJsonBody), + jsonSerializerContext) as global::tryAGI.OpenAI.CreateVideoExtendJsonBody; + } + + /// + /// Deserializes a JSON string using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public static global::tryAGI.OpenAI.CreateVideoExtendJsonBody? FromJson( + string json, + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.Deserialize( + json, + jsonSerializerOptions); + } + + /// + /// Deserializes a JSON stream using the provided JsonSerializerContext. + /// + public static async global::System.Threading.Tasks.ValueTask FromJsonStreamAsync( + global::System.IO.Stream jsonStream, + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return (await global::System.Text.Json.JsonSerializer.DeserializeAsync( + jsonStream, + typeof(global::tryAGI.OpenAI.CreateVideoExtendJsonBody), + jsonSerializerContext).ConfigureAwait(false)) as global::tryAGI.OpenAI.CreateVideoExtendJsonBody; + } + + /// + /// Deserializes a JSON stream using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public static global::System.Threading.Tasks.ValueTask FromJsonStreamAsync( + global::System.IO.Stream jsonStream, + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.DeserializeAsync( + jsonStream, + jsonSerializerOptions); + } + } +} diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoExtendJsonBody.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoExtendJsonBody.g.cs new file mode 100644 index 00000000..455ee833 --- /dev/null +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoExtendJsonBody.g.cs @@ -0,0 +1,71 @@ + +#nullable enable + +namespace tryAGI.OpenAI +{ + /// + /// JSON parameters for extending an existing generated video. + /// + public sealed partial class CreateVideoExtendJsonBody + { + /// + /// Reference to the completed video to extend. + /// + [global::System.Text.Json.Serialization.JsonPropertyName("video")] + [global::System.Text.Json.Serialization.JsonRequired] + public required global::tryAGI.OpenAI.VideoReferenceInputParam Video { get; set; } + + /// + /// Updated text prompt that directs the extension generation. + /// + [global::System.Text.Json.Serialization.JsonPropertyName("prompt")] + [global::System.Text.Json.Serialization.JsonRequired] + public required string Prompt { get; set; } + + /// + /// Length of the newly generated extension segment in seconds (allowed values: 4, 8, 12, 16, 20). + /// + [global::System.Text.Json.Serialization.JsonPropertyName("seconds")] + [global::System.Text.Json.Serialization.JsonConverter(typeof(global::tryAGI.OpenAI.JsonConverters.VideoSecondsJsonConverter))] + [global::System.Text.Json.Serialization.JsonRequired] + public required global::tryAGI.OpenAI.VideoSeconds Seconds { get; set; } + + /// + /// Additional properties that are not explicitly defined in the schema + /// + [global::System.Text.Json.Serialization.JsonExtensionData] + public global::System.Collections.Generic.IDictionary AdditionalProperties { get; set; } = new global::System.Collections.Generic.Dictionary(); + + /// + /// Initializes a new instance of the class. + /// + /// + /// Reference to the completed video to extend. + /// + /// + /// Updated text prompt that directs the extension generation. + /// + /// + /// Length of the newly generated extension segment in seconds (allowed values: 4, 8, 12, 16, 20). + /// +#if NET7_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.SetsRequiredMembers] +#endif + public CreateVideoExtendJsonBody( + global::tryAGI.OpenAI.VideoReferenceInputParam video, + string prompt, + global::tryAGI.OpenAI.VideoSeconds seconds) + { + this.Video = video ?? throw new global::System.ArgumentNullException(nameof(video)); + this.Prompt = prompt ?? throw new global::System.ArgumentNullException(nameof(prompt)); + this.Seconds = seconds; + } + + /// + /// Initializes a new instance of the class. + /// + public CreateVideoExtendJsonBody() + { + } + } +} \ No newline at end of file diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoExtendMultipartBody.Json.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoExtendMultipartBody.Json.g.cs new file mode 100644 index 00000000..6c8e5a86 --- /dev/null +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoExtendMultipartBody.Json.g.cs @@ -0,0 +1,92 @@ +#nullable enable + +namespace tryAGI.OpenAI +{ + public sealed partial class CreateVideoExtendMultipartBody + { + /// + /// Serializes the current instance to a JSON string using the provided JsonSerializerContext. + /// + public string ToJson( + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return global::System.Text.Json.JsonSerializer.Serialize( + this, + this.GetType(), + jsonSerializerContext); + } + + /// + /// Serializes the current instance to a JSON string using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public string ToJson( + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.Serialize( + this, + jsonSerializerOptions); + } + + /// + /// Deserializes a JSON string using the provided JsonSerializerContext. + /// + public static global::tryAGI.OpenAI.CreateVideoExtendMultipartBody? FromJson( + string json, + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return global::System.Text.Json.JsonSerializer.Deserialize( + json, + typeof(global::tryAGI.OpenAI.CreateVideoExtendMultipartBody), + jsonSerializerContext) as global::tryAGI.OpenAI.CreateVideoExtendMultipartBody; + } + + /// + /// Deserializes a JSON string using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public static global::tryAGI.OpenAI.CreateVideoExtendMultipartBody? FromJson( + string json, + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.Deserialize( + json, + jsonSerializerOptions); + } + + /// + /// Deserializes a JSON stream using the provided JsonSerializerContext. + /// + public static async global::System.Threading.Tasks.ValueTask FromJsonStreamAsync( + global::System.IO.Stream jsonStream, + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return (await global::System.Text.Json.JsonSerializer.DeserializeAsync( + jsonStream, + typeof(global::tryAGI.OpenAI.CreateVideoExtendMultipartBody), + jsonSerializerContext).ConfigureAwait(false)) as global::tryAGI.OpenAI.CreateVideoExtendMultipartBody; + } + + /// + /// Deserializes a JSON stream using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public static global::System.Threading.Tasks.ValueTask FromJsonStreamAsync( + global::System.IO.Stream jsonStream, + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.DeserializeAsync( + jsonStream, + jsonSerializerOptions); + } + } +} diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoExtendMultipartBody.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoExtendMultipartBody.g.cs new file mode 100644 index 00000000..3c3e6130 --- /dev/null +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoExtendMultipartBody.g.cs @@ -0,0 +1,72 @@ + +#pragma warning disable CS0618 // Type or member is obsolete + +#nullable enable + +namespace tryAGI.OpenAI +{ + /// + /// Multipart parameters for extending an existing generated video. + /// + public sealed partial class CreateVideoExtendMultipartBody + { + /// + /// + /// + [global::System.Text.Json.Serialization.JsonPropertyName("video")] + [global::System.Text.Json.Serialization.JsonConverter(typeof(global::tryAGI.OpenAI.JsonConverters.OneOfJsonConverter))] + [global::System.Text.Json.Serialization.JsonRequired] + public required global::tryAGI.OpenAI.OneOf Video { get; set; } + + /// + /// Updated text prompt that directs the extension generation. + /// + [global::System.Text.Json.Serialization.JsonPropertyName("prompt")] + [global::System.Text.Json.Serialization.JsonRequired] + public required string Prompt { get; set; } + + /// + /// Length of the newly generated extension segment in seconds (allowed values: 4, 8, 12, 16, 20). + /// + [global::System.Text.Json.Serialization.JsonPropertyName("seconds")] + [global::System.Text.Json.Serialization.JsonConverter(typeof(global::tryAGI.OpenAI.JsonConverters.VideoSecondsJsonConverter))] + [global::System.Text.Json.Serialization.JsonRequired] + public required global::tryAGI.OpenAI.VideoSeconds Seconds { get; set; } + + /// + /// Additional properties that are not explicitly defined in the schema + /// + [global::System.Text.Json.Serialization.JsonExtensionData] + public global::System.Collections.Generic.IDictionary AdditionalProperties { get; set; } = new global::System.Collections.Generic.Dictionary(); + + /// + /// Initializes a new instance of the class. + /// + /// + /// + /// Updated text prompt that directs the extension generation. + /// + /// + /// Length of the newly generated extension segment in seconds (allowed values: 4, 8, 12, 16, 20). + /// +#if NET7_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.SetsRequiredMembers] +#endif + public CreateVideoExtendMultipartBody( + global::tryAGI.OpenAI.OneOf video, + string prompt, + global::tryAGI.OpenAI.VideoSeconds seconds) + { + this.Video = video; + this.Prompt = prompt ?? throw new global::System.ArgumentNullException(nameof(prompt)); + this.Seconds = seconds; + } + + /// + /// Initializes a new instance of the class. + /// + public CreateVideoExtendMultipartBody() + { + } + } +} \ No newline at end of file diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoBody.Json.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoJsonBody.Json.g.cs similarity index 87% rename from src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoBody.Json.g.cs rename to src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoJsonBody.Json.g.cs index b67212b0..f18cba21 100644 --- a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoBody.Json.g.cs +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoJsonBody.Json.g.cs @@ -2,7 +2,7 @@ namespace tryAGI.OpenAI { - public sealed partial class CreateVideoBody + public sealed partial class CreateVideoJsonBody { /// /// Serializes the current instance to a JSON string using the provided JsonSerializerContext. @@ -34,14 +34,14 @@ public string ToJson( /// /// Deserializes a JSON string using the provided JsonSerializerContext. /// - public static global::tryAGI.OpenAI.CreateVideoBody? FromJson( + public static global::tryAGI.OpenAI.CreateVideoJsonBody? FromJson( string json, global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) { return global::System.Text.Json.JsonSerializer.Deserialize( json, - typeof(global::tryAGI.OpenAI.CreateVideoBody), - jsonSerializerContext) as global::tryAGI.OpenAI.CreateVideoBody; + typeof(global::tryAGI.OpenAI.CreateVideoJsonBody), + jsonSerializerContext) as global::tryAGI.OpenAI.CreateVideoJsonBody; } /// @@ -51,11 +51,11 @@ public string ToJson( [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] #endif - public static global::tryAGI.OpenAI.CreateVideoBody? FromJson( + public static global::tryAGI.OpenAI.CreateVideoJsonBody? FromJson( string json, global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) { - return global::System.Text.Json.JsonSerializer.Deserialize( + return global::System.Text.Json.JsonSerializer.Deserialize( json, jsonSerializerOptions); } @@ -63,14 +63,14 @@ public string ToJson( /// /// Deserializes a JSON stream using the provided JsonSerializerContext. /// - public static async global::System.Threading.Tasks.ValueTask FromJsonStreamAsync( + public static async global::System.Threading.Tasks.ValueTask FromJsonStreamAsync( global::System.IO.Stream jsonStream, global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) { return (await global::System.Text.Json.JsonSerializer.DeserializeAsync( jsonStream, - typeof(global::tryAGI.OpenAI.CreateVideoBody), - jsonSerializerContext).ConfigureAwait(false)) as global::tryAGI.OpenAI.CreateVideoBody; + typeof(global::tryAGI.OpenAI.CreateVideoJsonBody), + jsonSerializerContext).ConfigureAwait(false)) as global::tryAGI.OpenAI.CreateVideoJsonBody; } /// @@ -80,11 +80,11 @@ public string ToJson( [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] #endif - public static global::System.Threading.Tasks.ValueTask FromJsonStreamAsync( + public static global::System.Threading.Tasks.ValueTask FromJsonStreamAsync( global::System.IO.Stream jsonStream, global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) { - return global::System.Text.Json.JsonSerializer.DeserializeAsync( + return global::System.Text.Json.JsonSerializer.DeserializeAsync( jsonStream, jsonSerializerOptions); } diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoBody.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoJsonBody.g.cs similarity index 70% rename from src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoBody.g.cs rename to src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoJsonBody.g.cs index 58c58c06..347a9796 100644 --- a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoBody.g.cs +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoJsonBody.g.cs @@ -4,9 +4,9 @@ namespace tryAGI.OpenAI { /// - /// Parameters for creating a new video generation job. + /// JSON parameters for creating a new video generation job. /// - public sealed partial class CreateVideoBody + public sealed partial class CreateVideoJsonBody { /// /// The video generation model to use (allowed values: sora-2, sora-2-pro). Defaults to `sora-2`. @@ -23,22 +23,10 @@ public sealed partial class CreateVideoBody public required string Prompt { get; set; } /// - /// Optional multipart reference asset that guides generation. + /// Optional reference object that guides generation. Provide exactly one of `image_url` or `file_id`. /// [global::System.Text.Json.Serialization.JsonPropertyName("input_reference")] - public byte[]? InputReference { get; set; } - - /// - /// Optional multipart reference asset that guides generation. - /// - [global::System.Text.Json.Serialization.JsonPropertyName("input_referencename")] - public string? InputReferencename { get; set; } - - /// - /// Optional JSON-safe image reference that guides generation. Provide exactly one of `image_url` or `file_id`. - /// - [global::System.Text.Json.Serialization.JsonPropertyName("image_reference")] - public global::tryAGI.OpenAI.ImageRefParam2? ImageReference { get; set; } + public global::tryAGI.OpenAI.ImageRefParam2? InputReference { get; set; } /// /// Clip duration in seconds (allowed values: 4, 8, 12). Defaults to 4 seconds. @@ -61,7 +49,7 @@ public sealed partial class CreateVideoBody public global::System.Collections.Generic.IDictionary AdditionalProperties { get; set; } = new global::System.Collections.Generic.Dictionary(); /// - /// Initializes a new instance of the class. + /// Initializes a new instance of the class. /// /// /// The video generation model to use (allowed values: sora-2, sora-2-pro). Defaults to `sora-2`. @@ -70,13 +58,7 @@ public sealed partial class CreateVideoBody /// Text prompt that describes the video to generate. /// /// - /// Optional multipart reference asset that guides generation. - /// - /// - /// Optional multipart reference asset that guides generation. - /// - /// - /// Optional JSON-safe image reference that guides generation. Provide exactly one of `image_url` or `file_id`. + /// Optional reference object that guides generation. Provide exactly one of `image_url` or `file_id`. /// /// /// Clip duration in seconds (allowed values: 4, 8, 12). Defaults to 4 seconds. @@ -87,28 +69,24 @@ public sealed partial class CreateVideoBody #if NET7_0_OR_GREATER [global::System.Diagnostics.CodeAnalysis.SetsRequiredMembers] #endif - public CreateVideoBody( + public CreateVideoJsonBody( string prompt, global::tryAGI.OpenAI.VideoModel? model, - byte[]? inputReference, - string? inputReferencename, - global::tryAGI.OpenAI.ImageRefParam2? imageReference, + global::tryAGI.OpenAI.ImageRefParam2? inputReference, global::tryAGI.OpenAI.VideoSeconds? seconds, global::tryAGI.OpenAI.VideoSize? size) { this.Prompt = prompt ?? throw new global::System.ArgumentNullException(nameof(prompt)); this.Model = model; this.InputReference = inputReference; - this.InputReferencename = inputReferencename; - this.ImageReference = imageReference; this.Seconds = seconds; this.Size = size; } /// - /// Initializes a new instance of the class. + /// Initializes a new instance of the class. /// - public CreateVideoBody() + public CreateVideoJsonBody() { } } diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoMultipartBody.Json.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoMultipartBody.Json.g.cs new file mode 100644 index 00000000..deadfb56 --- /dev/null +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoMultipartBody.Json.g.cs @@ -0,0 +1,92 @@ +#nullable enable + +namespace tryAGI.OpenAI +{ + public sealed partial class CreateVideoMultipartBody + { + /// + /// Serializes the current instance to a JSON string using the provided JsonSerializerContext. + /// + public string ToJson( + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return global::System.Text.Json.JsonSerializer.Serialize( + this, + this.GetType(), + jsonSerializerContext); + } + + /// + /// Serializes the current instance to a JSON string using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public string ToJson( + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.Serialize( + this, + jsonSerializerOptions); + } + + /// + /// Deserializes a JSON string using the provided JsonSerializerContext. + /// + public static global::tryAGI.OpenAI.CreateVideoMultipartBody? FromJson( + string json, + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return global::System.Text.Json.JsonSerializer.Deserialize( + json, + typeof(global::tryAGI.OpenAI.CreateVideoMultipartBody), + jsonSerializerContext) as global::tryAGI.OpenAI.CreateVideoMultipartBody; + } + + /// + /// Deserializes a JSON string using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public static global::tryAGI.OpenAI.CreateVideoMultipartBody? FromJson( + string json, + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.Deserialize( + json, + jsonSerializerOptions); + } + + /// + /// Deserializes a JSON stream using the provided JsonSerializerContext. + /// + public static async global::System.Threading.Tasks.ValueTask FromJsonStreamAsync( + global::System.IO.Stream jsonStream, + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return (await global::System.Text.Json.JsonSerializer.DeserializeAsync( + jsonStream, + typeof(global::tryAGI.OpenAI.CreateVideoMultipartBody), + jsonSerializerContext).ConfigureAwait(false)) as global::tryAGI.OpenAI.CreateVideoMultipartBody; + } + + /// + /// Deserializes a JSON stream using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public static global::System.Threading.Tasks.ValueTask FromJsonStreamAsync( + global::System.IO.Stream jsonStream, + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.DeserializeAsync( + jsonStream, + jsonSerializerOptions); + } + } +} diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoMultipartBody.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoMultipartBody.g.cs new file mode 100644 index 00000000..de867c8a --- /dev/null +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.CreateVideoMultipartBody.g.cs @@ -0,0 +1,94 @@ + +#pragma warning disable CS0618 // Type or member is obsolete + +#nullable enable + +namespace tryAGI.OpenAI +{ + /// + /// Multipart parameters for creating a new video generation job. + /// + public sealed partial class CreateVideoMultipartBody + { + /// + /// The video generation model to use (allowed values: sora-2, sora-2-pro). Defaults to `sora-2`. + /// + [global::System.Text.Json.Serialization.JsonPropertyName("model")] + [global::System.Text.Json.Serialization.JsonConverter(typeof(global::tryAGI.OpenAI.JsonConverters.VideoModelJsonConverter))] + public global::tryAGI.OpenAI.VideoModel? Model { get; set; } + + /// + /// Text prompt that describes the video to generate. + /// + [global::System.Text.Json.Serialization.JsonPropertyName("prompt")] + [global::System.Text.Json.Serialization.JsonRequired] + public required string Prompt { get; set; } + + /// + /// + /// + [global::System.Text.Json.Serialization.JsonPropertyName("input_reference")] + [global::System.Text.Json.Serialization.JsonConverter(typeof(global::tryAGI.OpenAI.JsonConverters.OneOfJsonConverter))] + public global::tryAGI.OpenAI.OneOf? InputReference { get; set; } + + /// + /// Clip duration in seconds (allowed values: 4, 8, 12). Defaults to 4 seconds. + /// + [global::System.Text.Json.Serialization.JsonPropertyName("seconds")] + [global::System.Text.Json.Serialization.JsonConverter(typeof(global::tryAGI.OpenAI.JsonConverters.VideoSecondsJsonConverter))] + public global::tryAGI.OpenAI.VideoSeconds? Seconds { get; set; } + + /// + /// Output resolution formatted as width x height (allowed values: 720x1280, 1280x720, 1024x1792, 1792x1024). Defaults to 720x1280. + /// + [global::System.Text.Json.Serialization.JsonPropertyName("size")] + [global::System.Text.Json.Serialization.JsonConverter(typeof(global::tryAGI.OpenAI.JsonConverters.VideoSizeJsonConverter))] + public global::tryAGI.OpenAI.VideoSize? Size { get; set; } + + /// + /// Additional properties that are not explicitly defined in the schema + /// + [global::System.Text.Json.Serialization.JsonExtensionData] + public global::System.Collections.Generic.IDictionary AdditionalProperties { get; set; } = new global::System.Collections.Generic.Dictionary(); + + /// + /// Initializes a new instance of the class. + /// + /// + /// The video generation model to use (allowed values: sora-2, sora-2-pro). Defaults to `sora-2`. + /// + /// + /// Text prompt that describes the video to generate. + /// + /// + /// + /// Clip duration in seconds (allowed values: 4, 8, 12). Defaults to 4 seconds. + /// + /// + /// Output resolution formatted as width x height (allowed values: 720x1280, 1280x720, 1024x1792, 1792x1024). Defaults to 720x1280. + /// +#if NET7_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.SetsRequiredMembers] +#endif + public CreateVideoMultipartBody( + string prompt, + global::tryAGI.OpenAI.VideoModel? model, + global::tryAGI.OpenAI.OneOf? inputReference, + global::tryAGI.OpenAI.VideoSeconds? seconds, + global::tryAGI.OpenAI.VideoSize? size) + { + this.Prompt = prompt ?? throw new global::System.ArgumentNullException(nameof(prompt)); + this.Model = model; + this.InputReference = inputReference; + this.Seconds = seconds; + this.Size = size; + } + + /// + /// Initializes a new instance of the class. + /// + public CreateVideoMultipartBody() + { + } + } +} \ No newline at end of file diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.VideoCharacterResource.Json.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.VideoCharacterResource.Json.g.cs new file mode 100644 index 00000000..86859ddb --- /dev/null +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.VideoCharacterResource.Json.g.cs @@ -0,0 +1,92 @@ +#nullable enable + +namespace tryAGI.OpenAI +{ + public sealed partial class VideoCharacterResource + { + /// + /// Serializes the current instance to a JSON string using the provided JsonSerializerContext. + /// + public string ToJson( + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return global::System.Text.Json.JsonSerializer.Serialize( + this, + this.GetType(), + jsonSerializerContext); + } + + /// + /// Serializes the current instance to a JSON string using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public string ToJson( + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.Serialize( + this, + jsonSerializerOptions); + } + + /// + /// Deserializes a JSON string using the provided JsonSerializerContext. + /// + public static global::tryAGI.OpenAI.VideoCharacterResource? FromJson( + string json, + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return global::System.Text.Json.JsonSerializer.Deserialize( + json, + typeof(global::tryAGI.OpenAI.VideoCharacterResource), + jsonSerializerContext) as global::tryAGI.OpenAI.VideoCharacterResource; + } + + /// + /// Deserializes a JSON string using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public static global::tryAGI.OpenAI.VideoCharacterResource? FromJson( + string json, + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.Deserialize( + json, + jsonSerializerOptions); + } + + /// + /// Deserializes a JSON stream using the provided JsonSerializerContext. + /// + public static async global::System.Threading.Tasks.ValueTask FromJsonStreamAsync( + global::System.IO.Stream jsonStream, + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return (await global::System.Text.Json.JsonSerializer.DeserializeAsync( + jsonStream, + typeof(global::tryAGI.OpenAI.VideoCharacterResource), + jsonSerializerContext).ConfigureAwait(false)) as global::tryAGI.OpenAI.VideoCharacterResource; + } + + /// + /// Deserializes a JSON stream using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public static global::System.Threading.Tasks.ValueTask FromJsonStreamAsync( + global::System.IO.Stream jsonStream, + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.DeserializeAsync( + jsonStream, + jsonSerializerOptions); + } + } +} diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.VideoCharacterResource.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.VideoCharacterResource.g.cs new file mode 100644 index 00000000..f9d9f7a5 --- /dev/null +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.VideoCharacterResource.g.cs @@ -0,0 +1,65 @@ + +#nullable enable + +namespace tryAGI.OpenAI +{ + /// + /// + /// + public sealed partial class VideoCharacterResource + { + /// + /// + /// + [global::System.Text.Json.Serialization.JsonPropertyName("id")] + public string? Id { get; set; } + + /// + /// + /// + [global::System.Text.Json.Serialization.JsonPropertyName("name")] + public string? Name { get; set; } + + /// + /// Unix timestamp (in seconds) when the character was created. + /// + [global::System.Text.Json.Serialization.JsonPropertyName("created_at")] + [global::System.Text.Json.Serialization.JsonConverter(typeof(global::tryAGI.OpenAI.JsonConverters.UnixTimestampJsonConverter))] + [global::System.Text.Json.Serialization.JsonRequired] + public required global::System.DateTimeOffset CreatedAt { get; set; } + + /// + /// Additional properties that are not explicitly defined in the schema + /// + [global::System.Text.Json.Serialization.JsonExtensionData] + public global::System.Collections.Generic.IDictionary AdditionalProperties { get; set; } = new global::System.Collections.Generic.Dictionary(); + + /// + /// Initializes a new instance of the class. + /// + /// + /// + /// + /// Unix timestamp (in seconds) when the character was created. + /// +#if NET7_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.SetsRequiredMembers] +#endif + public VideoCharacterResource( + global::System.DateTimeOffset createdAt, + string? id, + string? name) + { + this.CreatedAt = createdAt; + this.Id = id; + this.Name = name; + } + + /// + /// Initializes a new instance of the class. + /// + public VideoCharacterResource() + { + } + } +} \ No newline at end of file diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.VideoCharacterResourceId.Json.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.VideoCharacterResourceId.Json.g.cs new file mode 100644 index 00000000..af4c52c1 --- /dev/null +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.VideoCharacterResourceId.Json.g.cs @@ -0,0 +1,92 @@ +#nullable enable + +namespace tryAGI.OpenAI +{ + public sealed partial class VideoCharacterResourceId + { + /// + /// Serializes the current instance to a JSON string using the provided JsonSerializerContext. + /// + public string ToJson( + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return global::System.Text.Json.JsonSerializer.Serialize( + this, + this.GetType(), + jsonSerializerContext); + } + + /// + /// Serializes the current instance to a JSON string using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public string ToJson( + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.Serialize( + this, + jsonSerializerOptions); + } + + /// + /// Deserializes a JSON string using the provided JsonSerializerContext. + /// + public static global::tryAGI.OpenAI.VideoCharacterResourceId? FromJson( + string json, + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return global::System.Text.Json.JsonSerializer.Deserialize( + json, + typeof(global::tryAGI.OpenAI.VideoCharacterResourceId), + jsonSerializerContext) as global::tryAGI.OpenAI.VideoCharacterResourceId; + } + + /// + /// Deserializes a JSON string using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public static global::tryAGI.OpenAI.VideoCharacterResourceId? FromJson( + string json, + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.Deserialize( + json, + jsonSerializerOptions); + } + + /// + /// Deserializes a JSON stream using the provided JsonSerializerContext. + /// + public static async global::System.Threading.Tasks.ValueTask FromJsonStreamAsync( + global::System.IO.Stream jsonStream, + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return (await global::System.Text.Json.JsonSerializer.DeserializeAsync( + jsonStream, + typeof(global::tryAGI.OpenAI.VideoCharacterResourceId), + jsonSerializerContext).ConfigureAwait(false)) as global::tryAGI.OpenAI.VideoCharacterResourceId; + } + + /// + /// Deserializes a JSON stream using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public static global::System.Threading.Tasks.ValueTask FromJsonStreamAsync( + global::System.IO.Stream jsonStream, + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.DeserializeAsync( + jsonStream, + jsonSerializerOptions); + } + } +} diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.VideoCharacterResourceId.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.VideoCharacterResourceId.g.cs new file mode 100644 index 00000000..cf08b44e --- /dev/null +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.VideoCharacterResourceId.g.cs @@ -0,0 +1,18 @@ + +#nullable enable + +namespace tryAGI.OpenAI +{ + /// + /// + /// + public sealed partial class VideoCharacterResourceId + { + + /// + /// Additional properties that are not explicitly defined in the schema + /// + [global::System.Text.Json.Serialization.JsonExtensionData] + public global::System.Collections.Generic.IDictionary AdditionalProperties { get; set; } = new global::System.Collections.Generic.Dictionary(); + } +} \ No newline at end of file diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.VideoCharacterResourceName.Json.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.VideoCharacterResourceName.Json.g.cs new file mode 100644 index 00000000..d183dc24 --- /dev/null +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.VideoCharacterResourceName.Json.g.cs @@ -0,0 +1,92 @@ +#nullable enable + +namespace tryAGI.OpenAI +{ + public sealed partial class VideoCharacterResourceName + { + /// + /// Serializes the current instance to a JSON string using the provided JsonSerializerContext. + /// + public string ToJson( + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return global::System.Text.Json.JsonSerializer.Serialize( + this, + this.GetType(), + jsonSerializerContext); + } + + /// + /// Serializes the current instance to a JSON string using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public string ToJson( + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.Serialize( + this, + jsonSerializerOptions); + } + + /// + /// Deserializes a JSON string using the provided JsonSerializerContext. + /// + public static global::tryAGI.OpenAI.VideoCharacterResourceName? FromJson( + string json, + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return global::System.Text.Json.JsonSerializer.Deserialize( + json, + typeof(global::tryAGI.OpenAI.VideoCharacterResourceName), + jsonSerializerContext) as global::tryAGI.OpenAI.VideoCharacterResourceName; + } + + /// + /// Deserializes a JSON string using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public static global::tryAGI.OpenAI.VideoCharacterResourceName? FromJson( + string json, + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.Deserialize( + json, + jsonSerializerOptions); + } + + /// + /// Deserializes a JSON stream using the provided JsonSerializerContext. + /// + public static async global::System.Threading.Tasks.ValueTask FromJsonStreamAsync( + global::System.IO.Stream jsonStream, + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return (await global::System.Text.Json.JsonSerializer.DeserializeAsync( + jsonStream, + typeof(global::tryAGI.OpenAI.VideoCharacterResourceName), + jsonSerializerContext).ConfigureAwait(false)) as global::tryAGI.OpenAI.VideoCharacterResourceName; + } + + /// + /// Deserializes a JSON stream using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public static global::System.Threading.Tasks.ValueTask FromJsonStreamAsync( + global::System.IO.Stream jsonStream, + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.DeserializeAsync( + jsonStream, + jsonSerializerOptions); + } + } +} diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.VideoCharacterResourceName.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.VideoCharacterResourceName.g.cs new file mode 100644 index 00000000..5c2127a2 --- /dev/null +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.VideoCharacterResourceName.g.cs @@ -0,0 +1,18 @@ + +#nullable enable + +namespace tryAGI.OpenAI +{ + /// + /// + /// + public sealed partial class VideoCharacterResourceName + { + + /// + /// Additional properties that are not explicitly defined in the schema + /// + [global::System.Text.Json.Serialization.JsonExtensionData] + public global::System.Collections.Generic.IDictionary AdditionalProperties { get; set; } = new global::System.Collections.Generic.Dictionary(); + } +} \ No newline at end of file diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.VideoReferenceInputParam.Json.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.VideoReferenceInputParam.Json.g.cs new file mode 100644 index 00000000..7f53885e --- /dev/null +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.VideoReferenceInputParam.Json.g.cs @@ -0,0 +1,92 @@ +#nullable enable + +namespace tryAGI.OpenAI +{ + public sealed partial class VideoReferenceInputParam + { + /// + /// Serializes the current instance to a JSON string using the provided JsonSerializerContext. + /// + public string ToJson( + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return global::System.Text.Json.JsonSerializer.Serialize( + this, + this.GetType(), + jsonSerializerContext); + } + + /// + /// Serializes the current instance to a JSON string using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public string ToJson( + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.Serialize( + this, + jsonSerializerOptions); + } + + /// + /// Deserializes a JSON string using the provided JsonSerializerContext. + /// + public static global::tryAGI.OpenAI.VideoReferenceInputParam? FromJson( + string json, + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return global::System.Text.Json.JsonSerializer.Deserialize( + json, + typeof(global::tryAGI.OpenAI.VideoReferenceInputParam), + jsonSerializerContext) as global::tryAGI.OpenAI.VideoReferenceInputParam; + } + + /// + /// Deserializes a JSON string using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public static global::tryAGI.OpenAI.VideoReferenceInputParam? FromJson( + string json, + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.Deserialize( + json, + jsonSerializerOptions); + } + + /// + /// Deserializes a JSON stream using the provided JsonSerializerContext. + /// + public static async global::System.Threading.Tasks.ValueTask FromJsonStreamAsync( + global::System.IO.Stream jsonStream, + global::System.Text.Json.Serialization.JsonSerializerContext jsonSerializerContext) + { + return (await global::System.Text.Json.JsonSerializer.DeserializeAsync( + jsonStream, + typeof(global::tryAGI.OpenAI.VideoReferenceInputParam), + jsonSerializerContext).ConfigureAwait(false)) as global::tryAGI.OpenAI.VideoReferenceInputParam; + } + + /// + /// Deserializes a JSON stream using the provided JsonSerializerOptions. + /// +#if NET8_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.RequiresUnreferencedCode("JSON serialization and deserialization might require types that cannot be statically analyzed. Use the overload that takes a JsonTypeInfo or JsonSerializerContext, or make sure all of the required types are preserved.")] + [global::System.Diagnostics.CodeAnalysis.RequiresDynamicCode("JSON serialization and deserialization might require types that cannot be statically analyzed and might need runtime code generation. Use System.Text.Json source generation for native AOT applications.")] +#endif + public static global::System.Threading.Tasks.ValueTask FromJsonStreamAsync( + global::System.IO.Stream jsonStream, + global::System.Text.Json.JsonSerializerOptions? jsonSerializerOptions = null) + { + return global::System.Text.Json.JsonSerializer.DeserializeAsync( + jsonStream, + jsonSerializerOptions); + } + } +} diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.VideoReferenceInputParam.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.VideoReferenceInputParam.g.cs new file mode 100644 index 00000000..30c1e0c0 --- /dev/null +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.Models.VideoReferenceInputParam.g.cs @@ -0,0 +1,49 @@ + +#nullable enable + +namespace tryAGI.OpenAI +{ + /// + /// Reference to the completed video. + /// + public sealed partial class VideoReferenceInputParam + { + /// + /// The identifier of the completed video.
+ /// Example: video_123 + ///
+ /// video_123 + [global::System.Text.Json.Serialization.JsonPropertyName("id")] + [global::System.Text.Json.Serialization.JsonRequired] + public required string Id { get; set; } + + /// + /// Additional properties that are not explicitly defined in the schema + /// + [global::System.Text.Json.Serialization.JsonExtensionData] + public global::System.Collections.Generic.IDictionary AdditionalProperties { get; set; } = new global::System.Collections.Generic.Dictionary(); + + /// + /// Initializes a new instance of the class. + /// + /// + /// The identifier of the completed video.
+ /// Example: video_123 + /// +#if NET7_0_OR_GREATER + [global::System.Diagnostics.CodeAnalysis.SetsRequiredMembers] +#endif + public VideoReferenceInputParam( + string id) + { + this.Id = id ?? throw new global::System.ArgumentNullException(nameof(id)); + } + + /// + /// Initializes a new instance of the class. + /// + public VideoReferenceInputParam() + { + } + } +} \ No newline at end of file diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.VideosClient.CreateVideo.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.VideosClient.CreateVideo.g.cs index 5d70be9d..0526b1df 100644 --- a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.VideosClient.CreateVideo.g.cs +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.VideosClient.CreateVideo.g.cs @@ -7,11 +7,11 @@ public partial class VideosClient { partial void PrepareCreateVideoArguments( global::System.Net.Http.HttpClient httpClient, - global::tryAGI.OpenAI.CreateVideoBody request); + global::tryAGI.OpenAI.CreateVideoMultipartBody request); partial void PrepareCreateVideoRequest( global::System.Net.Http.HttpClient httpClient, global::System.Net.Http.HttpRequestMessage httpRequestMessage, - global::tryAGI.OpenAI.CreateVideoBody request); + global::tryAGI.OpenAI.CreateVideoMultipartBody request); partial void ProcessCreateVideoResponse( global::System.Net.Http.HttpClient httpClient, global::System.Net.Http.HttpResponseMessage httpResponseMessage); @@ -29,7 +29,7 @@ partial void ProcessCreateVideoResponseContent( /// public async global::System.Threading.Tasks.Task CreateVideoAsync( - global::tryAGI.OpenAI.CreateVideoBody request, + global::tryAGI.OpenAI.CreateVideoMultipartBody request, global::System.Threading.CancellationToken cancellationToken = default) { request = request ?? throw new global::System.ArgumentNullException(nameof(request)); @@ -81,22 +81,9 @@ partial void ProcessCreateVideoResponseContent( if (request.InputReference != default) { - var __contentInputReference = new global::System.Net.Http.ByteArrayContent(request.InputReference ?? global::System.Array.Empty()); __httpRequestContent.Add( - content: __contentInputReference, - name: "\"input_reference\"", - fileName: request.InputReferencename != null ? $"\"{request.InputReferencename}\"" : string.Empty); - if (__contentInputReference.Headers.ContentDisposition != null) - { - __contentInputReference.Headers.ContentDisposition.FileNameStar = null; - } - } - if (request.ImageReference != default) - { - - __httpRequestContent.Add( - content: new global::System.Net.Http.StringContent($"{request.ImageReference}"), - name: "\"image_reference\""); + content: new global::System.Net.Http.StringContent(request.InputReference?.ToString() ?? string.Empty), + name: "\"input_reference\""); } if (request.Seconds != default) { @@ -215,15 +202,7 @@ partial void ProcessCreateVideoResponseContent( /// /// Text prompt that describes the video to generate. /// - /// - /// Optional multipart reference asset that guides generation. - /// - /// - /// Optional multipart reference asset that guides generation. - /// - /// - /// Optional JSON-safe image reference that guides generation. Provide exactly one of `image_url` or `file_id`. - /// + /// /// /// Clip duration in seconds (allowed values: 4, 8, 12). Defaults to 4 seconds. /// @@ -235,20 +214,16 @@ partial void ProcessCreateVideoResponseContent( public async global::System.Threading.Tasks.Task CreateVideoAsync( string prompt, global::tryAGI.OpenAI.VideoModel? model = default, - byte[]? inputReference = default, - string? inputReferencename = default, - global::tryAGI.OpenAI.ImageRefParam2? imageReference = default, + global::tryAGI.OpenAI.OneOf? inputReference = default, global::tryAGI.OpenAI.VideoSeconds? seconds = default, global::tryAGI.OpenAI.VideoSize? size = default, global::System.Threading.CancellationToken cancellationToken = default) { - var __request = new global::tryAGI.OpenAI.CreateVideoBody + var __request = new global::tryAGI.OpenAI.CreateVideoMultipartBody { Model = model, Prompt = prompt, InputReference = inputReference, - InputReferencename = inputReferencename, - ImageReference = imageReference, Seconds = seconds, Size = size, }; diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.VideosClient.CreateVideoCharacter.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.VideosClient.CreateVideoCharacter.g.cs new file mode 100644 index 00000000..90a41b5e --- /dev/null +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.VideosClient.CreateVideoCharacter.g.cs @@ -0,0 +1,209 @@ + +#nullable enable + +namespace tryAGI.OpenAI +{ + public partial class VideosClient + { + partial void PrepareCreateVideoCharacterArguments( + global::System.Net.Http.HttpClient httpClient, + global::tryAGI.OpenAI.CreateVideoCharacterBody request); + partial void PrepareCreateVideoCharacterRequest( + global::System.Net.Http.HttpClient httpClient, + global::System.Net.Http.HttpRequestMessage httpRequestMessage, + global::tryAGI.OpenAI.CreateVideoCharacterBody request); + partial void ProcessCreateVideoCharacterResponse( + global::System.Net.Http.HttpClient httpClient, + global::System.Net.Http.HttpResponseMessage httpResponseMessage); + + partial void ProcessCreateVideoCharacterResponseContent( + global::System.Net.Http.HttpClient httpClient, + global::System.Net.Http.HttpResponseMessage httpResponseMessage, + ref string content); + + /// + /// Create a character from an uploaded video. + /// + /// + /// The token to cancel the operation with + /// + public async global::System.Threading.Tasks.Task CreateVideoCharacterAsync( + + global::tryAGI.OpenAI.CreateVideoCharacterBody request, + global::System.Threading.CancellationToken cancellationToken = default) + { + request = request ?? throw new global::System.ArgumentNullException(nameof(request)); + + PrepareArguments( + client: HttpClient); + PrepareCreateVideoCharacterArguments( + httpClient: HttpClient, + request: request); + + var __pathBuilder = new global::tryAGI.OpenAI.PathBuilder( + path: "/videos/characters", + baseUri: HttpClient.BaseAddress); + var __path = __pathBuilder.ToString(); + using var __httpRequest = new global::System.Net.Http.HttpRequestMessage( + method: global::System.Net.Http.HttpMethod.Post, + requestUri: new global::System.Uri(__path, global::System.UriKind.RelativeOrAbsolute)); +#if NET6_0_OR_GREATER + __httpRequest.Version = global::System.Net.HttpVersion.Version11; + __httpRequest.VersionPolicy = global::System.Net.Http.HttpVersionPolicy.RequestVersionOrHigher; +#endif + + foreach (var __authorization in Authorizations) + { + if (__authorization.Type == "Http" || + __authorization.Type == "OAuth2") + { + __httpRequest.Headers.Authorization = new global::System.Net.Http.Headers.AuthenticationHeaderValue( + scheme: __authorization.Name, + parameter: __authorization.Value); + } + else if (__authorization.Type == "ApiKey" && + __authorization.Location == "Header") + { + __httpRequest.Headers.Add(__authorization.Name, __authorization.Value); + } + } + using var __httpRequestContent = new global::System.Net.Http.MultipartFormDataContent(); + var __contentVideo = new global::System.Net.Http.ByteArrayContent(request.Video ?? global::System.Array.Empty()); + __httpRequestContent.Add( + content: __contentVideo, + name: "\"video\"", + fileName: request.Videoname != null ? $"\"{request.Videoname}\"" : string.Empty); + if (__contentVideo.Headers.ContentDisposition != null) + { + __contentVideo.Headers.ContentDisposition.FileNameStar = null; + } + __httpRequestContent.Add( + content: new global::System.Net.Http.StringContent($"{request.Name}"), + name: "\"name\""); + __httpRequest.Content = __httpRequestContent; + + PrepareRequest( + client: HttpClient, + request: __httpRequest); + PrepareCreateVideoCharacterRequest( + httpClient: HttpClient, + httpRequestMessage: __httpRequest, + request: request); + + using var __response = await HttpClient.SendAsync( + request: __httpRequest, + completionOption: global::System.Net.Http.HttpCompletionOption.ResponseContentRead, + cancellationToken: cancellationToken).ConfigureAwait(false); + + ProcessResponse( + client: HttpClient, + response: __response); + ProcessCreateVideoCharacterResponse( + httpClient: HttpClient, + httpResponseMessage: __response); + + if (ReadResponseAsString) + { + var __content = await __response.Content.ReadAsStringAsync( +#if NET5_0_OR_GREATER + cancellationToken +#endif + ).ConfigureAwait(false); + + ProcessResponseContent( + client: HttpClient, + response: __response, + content: ref __content); + ProcessCreateVideoCharacterResponseContent( + httpClient: HttpClient, + httpResponseMessage: __response, + content: ref __content); + + try + { + __response.EnsureSuccessStatusCode(); + + return + global::tryAGI.OpenAI.VideoCharacterResource.FromJson(__content, JsonSerializerContext) ?? + throw new global::System.InvalidOperationException($"Response deserialization failed for \"{__content}\" "); + } + catch (global::System.Exception __ex) + { + throw new global::tryAGI.OpenAI.ApiException( + message: __content ?? __response.ReasonPhrase ?? string.Empty, + innerException: __ex, + statusCode: __response.StatusCode) + { + ResponseBody = __content, + ResponseHeaders = global::System.Linq.Enumerable.ToDictionary( + __response.Headers, + h => h.Key, + h => h.Value), + }; + } + } + else + { + try + { + __response.EnsureSuccessStatusCode(); + + using var __content = await __response.Content.ReadAsStreamAsync( +#if NET5_0_OR_GREATER + cancellationToken +#endif + ).ConfigureAwait(false); + + return + await global::tryAGI.OpenAI.VideoCharacterResource.FromJsonStreamAsync(__content, JsonSerializerContext).ConfigureAwait(false) ?? + throw new global::System.InvalidOperationException("Response deserialization failed."); + } + catch (global::System.Exception __ex) + { + throw new global::tryAGI.OpenAI.ApiException( + message: __response.ReasonPhrase ?? string.Empty, + innerException: __ex, + statusCode: __response.StatusCode) + { + ResponseHeaders = global::System.Linq.Enumerable.ToDictionary( + __response.Headers, + h => h.Key, + h => h.Value), + }; + } + } + } + + /// + /// Create a character from an uploaded video. + /// + /// + /// Video file used to create a character. + /// + /// + /// Video file used to create a character. + /// + /// + /// Display name for this API character. + /// + /// The token to cancel the operation with + /// + public async global::System.Threading.Tasks.Task CreateVideoCharacterAsync( + byte[] video, + string videoname, + string name, + global::System.Threading.CancellationToken cancellationToken = default) + { + var __request = new global::tryAGI.OpenAI.CreateVideoCharacterBody + { + Video = video, + Videoname = videoname, + Name = name, + }; + + return await CreateVideoCharacterAsync( + request: __request, + cancellationToken: cancellationToken).ConfigureAwait(false); + } + } +} \ No newline at end of file diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.VideosClient.CreateVideoEdit.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.VideosClient.CreateVideoEdit.g.cs new file mode 100644 index 00000000..06eeb8be --- /dev/null +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.VideosClient.CreateVideoEdit.g.cs @@ -0,0 +1,196 @@ + +#nullable enable + +namespace tryAGI.OpenAI +{ + public partial class VideosClient + { + partial void PrepareCreateVideoEditArguments( + global::System.Net.Http.HttpClient httpClient, + global::tryAGI.OpenAI.CreateVideoEditMultipartBody request); + partial void PrepareCreateVideoEditRequest( + global::System.Net.Http.HttpClient httpClient, + global::System.Net.Http.HttpRequestMessage httpRequestMessage, + global::tryAGI.OpenAI.CreateVideoEditMultipartBody request); + partial void ProcessCreateVideoEditResponse( + global::System.Net.Http.HttpClient httpClient, + global::System.Net.Http.HttpResponseMessage httpResponseMessage); + + partial void ProcessCreateVideoEditResponseContent( + global::System.Net.Http.HttpClient httpClient, + global::System.Net.Http.HttpResponseMessage httpResponseMessage, + ref string content); + + /// + /// Create a new video generation job by editing a source video or existing generated video. + /// + /// + /// The token to cancel the operation with + /// + public async global::System.Threading.Tasks.Task CreateVideoEditAsync( + + global::tryAGI.OpenAI.CreateVideoEditMultipartBody request, + global::System.Threading.CancellationToken cancellationToken = default) + { + request = request ?? throw new global::System.ArgumentNullException(nameof(request)); + + PrepareArguments( + client: HttpClient); + PrepareCreateVideoEditArguments( + httpClient: HttpClient, + request: request); + + var __pathBuilder = new global::tryAGI.OpenAI.PathBuilder( + path: "/videos/edits", + baseUri: HttpClient.BaseAddress); + var __path = __pathBuilder.ToString(); + using var __httpRequest = new global::System.Net.Http.HttpRequestMessage( + method: global::System.Net.Http.HttpMethod.Post, + requestUri: new global::System.Uri(__path, global::System.UriKind.RelativeOrAbsolute)); +#if NET6_0_OR_GREATER + __httpRequest.Version = global::System.Net.HttpVersion.Version11; + __httpRequest.VersionPolicy = global::System.Net.Http.HttpVersionPolicy.RequestVersionOrHigher; +#endif + + foreach (var __authorization in Authorizations) + { + if (__authorization.Type == "Http" || + __authorization.Type == "OAuth2") + { + __httpRequest.Headers.Authorization = new global::System.Net.Http.Headers.AuthenticationHeaderValue( + scheme: __authorization.Name, + parameter: __authorization.Value); + } + else if (__authorization.Type == "ApiKey" && + __authorization.Location == "Header") + { + __httpRequest.Headers.Add(__authorization.Name, __authorization.Value); + } + } + using var __httpRequestContent = new global::System.Net.Http.MultipartFormDataContent(); + __httpRequestContent.Add( + content: new global::System.Net.Http.StringContent(request.Video.ToString() ?? string.Empty), + name: "\"video\""); + __httpRequestContent.Add( + content: new global::System.Net.Http.StringContent($"{request.Prompt}"), + name: "\"prompt\""); + __httpRequest.Content = __httpRequestContent; + + PrepareRequest( + client: HttpClient, + request: __httpRequest); + PrepareCreateVideoEditRequest( + httpClient: HttpClient, + httpRequestMessage: __httpRequest, + request: request); + + using var __response = await HttpClient.SendAsync( + request: __httpRequest, + completionOption: global::System.Net.Http.HttpCompletionOption.ResponseContentRead, + cancellationToken: cancellationToken).ConfigureAwait(false); + + ProcessResponse( + client: HttpClient, + response: __response); + ProcessCreateVideoEditResponse( + httpClient: HttpClient, + httpResponseMessage: __response); + + if (ReadResponseAsString) + { + var __content = await __response.Content.ReadAsStringAsync( +#if NET5_0_OR_GREATER + cancellationToken +#endif + ).ConfigureAwait(false); + + ProcessResponseContent( + client: HttpClient, + response: __response, + content: ref __content); + ProcessCreateVideoEditResponseContent( + httpClient: HttpClient, + httpResponseMessage: __response, + content: ref __content); + + try + { + __response.EnsureSuccessStatusCode(); + + return + global::tryAGI.OpenAI.VideoResource.FromJson(__content, JsonSerializerContext) ?? + throw new global::System.InvalidOperationException($"Response deserialization failed for \"{__content}\" "); + } + catch (global::System.Exception __ex) + { + throw new global::tryAGI.OpenAI.ApiException( + message: __content ?? __response.ReasonPhrase ?? string.Empty, + innerException: __ex, + statusCode: __response.StatusCode) + { + ResponseBody = __content, + ResponseHeaders = global::System.Linq.Enumerable.ToDictionary( + __response.Headers, + h => h.Key, + h => h.Value), + }; + } + } + else + { + try + { + __response.EnsureSuccessStatusCode(); + + using var __content = await __response.Content.ReadAsStreamAsync( +#if NET5_0_OR_GREATER + cancellationToken +#endif + ).ConfigureAwait(false); + + return + await global::tryAGI.OpenAI.VideoResource.FromJsonStreamAsync(__content, JsonSerializerContext).ConfigureAwait(false) ?? + throw new global::System.InvalidOperationException("Response deserialization failed."); + } + catch (global::System.Exception __ex) + { + throw new global::tryAGI.OpenAI.ApiException( + message: __response.ReasonPhrase ?? string.Empty, + innerException: __ex, + statusCode: __response.StatusCode) + { + ResponseHeaders = global::System.Linq.Enumerable.ToDictionary( + __response.Headers, + h => h.Key, + h => h.Value), + }; + } + } + } + + /// + /// Create a new video generation job by editing a source video or existing generated video. + /// + /// + /// + /// Text prompt that describes how to edit the source video. + /// + /// The token to cancel the operation with + /// + public async global::System.Threading.Tasks.Task CreateVideoEditAsync( + global::tryAGI.OpenAI.OneOf video, + string prompt, + global::System.Threading.CancellationToken cancellationToken = default) + { + var __request = new global::tryAGI.OpenAI.CreateVideoEditMultipartBody + { + Video = video, + Prompt = prompt, + }; + + return await CreateVideoEditAsync( + request: __request, + cancellationToken: cancellationToken).ConfigureAwait(false); + } + } +} \ No newline at end of file diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.VideosClient.CreateVideoExtend.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.VideosClient.CreateVideoExtend.g.cs new file mode 100644 index 00000000..59a6329a --- /dev/null +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.VideosClient.CreateVideoExtend.g.cs @@ -0,0 +1,204 @@ + +#nullable enable + +namespace tryAGI.OpenAI +{ + public partial class VideosClient + { + partial void PrepareCreateVideoExtendArguments( + global::System.Net.Http.HttpClient httpClient, + global::tryAGI.OpenAI.CreateVideoExtendMultipartBody request); + partial void PrepareCreateVideoExtendRequest( + global::System.Net.Http.HttpClient httpClient, + global::System.Net.Http.HttpRequestMessage httpRequestMessage, + global::tryAGI.OpenAI.CreateVideoExtendMultipartBody request); + partial void ProcessCreateVideoExtendResponse( + global::System.Net.Http.HttpClient httpClient, + global::System.Net.Http.HttpResponseMessage httpResponseMessage); + + partial void ProcessCreateVideoExtendResponseContent( + global::System.Net.Http.HttpClient httpClient, + global::System.Net.Http.HttpResponseMessage httpResponseMessage, + ref string content); + + /// + /// Create an extension of a completed video. + /// + /// + /// The token to cancel the operation with + /// + public async global::System.Threading.Tasks.Task CreateVideoExtendAsync( + + global::tryAGI.OpenAI.CreateVideoExtendMultipartBody request, + global::System.Threading.CancellationToken cancellationToken = default) + { + request = request ?? throw new global::System.ArgumentNullException(nameof(request)); + + PrepareArguments( + client: HttpClient); + PrepareCreateVideoExtendArguments( + httpClient: HttpClient, + request: request); + + var __pathBuilder = new global::tryAGI.OpenAI.PathBuilder( + path: "/videos/extensions", + baseUri: HttpClient.BaseAddress); + var __path = __pathBuilder.ToString(); + using var __httpRequest = new global::System.Net.Http.HttpRequestMessage( + method: global::System.Net.Http.HttpMethod.Post, + requestUri: new global::System.Uri(__path, global::System.UriKind.RelativeOrAbsolute)); +#if NET6_0_OR_GREATER + __httpRequest.Version = global::System.Net.HttpVersion.Version11; + __httpRequest.VersionPolicy = global::System.Net.Http.HttpVersionPolicy.RequestVersionOrHigher; +#endif + + foreach (var __authorization in Authorizations) + { + if (__authorization.Type == "Http" || + __authorization.Type == "OAuth2") + { + __httpRequest.Headers.Authorization = new global::System.Net.Http.Headers.AuthenticationHeaderValue( + scheme: __authorization.Name, + parameter: __authorization.Value); + } + else if (__authorization.Type == "ApiKey" && + __authorization.Location == "Header") + { + __httpRequest.Headers.Add(__authorization.Name, __authorization.Value); + } + } + using var __httpRequestContent = new global::System.Net.Http.MultipartFormDataContent(); + __httpRequestContent.Add( + content: new global::System.Net.Http.StringContent(request.Video.ToString() ?? string.Empty), + name: "\"video\""); + __httpRequestContent.Add( + content: new global::System.Net.Http.StringContent($"{request.Prompt}"), + name: "\"prompt\""); + __httpRequestContent.Add( + content: new global::System.Net.Http.StringContent($"{request.Seconds.ToValueString()}"), + name: "\"seconds\""); + __httpRequest.Content = __httpRequestContent; + + PrepareRequest( + client: HttpClient, + request: __httpRequest); + PrepareCreateVideoExtendRequest( + httpClient: HttpClient, + httpRequestMessage: __httpRequest, + request: request); + + using var __response = await HttpClient.SendAsync( + request: __httpRequest, + completionOption: global::System.Net.Http.HttpCompletionOption.ResponseContentRead, + cancellationToken: cancellationToken).ConfigureAwait(false); + + ProcessResponse( + client: HttpClient, + response: __response); + ProcessCreateVideoExtendResponse( + httpClient: HttpClient, + httpResponseMessage: __response); + + if (ReadResponseAsString) + { + var __content = await __response.Content.ReadAsStringAsync( +#if NET5_0_OR_GREATER + cancellationToken +#endif + ).ConfigureAwait(false); + + ProcessResponseContent( + client: HttpClient, + response: __response, + content: ref __content); + ProcessCreateVideoExtendResponseContent( + httpClient: HttpClient, + httpResponseMessage: __response, + content: ref __content); + + try + { + __response.EnsureSuccessStatusCode(); + + return + global::tryAGI.OpenAI.VideoResource.FromJson(__content, JsonSerializerContext) ?? + throw new global::System.InvalidOperationException($"Response deserialization failed for \"{__content}\" "); + } + catch (global::System.Exception __ex) + { + throw new global::tryAGI.OpenAI.ApiException( + message: __content ?? __response.ReasonPhrase ?? string.Empty, + innerException: __ex, + statusCode: __response.StatusCode) + { + ResponseBody = __content, + ResponseHeaders = global::System.Linq.Enumerable.ToDictionary( + __response.Headers, + h => h.Key, + h => h.Value), + }; + } + } + else + { + try + { + __response.EnsureSuccessStatusCode(); + + using var __content = await __response.Content.ReadAsStreamAsync( +#if NET5_0_OR_GREATER + cancellationToken +#endif + ).ConfigureAwait(false); + + return + await global::tryAGI.OpenAI.VideoResource.FromJsonStreamAsync(__content, JsonSerializerContext).ConfigureAwait(false) ?? + throw new global::System.InvalidOperationException("Response deserialization failed."); + } + catch (global::System.Exception __ex) + { + throw new global::tryAGI.OpenAI.ApiException( + message: __response.ReasonPhrase ?? string.Empty, + innerException: __ex, + statusCode: __response.StatusCode) + { + ResponseHeaders = global::System.Linq.Enumerable.ToDictionary( + __response.Headers, + h => h.Key, + h => h.Value), + }; + } + } + } + + /// + /// Create an extension of a completed video. + /// + /// + /// + /// Updated text prompt that directs the extension generation. + /// + /// + /// Length of the newly generated extension segment in seconds (allowed values: 4, 8, 12, 16, 20). + /// + /// The token to cancel the operation with + /// + public async global::System.Threading.Tasks.Task CreateVideoExtendAsync( + global::tryAGI.OpenAI.OneOf video, + string prompt, + global::tryAGI.OpenAI.VideoSeconds seconds, + global::System.Threading.CancellationToken cancellationToken = default) + { + var __request = new global::tryAGI.OpenAI.CreateVideoExtendMultipartBody + { + Video = video, + Prompt = prompt, + Seconds = seconds, + }; + + return await CreateVideoExtendAsync( + request: __request, + cancellationToken: cancellationToken).ConfigureAwait(false); + } + } +} \ No newline at end of file diff --git a/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.VideosClient.GetVideoCharacter.g.cs b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.VideosClient.GetVideoCharacter.g.cs new file mode 100644 index 00000000..2d513e31 --- /dev/null +++ b/src/libs/tryAGI.OpenAI/Generated/tryAGI.OpenAI.VideosClient.GetVideoCharacter.g.cs @@ -0,0 +1,162 @@ + +#nullable enable + +namespace tryAGI.OpenAI +{ + public partial class VideosClient + { + partial void PrepareGetVideoCharacterArguments( + global::System.Net.Http.HttpClient httpClient, + ref string characterId); + partial void PrepareGetVideoCharacterRequest( + global::System.Net.Http.HttpClient httpClient, + global::System.Net.Http.HttpRequestMessage httpRequestMessage, + string characterId); + partial void ProcessGetVideoCharacterResponse( + global::System.Net.Http.HttpClient httpClient, + global::System.Net.Http.HttpResponseMessage httpResponseMessage); + + partial void ProcessGetVideoCharacterResponseContent( + global::System.Net.Http.HttpClient httpClient, + global::System.Net.Http.HttpResponseMessage httpResponseMessage, + ref string content); + + /// + /// Fetch a character. + /// + /// + /// Example: char_123 + /// + /// The token to cancel the operation with + /// + public async global::System.Threading.Tasks.Task GetVideoCharacterAsync( + string characterId, + global::System.Threading.CancellationToken cancellationToken = default) + { + PrepareArguments( + client: HttpClient); + PrepareGetVideoCharacterArguments( + httpClient: HttpClient, + characterId: ref characterId); + + var __pathBuilder = new global::tryAGI.OpenAI.PathBuilder( + path: $"/videos/characters/{characterId}", + baseUri: HttpClient.BaseAddress); + var __path = __pathBuilder.ToString(); + using var __httpRequest = new global::System.Net.Http.HttpRequestMessage( + method: global::System.Net.Http.HttpMethod.Get, + requestUri: new global::System.Uri(__path, global::System.UriKind.RelativeOrAbsolute)); +#if NET6_0_OR_GREATER + __httpRequest.Version = global::System.Net.HttpVersion.Version11; + __httpRequest.VersionPolicy = global::System.Net.Http.HttpVersionPolicy.RequestVersionOrHigher; +#endif + + foreach (var __authorization in Authorizations) + { + if (__authorization.Type == "Http" || + __authorization.Type == "OAuth2") + { + __httpRequest.Headers.Authorization = new global::System.Net.Http.Headers.AuthenticationHeaderValue( + scheme: __authorization.Name, + parameter: __authorization.Value); + } + else if (__authorization.Type == "ApiKey" && + __authorization.Location == "Header") + { + __httpRequest.Headers.Add(__authorization.Name, __authorization.Value); + } + } + + PrepareRequest( + client: HttpClient, + request: __httpRequest); + PrepareGetVideoCharacterRequest( + httpClient: HttpClient, + httpRequestMessage: __httpRequest, + characterId: characterId); + + using var __response = await HttpClient.SendAsync( + request: __httpRequest, + completionOption: global::System.Net.Http.HttpCompletionOption.ResponseContentRead, + cancellationToken: cancellationToken).ConfigureAwait(false); + + ProcessResponse( + client: HttpClient, + response: __response); + ProcessGetVideoCharacterResponse( + httpClient: HttpClient, + httpResponseMessage: __response); + + if (ReadResponseAsString) + { + var __content = await __response.Content.ReadAsStringAsync( +#if NET5_0_OR_GREATER + cancellationToken +#endif + ).ConfigureAwait(false); + + ProcessResponseContent( + client: HttpClient, + response: __response, + content: ref __content); + ProcessGetVideoCharacterResponseContent( + httpClient: HttpClient, + httpResponseMessage: __response, + content: ref __content); + + try + { + __response.EnsureSuccessStatusCode(); + + return + global::tryAGI.OpenAI.VideoCharacterResource.FromJson(__content, JsonSerializerContext) ?? + throw new global::System.InvalidOperationException($"Response deserialization failed for \"{__content}\" "); + } + catch (global::System.Exception __ex) + { + throw new global::tryAGI.OpenAI.ApiException( + message: __content ?? __response.ReasonPhrase ?? string.Empty, + innerException: __ex, + statusCode: __response.StatusCode) + { + ResponseBody = __content, + ResponseHeaders = global::System.Linq.Enumerable.ToDictionary( + __response.Headers, + h => h.Key, + h => h.Value), + }; + } + } + else + { + try + { + __response.EnsureSuccessStatusCode(); + + using var __content = await __response.Content.ReadAsStreamAsync( +#if NET5_0_OR_GREATER + cancellationToken +#endif + ).ConfigureAwait(false); + + return + await global::tryAGI.OpenAI.VideoCharacterResource.FromJsonStreamAsync(__content, JsonSerializerContext).ConfigureAwait(false) ?? + throw new global::System.InvalidOperationException("Response deserialization failed."); + } + catch (global::System.Exception __ex) + { + throw new global::tryAGI.OpenAI.ApiException( + message: __response.ReasonPhrase ?? string.Empty, + innerException: __ex, + statusCode: __response.StatusCode) + { + ResponseHeaders = global::System.Linq.Enumerable.ToDictionary( + __response.Headers, + h => h.Key, + h => h.Value), + }; + } + } + } + } +} \ No newline at end of file diff --git a/src/libs/tryAGI.OpenAI/openapi.yaml b/src/libs/tryAGI.OpenAI/openapi.yaml index 6435f3ac..e6dfcdc3 100644 --- a/src/libs/tryAGI.OpenAI/openapi.yaml +++ b/src/libs/tryAGI.OpenAI/openapi.yaml @@ -9094,10 +9094,10 @@ paths: content: multipart/form-data: schema: - $ref: '#/components/schemas/CreateVideoBody' + $ref: '#/components/schemas/CreateVideoMultipartBody' application/json: schema: - $ref: '#/components/schemas/CreateVideoBody' + $ref: '#/components/schemas/CreateVideoJsonBody' responses: '200': description: Success @@ -9164,6 +9164,123 @@ paths: java: "package com.openai.example;\n\nimport com.openai.client.OpenAIClient;\nimport com.openai.client.okhttp.OpenAIOkHttpClient;\nimport com.openai.models.videos.VideoListPage;\nimport com.openai.models.videos.VideoListParams;\n\npublic final class Main {\n private Main() {}\n\n public static void main(String[] args) {\n OpenAIClient client = OpenAIOkHttpClient.fromEnv();\n\n VideoListPage page = client.videos().list();\n }\n}" node.js: "import OpenAI from 'openai';\n\nconst client = new OpenAI({\n apiKey: process.env['OPENAI_API_KEY'], // This is the default and can be omitted\n});\n\n// Automatically fetches more pages as needed.\nfor await (const video of client.videos.list()) {\n console.log(video.id);\n}" response: "{\n \"data\": [\n {\n \"id\": \"video_123\",\n \"object\": \"video\",\n \"model\": \"sora-2\",\n \"status\": \"completed\"\n }\n ],\n \"object\": \"list\"\n}\n" + /videos/characters: + post: + tags: + - Videos + summary: Create a character from an uploaded video. + operationId: CreateVideoCharacter + requestBody: + content: + multipart/form-data: + schema: + $ref: '#/components/schemas/CreateVideoCharacterBody' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/VideoCharacterResource' + x-oaiMeta: + examples: + response: '' + request: + node.js: "import OpenAI from 'openai';\n\nconst client = new OpenAI({\n apiKey: process.env['OPENAI_API_KEY'], // This is the default and can be omitted\n});\n\nconst character = await client.videos.character.create({\n name: 'x',\n video: fs.createReadStream('path/to/file'),\n});\n\nconsole.log(character.id);" + python: "import os\nfrom openai import OpenAI\n\nclient = OpenAI(\n api_key=os.environ.get(\"OPENAI_API_KEY\"), # This is the default and can be omitted\n)\ncharacter = client.videos.character.create(\n name=\"x\",\n video=b\"Example data\",\n)\nprint(character.id)" + go: "package main\n\nimport (\n\t\"bytes\"\n\t\"context\"\n\t\"fmt\"\n\t\"io\"\n\n\t\"github.com/openai/openai-go\"\n\t\"github.com/openai/openai-go/option\"\n)\n\nfunc main() {\n\tclient := openai.NewClient(\n\t\toption.WithAPIKey(\"My API Key\"),\n\t)\n\tcharacter, err := client.Videos.Character.New(context.TODO(), openai.VideoCharacterNewParams{\n\t\tName: \"x\",\n\t\tVideo: io.Reader(bytes.NewBuffer([]byte(\"Example data\"))),\n\t})\n\tif err != nil {\n\t\tpanic(err.Error())\n\t}\n\tfmt.Printf(\"%+v\\n\", character.ID)\n}\n" + java: "package com.openai.example;\n\nimport com.openai.client.OpenAIClient;\nimport com.openai.client.okhttp.OpenAIOkHttpClient;\nimport com.openai.models.videos.character.CharacterCreateParams;\nimport com.openai.models.videos.character.CharacterCreateResponse;\nimport java.io.ByteArrayInputStream;\n\npublic final class Main {\n private Main() {}\n\n public static void main(String[] args) {\n OpenAIClient client = OpenAIOkHttpClient.fromEnv();\n\n CharacterCreateParams params = CharacterCreateParams.builder()\n .name(\"x\")\n .video(ByteArrayInputStream(\"Example data\".getBytes()))\n .build();\n CharacterCreateResponse character = client.videos().character().create(params);\n }\n}" + ruby: "require \"openai\"\n\nopenai = OpenAI::Client.new(api_key: \"My API Key\")\n\ncharacter = openai.videos.character.create(name: \"x\", video: StringIO.new(\"Example data\"))\n\nputs(character)" + '/videos/characters/{character_id}': + get: + tags: + - Videos + summary: Fetch a character. + operationId: GetVideoCharacter + parameters: + - name: character_id + in: path + description: The identifier of the character to retrieve. + required: true + schema: + type: string + example: char_123 + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/VideoCharacterResource' + x-oaiMeta: + examples: + response: '' + request: + node.js: "import OpenAI from 'openai';\n\nconst client = new OpenAI({\n apiKey: process.env['OPENAI_API_KEY'], // This is the default and can be omitted\n});\n\nconst character = await client.videos.character.get('char_123');\n\nconsole.log(character.id);" + python: "import os\nfrom openai import OpenAI\n\nclient = OpenAI(\n api_key=os.environ.get(\"OPENAI_API_KEY\"), # This is the default and can be omitted\n)\ncharacter = client.videos.character.get(\n \"char_123\",\n)\nprint(character.id)" + go: "package main\n\nimport (\n\t\"context\"\n\t\"fmt\"\n\n\t\"github.com/openai/openai-go\"\n\t\"github.com/openai/openai-go/option\"\n)\n\nfunc main() {\n\tclient := openai.NewClient(\n\t\toption.WithAPIKey(\"My API Key\"),\n\t)\n\tcharacter, err := client.Videos.Character.Get(context.TODO(), \"char_123\")\n\tif err != nil {\n\t\tpanic(err.Error())\n\t}\n\tfmt.Printf(\"%+v\\n\", character.ID)\n}\n" + java: "package com.openai.example;\n\nimport com.openai.client.OpenAIClient;\nimport com.openai.client.okhttp.OpenAIOkHttpClient;\nimport com.openai.models.videos.character.CharacterGetParams;\nimport com.openai.models.videos.character.CharacterGetResponse;\n\npublic final class Main {\n private Main() {}\n\n public static void main(String[] args) {\n OpenAIClient client = OpenAIOkHttpClient.fromEnv();\n\n CharacterGetResponse character = client.videos().character().get(\"char_123\");\n }\n}" + ruby: "require \"openai\"\n\nopenai = OpenAI::Client.new(api_key: \"My API Key\")\n\ncharacter = openai.videos.character.get(\"char_123\")\n\nputs(character)" + /videos/edits: + post: + tags: + - Videos + summary: Create a new video generation job by editing a source video or existing generated video. + operationId: CreateVideoEdit + requestBody: + content: + multipart/form-data: + schema: + $ref: '#/components/schemas/CreateVideoEditMultipartBody' + application/json: + schema: + $ref: '#/components/schemas/CreateVideoEditJsonBody' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/VideoResource' + x-oaiMeta: + examples: + response: '' + request: + node.js: "import OpenAI from 'openai';\n\nconst client = new OpenAI({\n apiKey: process.env['OPENAI_API_KEY'], // This is the default and can be omitted\n});\n\nconst video = await client.videos.edit({ prompt: 'x', video: fs.createReadStream('path/to/file') });\n\nconsole.log(video.id);" + python: "import os\nfrom openai import OpenAI\n\nclient = OpenAI(\n api_key=os.environ.get(\"OPENAI_API_KEY\"), # This is the default and can be omitted\n)\nvideo = client.videos.edit(\n prompt=\"x\",\n video=b\"Example data\",\n)\nprint(video.id)" + go: "package main\n\nimport (\n\t\"bytes\"\n\t\"context\"\n\t\"fmt\"\n\t\"io\"\n\n\t\"github.com/openai/openai-go\"\n\t\"github.com/openai/openai-go/option\"\n)\n\nfunc main() {\n\tclient := openai.NewClient(\n\t\toption.WithAPIKey(\"My API Key\"),\n\t)\n\tvideo, err := client.Videos.Edit(context.TODO(), openai.VideoEditParams{\n\t\tPrompt: \"x\",\n\t\tVideo: openai.VideoEditParamsVideoUnion{\n\t\t\tOfFile: io.Reader(bytes.NewBuffer([]byte(\"Example data\"))),\n\t\t},\n\t})\n\tif err != nil {\n\t\tpanic(err.Error())\n\t}\n\tfmt.Printf(\"%+v\\n\", video.ID)\n}\n" + java: "package com.openai.example;\n\nimport com.openai.client.OpenAIClient;\nimport com.openai.client.okhttp.OpenAIOkHttpClient;\nimport com.openai.models.videos.Video;\nimport com.openai.models.videos.VideoEditParams;\nimport java.io.ByteArrayInputStream;\nimport java.io.InputStream;\n\npublic final class Main {\n private Main() {}\n\n public static void main(String[] args) {\n OpenAIClient client = OpenAIOkHttpClient.fromEnv();\n\n VideoEditParams params = VideoEditParams.builder()\n .prompt(\"x\")\n .video(ByteArrayInputStream(\"Example data\".getBytes()))\n .build();\n Video video = client.videos().edit(params);\n }\n}" + ruby: "require \"openai\"\n\nopenai = OpenAI::Client.new(api_key: \"My API Key\")\n\nvideo = openai.videos.edit(prompt: \"x\", video: StringIO.new(\"Example data\"))\n\nputs(video)" + /videos/extensions: + post: + tags: + - Videos + summary: Create an extension of a completed video. + operationId: CreateVideoExtend + requestBody: + content: + multipart/form-data: + schema: + $ref: '#/components/schemas/CreateVideoExtendMultipartBody' + application/json: + schema: + $ref: '#/components/schemas/CreateVideoExtendJsonBody' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/VideoResource' + x-oaiMeta: + examples: + response: '' + request: + node.js: "import OpenAI from 'openai';\n\nconst client = new OpenAI({\n apiKey: process.env['OPENAI_API_KEY'], // This is the default and can be omitted\n});\n\nconst video = await client.videos.extend({\n prompt: 'x',\n seconds: '4',\n video: { id: 'video_123' },\n});\n\nconsole.log(video.id);" + python: "import os\nfrom openai import OpenAI\n\nclient = OpenAI(\n api_key=os.environ.get(\"OPENAI_API_KEY\"), # This is the default and can be omitted\n)\nvideo = client.videos.extend(\n prompt=\"x\",\n seconds=\"4\",\n video={\n \"id\": \"video_123\"\n },\n)\nprint(video.id)" + go: "package main\n\nimport (\n\t\"context\"\n\t\"fmt\"\n\n\t\"github.com/openai/openai-go\"\n\t\"github.com/openai/openai-go/option\"\n)\n\nfunc main() {\n\tclient := openai.NewClient(\n\t\toption.WithAPIKey(\"My API Key\"),\n\t)\n\tvideo, err := client.Videos.Extend(context.TODO(), openai.VideoExtendParams{\n\t\tPrompt: \"x\",\n\t\tSeconds: openai.VideoSeconds4,\n\t\tVideo: openai.VideoExtendParamsVideoUnion{\n\t\t\tOfVideoExtendsVideoVideoReferenceInputParam: &openai.VideoExtendParamsVideoVideoReferenceInputParam{\n\t\t\t\tID: \"video_123\",\n\t\t\t},\n\t\t},\n\t})\n\tif err != nil {\n\t\tpanic(err.Error())\n\t}\n\tfmt.Printf(\"%+v\\n\", video.ID)\n}\n" + java: "package com.openai.example;\n\nimport com.openai.client.OpenAIClient;\nimport com.openai.client.okhttp.OpenAIOkHttpClient;\nimport com.openai.models.videos.Video;\nimport com.openai.models.videos.VideoExtendParams;\nimport com.openai.models.videos.VideoSeconds;\n\npublic final class Main {\n private Main() {}\n\n public static void main(String[] args) {\n OpenAIClient client = OpenAIOkHttpClient.fromEnv();\n\n VideoExtendParams params = VideoExtendParams.builder()\n .prompt(\"x\")\n .seconds(VideoSeconds._4)\n .video(VideoExtendParams.Video.VideoReferenceInputParam.builder()\n .id(\"video_123\")\n .build())\n .build();\n Video video = client.videos().extend(params);\n }\n}" + ruby: "require \"openai\"\n\nopenai = OpenAI::Client.new(api_key: \"My API Key\")\n\nvideo = openai.videos.extend_(prompt: \"x\", seconds: :\"4\", video: {id: \"video_123\"})\n\nputs(video)" '/videos/{video_id}': get: tags: @@ -37096,8 +37213,8 @@ components: - '8' - '12' type: string - CreateVideoBody: - title: Create video request + CreateVideoMultipartBody: + title: Create video multipart request required: - prompt type: object @@ -37111,11 +37228,34 @@ components: type: string description: Text prompt that describes the video to generate. input_reference: + oneOf: + - type: string + description: Optional reference asset upload or reference object that guides generation. + format: binary + - $ref: '#/components/schemas/ImageRefParam-2' + seconds: + description: 'Clip duration in seconds (allowed values: 4, 8, 12). Defaults to 4 seconds.' + $ref: '#/components/schemas/VideoSeconds' + size: + description: 'Output resolution formatted as width x height (allowed values: 720x1280, 1280x720, 1024x1792, 1792x1024). Defaults to 720x1280.' + $ref: '#/components/schemas/VideoSize' + description: Multipart parameters for creating a new video generation job. + CreateVideoJsonBody: + title: Create video JSON request + required: + - prompt + type: object + properties: + model: + description: 'The video generation model to use (allowed values: sora-2, sora-2-pro). Defaults to `sora-2`.' + $ref: '#/components/schemas/VideoModel' + prompt: + maxLength: 32000 + minLength: 1 type: string - description: Optional multipart reference asset that guides generation. - format: binary - image_reference: - description: Optional JSON-safe image reference that guides generation. Provide exactly one of `image_url` or `file_id`. + description: Text prompt that describes the video to generate. + input_reference: + description: Optional reference object that guides generation. Provide exactly one of `image_url` or `file_id`. $ref: '#/components/schemas/ImageRefParam-2' seconds: description: 'Clip duration in seconds (allowed values: 4, 8, 12). Defaults to 4 seconds.' @@ -37123,7 +37263,132 @@ components: size: description: 'Output resolution formatted as width x height (allowed values: 720x1280, 1280x720, 1024x1792, 1792x1024). Defaults to 720x1280.' $ref: '#/components/schemas/VideoSize' - description: Parameters for creating a new video generation job. + description: JSON parameters for creating a new video generation job. + CreateVideoCharacterBody: + title: Create character request + required: + - video + - name + type: object + properties: + video: + type: string + description: Video file used to create a character. + format: binary + name: + maxLength: 80 + minLength: 1 + type: string + description: Display name for this API character. + description: Parameters for creating a character from an uploaded video. + VideoCharacterResource: + required: + - id + - name + - created_at + type: object + properties: + id: + anyOf: + - type: string + description: Identifier for the character creation cameo. + - type: 'null' + name: + anyOf: + - type: string + description: Display name for the character. + - type: 'null' + created_at: + type: integer + description: Unix timestamp (in seconds) when the character was created. + VideoReferenceInputParam: + required: + - id + type: object + properties: + id: + type: string + description: The identifier of the completed video. + example: video_123 + description: Reference to the completed video. + CreateVideoEditMultipartBody: + title: Create video edit multipart request + required: + - video + - prompt + type: object + properties: + video: + oneOf: + - type: string + description: Reference to the completed video to edit. + format: binary + - $ref: '#/components/schemas/VideoReferenceInputParam' + prompt: + maxLength: 32000 + minLength: 1 + type: string + description: Text prompt that describes how to edit the source video. + description: Parameters for editing an existing generated video. + CreateVideoEditJsonBody: + title: Create video edit JSON request + required: + - video + - prompt + type: object + properties: + video: + description: Reference to the completed video to edit. + $ref: '#/components/schemas/VideoReferenceInputParam' + prompt: + maxLength: 32000 + minLength: 1 + type: string + description: Text prompt that describes how to edit the source video. + description: JSON parameters for editing an existing generated video. + CreateVideoExtendMultipartBody: + title: Create video extension multipart request + required: + - video + - prompt + - seconds + type: object + properties: + video: + oneOf: + - $ref: '#/components/schemas/VideoReferenceInputParam' + - type: string + description: Reference to the completed video to extend. + format: binary + prompt: + maxLength: 32000 + minLength: 1 + type: string + description: Updated text prompt that directs the extension generation. + seconds: + description: 'Length of the newly generated extension segment in seconds (allowed values: 4, 8, 12, 16, 20).' + $ref: '#/components/schemas/VideoSeconds' + description: Multipart parameters for extending an existing generated video. + CreateVideoExtendJsonBody: + title: Create video extension JSON request + required: + - video + - prompt + - seconds + type: object + properties: + video: + description: Reference to the completed video to extend. + $ref: '#/components/schemas/VideoReferenceInputParam' + prompt: + maxLength: 32000 + minLength: 1 + type: string + description: Updated text prompt that directs the extension generation. + seconds: + description: 'Length of the newly generated extension segment in seconds (allowed values: 4, 8, 12, 16, 20).' + $ref: '#/components/schemas/VideoSeconds' + description: JSON parameters for extending an existing generated video. DeletedVideoResource: title: Deleted video response required: