From 380cc9e21d647cd2a829e985c1c61bd607d4d7a6 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" Date: Wed, 10 Sep 2025 12:37:44 +0000 Subject: [PATCH] feat: Updated OpenAPI spec --- .../Generated/AssemblyAI.Models.LemurBaseParams.g.cs | 4 ++-- .../Generated/AssemblyAI.Models.LemurTaskParamsVariant1.g.cs | 4 ++-- src/libs/AssemblyAI/openapi.yaml | 4 ++-- 3 files changed, 6 insertions(+), 6 deletions(-) diff --git a/src/libs/AssemblyAI/Generated/AssemblyAI.Models.LemurBaseParams.g.cs b/src/libs/AssemblyAI/Generated/AssemblyAI.Models.LemurBaseParams.g.cs index c98f20d..bfc02e7 100644 --- a/src/libs/AssemblyAI/Generated/AssemblyAI.Models.LemurBaseParams.g.cs +++ b/src/libs/AssemblyAI/Generated/AssemblyAI.Models.LemurBaseParams.g.cs @@ -35,7 +35,7 @@ public sealed partial class LemurBaseParams public string? InputText { get; set; } /// - /// Max output size in tokens, up to 4000
+ /// Maximum output size in tokens, up to the `final_model`'s max [(see chart)](/docs/lemur/customize-parameters#change-the-maximum-output-size).
/// Default Value: 2000 ///
[global::System.Text.Json.Serialization.JsonPropertyName("max_output_size")] @@ -78,7 +78,7 @@ public sealed partial class LemurBaseParams /// Use either transcript_ids or input_text as input into LeMUR. /// /// - /// Max output size in tokens, up to 4000
+ /// Maximum output size in tokens, up to the `final_model`'s max [(see chart)](/docs/lemur/customize-parameters#change-the-maximum-output-size).
/// Default Value: 2000 /// /// diff --git a/src/libs/AssemblyAI/Generated/AssemblyAI.Models.LemurTaskParamsVariant1.g.cs b/src/libs/AssemblyAI/Generated/AssemblyAI.Models.LemurTaskParamsVariant1.g.cs index a1eada6..3ec468f 100644 --- a/src/libs/AssemblyAI/Generated/AssemblyAI.Models.LemurTaskParamsVariant1.g.cs +++ b/src/libs/AssemblyAI/Generated/AssemblyAI.Models.LemurTaskParamsVariant1.g.cs @@ -25,7 +25,7 @@ public sealed partial class LemurTaskParamsVariant1 public string? InputText { get; set; } /// - /// Max output size in tokens, up to 4000
+ /// Maximum output size in tokens, up to the `final_model`'s max [(see chart)](/docs/lemur/customize-parameters#change-the-maximum-output-size).
/// Default Value: 2000 ///
[global::System.Text.Json.Serialization.JsonPropertyName("max_output_size")] @@ -72,7 +72,7 @@ public sealed partial class LemurTaskParamsVariant1 /// Use either transcript_ids or input_text as input into LeMUR. /// /// - /// Max output size in tokens, up to 4000
+ /// Maximum output size in tokens, up to the `final_model`'s max [(see chart)](/docs/lemur/customize-parameters#change-the-maximum-output-size).
/// Default Value: 2000 /// /// diff --git a/src/libs/AssemblyAI/openapi.yaml b/src/libs/AssemblyAI/openapi.yaml index ea22fe4..a3ce96d 100644 --- a/src/libs/AssemblyAI/openapi.yaml +++ b/src/libs/AssemblyAI/openapi.yaml @@ -5610,7 +5610,7 @@ components: x-go-type: LeMURModel max_output_size: x-label: Maximum output size - description: Max output size in tokens, up to 4000 + description: Maximum output size in tokens, up to the `final_model`'s max [(see chart)](/docs/lemur/customize-parameters#change-the-maximum-output-size). type: integer default: 2000 temperature: @@ -5674,7 +5674,7 @@ components: - $ref: "#/components/schemas/LemurModel" max_output_size: x-label: Maximum output size - description: Max output size in tokens, up to 4000 + description: Maximum output size in tokens, up to the `final_model`'s max [(see chart)](/docs/lemur/customize-parameters#change-the-maximum-output-size). type: integer default: 2000 temperature: