From 1854241c1c59d0e3812469d0babce580a59d7ee0 Mon Sep 17 00:00:00 2001 From: AssemblyAI Date: Tue, 13 Feb 2024 09:48:04 -0500 Subject: [PATCH] Project import generated by Copybara. GitOrigin-RevId: 7826d32b11c9a7583b8697d6ce60f08020f36486 --- package.json | 2 +- src/types/openapi.generated.ts | 18 ++++++++++++++++++ 2 files changed, 19 insertions(+), 1 deletion(-) diff --git a/package.json b/package.json index 5887215..a4e1758 100644 --- a/package.json +++ b/package.json @@ -1,6 +1,6 @@ { "name": "assemblyai", - "version": "4.2.2", + "version": "4.2.3", "description": "The AssemblyAI JavaScript SDK provides an easy-to-use interface for interacting with the AssemblyAI API, which supports async and real-time transcription, as well as the latest LeMUR models.", "engines": { "node": ">=18" diff --git a/src/types/openapi.generated.ts b/src/types/openapi.generated.ts index 0f648d7..3e4a9f9 100644 --- a/src/types/openapi.generated.ts +++ b/src/types/openapi.generated.ts @@ -1102,6 +1102,12 @@ export type SeverityScoreSummary = { medium: number; }; +/** + * @description The speech model to use for the transcription. + * @enum {string} + */ +export type SpeechModel = "nano"; + /** * @description The replacement logic for detected PII, can be "entity_type" or "hash". See [PII redaction](https://www.assemblyai.com/docs/models/pii-redaction) for more details. * @enum {string} @@ -1310,6 +1316,7 @@ export type TopicDetectionResult = { * @description A transcript object * @example { * "id": "9ea68fd3-f953-42c1-9742-976c447fb463", + * "speech_model": null, * "language_model": "assemblyai_default", * "acoustic_model": "assemblyai_default", * "language_code": "en_us", @@ -2112,6 +2119,11 @@ export type Transcript = { speaker_labels?: boolean | null; /** @description Tell the speaker label model how many speakers it should attempt to identify, up to 10. See [Speaker diarization](https://www.assemblyai.com/docs/models/speaker-diarization) for more details. */ speakers_expected?: number | null; + /** + * @description The speech model used for the transcription. When `null`, the default model is used. + * @default null + */ + speech_model: SpeechModel | null; /** * Format: float * @description Defaults to null. Reject audio files that contain less than this fraction of speech. @@ -2272,6 +2284,7 @@ export type TranscriptListItem = { /** * @description The parameters for creating a transcript * @example { + * "speech_model": null, * "language_code": "en_us", * "punctuate": true, * "format_text": true, @@ -2377,6 +2390,11 @@ export type TranscriptOptionalParams = { * @default null */ speakers_expected?: number | null; + /** + * @description The speech model to use for the transcription. When `null`, the default model is used. + * @default null + */ + speech_model?: SpeechModel | null; /** * Format: float * @description Reject audio files that contain less than this fraction of speech.