diff --git a/fern/pages/audio-intelligence/content-moderation.mdx b/fern/pages/audio-intelligence/content-moderation.mdx index 8fbdd89d..d5041bab 100644 --- a/fern/pages/audio-intelligence/content-moderation.mdx +++ b/fern/pages/audio-intelligence/content-moderation.mdx @@ -91,7 +91,7 @@ response = requests.post(url, json=data, headers=headers) transcript_id = response.json()['id'] polling_endpoint = base_url + "/v2/transcript/" + transcript_id -print(f"Transcript ID:", transcript_id) +print(f"Transcript ID: {transcript_id}") while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() @@ -143,7 +143,7 @@ config = aai.TranscriptionConfig( transcript = aai.Transcriber().transcribe(audio_file, config) -print(f"Transcript ID:", transcript.id) +print(f"Transcript ID: {transcript.id}") for result in transcript.content_safety.results: print(result.text) diff --git a/fern/pages/audio-intelligence/entity-detection.mdx b/fern/pages/audio-intelligence/entity-detection.mdx index 2f5eeea4..2fc2a7bf 100644 --- a/fern/pages/audio-intelligence/entity-detection.mdx +++ b/fern/pages/audio-intelligence/entity-detection.mdx @@ -146,7 +146,7 @@ response = requests.post(url, json=data, headers=headers) transcript_id = response.json()['id'] polling_endpoint = base_url + "/v2/transcript/" + transcript_id -print(f"Transcript ID:", transcript_id) +print(f"Transcript ID: {transcript_id}") while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() @@ -183,7 +183,7 @@ config = aai.TranscriptionConfig( ) transcript = aai.Transcriber().transcribe(audio_file, config) -print(f"Transcript ID:", transcript.id) +print(f"Transcript ID: {transcript.id}") for entity in transcript.entities: print(entity.text) diff --git a/fern/pages/audio-intelligence/key-phrases.mdx b/fern/pages/audio-intelligence/key-phrases.mdx index 49351b30..07c1f8d2 100644 --- a/fern/pages/audio-intelligence/key-phrases.mdx +++ b/fern/pages/audio-intelligence/key-phrases.mdx @@ -92,7 +92,7 @@ response = requests.post(url, json=data, headers=headers) transcript_id = response.json()['id'] polling_endpoint = base_url + "/v2/transcript/" + transcript_id -print(f"Transcript ID:", transcript_id) +print(f"Transcript ID: {transcript_id}") while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() @@ -127,7 +127,7 @@ config = aai.TranscriptionConfig( ) transcript = aai.Transcriber().transcribe(audio_file, config) -print(f"Transcript ID:", transcript.id) +print(f"Transcript ID: {transcript.id}") for result in transcript.auto_highlights.results: print(f"Highlight: {result.text}, Count: {result.count}, Rank: {result.rank}, Timestamps: {result.timestamps}") diff --git a/fern/pages/audio-intelligence/pii-redaction.mdx b/fern/pages/audio-intelligence/pii-redaction.mdx index ed24a3b6..bd748aee 100644 --- a/fern/pages/audio-intelligence/pii-redaction.mdx +++ b/fern/pages/audio-intelligence/pii-redaction.mdx @@ -150,7 +150,7 @@ response = requests.post(url, json=data, headers=headers) transcript_id = response.json()['id'] polling_endpoint = base_url + "/v2/transcript/" + transcript_id -print(f"Transcript ID:", transcript_id) +print(f"Transcript ID: {transcript_id}") while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() @@ -192,7 +192,7 @@ config = aai.TranscriptionConfig( ) transcript = aai.Transcriber().transcribe(audio_file, config) -print(f"Transcript ID:", transcript.id) +print(f"Transcript ID: {transcript.id}") print(transcript.text) ``` diff --git a/fern/pages/audio-intelligence/sentiment-analysis.mdx b/fern/pages/audio-intelligence/sentiment-analysis.mdx index 54956e9b..a481c90c 100644 --- a/fern/pages/audio-intelligence/sentiment-analysis.mdx +++ b/fern/pages/audio-intelligence/sentiment-analysis.mdx @@ -84,7 +84,7 @@ response = requests.post(url, json=data, headers=headers) transcript_id = response.json()['id'] polling_endpoint = base_url + "/v2/transcript/" + transcript_id -print(f"Transcript ID:", transcript_id) +print(f"Transcript ID: {transcript_id}") while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() @@ -122,7 +122,7 @@ config = aai.TranscriptionConfig( ) transcript = aai.Transcriber().transcribe(audio_file, config) -print(f"Transcript ID:", transcript.id) +print(f"Transcript ID: {transcript.id}") for sentiment_result in transcript.sentiment_analysis: print(sentiment_result.text) diff --git a/fern/pages/audio-intelligence/topic-detection.mdx b/fern/pages/audio-intelligence/topic-detection.mdx index b78f4e55..fe01f81e 100644 --- a/fern/pages/audio-intelligence/topic-detection.mdx +++ b/fern/pages/audio-intelligence/topic-detection.mdx @@ -89,7 +89,7 @@ response = requests.post(url, json=data, headers=headers) transcript_id = response.json()['id'] polling_endpoint = base_url + "/v2/transcript/" + transcript_id -print(f"Transcript ID:", transcript_id) +print(f"Transcript ID: {transcript_id}") while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() @@ -133,7 +133,7 @@ config = aai.TranscriptionConfig( ) transcript = aai.Transcriber().transcribe(audio_file, config) -print(f"Transcript ID:", transcript.id) +print(f"Transcript ID: {transcript.id}") # Get the parts of the transcript that were tagged with topics for result in transcript.iab_categories.results: diff --git a/fern/pages/guides/cookbooks/core-transcription/automatic-language-detection-route-default-language.mdx b/fern/pages/guides/cookbooks/core-transcription/automatic-language-detection-route-default-language.mdx index c6baa2e1..24f8f62e 100644 --- a/fern/pages/guides/cookbooks/core-transcription/automatic-language-detection-route-default-language.mdx +++ b/fern/pages/guides/cookbooks/core-transcription/automatic-language-detection-route-default-language.mdx @@ -18,7 +18,7 @@ Install the SDK: npm install assemblyai ``` -```python +```bash pip install assemblyai ``` diff --git a/fern/pages/guides/cookbooks/core-transcription/batch_transcription.mdx b/fern/pages/guides/cookbooks/core-transcription/batch_transcription.mdx index 194d07b4..97d0ae22 100644 --- a/fern/pages/guides/cookbooks/core-transcription/batch_transcription.mdx +++ b/fern/pages/guides/cookbooks/core-transcription/batch_transcription.mdx @@ -46,7 +46,7 @@ Before we begin, make sure you have an AssemblyAI account and an API key. You ca Install the SDK. -```python +```bash pip install -U assemblyai ``` diff --git a/fern/pages/guides/cookbooks/core-transcription/do-more-with-sdk.mdx b/fern/pages/guides/cookbooks/core-transcription/do-more-with-sdk.mdx index 57409c86..e8f24ee3 100644 --- a/fern/pages/guides/cookbooks/core-transcription/do-more-with-sdk.mdx +++ b/fern/pages/guides/cookbooks/core-transcription/do-more-with-sdk.mdx @@ -20,19 +20,19 @@ pip show assemblyai If this version is not the same as the current version of the [Python SDK](https://github.com/AssemblyAI/assemblyai-python-sdk) then you can update your version by typing this code in your terminal: -```python +```bash pip install assemblyai --upgrade ``` -```js +```bash npm info assemblyai version ``` If this version is not the same as the current version of the [JavaScript SDK](https://github.com/AssemblyAI/assemblyai-node-sdk) then you can update your version by typing this code in your terminal: -```js +```bash npm update assemblyai --save ``` diff --git a/fern/pages/guides/cookbooks/core-transcription/gladia_to_aai.mdx b/fern/pages/guides/cookbooks/core-transcription/gladia_to_aai.mdx index 125659c1..333f54af 100644 --- a/fern/pages/guides/cookbooks/core-transcription/gladia_to_aai.mdx +++ b/fern/pages/guides/cookbooks/core-transcription/gladia_to_aai.mdx @@ -57,7 +57,7 @@ while True: transcript = requests.get(polling_endpoint, headers=headers).json() if transcript['status'] == 'done': - print(f"Full Transcript:", transcript['result']['transcription']['full_transcript']) + print(f"Full Transcript: {transcript['result']['transcription']['full_transcript']}") break elif transcript['status'] == 'error': @@ -103,7 +103,7 @@ while True: transcript = requests.get(polling_endpoint, headers=headers).json() if transcript['status'] == 'completed': - print(f"Full Transcript:", transcript['text']) + print(f"Full Transcript: {transcript['text']}") break elif transcript['status'] == 'error': @@ -248,7 +248,7 @@ while True: transcript = requests.get(polling_endpoint, headers=headers).json() if transcript['status'] == 'done': - print(f"Full Transcript:", transcript['result']['transcription']['full_transcript']) + print(f"Full Transcript: {transcript['result']['transcription']['full_transcript']}") break elif transcript['status'] == 'error': @@ -284,7 +284,7 @@ while True: transcript = requests.get(polling_endpoint, headers=headers).json() if transcript['status'] == 'completed': - print(f"Full Transcript:", transcript['text']) + print(f"Full Transcript: {transcript['text']}") break elif transcript['status'] == 'error': diff --git a/fern/pages/guides/cookbooks/core-transcription/subtitle_creation_by_word_count.mdx b/fern/pages/guides/cookbooks/core-transcription/subtitle_creation_by_word_count.mdx index fbe197b8..1d8b8b3f 100644 --- a/fern/pages/guides/cookbooks/core-transcription/subtitle_creation_by_word_count.mdx +++ b/fern/pages/guides/cookbooks/core-transcription/subtitle_creation_by_word_count.mdx @@ -59,7 +59,7 @@ print("SRT file generated.") ## Step-by-Step Instructions -```python +```bash pip install -U assemblyai ``` diff --git a/fern/pages/guides/cookbooks/core-transcription/subtitles.mdx b/fern/pages/guides/cookbooks/core-transcription/subtitles.mdx index 347957e0..8af822c5 100644 --- a/fern/pages/guides/cookbooks/core-transcription/subtitles.mdx +++ b/fern/pages/guides/cookbooks/core-transcription/subtitles.mdx @@ -44,7 +44,7 @@ The entire source code of this guide can be viewed [here](https://github.com/Ass Install the SDK. -```python +```bash pip install -U assemblyai ``` diff --git a/fern/pages/guides/cookbooks/lemur/input-text-chapters.mdx b/fern/pages/guides/cookbooks/lemur/input-text-chapters.mdx index f1d91ce4..55af83c7 100644 --- a/fern/pages/guides/cookbooks/lemur/input-text-chapters.mdx +++ b/fern/pages/guides/cookbooks/lemur/input-text-chapters.mdx @@ -117,7 +117,7 @@ polling_endpoint = base_url + "/v2/transcript/" + transcript_id while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() if transcription_result['status'] == 'completed': - print(f"Transcript ID:", transcript_id) + print(f"Transcript ID: {transcript_id}") break elif transcription_result['status'] == 'error': raise RuntimeError(f"Transcription failed: {transcription_result['error']}") diff --git a/fern/pages/guides/cookbooks/lemur/input-text-speaker-labels.mdx b/fern/pages/guides/cookbooks/lemur/input-text-speaker-labels.mdx index 8522a9b6..ffb5988b 100644 --- a/fern/pages/guides/cookbooks/lemur/input-text-speaker-labels.mdx +++ b/fern/pages/guides/cookbooks/lemur/input-text-speaker-labels.mdx @@ -54,7 +54,7 @@ In this guide, we will include Speaker Labels to the conversation by using the ` Install the SDK. -```python +```bash pip install -U assemblyai ``` diff --git a/fern/pages/guides/cookbooks/lemur/phone-call-segmentation.mdx b/fern/pages/guides/cookbooks/lemur/phone-call-segmentation.mdx index 20949d29..283bde5e 100644 --- a/fern/pages/guides/cookbooks/lemur/phone-call-segmentation.mdx +++ b/fern/pages/guides/cookbooks/lemur/phone-call-segmentation.mdx @@ -132,7 +132,7 @@ while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() if transcription_result['status'] == 'completed': - print(f"Transcript ID:", transcript_id) + print(f"Transcript ID: {transcript_id}") break elif transcription_result['status'] == 'error': diff --git a/fern/pages/guides/cookbooks/lemur/soap-note-generation.mdx b/fern/pages/guides/cookbooks/lemur/soap-note-generation.mdx index 1e79daca..54991026 100644 --- a/fern/pages/guides/cookbooks/lemur/soap-note-generation.mdx +++ b/fern/pages/guides/cookbooks/lemur/soap-note-generation.mdx @@ -50,7 +50,7 @@ response = requests.post(url, json=data, headers=headers) transcript_id = response.json()['id'] polling_endpoint = base_url + "/v2/transcript/" + transcript_id -print(f"Transcript ID:", transcript_id) +print(f"Transcript ID: {transcript_id}") while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() diff --git a/fern/pages/guides/generate-meeting-action-items-with-lemur.mdx b/fern/pages/guides/generate-meeting-action-items-with-lemur.mdx index b8889ec2..509ee45a 100644 --- a/fern/pages/guides/generate-meeting-action-items-with-lemur.mdx +++ b/fern/pages/guides/generate-meeting-action-items-with-lemur.mdx @@ -24,7 +24,7 @@ In this guide, you'll use AssemblyAI to transcribe a meeting recording. Then, yo Install the SDK. -```python +```bash pip install -U assemblyai ``` @@ -33,7 +33,7 @@ pip install -U assemblyai Install the SDK. -```javascript +```bash npm install assemblyai ``` diff --git a/fern/pages/guides/generating-subtitles-for-videos.mdx b/fern/pages/guides/generating-subtitles-for-videos.mdx index 9463bbf3..8cf8d475 100644 --- a/fern/pages/guides/generating-subtitles-for-videos.mdx +++ b/fern/pages/guides/generating-subtitles-for-videos.mdx @@ -65,9 +65,8 @@ request. ```python - - - +import requests +import time ``` @@ -75,7 +74,7 @@ request. Install the SDK. -```python +```bash pip install -U assemblyai ``` @@ -83,7 +82,8 @@ pip install -U assemblyai ```javascript - +import axios from "axios"; +import fs from "fs-extra"; ``` @@ -105,7 +105,7 @@ key. ```python -base_url = "https://api.assemblyai.com/v2" +base_url = "https://api.assemblyai.com" headers = { "authorization": "" @@ -125,7 +125,7 @@ aai.settings.api_key = "" ```javascript -const baseUrl = "https://api.assemblyai.com/v2"; +const baseUrl = "https://api.assemblyai.com"; const headers = { authorization: "", @@ -156,7 +156,7 @@ Create a `Transcriber` object. ```python with open("./my-audio.mp3", "rb") as f: - response = requests.post(base_url + "/upload", + response = requests.post(base_url + "/v2/upload", headers=headers, data=f) @@ -176,7 +176,7 @@ transcriber = aai.Transcriber() ```javascript const path = "./my-audio.mp3"; const audioData = await fs.readFile(path); -const uploadResponse = await axios.post(`${baseUrl}/upload`, audioData, { +const uploadResponse = await axios.post(`${baseUrl}/v2/upload`, audioData, { headers, }); const uploadUrl = uploadResponse.data.upload_url; @@ -252,7 +252,7 @@ headers. ```python -url = base_url + "/transcript" +url = base_url + "/v2/transcript" response = requests.post(url, json=data, headers=headers) ``` @@ -267,7 +267,7 @@ transcript = transcriber.transcribe("https://example.org/audio.mp3") ```javascript -const url = `${baseUrl}/transcript`; +const url = `${baseUrl}/v2/transcript`; const response = await axios.post(url, data, { headers: headers }); ``` @@ -398,10 +398,7 @@ print(transcript.export_subtitles_vtt()) ```javascript -async function getSubtitleFile( - transcriptId: string, - fileFormat: "srt" -): Promise { +async function getSubtitleFile(transcriptId, fileFormat) { if (!["srt", "vtt"].includes(fileFormat)) { throw new Error( `Unsupported file format: ${fileFormat}. Please specify 'srt' or 'vtt'.` @@ -411,9 +408,9 @@ async function getSubtitleFile( const url = `https://api.assemblyai.com/v2/transcript/${transcriptId}/${fileFormat}`; try { - const response = await axios.get(url, { headers }); + const response = await axios.get(url, { headers }); return response.data; - } catch (error: any) { + } catch (error) { throw new Error( `Failed to retrieve ${fileFormat.toUpperCase()} file: ${ error.response?.status diff --git a/fern/pages/guides/identifying-hate-speech-in-audio-or-video-files.mdx b/fern/pages/guides/identifying-hate-speech-in-audio-or-video-files.mdx index 6c2d2aa7..6f55762d 100644 --- a/fern/pages/guides/identifying-hate-speech-in-audio-or-video-files.mdx +++ b/fern/pages/guides/identifying-hate-speech-in-audio-or-video-files.mdx @@ -41,7 +41,7 @@ import time Install the SDK. -```python +```bash pip install -U assemblyai ``` @@ -69,7 +69,7 @@ import fs from "fs-extra"; key. ```python -base_url = "https://api.assemblyai.com/v2" +base_url = "https://api.assemblyai.com" headers = { "authorization": "" @@ -93,7 +93,7 @@ aai.settings.api_key = "" key. ```javascript -const baseUrl = "https://api.assemblyai.com/v2"; +const baseUrl = "https://api.assemblyai.com"; const headers = { authorization: "", @@ -112,7 +112,7 @@ const headers = { ```python with open("./my-audio.mp3", "rb") as f: - response = requests.post(base_url + "/upload", + response = requests.post(base_url + "/v2/upload", headers=headers, data=f) @@ -135,7 +135,7 @@ config = aai.TranscriptionConfig(content_safety=True) ```javascript const path = "./my-audio.mp3"; const audioData = await fs.readFile(path); -const urlResponse = await axios.post(`${baseUrl}/upload`, audioData, { +const urlResponse = await axios.post(`${baseUrl}/v2/upload`, audioData, { headers, }); const uploadUrl = urlResponse.data.upload_url; @@ -193,7 +193,7 @@ const data = { headers. ```python -url = base_url + "/transcript" +url = base_url + "/v2/transcript" response = requests.post(url, json=data, headers=headers) ``` @@ -214,7 +214,7 @@ transcript = transcriber.transcribe(FILE_URL) headers. ```javascript -const url = `${baseUrl}/transcript`; +const url = `${baseUrl}/v2/transcript`; const response = await axios.post(url, data, { headers: headers }); ``` diff --git a/fern/pages/guides/identifying-highlights-in-audio-or-video-files.mdx b/fern/pages/guides/identifying-highlights-in-audio-or-video-files.mdx index 56891d42..97cadfc1 100644 --- a/fern/pages/guides/identifying-highlights-in-audio-or-video-files.mdx +++ b/fern/pages/guides/identifying-highlights-in-audio-or-video-files.mdx @@ -41,7 +41,7 @@ import time Install the SDK. -```python +```bash pip install -U assemblyai ``` @@ -67,7 +67,7 @@ import fs from "fs-extra"; key. ```python -base_url = "https://api.assemblyai.com/v2" +base_url = "https://api.assemblyai.com" headers = { "authorization": "" @@ -90,7 +90,7 @@ aai.settings.api_key = "" key. ```javascript -const baseUrl = "https://api.assemblyai.com/v2"; +const baseUrl = "https://api.assemblyai.com"; const headers = { authorization: "", @@ -109,7 +109,7 @@ const headers = { ```python with open("./my-audio.mp3", "rb") as f: - response = requests.post(base_url + "/upload", + response = requests.post(base_url + "/v2/upload", headers=headers, data=f) @@ -132,7 +132,7 @@ config = aai.TranscriptionConfig(auto_highlights=True) ```javascript const path = "./my-audio.mp3"; const audioData = await fs.readFile(path); -const uploadResponse = await axios.post(`${baseUrl}/upload`, audioData, { +const uploadResponse = await axios.post(`${baseUrl}/v2/upload`, audioData, { headers, }); const uploadUrl = uploadResponse.data.upload_url; @@ -190,7 +190,7 @@ const data = { headers. ```python -url = base_url + "/transcript" +url = base_url + "/v2/transcript" response = requests.post(url, json=data, headers=headers) ``` @@ -210,7 +210,7 @@ transcript = transcriber.transcribe(FILE_URL) headers. ```javascript -const url = `${baseUrl}/transcript`; +const url = `${baseUrl}/v2/transcript`; const response = await axios.post(url, data, { headers: headers }); ``` diff --git a/fern/pages/guides/identifying-speakers-in-audio-recordings.mdx b/fern/pages/guides/identifying-speakers-in-audio-recordings.mdx index 3aed9608..39d7d42e 100644 --- a/fern/pages/guides/identifying-speakers-in-audio-recordings.mdx +++ b/fern/pages/guides/identifying-speakers-in-audio-recordings.mdx @@ -42,15 +42,14 @@ request. ```python - - - +import requests +import time ``` -```python +```bash pip install -U assemblyai ``` @@ -58,7 +57,8 @@ pip install -U assemblyai ```javascript - +import axios from "axios"; +import fs from "fs-extra"; ``` @@ -84,7 +84,7 @@ key. ```python -base_url = "https://api.assemblyai.com/v2" +base_url = "https://api.assemblyai.com" headers = { "authorization": "" @@ -104,7 +104,7 @@ aai.settings.api_key = "" ```javascript -const baseUrl = "https://api.assemblyai.com/v2"; +const baseUrl = "https://api.assemblyai.com"; const headers = { authorization: "", @@ -134,7 +134,7 @@ Upload your local file to the AssemblyAI API. ```python with open("./my-audio.mp3", "rb") as f: - response = requests.post(base_url + "/upload", + response = requests.post(base_url + "/v2/upload", headers=headers, data=f) @@ -155,7 +155,7 @@ config = aai.TranscriptionConfig(speaker_labels=True) ```javascript const path = "./my-audio.mp3"; const audioData = await fs.readFile(path); -const uploadResponse = await axios.post(`${baseUrl}/upload`, audioData, { +const uploadResponse = await axios.post(`${baseUrl}/v2/upload`, audioData, { headers, }); const uploadUrl = uploadResponse.data.upload_url; @@ -233,7 +233,7 @@ headers. ```python -url = base_url + "/transcript" +url = base_url + "/v2/transcript" response = requests.post(url, json=data, headers=headers) ``` @@ -250,7 +250,7 @@ transcript = transcriber.transcribe(FILE_URL) ```javascript -const url = `${baseUrl}/transcript`; +const url = `${baseUrl}/v2/transcript`; const response = await axios.post(url, data, { headers: headers }); ``` diff --git a/fern/pages/guides/transcribing-an-audio-file.mdx b/fern/pages/guides/transcribing-an-audio-file.mdx index 2a9783aa..7b502b9a 100644 --- a/fern/pages/guides/transcribing-an-audio-file.mdx +++ b/fern/pages/guides/transcribing-an-audio-file.mdx @@ -216,7 +216,7 @@ After making the request, you'll receive an ID for the transcription. Use it to ```python transcript_id = response.json()['id'] -print(f"Transcript ID: ", transcript_id) +print(f"Transcript ID: {transcript_id}") polling_endpoint = f"https://api.assemblyai.com/v2/transcript/{transcript_id}" @@ -240,7 +240,7 @@ while True: You can access the transcription ID and results through the `Transcriber` object's `id` and `text` attributes. ```python -print(f"Transcript ID: ", transcript.id) +print(f"Transcript ID: {transcript.id}") print(transcript.text) ``` diff --git a/fern/pages/guides/webhooks.mdx b/fern/pages/guides/webhooks.mdx index deeee44e..ab2b6bf2 100644 --- a/fern/pages/guides/webhooks.mdx +++ b/fern/pages/guides/webhooks.mdx @@ -203,7 +203,7 @@ polling_endpoint = f"https://api.assemblyai.com/v2/transcript/{transcript_id}" transcription_result = requests.get(polling_endpoint, headers=headers).json() if transcription_result['status'] == 'completed': - print(f"Transcript ID:", transcript_id) + print(f"Transcript ID: {transcript_id}") elif transcription_result['status'] == 'error': raise RuntimeError(f"Transcription failed: {transcription_result['error']}") diff --git a/fern/pages/speech-to-text/pre-recorded-audio/automatic-language-detection.mdx b/fern/pages/speech-to-text/pre-recorded-audio/automatic-language-detection.mdx index f4ed9cbd..3044d242 100644 --- a/fern/pages/speech-to-text/pre-recorded-audio/automatic-language-detection.mdx +++ b/fern/pages/speech-to-text/pre-recorded-audio/automatic-language-detection.mdx @@ -207,9 +207,9 @@ while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() if transcription_result['status'] == 'completed': - print(f"Transcript ID:", transcript_id) - print(f"Language Code:", transcription_result['language_code']) - print(f"Text:", transcription_result['text']) + print(f"Transcript ID: {transcript_id}") + print(f"Language Code: {transcription_result['language_code']}") + print(f"Text: {transcription_result['text']}") break elif transcription_result['status'] == 'error': @@ -366,9 +366,9 @@ while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() if transcription_result['status'] == 'completed': - print(f"Transcript ID:", transcript_id) - print(f"Language Code:", transcription_result['language_code']) - print(f"Text:", transcription_result['text']) + print(f"Transcript ID: {transcript_id}") + print(f"Language Code: {transcription_result['language_code']}") + print(f"Text: {transcription_result['text']}") break elif transcription_result['status'] == 'error': @@ -534,9 +534,9 @@ while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() if transcription_result['status'] == 'completed': - print(f"Transcript ID:", transcript_id) - print(f"Language Code:", transcription_result['language_code']) - print(f"Text:", transcription_result['text']) + print(f"Transcript ID: {transcript_id}") + print(f"Language Code: {transcription_result['language_code']}") + print(f"Text: {transcription_result['text']}") break elif transcription_result['status'] == 'error': @@ -688,9 +688,9 @@ while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() if transcription_result['status'] == 'completed': - print(f"Transcript ID:", transcript_id) - print(f"Language Confidence:", transcription_result['language_confidence']) - print(f"Text:", transcription_result['text']) + print(f"Transcript ID: {transcript_id}") + print(f"Language Confidence: {transcription_result['language_confidence']}") + print(f"Text: {transcription_result['text']}") break elif transcription_result['status'] == 'error': @@ -840,8 +840,8 @@ while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() if transcription_result['status'] == 'completed': - print(f"Transcript ID:", transcript_id) - print(f"Text:", transcription_result['text']) + print(f"Transcript ID: {transcript_id}") + print(f"Text: {transcription_result['text']}") break elif transcription_result['status'] == 'error': diff --git a/fern/pages/speech-to-text/pre-recorded-audio/code-switching-docs.mdx b/fern/pages/speech-to-text/pre-recorded-audio/code-switching-docs.mdx index a630ffe1..dc2f3f8b 100644 --- a/fern/pages/speech-to-text/pre-recorded-audio/code-switching-docs.mdx +++ b/fern/pages/speech-to-text/pre-recorded-audio/code-switching-docs.mdx @@ -209,7 +209,7 @@ while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() if transcription_result['status'] == 'completed': - print(f"Transcript:", transcription_result['text']) + print(f"Transcript: {transcription_result['text']}") break elif transcription_result['status'] == 'error': @@ -391,7 +391,7 @@ while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() if transcription_result['status'] == 'completed': - print(f"Transcript:", transcription_result['text']) + print(f"Transcript: {transcription_result['text']}") break elif transcription_result['status'] == 'error': @@ -566,7 +566,7 @@ while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() if transcription_result['status'] == 'completed': - print(f"Transcript:", transcription_result['text']) + print(f"Transcript: {transcription_result['text']}") break elif transcription_result['status'] == 'error': diff --git a/fern/pages/speech-to-text/pre-recorded-audio/custom-spelling.mdx b/fern/pages/speech-to-text/pre-recorded-audio/custom-spelling.mdx index 17aabcf3..b6782bcd 100644 --- a/fern/pages/speech-to-text/pre-recorded-audio/custom-spelling.mdx +++ b/fern/pages/speech-to-text/pre-recorded-audio/custom-spelling.mdx @@ -64,7 +64,7 @@ while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() if transcription_result['status'] == 'completed': - print(f"Transcript ID:", transcript_id) + print(f"Transcript ID: {transcript_id}") break elif transcription_result['status'] == 'error': diff --git a/fern/pages/speech-to-text/pre-recorded-audio/custom-vocabulary.mdx b/fern/pages/speech-to-text/pre-recorded-audio/custom-vocabulary.mdx index bf5f3f3d..b1196149 100644 --- a/fern/pages/speech-to-text/pre-recorded-audio/custom-vocabulary.mdx +++ b/fern/pages/speech-to-text/pre-recorded-audio/custom-vocabulary.mdx @@ -71,7 +71,7 @@ while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() if transcription_result['status'] == 'completed': - print(f"Transcript ID:", transcript_id) + print(f"Transcript ID: {transcript_id}") break elif transcription_result['status'] == 'error': diff --git a/fern/pages/speech-to-text/pre-recorded-audio/delete-transcripts.mdx b/fern/pages/speech-to-text/pre-recorded-audio/delete-transcripts.mdx index 5ffb2a51..b9fed09f 100644 --- a/fern/pages/speech-to-text/pre-recorded-audio/delete-transcripts.mdx +++ b/fern/pages/speech-to-text/pre-recorded-audio/delete-transcripts.mdx @@ -72,7 +72,7 @@ while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() if transcription_result['status'] == 'completed': - print(f"Transcript ID:", transcript_id) + print(f"Transcript ID: {transcript_id}") break elif transcription_result['status'] == 'error': diff --git a/fern/pages/speech-to-text/pre-recorded-audio/filler-words.mdx b/fern/pages/speech-to-text/pre-recorded-audio/filler-words.mdx index a9735c43..48fd99f7 100644 --- a/fern/pages/speech-to-text/pre-recorded-audio/filler-words.mdx +++ b/fern/pages/speech-to-text/pre-recorded-audio/filler-words.mdx @@ -120,7 +120,7 @@ while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() if transcription_result['status'] == 'completed': - print(f"Transcript ID:", transcript_id) + print(f"Transcript ID: {transcript_id}") break elif transcription_result['status'] == 'error': diff --git a/fern/pages/speech-to-text/pre-recorded-audio/multichannel-transcription.mdx b/fern/pages/speech-to-text/pre-recorded-audio/multichannel-transcription.mdx index 65f8d3a5..7a8117d3 100644 --- a/fern/pages/speech-to-text/pre-recorded-audio/multichannel-transcription.mdx +++ b/fern/pages/speech-to-text/pre-recorded-audio/multichannel-transcription.mdx @@ -84,7 +84,7 @@ while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() if transcription_result['status'] == 'completed': - print(f"Transcript ID:", transcript_id) + print(f"Transcript ID: {transcript_id}") break elif transcription_result['status'] == 'error': @@ -257,7 +257,7 @@ while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() if transcription_result['status'] == 'completed': - print(f"Transcript ID:", transcript_id) + print(f"Transcript ID: {transcript_id}") break elif transcription_result['status'] == 'error': diff --git a/fern/pages/speech-to-text/pre-recorded-audio/profanity-filtering.mdx b/fern/pages/speech-to-text/pre-recorded-audio/profanity-filtering.mdx index 8153dce2..de0c794b 100644 --- a/fern/pages/speech-to-text/pre-recorded-audio/profanity-filtering.mdx +++ b/fern/pages/speech-to-text/pre-recorded-audio/profanity-filtering.mdx @@ -114,7 +114,7 @@ while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() if transcription_result['status'] == 'completed': - print(f"Transcript ID:", transcript_id) + print(f"Transcript ID: {transcript_id}") break elif transcription_result['status'] == 'error': diff --git a/fern/pages/speech-to-text/pre-recorded-audio/select-the-region.mdx b/fern/pages/speech-to-text/pre-recorded-audio/select-the-region.mdx index 4649be49..eb32e577 100644 --- a/fern/pages/speech-to-text/pre-recorded-audio/select-the-region.mdx +++ b/fern/pages/speech-to-text/pre-recorded-audio/select-the-region.mdx @@ -98,7 +98,7 @@ while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() if transcription_result['status'] == 'completed': - print(f"Transcript ID:", transcript_id) + print(f"Transcript ID: {transcript_id}") break elif transcription_result['status'] == 'error': @@ -246,7 +246,7 @@ while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() if transcription_result['status'] == 'completed': - print(f"Transcript ID:", transcript_id) + print(f"Transcript ID: {transcript_id}") break elif transcription_result['status'] == 'error': diff --git a/fern/pages/speech-to-text/pre-recorded-audio/set-language-manually.mdx b/fern/pages/speech-to-text/pre-recorded-audio/set-language-manually.mdx index ca9ae6e7..414275e7 100644 --- a/fern/pages/speech-to-text/pre-recorded-audio/set-language-manually.mdx +++ b/fern/pages/speech-to-text/pre-recorded-audio/set-language-manually.mdx @@ -92,7 +92,7 @@ while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() if transcription_result['status'] == 'completed': - print(f"Transcript ID:", transcript_id) + print(f"Transcript ID: {transcript_id}") break elif transcription_result['status'] == 'error': diff --git a/fern/pages/speech-to-text/pre-recorded-audio/set-the-start-and-end-of-the-transcript.mdx b/fern/pages/speech-to-text/pre-recorded-audio/set-the-start-and-end-of-the-transcript.mdx index a8eacd48..774dfcde 100644 --- a/fern/pages/speech-to-text/pre-recorded-audio/set-the-start-and-end-of-the-transcript.mdx +++ b/fern/pages/speech-to-text/pre-recorded-audio/set-the-start-and-end-of-the-transcript.mdx @@ -68,7 +68,7 @@ while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() if transcription_result['status'] == 'completed': - print(f"Transcript ID:", transcript_id) + print(f"Transcript ID: {transcript_id}") break elif transcription_result['status'] == 'error': diff --git a/fern/pages/speech-to-text/pre-recorded-audio/speaker-diarization.mdx b/fern/pages/speech-to-text/pre-recorded-audio/speaker-diarization.mdx index ba127f4b..0609966c 100644 --- a/fern/pages/speech-to-text/pre-recorded-audio/speaker-diarization.mdx +++ b/fern/pages/speech-to-text/pre-recorded-audio/speaker-diarization.mdx @@ -199,7 +199,7 @@ while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() if transcription_result['status'] == 'completed': - print(f"Transcript ID:", transcript_id) + print(f"Transcript ID: {transcript_id}") break elif transcription_result['status'] == 'error': diff --git a/fern/pages/speech-to-text/pre-recorded-audio/speech-threshold.mdx b/fern/pages/speech-to-text/pre-recorded-audio/speech-threshold.mdx index 3fe363ec..ddcf00fa 100644 --- a/fern/pages/speech-to-text/pre-recorded-audio/speech-threshold.mdx +++ b/fern/pages/speech-to-text/pre-recorded-audio/speech-threshold.mdx @@ -85,7 +85,7 @@ while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() if transcription_result['status'] == 'completed': - print(f"Transcript ID:", transcript_id) + print(f"Transcript ID: {transcript_id}") break elif transcription_result['status'] == 'error': diff --git a/fern/pages/speech-to-text/pre-recorded-audio/transcript-export-options.mdx b/fern/pages/speech-to-text/pre-recorded-audio/transcript-export-options.mdx index 346db96e..19250399 100644 --- a/fern/pages/speech-to-text/pre-recorded-audio/transcript-export-options.mdx +++ b/fern/pages/speech-to-text/pre-recorded-audio/transcript-export-options.mdx @@ -87,7 +87,7 @@ while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() if transcription_result['status'] == 'completed': - print(f"Transcript ID:", transcript_id) + print(f"Transcript ID: {transcript_id}") break elif transcription_result['status'] == 'error': @@ -261,7 +261,7 @@ while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() if transcription_result['status'] == 'completed': - print(f"Transcript ID:", transcript_id) + print(f"Transcript ID: {transcript_id}") break elif transcription_result['status'] == 'error': @@ -424,7 +424,7 @@ while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() if transcription_result['status'] == 'completed': - print(f"Transcript ID:", transcript_id) + print(f"Transcript ID: {transcript_id}") break elif transcription_result['status'] == 'error': diff --git a/fern/pages/speech-to-text/pre-recorded-audio/word-search.mdx b/fern/pages/speech-to-text/pre-recorded-audio/word-search.mdx index b36abb35..6aaeeee1 100644 --- a/fern/pages/speech-to-text/pre-recorded-audio/word-search.mdx +++ b/fern/pages/speech-to-text/pre-recorded-audio/word-search.mdx @@ -79,7 +79,7 @@ while True: transcription_result = requests.get(polling_endpoint, headers=headers).json() if transcription_result['status'] == 'completed': - print(f"Transcript ID:", transcript_id) + print(f"Transcript ID: {transcript_id}") break elif transcription_result['status'] == 'error':