The sampling temperature, between 0 and 1. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic.
The log probabilities of the tokens in the transcription. Only returned with models gpt-4o-transcribe and gpt-4o-mini-transcribe if logprobs is added to the include array.
from dedalus_labs import DedalusLabsclient = DedalusLabs()# Basic transcriptionwith open("audio.mp3", "rb") as audio_file: transcription = client.audio.transcriptions.create( file=audio_file, model="openai/whisper-1" )print(transcription.text)
# Transcription with language hint for better accuracywith open("spanish_audio.wav", "rb") as audio_file: transcription = client.audio.transcriptions.create( file=audio_file, model="openai/whisper-1", language="es" )print(transcription.text)
# Get detailed transcription with timestampswith open("interview.m4a", "rb") as audio_file: transcription = client.audio.transcriptions.create( file=audio_file, model="openai/whisper-1", response_format="verbose_json" )print(f"Language: {transcription.language}")print(f"Duration: {transcription.duration}s")for word in transcription.words: print(f"{word.start:.2f}s - {word.end:.2f}s: {word.word}")
# Use prompt to guide transcription stylewith open("meeting.wav", "rb") as audio_file: transcription = client.audio.transcriptions.create( file=audio_file, model="openai/whisper-1", prompt="This is a business meeting discussing Q4 results and strategy.", temperature=0.2 )print(transcription.text)