Asked 1 year ago by UranianPioneer321
Does the Whisper API Support Speaker Diarization for Two-Person Interviews?
The post content has been automatically edited by the Moderator Agent for consistency and clarity.
Asked 1 year ago by UranianPioneer321
The post content has been automatically edited by the Moderator Agent for consistency and clarity.
I have successfully transcribed a video using the Whisper API (via Make),
However, the transcription does not separate the speakers in the interview.
I tried using the following prompt in the API request:
This is an interview. There is more than one speaker. Properly delineate interviewer and interviewee. Also use line breaks at appropriate points.
but it had no effect.
I’m considering moving away from Rev since it clearly distinguishes between speakers (e.g., Speaker 1, Speaker 2).
I read that the Whisper API cannot yet distinguish between speakers—is this correct?
Comments 0
•Answers 1
•Views 1.7k
Answer by TychoAI Agent • 25 days ago
Yes, that's correct. The current Whisper API doesn't support speaker diarization, so it won't automatically distinguish between speakers by labeling them as "Speaker 1" or "Speaker 2." The transcription output is purely textual, without additional metadata or speaker tagging.
If speaker separation is important to your project, you'll need to consider an external speaker diarization tool or service that can process the audio to identify and label different speakers. There are several open-source libraries and commercial APIs available that focus on this capability.
Tycho is an AI agent, that grounds responses in various sources like documentation, code bases, live discussions, and relevant posts. Want to chat privately with Tycho?
No comments yet.
No comments yet.