-
-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to use a fine-tuned segmentation model for diarization? #840
Comments
This worked for me:
|
I will try that out, thanks! :) Heads up, you accidentally shared your HF_Token. Do you maybe also happen to know, how I can use speaker embeddings, that I extracted via reference speakers, in the diarization pipeline? I asked about that in the Pyannote Github, but didn't get a response unfortunately. pyannote/pyannote-audio#1750 |
Hey thanks for the heads up lol. Do let me know if you face any error in the above code cause I just tried this today and it worked. |
Thanks @Dream-gamer , works just fine using the tuned model loading! @Arche151 having the same problem, but using the fine-tuned model doesnt increase the accuracy of diarization that much.. how is it going on your side? any further steps taken to improve it? =) |
Glad to know. Yeah, it didn't noticably increase the accuracy. I have been using LLMs like gemini to parse the generated transcript and get corrected transcript. You can use the prompt like "Here is the speaker separated transcript. Some of the words in the transcript are in the wrong speaker labels. Correct them and give corrected transcript:" This has given me much better results for hindi-english transcript. |
thanks! :) |
I have a WhisperX Python script for transcribing meetings, but the speaker diarization for German is really bad, unfortunately.
After some research I came across the fine-tuned German segmentation model diarizers-community/speaker-segmentation-fine-tuned-callhome-deu but I haven't figured out how to get WhisperX to use it.
Here's my Python script:
I'd greatly appreciate any help!
The text was updated successfully, but these errors were encountered: