-
Notifications
You must be signed in to change notification settings - Fork 439
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to get embeddings of audio data streaming from microphone. #56
Comments
The difficult part of the implementation is to get a reliable system for receiving these chunks and for triggering a function call when enough chunks are gathered to compute an embedding. If you have that already, that's great. Take a look at |
To do that I'm using code provided by Google for streaming speech recognition on an audio stream I am getting embeddings but I believe that I'm doing something wrong since the clustering algo is producing a single class (cluster) while trying to perform speaker diarization on the extracted embedding Here's what my code looks like :
|
How to avoid losing information when you split a file into chunks. |
I am also trying to implement this function, have you implemented it, or have any good suggestions?My email is [email protected],hope your reply.Thanks. |
I am using resemblyzer to create embeddings for speaker diarization.
It works fine when a whole wave file is loaded into the resemblyzer.
Now I want to try out real-time speaker diarization using data streaming from microphone using pyaudio (in form of chunks).
A chunk is essentially a frame of fixed size (100 ms in my case).
How do I get separate embedding for each chunk using resemblyzer?
The text was updated successfully, but these errors were encountered: