Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Can i use this sdk for multispeaker(2)? #63

Open
aj7tesh opened this issue Feb 9, 2022 · 2 comments
Open

Can i use this sdk for multispeaker(2)? #63

aj7tesh opened this issue Feb 9, 2022 · 2 comments

Comments

@aj7tesh
Copy link

aj7tesh commented Feb 9, 2022

I have been trying to run this sdk for multispeaker transcript generation from a wav file. However the results are really poor, not even a single word is identified correctly. Am i missing something or this sdk doesnt support 2 speakers

@aj7tesh
Copy link
Author

aj7tesh commented Feb 9, 2022

is there a way to generae transcript when we have 2 speaker speaking in 2 languages where speaker 1 speaks in 2 languages. i.e code switching asr

@DaaS-20xx
Copy link

Same issue here. Using only parameter "show_speaker_label=True" in input to client.start_stream_transcription, does not work. Actually the output is generated and the "transcript" parameter within the "alternative[0]" objects is returned, as well as the "speaker" parameter from "alternatives[0].items[0].speaker". But, as [aj7tesh said, there is NO match at all with the speech. It seems a totally casual and different audio at all.
Is there any other parameters configuration to be given in input?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants