SpeakerMap Model with Voice Signatures

Together with voice signatures created using the voice signatures feature, the SpeakerMap can be used to detect known speakers. Unknown new speakers will still be detected as well. For example usage with the sdk see the Voice Signatures Recipes page under the Usage ➜ Recipes section on the sidebar.

The Voice Signatures are created from audio containing known speakers. Then, the voice signatures can be passed to the SpeakerMap model as an additional input.

Voice signatures are stored in an object mapping the speaker_id to the voice signature that contains the base64 serialized speakers identity data:

{
"pedro": {
"version": 1,
"data": "U3VwZXJkdXBlcm1lZ2FzdGFydm9pY2VzaWduYXR1cmVzMQ..." # base64 encoded voice signature
},
"mariyana": {
"version": 1,
"data": "T1RPIGlzIGdyZWF0ISBJdCdzIHRydWUu..." # base64 encoded voice signature
},
}

Same as without Voice Signatures, the SpeakerMap classifies audio with a speaker label. For known speakers it will return the speaker_id of the voice signature. For unknown speakers a default label speaker_1, speaker_2, etc is returned. unknown is returned, if the speaker could not be identified.

All the outputs of the SpeakerMap model remain the same as when used without Voice Signatures, except that the speaker labels can be user defined identifiers for known speakers.