objects76 commited on
Commit
c021ae5
1 Parent(s): 445a0de

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +129 -0
README.md ADDED
@@ -0,0 +1,129 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - pyannote
4
+ - pyannote-audio
5
+ - pyannote-audio-model
6
+ - audio
7
+ - voice
8
+ - speech
9
+ - speaker
10
+ - speaker-diarization
11
+ - speaker-change-detection
12
+ - speaker-segmentation
13
+ - voice-activity-detection
14
+ - overlapped-speech-detection
15
+ - resegmentation
16
+ license: mit
17
+ inference: false
18
+ extra_gated_prompt: "The collected information will help acquire a better knowledge of pyannote.audio userbase and help its maintainers improve it further. Though this model uses MIT license and will always remain open-source, we will occasionnally email you about premium models and paid services around pyannote."
19
+ extra_gated_fields:
20
+ Company/university: text
21
+ Website: text
22
+ ---
23
+
24
+ Using this open-source model in production?
25
+ Consider switching to [pyannoteAI](https://www.pyannote.ai) for better and faster options.
26
+
27
+ # 🎹 "Powerset" speaker segmentation
28
+
29
+ This model ingests 10 seconds of mono audio sampled at 16kHz and outputs speaker diarization as a (num_frames, num_classes) matrix where the 7 classes are _non-speech_, _speaker #1_, _speaker #2_, _speaker #3_, _speakers #1 and #2_, _speakers #1 and #3_, and _speakers #2 and #3_.
30
+
31
+ ![Example output](example.png)
32
+
33
+ ```python
34
+ # waveform (first row)
35
+ duration, sample_rate, num_channels = 10, 16000, 1
36
+ waveform = torch.randn(batch_size, num_channels, duration * sample_rate)
37
+
38
+ # powerset multi-class encoding (second row)
39
+ powerset_encoding = model(waveform)
40
+
41
+ # multi-label encoding (third row)
42
+ from pyannote.audio.utils.powerset import Powerset
43
+ max_speakers_per_chunk, max_speakers_per_frame = 3, 2
44
+ to_multilabel = Powerset(
45
+ max_speakers_per_chunk,
46
+ max_speakers_per_frame).to_multilabel
47
+ multilabel_encoding = to_multilabel(powerset_encoding)
48
+ ```
49
+
50
+ The various concepts behind this model are described in details in this [paper](https://www.isca-speech.org/archive/interspeech_2023/plaquet23_interspeech.html).
51
+
52
+ It has been trained by Séverin Baroudi with [pyannote.audio](https://github.com/pyannote/pyannote-audio) `3.0.0` using the combination of the training sets of AISHELL, AliMeeting, AMI, AVA-AVD, DIHARD, Ego4D, MSDWild, REPERE, and VoxConverse.
53
+
54
+ This [companion repository](https://github.com/FrenchKrab/IS2023-powerset-diarization/) by [Alexis Plaquet](https://frenchkrab.github.io/) also provides instructions on how to train or finetune such a model on your own data.
55
+
56
+ ## Requirements
57
+
58
+ 1. Install [`pyannote.audio`](https://github.com/pyannote/pyannote-audio) `3.0` with `pip install pyannote.audio`
59
+ 2. Accept [`pyannote/segmentation-3.0`](https://hf.co/pyannote/segmentation-3.0) user conditions
60
+ 3. Create access token at [`hf.co/settings/tokens`](https://hf.co/settings/tokens).
61
+
62
+
63
+ ## Usage
64
+
65
+ ```python
66
+ # instantiate the model
67
+ from pyannote.audio import Model
68
+ model = Model.from_pretrained(
69
+ "pyannote/segmentation-3.0",
70
+ use_auth_token="HUGGINGFACE_ACCESS_TOKEN_GOES_HERE")
71
+ ```
72
+
73
+ ### Speaker diarization
74
+
75
+ This model cannot be used to perform speaker diarization of full recordings on its own (it only processes 10s chunks).
76
+
77
+ See [pyannote/speaker-diarization-3.0](https://hf.co/pyannote/speaker-diarization-3.0) pipeline that uses an additional speaker embedding model to perform full recording speaker diarization.
78
+
79
+ ### Voice activity detection
80
+
81
+ ```python
82
+ from pyannote.audio.pipelines import VoiceActivityDetection
83
+ pipeline = VoiceActivityDetection(segmentation=model)
84
+ HYPER_PARAMETERS = {
85
+ # remove speech regions shorter than that many seconds.
86
+ "min_duration_on": 0.0,
87
+ # fill non-speech regions shorter than that many seconds.
88
+ "min_duration_off": 0.0
89
+ }
90
+ pipeline.instantiate(HYPER_PARAMETERS)
91
+ vad = pipeline("audio.wav")
92
+ # `vad` is a pyannote.core.Annotation instance containing speech regions
93
+ ```
94
+
95
+ ### Overlapped speech detection
96
+
97
+ ```python
98
+ from pyannote.audio.pipelines import OverlappedSpeechDetection
99
+ pipeline = OverlappedSpeechDetection(segmentation=model)
100
+ HYPER_PARAMETERS = {
101
+ # remove overlapped speech regions shorter than that many seconds.
102
+ "min_duration_on": 0.0,
103
+ # fill non-overlapped speech regions shorter than that many seconds.
104
+ "min_duration_off": 0.0
105
+ }
106
+ pipeline.instantiate(HYPER_PARAMETERS)
107
+ osd = pipeline("audio.wav")
108
+ # `osd` is a pyannote.core.Annotation instance containing overlapped speech regions
109
+ ```
110
+
111
+ ## Citations
112
+
113
+ ```bibtex
114
+ @inproceedings{Plaquet23,
115
+ author={Alexis Plaquet and Hervé Bredin},
116
+ title={{Powerset multi-class cross entropy loss for neural speaker diarization}},
117
+ year=2023,
118
+ booktitle={Proc. INTERSPEECH 2023},
119
+ }
120
+ ```
121
+
122
+ ```bibtex
123
+ @inproceedings{Bredin23,
124
+ author={Hervé Bredin},
125
+ title={{pyannote.audio 2.1 speaker diarization pipeline: principle, benchmark, and recipe}},
126
+ year=2023,
127
+ booktitle={Proc. INTERSPEECH 2023},
128
+ }
129
+ ```