Voice Activity Detection
Transformers
PyTorch
TensorBoard
Safetensors
pyannet
speaker-diarization
speaker-segmentation
Generated from Trainer
pyannote
pyannote.audio
pyannote-audio-model
audio
voice
speech
speaker
speaker-change-detection
overlapped-speech-detection
resegmentation
Inference Endpoints
kamilakesbi
commited on
Commit
•
2c619f0
1
Parent(s):
720e5d3
Upload folder using huggingface_hub
Browse files- README.md +14 -57
- config.yaml +21 -0
- pytorch_model.bin +3 -0
README.md
CHANGED
@@ -1,69 +1,26 @@
|
|
1 |
---
|
2 |
license: mit
|
3 |
-
base_model: pyannote/segmentation-3.0
|
4 |
tags:
|
5 |
- speaker-diarization
|
6 |
- speaker-segmentation
|
7 |
- generated_from_trainer
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
8 |
datasets:
|
9 |
- diarizers-community/callhome
|
|
|
10 |
model-index:
|
11 |
- name: speaker-segmentation-fine-tuned-callhome-jpn
|
12 |
results: []
|
13 |
---
|
14 |
-
|
15 |
-
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
16 |
-
should probably proofread and complete it, then remove this comment. -->
|
17 |
-
|
18 |
-
# speaker-segmentation-fine-tuned-callhome-jpn
|
19 |
-
|
20 |
-
This model is a fine-tuned version of [pyannote/segmentation-3.0](https://huggingface.co/pyannote/segmentation-3.0) on the diarizers-community/callhome jpn dataset.
|
21 |
-
It achieves the following results on the evaluation set:
|
22 |
-
- Loss: 0.7433
|
23 |
-
- Der: 0.2234
|
24 |
-
- False Alarm: 0.0478
|
25 |
-
- Missed Detection: 0.1328
|
26 |
-
- Confusion: 0.0428
|
27 |
-
|
28 |
-
## Model description
|
29 |
-
|
30 |
-
More information needed
|
31 |
-
|
32 |
-
## Intended uses & limitations
|
33 |
-
|
34 |
-
More information needed
|
35 |
-
|
36 |
-
## Training and evaluation data
|
37 |
-
|
38 |
-
More information needed
|
39 |
-
|
40 |
-
## Training procedure
|
41 |
-
|
42 |
-
### Training hyperparameters
|
43 |
-
|
44 |
-
The following hyperparameters were used during training:
|
45 |
-
- learning_rate: 0.001
|
46 |
-
- train_batch_size: 32
|
47 |
-
- eval_batch_size: 32
|
48 |
-
- seed: 42
|
49 |
-
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
50 |
-
- lr_scheduler_type: cosine
|
51 |
-
- num_epochs: 5.0
|
52 |
-
|
53 |
-
### Training results
|
54 |
-
|
55 |
-
| Training Loss | Epoch | Step | Validation Loss | Der | False Alarm | Missed Detection | Confusion |
|
56 |
-
|:-------------:|:-----:|:----:|:---------------:|:------:|:-----------:|:----------------:|:---------:|
|
57 |
-
| 0.5771 | 1.0 | 328 | 0.7534 | 0.2321 | 0.0564 | 0.1261 | 0.0496 |
|
58 |
-
| 0.5388 | 2.0 | 656 | 0.7503 | 0.2261 | 0.0485 | 0.1347 | 0.0429 |
|
59 |
-
| 0.5061 | 3.0 | 984 | 0.7486 | 0.2248 | 0.0475 | 0.1350 | 0.0423 |
|
60 |
-
| 0.4883 | 4.0 | 1312 | 0.7374 | 0.2227 | 0.0492 | 0.1315 | 0.0421 |
|
61 |
-
| 0.493 | 5.0 | 1640 | 0.7433 | 0.2234 | 0.0478 | 0.1328 | 0.0428 |
|
62 |
-
|
63 |
-
|
64 |
-
### Framework versions
|
65 |
-
|
66 |
-
- Transformers 4.40.0
|
67 |
-
- Pytorch 2.2.2+cu121
|
68 |
-
- Datasets 2.18.0
|
69 |
-
- Tokenizers 0.19.1
|
|
|
1 |
---
|
2 |
license: mit
|
|
|
3 |
tags:
|
4 |
- speaker-diarization
|
5 |
- speaker-segmentation
|
6 |
- generated_from_trainer
|
7 |
+
- pyannote
|
8 |
+
- pyannote.audio
|
9 |
+
- pyannote-audio-model
|
10 |
+
- audio
|
11 |
+
- voice
|
12 |
+
- speech
|
13 |
+
- speaker
|
14 |
+
- speaker-change-detection
|
15 |
+
- voice-activity-detection
|
16 |
+
- overlapped-speech-detection
|
17 |
+
- resegmentation
|
18 |
+
base_model: pyannote/segmentation-3.0
|
19 |
datasets:
|
20 |
- diarizers-community/callhome
|
21 |
+
licence: mit
|
22 |
model-index:
|
23 |
- name: speaker-segmentation-fine-tuned-callhome-jpn
|
24 |
results: []
|
25 |
---
|
26 |
+
This is the model card of a pyannote model that has been pushed on the Hub. This model card has been automatically generated.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
config.yaml
ADDED
@@ -0,0 +1,21 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
model:
|
2 |
+
_target_: pyannote.audio.models.segmentation.PyanNet.PyanNet
|
3 |
+
linear:
|
4 |
+
hidden_size: 128
|
5 |
+
num_layers: 2
|
6 |
+
lstm:
|
7 |
+
batch_first: true
|
8 |
+
bidirectional: true
|
9 |
+
dropout: 0.0
|
10 |
+
hidden_size: 128
|
11 |
+
monolithic: true
|
12 |
+
num_layers: 4
|
13 |
+
num_channels: 1
|
14 |
+
sample_rate: 16000
|
15 |
+
sincnet:
|
16 |
+
sample_rate: 16000
|
17 |
+
stride: 10
|
18 |
+
task:
|
19 |
+
duration: 10.0
|
20 |
+
max_speakers_per_chunk: 3
|
21 |
+
max_speakers_per_frame: 2
|
pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cf667e302cb3ad72316803868e2cf007d35d506e4ac6daafdd527dfd69f3fa72
|
3 |
+
size 5912144
|