AustinXiao commited on
Commit
71bae53
·
verified ·
1 Parent(s): 408e6e3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +70 -2
README.md CHANGED
@@ -2,8 +2,76 @@
2
  tags:
3
  - model_hub_mixin
4
  - pytorch_model_hub_mixin
 
 
 
 
 
 
5
  ---
6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin) integration:
8
- - Library: [More Information Needed]
9
- - Docs: [More Information Needed]
 
2
  tags:
3
  - model_hub_mixin
4
  - pytorch_model_hub_mixin
5
+ license: mit
6
+ datasets:
7
+ - Bisher/ASVspoof_2019_LA
8
+ - Bisher/ASVspoof_2021_DF
9
+ language:
10
+ - en
11
  ---
12
 
13
+ ### 1. Introduction
14
+ Transformers and their variants have achieved great
15
+ success in speech processing. However, their multi-head selfattention mechanism is computationally expensive. Therefore, one
16
+ novel selective state space model, Mamba, has been proposed
17
+ as an alternative. Building on its success in automatic speech
18
+ recognition, we apply Mamba for spoofing attack detection.
19
+ Mamba is well-suited for this task as it can capture the artifacts
20
+ in spoofed speech signals by handling long-length sequences.
21
+ However, Mamba’s performance may suffer when it is trained
22
+ with limited labeled data. To mitigate this, we propose combining
23
+ a new structure of Mamba based on a dual-column architecture
24
+ with self-supervised learning, using the pre-trained wav2vec
25
+ 2.0 model. The experiments show that our proposed approach
26
+ achieves competitive results and faster inference on the ASVspoof
27
+ 2021 LA and DF datasets, and on the more challenging In-theWild dataset, it emerges as the strongest candidate for spoofing
28
+ attack detection.
29
+
30
+ ### 2. Setup Environment
31
+ You need to create the running environment by [Anaconda](https://www.anaconda.com/).
32
+ First, create and activate the environment:
33
+
34
+ ```bash
35
+ conda create -n XLSR_Mamba python=3.10
36
+ conda activate XLSR_Mamba
37
+ ```
38
+
39
+ Then install the requirements:
40
+
41
+ ```bash
42
+ pip install -r requirements.txt
43
+ ```
44
+
45
+ Install fairseq:
46
+
47
+ ```bash
48
+ git clone https://github.com/facebookresearch/fairseq.git fairseq_dir
49
+ cd fairseq_dir
50
+ git checkout a54021305d6b3c
51
+ pip install --editable ./
52
+ ```
53
+ ### 3. Pretrained Model
54
+ The pretrained model XLSR can be found at this [link](https://dl.fbaipublicfiles.com/fairseq/wav2vec/xlsr2_300m.pt).
55
+
56
+ ### 4. Results
57
+
58
+
59
+ Dataset | **EER (%)** | **min t-DCF** |
60
+ --------|--------------------|--------------------|
61
+ ASVspoof2021 LA | **0.93** | **0.208** |
62
+ ASVspoof2021 DF | **1.88** | **-** |
63
+ In-The-Wild | **6.71** | **-** |
64
+
65
+ ### 5. Citation
66
+ If you find our repository valuable for your work, please consider giving a star to this repo and citing our paper:
67
+ ```
68
+ @article{xiao2024xlsr,
69
+ title={{XLSR-Mamba}: A Dual-Column Bidirectional State Space Model for Spoofing Attack Detection},
70
+ author={Xiao, Yang and Das, Rohan Kumar},
71
+ journal={arXiv preprint arXiv:2411.10027},
72
+ year={2024}
73
+ }
74
+ ```
75
  This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin) integration:
76
+ - Library: [https://github.com/swagshaw/XLSR-Mamba]
77
+ - Paper: [https://arxiv.org/pdf/2411.10027]