salmanshahid
commited on
Upload folder using huggingface_hub
Browse files- .gitattributes +3 -11
- README.md +70 -3
- config.yaml +10 -0
.gitattributes
CHANGED
@@ -1,35 +1,27 @@
|
|
1 |
*.7z filter=lfs diff=lfs merge=lfs -text
|
2 |
*.arrow filter=lfs diff=lfs merge=lfs -text
|
3 |
*.bin filter=lfs diff=lfs merge=lfs -text
|
|
|
4 |
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
5 |
-
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
6 |
*.ftz filter=lfs diff=lfs merge=lfs -text
|
7 |
*.gz filter=lfs diff=lfs merge=lfs -text
|
8 |
*.h5 filter=lfs diff=lfs merge=lfs -text
|
9 |
*.joblib filter=lfs diff=lfs merge=lfs -text
|
10 |
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
11 |
-
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
12 |
*.model filter=lfs diff=lfs merge=lfs -text
|
13 |
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
14 |
-
*.npy filter=lfs diff=lfs merge=lfs -text
|
15 |
-
*.npz filter=lfs diff=lfs merge=lfs -text
|
16 |
*.onnx filter=lfs diff=lfs merge=lfs -text
|
17 |
*.ot filter=lfs diff=lfs merge=lfs -text
|
18 |
*.parquet filter=lfs diff=lfs merge=lfs -text
|
19 |
*.pb filter=lfs diff=lfs merge=lfs -text
|
20 |
-
*.pickle filter=lfs diff=lfs merge=lfs -text
|
21 |
-
*.pkl filter=lfs diff=lfs merge=lfs -text
|
22 |
*.pt filter=lfs diff=lfs merge=lfs -text
|
23 |
*.pth filter=lfs diff=lfs merge=lfs -text
|
24 |
*.rar filter=lfs diff=lfs merge=lfs -text
|
25 |
-
|
26 |
-
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
27 |
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
28 |
-
*.tar filter=lfs diff=lfs merge=lfs -text
|
29 |
*.tflite filter=lfs diff=lfs merge=lfs -text
|
30 |
*.tgz filter=lfs diff=lfs merge=lfs -text
|
31 |
-
*.wasm filter=lfs diff=lfs merge=lfs -text
|
32 |
*.xz filter=lfs diff=lfs merge=lfs -text
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
-
*.
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
1 |
*.7z filter=lfs diff=lfs merge=lfs -text
|
2 |
*.arrow filter=lfs diff=lfs merge=lfs -text
|
3 |
*.bin filter=lfs diff=lfs merge=lfs -text
|
4 |
+
*.bin.* filter=lfs diff=lfs merge=lfs -text
|
5 |
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
|
|
6 |
*.ftz filter=lfs diff=lfs merge=lfs -text
|
7 |
*.gz filter=lfs diff=lfs merge=lfs -text
|
8 |
*.h5 filter=lfs diff=lfs merge=lfs -text
|
9 |
*.joblib filter=lfs diff=lfs merge=lfs -text
|
10 |
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
|
|
11 |
*.model filter=lfs diff=lfs merge=lfs -text
|
12 |
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
13 |
*.onnx filter=lfs diff=lfs merge=lfs -text
|
14 |
*.ot filter=lfs diff=lfs merge=lfs -text
|
15 |
*.parquet filter=lfs diff=lfs merge=lfs -text
|
16 |
*.pb filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
17 |
*.pt filter=lfs diff=lfs merge=lfs -text
|
18 |
*.pth filter=lfs diff=lfs merge=lfs -text
|
19 |
*.rar filter=lfs diff=lfs merge=lfs -text
|
20 |
+
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
|
21 |
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
|
|
22 |
*.tflite filter=lfs diff=lfs merge=lfs -text
|
23 |
*.tgz filter=lfs diff=lfs merge=lfs -text
|
|
|
24 |
*.xz filter=lfs diff=lfs merge=lfs -text
|
25 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
26 |
+
*.zstandard filter=lfs diff=lfs merge=lfs -text
|
27 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
README.md
CHANGED
@@ -1,3 +1,70 @@
|
|
1 |
-
---
|
2 |
-
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
tags:
|
3 |
+
- pyannote
|
4 |
+
- pyannote-audio
|
5 |
+
- pyannote-audio-pipeline
|
6 |
+
- audio
|
7 |
+
- voice
|
8 |
+
- speech
|
9 |
+
- speaker
|
10 |
+
- voice-activity-detection
|
11 |
+
- automatic-speech-recognition
|
12 |
+
datasets:
|
13 |
+
- ami
|
14 |
+
- dihard
|
15 |
+
- voxconverse
|
16 |
+
license: mit
|
17 |
+
extra_gated_prompt: "The collected information will help acquire a better knowledge of pyannote.audio userbase and help its maintainers apply for grants to improve it further. If you are an academic researcher, please cite the relevant papers in your own publications using the model. If you work for a company, please consider contributing back to pyannote.audio development (e.g. through unrestricted gifts). We also provide scientific consulting services around speaker diarization and machine listening."
|
18 |
+
extra_gated_fields:
|
19 |
+
Company/university: text
|
20 |
+
Website: text
|
21 |
+
I plan to use this model for (task, type of audio data, etc): text
|
22 |
+
---
|
23 |
+
|
24 |
+
Using this open-source model in production?
|
25 |
+
Consider switching to [pyannoteAI](https://www.pyannote.ai) for better and faster options.
|
26 |
+
|
27 |
+
# 🎹 Voice activity detection
|
28 |
+
|
29 |
+
Relies on pyannote.audio 2.1: see [installation instructions](https://github.com/pyannote/pyannote-audio#installation).
|
30 |
+
|
31 |
+
|
32 |
+
```python
|
33 |
+
# 1. visit hf.co/pyannote/segmentation and accept user conditions
|
34 |
+
# 2. visit hf.co/settings/tokens to create an access token
|
35 |
+
# 3. instantiate pretrained voice activity detection pipeline
|
36 |
+
|
37 |
+
from pyannote.audio import Pipeline
|
38 |
+
pipeline = Pipeline.from_pretrained("pyannote/voice-activity-detection",
|
39 |
+
use_auth_token="ACCESS_TOKEN_GOES_HERE")
|
40 |
+
output = pipeline("audio.wav")
|
41 |
+
|
42 |
+
for speech in output.get_timeline().support():
|
43 |
+
# active speech between speech.start and speech.end
|
44 |
+
...
|
45 |
+
```
|
46 |
+
|
47 |
+
|
48 |
+
## Citation
|
49 |
+
|
50 |
+
```bibtex
|
51 |
+
@inproceedings{Bredin2021,
|
52 |
+
Title = {{End-to-end speaker segmentation for overlap-aware resegmentation}},
|
53 |
+
Author = {{Bredin}, Herv{\'e} and {Laurent}, Antoine},
|
54 |
+
Booktitle = {Proc. Interspeech 2021},
|
55 |
+
Address = {Brno, Czech Republic},
|
56 |
+
Month = {August},
|
57 |
+
Year = {2021},
|
58 |
+
}
|
59 |
+
```
|
60 |
+
|
61 |
+
```bibtex
|
62 |
+
@inproceedings{Bredin2020,
|
63 |
+
Title = {{pyannote.audio: neural building blocks for speaker diarization}},
|
64 |
+
Author = {{Bredin}, Herv{\'e} and {Yin}, Ruiqing and {Coria}, Juan Manuel and {Gelly}, Gregory and {Korshunov}, Pavel and {Lavechin}, Marvin and {Fustes}, Diego and {Titeux}, Hadrien and {Bouaziz}, Wassim and {Gill}, Marie-Philippe},
|
65 |
+
Booktitle = {ICASSP 2020, IEEE International Conference on Acoustics, Speech, and Signal Processing},
|
66 |
+
Address = {Barcelona, Spain},
|
67 |
+
Month = {May},
|
68 |
+
Year = {2020},
|
69 |
+
}
|
70 |
+
```
|
config.yaml
ADDED
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
pipeline:
|
2 |
+
name: pyannote.audio.pipelines.VoiceActivityDetection
|
3 |
+
params:
|
4 |
+
segmentation: pyannote/segmentation@Interspeech2021
|
5 |
+
|
6 |
+
params:
|
7 |
+
min_duration_off: 0.09791355693027545
|
8 |
+
min_duration_on: 0.05537587440407595
|
9 |
+
offset: 0.4806866463041527
|
10 |
+
onset: 0.8104268538848918
|