Update custom_interface_app.py
Browse files- custom_interface_app.py +5 -5
custom_interface_app.py
CHANGED
@@ -172,17 +172,17 @@ class ASR(Pretrained):
|
|
172 |
|
173 |
|
174 |
|
175 |
-
def classify_file_whisper_mkd(self,
|
176 |
# Load the audio file
|
177 |
-
waveform, sr = librosa.load(path, sr=16000)
|
178 |
|
179 |
# Get audio length in seconds
|
180 |
-
audio_length = len(waveform) /
|
181 |
|
182 |
if audio_length >= 20:
|
183 |
# split audio every 20 seconds
|
184 |
segments = []
|
185 |
-
max_duration = 20 *
|
186 |
num_segments = int(np.ceil(len(waveform) / max_duration))
|
187 |
start = 0
|
188 |
for i in range(num_segments):
|
@@ -190,7 +190,7 @@ class ASR(Pretrained):
|
|
190 |
if end > len(waveform):
|
191 |
end = len(waveform)
|
192 |
segment_part = waveform[start:end]
|
193 |
-
segment_len = len(segment_part) /
|
194 |
if segment_len < 1:
|
195 |
continue
|
196 |
segments.append(segment_part)
|
|
|
172 |
|
173 |
|
174 |
|
175 |
+
def classify_file_whisper_mkd(self, waveform, device):
|
176 |
# Load the audio file
|
177 |
+
# waveform, sr = librosa.load(path, sr=16000)
|
178 |
|
179 |
# Get audio length in seconds
|
180 |
+
audio_length = len(waveform) / 16000
|
181 |
|
182 |
if audio_length >= 20:
|
183 |
# split audio every 20 seconds
|
184 |
segments = []
|
185 |
+
max_duration = 20 * 16000 # Maximum segment duration in samples (20 seconds)
|
186 |
num_segments = int(np.ceil(len(waveform) / max_duration))
|
187 |
start = 0
|
188 |
for i in range(num_segments):
|
|
|
190 |
if end > len(waveform):
|
191 |
end = len(waveform)
|
192 |
segment_part = waveform[start:end]
|
193 |
+
segment_len = len(segment_part) / 16000
|
194 |
if segment_len < 1:
|
195 |
continue
|
196 |
segments.append(segment_part)
|