Fedir Zadniprovskyi commited on
Commit
1de8d65
·
1 Parent(s): 974e6f9

misc: add `vad_filter` form field

Browse files
src/faster_whisper_server/routers/stt.py CHANGED
@@ -136,6 +136,7 @@ def translate_file(
136
  response_format: Annotated[ResponseFormat | None, Form()] = None,
137
  temperature: Annotated[float, Form()] = 0.0,
138
  stream: Annotated[bool, Form()] = False,
 
139
  ) -> Response | StreamingResponse:
140
  if model is None:
141
  model = config.whisper.model
@@ -147,7 +148,7 @@ def translate_file(
147
  task=Task.TRANSLATE,
148
  initial_prompt=prompt,
149
  temperature=temperature,
150
- vad_filter=True,
151
  )
152
  segments = TranscriptionSegment.from_faster_whisper_segments(segments)
153
 
@@ -192,6 +193,7 @@ def transcribe_file(
192
  ] = ["segment"],
193
  stream: Annotated[bool, Form()] = False,
194
  hotwords: Annotated[str | None, Form()] = None,
 
195
  ) -> Response | StreamingResponse:
196
  if model is None:
197
  model = config.whisper.model
@@ -212,7 +214,7 @@ def transcribe_file(
212
  initial_prompt=prompt,
213
  word_timestamps="word" in timestamp_granularities,
214
  temperature=temperature,
215
- vad_filter=True,
216
  hotwords=hotwords,
217
  )
218
  segments = TranscriptionSegment.from_faster_whisper_segments(segments)
@@ -263,6 +265,7 @@ async def transcribe_stream(
263
  language: Annotated[Language | None, Query()] = None,
264
  response_format: Annotated[ResponseFormat | None, Query()] = None,
265
  temperature: Annotated[float, Query()] = 0.0,
 
266
  ) -> None:
267
  if model is None:
268
  model = config.whisper.model
@@ -274,7 +277,7 @@ async def transcribe_stream(
274
  transcribe_opts = {
275
  "language": language,
276
  "temperature": temperature,
277
- "vad_filter": True,
278
  "condition_on_previous_text": False,
279
  }
280
  whisper = model_manager.load_model(model)
 
136
  response_format: Annotated[ResponseFormat | None, Form()] = None,
137
  temperature: Annotated[float, Form()] = 0.0,
138
  stream: Annotated[bool, Form()] = False,
139
+ vad_filter: Annotated[bool, Form()] = False,
140
  ) -> Response | StreamingResponse:
141
  if model is None:
142
  model = config.whisper.model
 
148
  task=Task.TRANSLATE,
149
  initial_prompt=prompt,
150
  temperature=temperature,
151
+ vad_filter=vad_filter,
152
  )
153
  segments = TranscriptionSegment.from_faster_whisper_segments(segments)
154
 
 
193
  ] = ["segment"],
194
  stream: Annotated[bool, Form()] = False,
195
  hotwords: Annotated[str | None, Form()] = None,
196
+ vad_filter: Annotated[bool, Form()] = False,
197
  ) -> Response | StreamingResponse:
198
  if model is None:
199
  model = config.whisper.model
 
214
  initial_prompt=prompt,
215
  word_timestamps="word" in timestamp_granularities,
216
  temperature=temperature,
217
+ vad_filter=vad_filter,
218
  hotwords=hotwords,
219
  )
220
  segments = TranscriptionSegment.from_faster_whisper_segments(segments)
 
265
  language: Annotated[Language | None, Query()] = None,
266
  response_format: Annotated[ResponseFormat | None, Query()] = None,
267
  temperature: Annotated[float, Query()] = 0.0,
268
+ vad_filter: Annotated[bool, Query()] = False,
269
  ) -> None:
270
  if model is None:
271
  model = config.whisper.model
 
277
  transcribe_opts = {
278
  "language": language,
279
  "temperature": temperature,
280
+ "vad_filter": vad_filter,
281
  "condition_on_previous_text": False,
282
  }
283
  whisper = model_manager.load_model(model)