daihui.zhang
commited on
Commit
·
81e4670
1
Parent(s):
2c60bfa
remove vad model
Browse files
transcribe/transcription.py
CHANGED
|
@@ -182,7 +182,6 @@ class TranscriptionServer:
|
|
| 182 |
query_parameters_dict = dict(parse_qsl(urlparse(websocket.request.path).query))
|
| 183 |
from_lang, to_lang = query_parameters_dict.get('from'), query_parameters_dict.get('to')
|
| 184 |
|
| 185 |
-
|
| 186 |
try:
|
| 187 |
logging.info("New client connected")
|
| 188 |
options = websocket.recv()
|
|
|
|
| 182 |
query_parameters_dict = dict(parse_qsl(urlparse(websocket.request.path).query))
|
| 183 |
from_lang, to_lang = query_parameters_dict.get('from'), query_parameters_dict.get('to')
|
| 184 |
|
|
|
|
| 185 |
try:
|
| 186 |
logging.info("New client connected")
|
| 187 |
options = websocket.recv()
|
transcribe/whisper_llm_serve.py
CHANGED
|
@@ -115,7 +115,7 @@ class SegmentManager:
|
|
| 115 |
class PywhisperInference:
|
| 116 |
whisper_model = None
|
| 117 |
llm_model = None
|
| 118 |
-
vad_model = None
|
| 119 |
|
| 120 |
@classmethod
|
| 121 |
def initializer(cls, event:mp.Event, warmup=True):
|
|
@@ -132,7 +132,7 @@ class PywhisperInference:
|
|
| 132 |
|
| 133 |
# init llamacpp
|
| 134 |
cls.llm_model = QwenTranslator(config.LLM_MODEL_PATH, config.LLM_SYS_PROMPT)
|
| 135 |
-
cls.vad_model = VoiceActivityDetector()
|
| 136 |
|
| 137 |
event.set()
|
| 138 |
|
|
@@ -172,10 +172,10 @@ class PywhisperInference:
|
|
| 172 |
def translate(cls, context: str, src_lang, dst_lang):
|
| 173 |
return cls.llm_model.translate(context, src_lang, dst_lang)
|
| 174 |
|
| 175 |
-
@classmethod
|
| 176 |
-
def voice_detect(cls, audio_buffer):
|
| 177 |
-
|
| 178 |
-
|
| 179 |
|
| 180 |
|
| 181 |
|
|
|
|
| 115 |
class PywhisperInference:
|
| 116 |
whisper_model = None
|
| 117 |
llm_model = None
|
| 118 |
+
# vad_model = None
|
| 119 |
|
| 120 |
@classmethod
|
| 121 |
def initializer(cls, event:mp.Event, warmup=True):
|
|
|
|
| 132 |
|
| 133 |
# init llamacpp
|
| 134 |
cls.llm_model = QwenTranslator(config.LLM_MODEL_PATH, config.LLM_SYS_PROMPT)
|
| 135 |
+
# cls.vad_model = VoiceActivityDetector()
|
| 136 |
|
| 137 |
event.set()
|
| 138 |
|
|
|
|
| 172 |
def translate(cls, context: str, src_lang, dst_lang):
|
| 173 |
return cls.llm_model.translate(context, src_lang, dst_lang)
|
| 174 |
|
| 175 |
+
# @classmethod
|
| 176 |
+
# def voice_detect(cls, audio_buffer):
|
| 177 |
+
# audio_buffer = np.frombuffer(audio_buffer, dtype=np.float32)
|
| 178 |
+
# return cls.vad_model(audio_buffer)
|
| 179 |
|
| 180 |
|
| 181 |
|