konieshadow commited on
Commit
5d8d9c8
·
1 Parent(s): b00f314

移除不必要的numpy导入,并在GemmaMLXChatCompletion类中重新组织mlx_lm模块的导入。调整了模型和分词器加载逻辑以优化代码结构。

Browse files
src/podcast_transcribe/asr/asr_distil_whisper_mlx.py CHANGED
@@ -6,7 +6,6 @@ import os
6
  from pydub import AudioSegment
7
  from typing import Dict, List, Union
8
  import logging
9
- import numpy as np
10
  import mlx_whisper
11
 
12
  # 导入基类
 
6
  from pydub import AudioSegment
7
  from typing import Dict, List, Union
8
  import logging
 
9
  import mlx_whisper
10
 
11
  # 导入基类
src/podcast_transcribe/llm/llm_gemma_mlx.py CHANGED
@@ -1,5 +1,3 @@
1
- from mlx_lm import load, generate
2
- from mlx_lm.sample_utils import make_sampler
3
  from typing import Dict, Union
4
  from .llm_base import BaseChatCompletion
5
 
@@ -14,6 +12,8 @@ class GemmaMLXChatCompletion(BaseChatCompletion):
14
  def _load_model_and_tokenizer(self):
15
  """加载 MLX 模型和分词器"""
16
  try:
 
 
17
  print(f"正在加载 MLX 模型: {self.model_name}")
18
  self.model, self.tokenizer = load(self.model_name)
19
  print(f"MLX 模型 {self.model_name} 加载成功")
@@ -32,6 +32,8 @@ class GemmaMLXChatCompletion(BaseChatCompletion):
32
  **kwargs
33
  ) -> str:
34
  """使用 MLX 生成响应"""
 
 
35
 
36
  # 为temperature和top_p创建一个采样器
37
  sampler = make_sampler(temp=temperature, top_p=top_p)
 
 
 
1
  from typing import Dict, Union
2
  from .llm_base import BaseChatCompletion
3
 
 
12
  def _load_model_and_tokenizer(self):
13
  """加载 MLX 模型和分词器"""
14
  try:
15
+ from mlx_lm import load
16
+
17
  print(f"正在加载 MLX 模型: {self.model_name}")
18
  self.model, self.tokenizer = load(self.model_name)
19
  print(f"MLX 模型 {self.model_name} 加载成功")
 
32
  **kwargs
33
  ) -> str:
34
  """使用 MLX 生成响应"""
35
+ from mlx_lm import load, generate
36
+ from mlx_lm.sample_utils import make_sampler
37
 
38
  # 为temperature和top_p创建一个采样器
39
  sampler = make_sampler(temp=temperature, top_p=top_p)