Spaces:
Running
Running
Commit
·
5d8d9c8
1
Parent(s):
b00f314
移除不必要的numpy导入,并在GemmaMLXChatCompletion类中重新组织mlx_lm模块的导入。调整了模型和分词器加载逻辑以优化代码结构。
Browse files
src/podcast_transcribe/asr/asr_distil_whisper_mlx.py
CHANGED
@@ -6,7 +6,6 @@ import os
|
|
6 |
from pydub import AudioSegment
|
7 |
from typing import Dict, List, Union
|
8 |
import logging
|
9 |
-
import numpy as np
|
10 |
import mlx_whisper
|
11 |
|
12 |
# 导入基类
|
|
|
6 |
from pydub import AudioSegment
|
7 |
from typing import Dict, List, Union
|
8 |
import logging
|
|
|
9 |
import mlx_whisper
|
10 |
|
11 |
# 导入基类
|
src/podcast_transcribe/llm/llm_gemma_mlx.py
CHANGED
@@ -1,5 +1,3 @@
|
|
1 |
-
from mlx_lm import load, generate
|
2 |
-
from mlx_lm.sample_utils import make_sampler
|
3 |
from typing import Dict, Union
|
4 |
from .llm_base import BaseChatCompletion
|
5 |
|
@@ -14,6 +12,8 @@ class GemmaMLXChatCompletion(BaseChatCompletion):
|
|
14 |
def _load_model_and_tokenizer(self):
|
15 |
"""加载 MLX 模型和分词器"""
|
16 |
try:
|
|
|
|
|
17 |
print(f"正在加载 MLX 模型: {self.model_name}")
|
18 |
self.model, self.tokenizer = load(self.model_name)
|
19 |
print(f"MLX 模型 {self.model_name} 加载成功")
|
@@ -32,6 +32,8 @@ class GemmaMLXChatCompletion(BaseChatCompletion):
|
|
32 |
**kwargs
|
33 |
) -> str:
|
34 |
"""使用 MLX 生成响应"""
|
|
|
|
|
35 |
|
36 |
# 为temperature和top_p创建一个采样器
|
37 |
sampler = make_sampler(temp=temperature, top_p=top_p)
|
|
|
|
|
|
|
1 |
from typing import Dict, Union
|
2 |
from .llm_base import BaseChatCompletion
|
3 |
|
|
|
12 |
def _load_model_and_tokenizer(self):
|
13 |
"""加载 MLX 模型和分词器"""
|
14 |
try:
|
15 |
+
from mlx_lm import load
|
16 |
+
|
17 |
print(f"正在加载 MLX 模型: {self.model_name}")
|
18 |
self.model, self.tokenizer = load(self.model_name)
|
19 |
print(f"MLX 模型 {self.model_name} 加载成功")
|
|
|
32 |
**kwargs
|
33 |
) -> str:
|
34 |
"""使用 MLX 生成响应"""
|
35 |
+
from mlx_lm import load, generate
|
36 |
+
from mlx_lm.sample_utils import make_sampler
|
37 |
|
38 |
# 为temperature和top_p创建一个采样器
|
39 |
sampler = make_sampler(temp=temperature, top_p=top_p)
|