Skip to content

Commit 50f5f7a

Browse files
authored
Merge pull request #205 from codelion/fix-mlx-model-id
Update MLX model patterns and reduce max_tokens in eval script
2 parents e004b2e + 0a6bc20 commit 50f5f7a

File tree

4 files changed

+5
-4
lines changed

4 files changed

+5
-4
lines changed

optillm/__init__.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
import os
33

44
# Version information
5-
__version__ = "0.1.17"
5+
__version__ = "0.1.18"
66

77
# Get the path to the root optillm.py
88
spec = util.spec_from_file_location(

optillm/inference.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -189,7 +189,8 @@ def should_use_mlx(model_id: str) -> bool:
189189
# Models that should use MLX
190190
mlx_patterns = [
191191
"mlx-community/",
192-
"mlx-"
192+
"mlx-",
193+
"-mlx-"
193194
]
194195

195196
# Known problematic models that should prefer MLX on Apple Silicon

scripts/eval_math500_benchmark.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -692,7 +692,7 @@ def get_llm_response(problem: str, model: str) -> str:
692692
messages=[
693693
{"role": "user", "content": SYSTEM_PROMPT + "\n" + problem}
694694
],
695-
max_tokens=32768, # for thinking models, we need to use a lot more tokens
695+
max_tokens=8192, # for thinking models, we need to use a lot more tokens
696696
# extra_body = {
697697
# "decoding" : "thinkdeeper",
698698
# }

setup.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33

44
setup(
55
name="optillm",
6-
version="0.1.17",
6+
version="0.1.18",
77
packages=find_packages(include=['optillm', 'optillm.*']), # This ensures all subpackages are included
88
py_modules=['optillm'],
99
package_data={

0 commit comments

Comments
 (0)