Skip to content

Commit 3b869e8

Browse files
committed
fix(infra): switch nvidia-modelopt from [all] to [hf] extra
The [all] extra includes [onnx] which pulls in cppimport — a package with broken build metadata (reports version 0.0.0 instead of 26.4.17). Since AReaL only needs HuggingFace integration (accelerate, peft, transformers, deepspeed) and not ONNX export tooling, switch to [hf]. This also removes ~400MB of unnecessary deps (onnxruntime-gpu, cupy-cuda12x, onnx, etc.) and the now-unneeded cppimport dependency-metadata override.
1 parent d770654 commit 3b869e8

File tree

4 files changed

+12
-400
lines changed

4 files changed

+12
-400
lines changed

pyproject.toml

Lines changed: 1 addition & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -169,7 +169,7 @@ cuda-train = [
169169
"areal[tms]",
170170
"areal[megatron]",
171171
"areal[kernels]",
172-
"nvidia-modelopt[all]; sys_platform == 'linux' and platform_machine == 'x86_64'",
172+
"nvidia-modelopt[hf]; sys_platform == 'linux' and platform_machine == 'x86_64'",
173173
]
174174
# Full CUDA setup: training packages + SGLang inference + flash-attn
175175
cuda = [
@@ -251,11 +251,6 @@ override-dependencies = [
251251

252252
# Static metadata so uv lock resolves flash-attn without downloading or building.
253253
# The actual pre-built wheel (variant-specific) is installed in the Dockerfile.
254-
[[tool.uv.dependency-metadata]]
255-
name = "cppimport"
256-
version = "26.4.17"
257-
requires-dist = []
258-
259254
[[tool.uv.dependency-metadata]]
260255
name = "flash-attn"
261256
version = "2.8.3"

pyproject.vllm.toml

Lines changed: 1 addition & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -180,7 +180,7 @@ cuda-train = [
180180
"areal[tms]",
181181
"areal[megatron]",
182182
"areal[kernels]",
183-
"nvidia-modelopt[all]; sys_platform == 'linux' and platform_machine == 'x86_64'",
183+
"nvidia-modelopt[hf]; sys_platform == 'linux' and platform_machine == 'x86_64'",
184184
]
185185
# Full CUDA setup: training packages + vLLM inference + flash-attn
186186
cuda = [
@@ -249,11 +249,6 @@ override-dependencies = [
249249
]
250250

251251
# flash-attn is a compiled CUDA extension — provide static metadata to avoid building.
252-
[[tool.uv.dependency-metadata]]
253-
name = "cppimport"
254-
version = "26.4.17"
255-
requires-dist = []
256-
257252
[[tool.uv.dependency-metadata]]
258253
name = "flash-attn"
259254
version = "2.8.3"

0 commit comments

Comments
 (0)