Skip to content

pass llama_cpp.llama.LLama() constructor arg overrides from cli#14

Open
kroll-j wants to merge 4 commits intoRafaelGodoyEbert:mainfrom
kroll-j:feature_llama_args
Open

pass llama_cpp.llama.LLama() constructor arg overrides from cli#14
kroll-j wants to merge 4 commits intoRafaelGodoyEbert:mainfrom
kroll-j:feature_llama_args

Conversation

@kroll-j
Copy link
Copy Markdown

@kroll-j kroll-j commented Mar 23, 2026

Hey, awesome project! I use it to create clips with local hardware. I need to pass additional arguments to the Llama() constructor to make the model fit and run on my GPU with decent speed. This PR adds a parameter to the CLI script which takes a JSON object. This overrides any arguments to Llama(). For example:

python main_improved.py [...] --ai-backend local --ai-model-name google_gemma-3-12b-it-Q6_K_L.gguf --llama-args '{ "n_gpu_layers":35, "n_ctx":8192, "kv_overrides":{"kv_unified":true}, "flash_attn": true, "seed": 1 }' --chunk-size 6000

This PR changes only the CLI, the GUI is untouched.

@kroll-j kroll-j marked this pull request as draft March 23, 2026 12:41
@kroll-j kroll-j marked this pull request as ready for review March 23, 2026 12:56
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant