Fix embedded provider initialization and logging

- Use global OnceLock for llama.cpp backend to prevent BackendAlreadyInitialized error
- Suppress verbose llama.cpp stderr logging during model loading
- Fix provider validation to accept "embedded.name" format (extract type before dot)
This commit is contained in:
Dhanji R. Prasanna
2026-01-28 10:33:10 +11:00
parent ba6e1f9896
commit e32c302023
2 changed files with 42 additions and 6 deletions

View File

@@ -139,9 +139,10 @@ pub fn load_config_with_cli_overrides(cli: &Cli) -> Result<Config> {
// Validate provider if specified
if let Some(ref provider) = cli.provider {
let valid_providers = ["anthropic", "databricks", "embedded", "openai"];
if !valid_providers.contains(&provider.as_str()) {
let provider_type = provider.split('.').next().unwrap_or(provider);
if !valid_providers.contains(&provider_type) {
return Err(anyhow::anyhow!(
"Invalid provider '{}'. Valid options: {:?}",
"Invalid provider '{}'. Provider type must be one of: {:?}",
provider,
valid_providers
));