From 74ea754d299cad71d6d597e319c718f883cc3fbc Mon Sep 17 00:00:00 2001 From: Yeachan-Heo Date: Tue, 28 Apr 2026 09:19:16 +0000 Subject: [PATCH] Restore Rust formatting compliance Run rustfmt from the Rust workspace so CI format checks pass without changing behavior. Constraint: Scope is formatting-only across tracked Rust files Confidence: high Scope-risk: narrow Tested: cd rust && cargo fmt --check Tested: git diff --check --- rust/crates/api/src/providers/mod.rs | 20 +- .../crates/api/src/providers/openai_compat.rs | 13 +- rust/crates/commands/src/lib.rs | 6 +- rust/crates/rusty-claude-cli/src/main.rs | 183 +++++++++++------- .../rusty-claude-cli/tests/compact_output.rs | 5 +- 5 files changed, 140 insertions(+), 87 deletions(-) diff --git a/rust/crates/api/src/providers/mod.rs b/rust/crates/api/src/providers/mod.rs index fb97900..86871a8 100644 --- a/rust/crates/api/src/providers/mod.rs +++ b/rust/crates/api/src/providers/mod.rs @@ -753,14 +753,14 @@ mod tests { #[test] fn returns_context_window_metadata_for_kimi_models() { // kimi-k2.5 - let k25_limit = model_token_limit("kimi-k2.5") - .expect("kimi-k2.5 should have token limit metadata"); + let k25_limit = + model_token_limit("kimi-k2.5").expect("kimi-k2.5 should have token limit metadata"); assert_eq!(k25_limit.max_output_tokens, 16_384); assert_eq!(k25_limit.context_window_tokens, 256_000); // kimi-k1.5 - let k15_limit = model_token_limit("kimi-k1.5") - .expect("kimi-k1.5 should have token limit metadata"); + let k15_limit = + model_token_limit("kimi-k1.5").expect("kimi-k1.5 should have token limit metadata"); assert_eq!(k15_limit.max_output_tokens, 16_384); assert_eq!(k15_limit.context_window_tokens, 256_000); } @@ -768,11 +768,13 @@ mod tests { #[test] fn kimi_alias_resolves_to_kimi_k25_token_limits() { // The "kimi" alias resolves to "kimi-k2.5" via resolve_model_alias() - let alias_limit = model_token_limit("kimi") - .expect("kimi alias should resolve to kimi-k2.5 limits"); - let direct_limit = model_token_limit("kimi-k2.5") - .expect("kimi-k2.5 should have limits"); - assert_eq!(alias_limit.max_output_tokens, direct_limit.max_output_tokens); + let alias_limit = + model_token_limit("kimi").expect("kimi alias should resolve to kimi-k2.5 limits"); + let direct_limit = model_token_limit("kimi-k2.5").expect("kimi-k2.5 should have limits"); + assert_eq!( + alias_limit.max_output_tokens, + direct_limit.max_output_tokens + ); assert_eq!( alias_limit.context_window_tokens, direct_limit.context_window_tokens diff --git a/rust/crates/api/src/providers/openai_compat.rs b/rust/crates/api/src/providers/openai_compat.rs index 4e4183b..a810502 100644 --- a/rust/crates/api/src/providers/openai_compat.rs +++ b/rust/crates/api/src/providers/openai_compat.rs @@ -2195,9 +2195,16 @@ mod tests { #[test] fn provider_specific_size_limits_are_correct() { - assert_eq!(OpenAiCompatConfig::dashscope().max_request_body_bytes, 6_291_456); // 6MB - assert_eq!(OpenAiCompatConfig::openai().max_request_body_bytes, 104_857_600); // 100MB - assert_eq!(OpenAiCompatConfig::xai().max_request_body_bytes, 52_428_800); // 50MB + assert_eq!( + OpenAiCompatConfig::dashscope().max_request_body_bytes, + 6_291_456 + ); // 6MB + assert_eq!( + OpenAiCompatConfig::openai().max_request_body_bytes, + 104_857_600 + ); // 100MB + assert_eq!(OpenAiCompatConfig::xai().max_request_body_bytes, 52_428_800); + // 50MB } #[test] diff --git a/rust/crates/commands/src/lib.rs b/rust/crates/commands/src/lib.rs index 9d69393..d4f1770 100644 --- a/rust/crates/commands/src/lib.rs +++ b/rust/crates/commands/src/lib.rs @@ -2623,10 +2623,8 @@ fn render_mcp_report_json_for( // runs, the existing serializer adds `status: "ok"` below. match loader.load() { Ok(runtime_config) => { - let mut value = render_mcp_summary_report_json( - cwd, - runtime_config.mcp().servers(), - ); + let mut value = + render_mcp_summary_report_json(cwd, runtime_config.mcp().servers()); if let Some(map) = value.as_object_mut() { map.insert("status".to_string(), Value::String("ok".to_string())); map.insert("config_load_error".to_string(), Value::Null); diff --git a/rust/crates/rusty-claude-cli/src/main.rs b/rust/crates/rusty-claude-cli/src/main.rs index 64ba249..9d9df4f 100644 --- a/rust/crates/rusty-claude-cli/src/main.rs +++ b/rust/crates/rusty-claude-cli/src/main.rs @@ -228,8 +228,10 @@ fn main() { // don't need to regex-scrape the prose. let kind = classify_error_kind(&message); if message.contains("`claw --help`") { - eprintln!("[error-kind: {kind}] -error: {message}"); + eprintln!( + "[error-kind: {kind}] +error: {message}" + ); } else { eprintln!( "[error-kind: {kind}] @@ -419,19 +421,17 @@ fn run() -> Result<(), Box> { CliAction::Config { section, output_format, - } => { - match output_format { - CliOutputFormat::Text => { - println!("{}", render_config_report(section.as_deref())?); - } - CliOutputFormat::Json => { - println!( - "{}", - serde_json::to_string_pretty(&render_config_json(section.as_deref())?)? - ); - } + } => match output_format { + CliOutputFormat::Text => { + println!("{}", render_config_report(section.as_deref())?); } - } + CliOutputFormat::Json => { + println!( + "{}", + serde_json::to_string_pretty(&render_config_json(section.as_deref())?)? + ); + } + }, CliAction::Diff { output_format } => match output_format { CliOutputFormat::Text => { println!("{}", render_diff_report()?); @@ -635,13 +635,7 @@ fn parse_args(args: &[String]) -> Result { } "--help" | "-h" if !rest.is_empty() - && matches!( - rest[0].as_str(), - "prompt" - | "commit" - | "pr" - | "issue" - ) => + && matches!(rest[0].as_str(), "prompt" | "commit" | "pr" | "issue") => { // `--help` following a subcommand that would otherwise forward // the arg to the API (e.g. `claw prompt --help`) should show @@ -1327,7 +1321,6 @@ fn suggest_closest_term<'a>(input: &str, candidates: &'a [&'a str]) -> Option<&' ranked_suggestions(input, candidates).into_iter().next() } - fn suggest_similar_subcommand(input: &str) -> Option> { const KNOWN_SUBCOMMANDS: &[&str] = &[ "help", @@ -1357,8 +1350,7 @@ fn suggest_similar_subcommand(input: &str) -> Option> { let prefix_match = common_prefix_len(&normalized_input, &normalized_candidate) >= 4; let substring_match = normalized_candidate.contains(&normalized_input) || normalized_input.contains(&normalized_candidate); - ((distance <= 2) || prefix_match || substring_match) - .then_some((distance, *candidate)) + ((distance <= 2) || prefix_match || substring_match).then_some((distance, *candidate)) }) .collect::>(); ranked.sort_by(|left, right| left.cmp(right).then_with(|| left.1.cmp(right.1))); @@ -1378,7 +1370,6 @@ fn common_prefix_len(left: &str, right: &str) -> usize { .count() } - fn looks_like_subcommand_typo(input: &str) -> bool { !input.is_empty() && input @@ -1487,13 +1478,11 @@ fn validate_model_syntax(model: &str) -> Result<(), String> { err_msg.push_str("\nDid you mean `openai/"); err_msg.push_str(trimmed); err_msg.push_str("`? (Requires OPENAI_API_KEY env var)"); - } - else if trimmed.starts_with("qwen") { + } else if trimmed.starts_with("qwen") { err_msg.push_str("\nDid you mean `qwen/"); err_msg.push_str(trimmed); err_msg.push_str("`? (Requires DASHSCOPE_API_KEY env var)"); - } - else if trimmed.starts_with("grok") { + } else if trimmed.starts_with("grok") { err_msg.push_str("\nDid you mean `xai/"); err_msg.push_str(trimmed); err_msg.push_str("`? (Requires XAI_API_KEY env var)"); @@ -4338,7 +4327,6 @@ impl LiveCli { Ok(()) } - fn run_prompt_compact_json(&mut self, input: &str) -> Result<(), Box> { let (mut runtime, hook_abort_monitor) = self.prepare_turn_runtime(false)?; let mut permission_prompter = CliPermissionPrompter::new(self.permission_mode); @@ -5457,7 +5445,13 @@ fn print_status_snapshot( match output_format { CliOutputFormat::Text => println!( "{}", - format_status_report(&provenance.resolved, usage, permission_mode.as_str(), &context, Some(&provenance)) + format_status_report( + &provenance.resolved, + usage, + permission_mode.as_str(), + &context, + Some(&provenance) + ) ), CliOutputFormat::Json => println!( "{}", @@ -9031,26 +9025,24 @@ fn print_help(output_format: CliOutputFormat) -> Result<(), Box