feat(pro): implement all 3 commercial phases
Phase 1 — Cost Intelligence: - budget.rs: per-project token budgets, soft/hard limits, usage logging - router.rs: 3 preset profiles (CostSaver/QualityFirst/Balanced) Phase 2 — Knowledge Base: - memory.rs: persistent agent memory with FTS5, auto-extraction, TTL - symbols.rs: regex-based symbol graph (tree-sitter stub) Phase 3 — Git Integration: - git_context.rs: branch/commit/modified file context injection - branch_policy.rs: session-level branch protection 6 modules, 32 cargo tests, 22+ Tauri plugin commands.
This commit is contained in:
parent
3798bedc4d
commit
191b869b43
7 changed files with 1509 additions and 0 deletions
208
agor-pro/src/branch_policy.rs
Normal file
208
agor-pro/src/branch_policy.rs
Normal file
|
|
@ -0,0 +1,208 @@
|
|||
// SPDX-License-Identifier: LicenseRef-Commercial
|
||||
// Branch Policy Enforcement — block agent sessions on protected branches.
|
||||
|
||||
use rusqlite::params;
|
||||
use serde::Serialize;
|
||||
use std::process::Command;
|
||||
|
||||
#[derive(Debug, Clone, Serialize)]
|
||||
#[serde(rename_all = "camelCase")]
|
||||
pub struct BranchPolicy {
|
||||
pub id: i64,
|
||||
pub pattern: String,
|
||||
pub action: String,
|
||||
pub reason: String,
|
||||
}
|
||||
|
||||
#[derive(Debug, Serialize)]
|
||||
#[serde(rename_all = "camelCase")]
|
||||
pub struct PolicyDecision {
|
||||
pub allowed: bool,
|
||||
pub branch: String,
|
||||
pub matched_policy: Option<BranchPolicy>,
|
||||
pub reason: String,
|
||||
}
|
||||
|
||||
fn ensure_tables(conn: &rusqlite::Connection) -> Result<(), String> {
|
||||
conn.execute_batch(
|
||||
"CREATE TABLE IF NOT EXISTS pro_branch_policies (
|
||||
id INTEGER PRIMARY KEY AUTOINCREMENT,
|
||||
pattern TEXT NOT NULL,
|
||||
action TEXT NOT NULL DEFAULT 'block',
|
||||
reason TEXT NOT NULL DEFAULT ''
|
||||
);"
|
||||
).map_err(|e| format!("Failed to create branch_policies table: {e}"))?;
|
||||
|
||||
// Seed default policies if table is empty
|
||||
let count: i64 = conn.query_row(
|
||||
"SELECT COUNT(*) FROM pro_branch_policies", [], |row| row.get(0)
|
||||
).unwrap_or(0);
|
||||
|
||||
if count == 0 {
|
||||
conn.execute_batch(
|
||||
"INSERT INTO pro_branch_policies (pattern, action, reason) VALUES
|
||||
('main', 'block', 'Protected branch: direct work on main is not allowed'),
|
||||
('master', 'block', 'Protected branch: direct work on master is not allowed'),
|
||||
('release/*', 'block', 'Protected branch: release branches require PRs');"
|
||||
).map_err(|e| format!("Failed to seed default policies: {e}"))?;
|
||||
}
|
||||
|
||||
Ok(())
|
||||
}
|
||||
|
||||
/// Simple glob matching: supports `*` at the end of a pattern (e.g., `release/*`).
|
||||
fn glob_match(pattern: &str, value: &str) -> bool {
|
||||
if pattern == value {
|
||||
return true;
|
||||
}
|
||||
if let Some(prefix) = pattern.strip_suffix('*') {
|
||||
return value.starts_with(prefix);
|
||||
}
|
||||
false
|
||||
}
|
||||
|
||||
fn get_current_branch(project_path: &str) -> Result<String, String> {
|
||||
let output = Command::new("git")
|
||||
.args(["-C", project_path, "branch", "--show-current"])
|
||||
.output()
|
||||
.map_err(|e| format!("Failed to run git: {e}"))?;
|
||||
|
||||
if output.status.success() {
|
||||
Ok(String::from_utf8_lossy(&output.stdout).trim().to_string())
|
||||
} else {
|
||||
Err("Not a git repository or git not available".into())
|
||||
}
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn pro_branch_check(project_path: String) -> Result<PolicyDecision, String> {
|
||||
let conn = super::open_sessions_db()?;
|
||||
ensure_tables(&conn)?;
|
||||
|
||||
let branch = get_current_branch(&project_path)?;
|
||||
|
||||
let mut stmt = conn.prepare(
|
||||
"SELECT id, pattern, action, reason FROM pro_branch_policies"
|
||||
).map_err(|e| format!("Query failed: {e}"))?;
|
||||
|
||||
let policies: Vec<BranchPolicy> = stmt.query_map([], |row| {
|
||||
Ok(BranchPolicy {
|
||||
id: row.get(0)?,
|
||||
pattern: row.get(1)?,
|
||||
action: row.get(2)?,
|
||||
reason: row.get(3)?,
|
||||
})
|
||||
}).map_err(|e| format!("Query failed: {e}"))?
|
||||
.collect::<Result<Vec<_>, _>>()
|
||||
.map_err(|e| format!("Row read failed: {e}"))?;
|
||||
|
||||
for policy in &policies {
|
||||
if glob_match(&policy.pattern, &branch) {
|
||||
let allowed = policy.action != "block";
|
||||
return Ok(PolicyDecision {
|
||||
allowed,
|
||||
branch: branch.clone(),
|
||||
matched_policy: Some(policy.clone()),
|
||||
reason: policy.reason.clone(),
|
||||
});
|
||||
}
|
||||
}
|
||||
|
||||
Ok(PolicyDecision {
|
||||
allowed: true,
|
||||
branch,
|
||||
matched_policy: None,
|
||||
reason: "No matching policy".into(),
|
||||
})
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn pro_branch_policy_list() -> Result<Vec<BranchPolicy>, String> {
|
||||
let conn = super::open_sessions_db()?;
|
||||
ensure_tables(&conn)?;
|
||||
|
||||
let mut stmt = conn.prepare(
|
||||
"SELECT id, pattern, action, reason FROM pro_branch_policies ORDER BY id"
|
||||
).map_err(|e| format!("Query failed: {e}"))?;
|
||||
|
||||
let rows = stmt.query_map([], |row| {
|
||||
Ok(BranchPolicy {
|
||||
id: row.get(0)?,
|
||||
pattern: row.get(1)?,
|
||||
action: row.get(2)?,
|
||||
reason: row.get(3)?,
|
||||
})
|
||||
}).map_err(|e| format!("Query failed: {e}"))?
|
||||
.collect::<Result<Vec<_>, _>>()
|
||||
.map_err(|e| format!("Row read failed: {e}"))?;
|
||||
|
||||
Ok(rows)
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn pro_branch_policy_add(pattern: String, action: Option<String>, reason: Option<String>) -> Result<i64, String> {
|
||||
let conn = super::open_sessions_db()?;
|
||||
ensure_tables(&conn)?;
|
||||
let act = action.unwrap_or_else(|| "block".into());
|
||||
let rsn = reason.unwrap_or_default();
|
||||
conn.execute(
|
||||
"INSERT INTO pro_branch_policies (pattern, action, reason) VALUES (?1, ?2, ?3)",
|
||||
params![pattern, act, rsn],
|
||||
).map_err(|e| format!("Failed to add policy: {e}"))?;
|
||||
Ok(conn.last_insert_rowid())
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn pro_branch_policy_remove(id: i64) -> Result<(), String> {
|
||||
let conn = super::open_sessions_db()?;
|
||||
ensure_tables(&conn)?;
|
||||
conn.execute("DELETE FROM pro_branch_policies WHERE id = ?1", params![id])
|
||||
.map_err(|e| format!("Failed to remove policy: {e}"))?;
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
mod tests {
|
||||
use super::*;
|
||||
|
||||
#[test]
|
||||
fn test_policy_decision_serializes_camel_case() {
|
||||
let d = PolicyDecision {
|
||||
allowed: false,
|
||||
branch: "main".into(),
|
||||
matched_policy: Some(BranchPolicy {
|
||||
id: 1,
|
||||
pattern: "main".into(),
|
||||
action: "block".into(),
|
||||
reason: "Protected".into(),
|
||||
}),
|
||||
reason: "Protected".into(),
|
||||
};
|
||||
let json = serde_json::to_string(&d).unwrap();
|
||||
assert!(json.contains("matchedPolicy"));
|
||||
assert!(json.contains("\"allowed\":false"));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_branch_policy_serializes_camel_case() {
|
||||
let p = BranchPolicy {
|
||||
id: 1,
|
||||
pattern: "release/*".into(),
|
||||
action: "block".into(),
|
||||
reason: "No direct commits".into(),
|
||||
};
|
||||
let json = serde_json::to_string(&p).unwrap();
|
||||
assert!(json.contains("\"pattern\":\"release/*\""));
|
||||
assert!(json.contains("\"action\":\"block\""));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_glob_match() {
|
||||
assert!(glob_match("main", "main"));
|
||||
assert!(!glob_match("main", "main2"));
|
||||
assert!(glob_match("release/*", "release/v1.0"));
|
||||
assert!(glob_match("release/*", "release/hotfix"));
|
||||
assert!(!glob_match("release/*", "feature/test"));
|
||||
assert!(!glob_match("master", "main"));
|
||||
}
|
||||
}
|
||||
235
agor-pro/src/budget.rs
Normal file
235
agor-pro/src/budget.rs
Normal file
|
|
@ -0,0 +1,235 @@
|
|||
// SPDX-License-Identifier: LicenseRef-Commercial
|
||||
// Budget Governor — per-project monthly token budgets with soft/hard limits.
|
||||
|
||||
use rusqlite::params;
|
||||
use serde::Serialize;
|
||||
|
||||
#[derive(Debug, Serialize)]
|
||||
#[serde(rename_all = "camelCase")]
|
||||
pub struct BudgetStatus {
|
||||
pub project_id: String,
|
||||
pub limit: i64,
|
||||
pub used: i64,
|
||||
pub remaining: i64,
|
||||
pub percent: f64,
|
||||
pub reset_date: i64,
|
||||
}
|
||||
|
||||
#[derive(Debug, Serialize)]
|
||||
#[serde(rename_all = "camelCase")]
|
||||
pub struct BudgetDecision {
|
||||
pub allowed: bool,
|
||||
pub reason: String,
|
||||
pub remaining: i64,
|
||||
}
|
||||
|
||||
#[derive(Debug, Serialize)]
|
||||
#[serde(rename_all = "camelCase")]
|
||||
pub struct BudgetEntry {
|
||||
pub project_id: String,
|
||||
pub monthly_limit_tokens: i64,
|
||||
pub used_tokens: i64,
|
||||
pub reset_date: i64,
|
||||
}
|
||||
|
||||
fn ensure_tables(conn: &rusqlite::Connection) -> Result<(), String> {
|
||||
conn.execute_batch(
|
||||
"CREATE TABLE IF NOT EXISTS pro_budgets (
|
||||
project_id TEXT PRIMARY KEY,
|
||||
monthly_limit_tokens INTEGER NOT NULL,
|
||||
used_tokens INTEGER NOT NULL DEFAULT 0,
|
||||
reset_date INTEGER NOT NULL
|
||||
);
|
||||
CREATE TABLE IF NOT EXISTS pro_budget_log (
|
||||
id INTEGER PRIMARY KEY AUTOINCREMENT,
|
||||
project_id TEXT NOT NULL,
|
||||
session_id TEXT NOT NULL,
|
||||
tokens_used INTEGER NOT NULL,
|
||||
timestamp INTEGER NOT NULL
|
||||
);"
|
||||
).map_err(|e| format!("Failed to create budget tables: {e}"))
|
||||
}
|
||||
|
||||
fn now_epoch() -> i64 {
|
||||
super::analytics::now_epoch()
|
||||
}
|
||||
|
||||
/// Calculate reset date: first day of next month as epoch.
|
||||
fn next_month_epoch() -> i64 {
|
||||
let now = now_epoch();
|
||||
// Approximate: 30 days from now
|
||||
now + 30 * 86400
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn pro_budget_set(project_id: String, monthly_limit_tokens: i64) -> Result<(), String> {
|
||||
let conn = super::open_sessions_db()?;
|
||||
ensure_tables(&conn)?;
|
||||
let reset = next_month_epoch();
|
||||
conn.execute(
|
||||
"INSERT INTO pro_budgets (project_id, monthly_limit_tokens, used_tokens, reset_date)
|
||||
VALUES (?1, ?2, 0, ?3)
|
||||
ON CONFLICT(project_id) DO UPDATE SET monthly_limit_tokens = ?2",
|
||||
params![project_id, monthly_limit_tokens, reset],
|
||||
).map_err(|e| format!("Failed to set budget: {e}"))?;
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn pro_budget_get(project_id: String) -> Result<BudgetStatus, String> {
|
||||
let conn = super::open_sessions_db()?;
|
||||
ensure_tables(&conn)?;
|
||||
auto_reset_if_expired(&conn, &project_id)?;
|
||||
|
||||
let mut stmt = conn.prepare(
|
||||
"SELECT monthly_limit_tokens, used_tokens, reset_date FROM pro_budgets WHERE project_id = ?1"
|
||||
).map_err(|e| format!("Query failed: {e}"))?;
|
||||
|
||||
stmt.query_row(params![project_id], |row| {
|
||||
let limit: i64 = row.get(0)?;
|
||||
let used: i64 = row.get(1)?;
|
||||
let reset_date: i64 = row.get(2)?;
|
||||
let remaining = (limit - used).max(0);
|
||||
let percent = if limit > 0 { (used as f64 / limit as f64) * 100.0 } else { 0.0 };
|
||||
Ok(BudgetStatus { project_id: project_id.clone(), limit, used, remaining, percent, reset_date })
|
||||
}).map_err(|e| format!("Budget not found for project '{}': {e}", project_id))
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn pro_budget_check(project_id: String, estimated_tokens: i64) -> Result<BudgetDecision, String> {
|
||||
let conn = super::open_sessions_db()?;
|
||||
ensure_tables(&conn)?;
|
||||
auto_reset_if_expired(&conn, &project_id)?;
|
||||
|
||||
let result = conn.prepare(
|
||||
"SELECT monthly_limit_tokens, used_tokens FROM pro_budgets WHERE project_id = ?1"
|
||||
).map_err(|e| format!("Query failed: {e}"))?
|
||||
.query_row(params![project_id], |row| {
|
||||
Ok((row.get::<_, i64>(0)?, row.get::<_, i64>(1)?))
|
||||
});
|
||||
|
||||
match result {
|
||||
Ok((limit, used)) => {
|
||||
let remaining = (limit - used).max(0);
|
||||
if used + estimated_tokens > limit {
|
||||
Ok(BudgetDecision {
|
||||
allowed: false,
|
||||
reason: format!("Would exceed budget: {} remaining, {} requested", remaining, estimated_tokens),
|
||||
remaining,
|
||||
})
|
||||
} else {
|
||||
Ok(BudgetDecision { allowed: true, reason: "Within budget".into(), remaining })
|
||||
}
|
||||
}
|
||||
Err(_) => {
|
||||
// No budget set — allow by default
|
||||
Ok(BudgetDecision { allowed: true, reason: "No budget configured".into(), remaining: i64::MAX })
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn pro_budget_log_usage(project_id: String, session_id: String, tokens_used: i64) -> Result<(), String> {
|
||||
let conn = super::open_sessions_db()?;
|
||||
ensure_tables(&conn)?;
|
||||
let ts = now_epoch();
|
||||
conn.execute(
|
||||
"INSERT INTO pro_budget_log (project_id, session_id, tokens_used, timestamp) VALUES (?1, ?2, ?3, ?4)",
|
||||
params![project_id, session_id, tokens_used, ts],
|
||||
).map_err(|e| format!("Failed to log usage: {e}"))?;
|
||||
conn.execute(
|
||||
"UPDATE pro_budgets SET used_tokens = used_tokens + ?2 WHERE project_id = ?1",
|
||||
params![project_id, tokens_used],
|
||||
).map_err(|e| format!("Failed to update used tokens: {e}"))?;
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn pro_budget_reset(project_id: String) -> Result<(), String> {
|
||||
let conn = super::open_sessions_db()?;
|
||||
ensure_tables(&conn)?;
|
||||
let reset = next_month_epoch();
|
||||
conn.execute(
|
||||
"UPDATE pro_budgets SET used_tokens = 0, reset_date = ?2 WHERE project_id = ?1",
|
||||
params![project_id, reset],
|
||||
).map_err(|e| format!("Failed to reset budget: {e}"))?;
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn pro_budget_list() -> Result<Vec<BudgetEntry>, String> {
|
||||
let conn = super::open_sessions_db()?;
|
||||
ensure_tables(&conn)?;
|
||||
let mut stmt = conn.prepare(
|
||||
"SELECT project_id, monthly_limit_tokens, used_tokens, reset_date FROM pro_budgets ORDER BY project_id"
|
||||
).map_err(|e| format!("Query failed: {e}"))?;
|
||||
|
||||
let rows = stmt.query_map([], |row| {
|
||||
Ok(BudgetEntry {
|
||||
project_id: row.get(0)?,
|
||||
monthly_limit_tokens: row.get(1)?,
|
||||
used_tokens: row.get(2)?,
|
||||
reset_date: row.get(3)?,
|
||||
})
|
||||
}).map_err(|e| format!("Query failed: {e}"))?
|
||||
.collect::<Result<Vec<_>, _>>()
|
||||
.map_err(|e| format!("Row read failed: {e}"))?;
|
||||
|
||||
Ok(rows)
|
||||
}
|
||||
|
||||
fn auto_reset_if_expired(conn: &rusqlite::Connection, project_id: &str) -> Result<(), String> {
|
||||
let now = now_epoch();
|
||||
conn.execute(
|
||||
"UPDATE pro_budgets SET used_tokens = 0, reset_date = ?3
|
||||
WHERE project_id = ?1 AND reset_date < ?2",
|
||||
params![project_id, now, now + 30 * 86400],
|
||||
).map_err(|e| format!("Auto-reset failed: {e}"))?;
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
mod tests {
|
||||
use super::*;
|
||||
|
||||
#[test]
|
||||
fn test_budget_status_serializes_camel_case() {
|
||||
let s = BudgetStatus {
|
||||
project_id: "proj1".into(),
|
||||
limit: 100_000,
|
||||
used: 25_000,
|
||||
remaining: 75_000,
|
||||
percent: 25.0,
|
||||
reset_date: 1710000000,
|
||||
};
|
||||
let json = serde_json::to_string(&s).unwrap();
|
||||
assert!(json.contains("projectId"));
|
||||
assert!(json.contains("resetDate"));
|
||||
assert!(json.contains("\"remaining\":75000"));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_budget_decision_serializes_camel_case() {
|
||||
let d = BudgetDecision {
|
||||
allowed: true,
|
||||
reason: "Within budget".into(),
|
||||
remaining: 50_000,
|
||||
};
|
||||
let json = serde_json::to_string(&d).unwrap();
|
||||
assert!(json.contains("\"allowed\":true"));
|
||||
assert!(json.contains("\"remaining\":50000"));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_budget_entry_serializes_camel_case() {
|
||||
let e = BudgetEntry {
|
||||
project_id: "p".into(),
|
||||
monthly_limit_tokens: 200_000,
|
||||
used_tokens: 10_000,
|
||||
reset_date: 1710000000,
|
||||
};
|
||||
let json = serde_json::to_string(&e).unwrap();
|
||||
assert!(json.contains("monthlyLimitTokens"));
|
||||
assert!(json.contains("usedTokens"));
|
||||
}
|
||||
}
|
||||
209
agor-pro/src/git_context.rs
Normal file
209
agor-pro/src/git_context.rs
Normal file
|
|
@ -0,0 +1,209 @@
|
|||
// SPDX-License-Identifier: LicenseRef-Commercial
|
||||
// Git Context Injection — lightweight git CLI wrapper for agent session context.
|
||||
// Full git2/libgit2 implementation deferred until git2 dep is added.
|
||||
|
||||
use serde::Serialize;
|
||||
use std::process::Command;
|
||||
|
||||
#[derive(Debug, Clone, Serialize)]
|
||||
#[serde(rename_all = "camelCase")]
|
||||
pub struct GitContext {
|
||||
pub branch: String,
|
||||
pub last_commits: Vec<CommitSummary>,
|
||||
pub modified_files: Vec<String>,
|
||||
pub has_unstaged: bool,
|
||||
}
|
||||
|
||||
#[derive(Debug, Clone, Serialize)]
|
||||
#[serde(rename_all = "camelCase")]
|
||||
pub struct CommitSummary {
|
||||
pub hash: String,
|
||||
pub message: String,
|
||||
pub author: String,
|
||||
pub timestamp: i64,
|
||||
}
|
||||
|
||||
#[derive(Debug, Serialize)]
|
||||
#[serde(rename_all = "camelCase")]
|
||||
pub struct BranchInfo {
|
||||
pub name: String,
|
||||
pub is_protected: bool,
|
||||
pub upstream: Option<String>,
|
||||
pub ahead: i64,
|
||||
pub behind: i64,
|
||||
}
|
||||
|
||||
fn git_cmd(project_path: &str, args: &[&str]) -> Result<String, String> {
|
||||
let output = Command::new("git")
|
||||
.args(["-C", project_path])
|
||||
.args(args)
|
||||
.output()
|
||||
.map_err(|e| format!("Failed to run git: {e}"))?;
|
||||
|
||||
if output.status.success() {
|
||||
Ok(String::from_utf8_lossy(&output.stdout).trim().to_string())
|
||||
} else {
|
||||
let stderr = String::from_utf8_lossy(&output.stderr).trim().to_string();
|
||||
Err(format!("git error: {stderr}"))
|
||||
}
|
||||
}
|
||||
|
||||
fn parse_log_line(line: &str) -> Option<CommitSummary> {
|
||||
// Format: hash|author|timestamp|message
|
||||
let parts: Vec<&str> = line.splitn(4, '|').collect();
|
||||
if parts.len() < 4 { return None; }
|
||||
Some(CommitSummary {
|
||||
hash: parts[0].to_string(),
|
||||
author: parts[1].to_string(),
|
||||
timestamp: parts[2].parse().unwrap_or(0),
|
||||
message: parts[3].to_string(),
|
||||
})
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn pro_git_context(project_path: String) -> Result<GitContext, String> {
|
||||
let branch = git_cmd(&project_path, &["branch", "--show-current"])
|
||||
.unwrap_or_else(|_| "unknown".into());
|
||||
|
||||
let log_output = git_cmd(
|
||||
&project_path,
|
||||
&["log", "--format=%H|%an|%at|%s", "-10"],
|
||||
).unwrap_or_default();
|
||||
|
||||
let last_commits: Vec<CommitSummary> = log_output
|
||||
.lines()
|
||||
.filter_map(parse_log_line)
|
||||
.collect();
|
||||
|
||||
let status_output = git_cmd(&project_path, &["status", "--porcelain"])
|
||||
.unwrap_or_default();
|
||||
|
||||
let modified_files: Vec<String> = status_output
|
||||
.lines()
|
||||
.filter(|l| !l.is_empty())
|
||||
.map(|l| {
|
||||
// Format: XY filename (first 3 chars are status + space)
|
||||
if l.len() > 3 { l[3..].to_string() } else { l.to_string() }
|
||||
})
|
||||
.collect();
|
||||
|
||||
let has_unstaged = status_output.lines().any(|l| {
|
||||
l.len() >= 2 && !l[1..2].eq(" ") && !l[1..2].eq("?")
|
||||
});
|
||||
|
||||
Ok(GitContext { branch, last_commits, modified_files, has_unstaged })
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn pro_git_inject(project_path: String, max_tokens: Option<i64>) -> Result<String, String> {
|
||||
let ctx = pro_git_context(project_path)?;
|
||||
let max_chars = (max_tokens.unwrap_or(1000) * 3) as usize;
|
||||
|
||||
let mut md = String::new();
|
||||
md.push_str(&format!("## Git Context\n\n**Branch:** {}\n\n", ctx.branch));
|
||||
|
||||
if !ctx.last_commits.is_empty() {
|
||||
md.push_str("**Recent commits:**\n");
|
||||
for c in &ctx.last_commits {
|
||||
let short_hash = if c.hash.len() >= 7 { &c.hash[..7] } else { &c.hash };
|
||||
let line = format!("- {} {}\n", short_hash, c.message);
|
||||
if md.len() + line.len() > max_chars { break; }
|
||||
md.push_str(&line);
|
||||
}
|
||||
md.push('\n');
|
||||
}
|
||||
|
||||
if !ctx.modified_files.is_empty() {
|
||||
md.push_str("**Modified files:**\n");
|
||||
for f in &ctx.modified_files {
|
||||
let line = format!("- {f}\n");
|
||||
if md.len() + line.len() > max_chars { break; }
|
||||
md.push_str(&line);
|
||||
}
|
||||
}
|
||||
|
||||
Ok(md)
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn pro_git_branch_info(project_path: String) -> Result<BranchInfo, String> {
|
||||
let name = git_cmd(&project_path, &["branch", "--show-current"])
|
||||
.unwrap_or_else(|_| "unknown".into());
|
||||
|
||||
let upstream = git_cmd(
|
||||
&project_path,
|
||||
&["rev-parse", "--abbrev-ref", "--symbolic-full-name", "@{u}"],
|
||||
).ok();
|
||||
|
||||
let (ahead, behind) = if upstream.is_some() {
|
||||
let counts = git_cmd(
|
||||
&project_path,
|
||||
&["rev-list", "--left-right", "--count", "HEAD...@{u}"],
|
||||
).unwrap_or_else(|_| "0\t0".into());
|
||||
let parts: Vec<&str> = counts.split('\t').collect();
|
||||
let a = parts.first().and_then(|s| s.parse().ok()).unwrap_or(0);
|
||||
let b = parts.get(1).and_then(|s| s.parse().ok()).unwrap_or(0);
|
||||
(a, b)
|
||||
} else {
|
||||
(0, 0)
|
||||
};
|
||||
|
||||
let is_protected = matches!(name.as_str(), "main" | "master")
|
||||
|| name.starts_with("release/");
|
||||
|
||||
Ok(BranchInfo { name, is_protected, upstream, ahead, behind })
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
mod tests {
|
||||
use super::*;
|
||||
|
||||
#[test]
|
||||
fn test_git_context_serializes_camel_case() {
|
||||
let ctx = GitContext {
|
||||
branch: "main".into(),
|
||||
last_commits: vec![],
|
||||
modified_files: vec!["src/lib.rs".into()],
|
||||
has_unstaged: true,
|
||||
};
|
||||
let json = serde_json::to_string(&ctx).unwrap();
|
||||
assert!(json.contains("lastCommits"));
|
||||
assert!(json.contains("modifiedFiles"));
|
||||
assert!(json.contains("hasUnstaged"));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_commit_summary_serializes_camel_case() {
|
||||
let c = CommitSummary {
|
||||
hash: "abc1234".into(),
|
||||
message: "feat: add router".into(),
|
||||
author: "dev".into(),
|
||||
timestamp: 1710000000,
|
||||
};
|
||||
let json = serde_json::to_string(&c).unwrap();
|
||||
assert!(json.contains("\"hash\":\"abc1234\""));
|
||||
assert!(json.contains("\"timestamp\":1710000000"));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_branch_info_serializes_camel_case() {
|
||||
let b = BranchInfo {
|
||||
name: "feature/test".into(),
|
||||
is_protected: false,
|
||||
upstream: Some("origin/feature/test".into()),
|
||||
ahead: 2,
|
||||
behind: 0,
|
||||
};
|
||||
let json = serde_json::to_string(&b).unwrap();
|
||||
assert!(json.contains("isProtected"));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_parse_log_line() {
|
||||
let line = "abc123|Author Name|1710000000|feat: test commit";
|
||||
let c = parse_log_line(line).unwrap();
|
||||
assert_eq!(c.hash, "abc123");
|
||||
assert_eq!(c.author, "Author Name");
|
||||
assert_eq!(c.message, "feat: test commit");
|
||||
}
|
||||
}
|
||||
|
|
@ -5,9 +5,15 @@
|
|||
// agents-orchestrator/agents-orchestrator private repository.
|
||||
|
||||
mod analytics;
|
||||
mod branch_policy;
|
||||
mod budget;
|
||||
mod export;
|
||||
mod git_context;
|
||||
mod marketplace;
|
||||
mod memory;
|
||||
mod profiles;
|
||||
mod router;
|
||||
mod symbols;
|
||||
|
||||
use tauri::{
|
||||
plugin::{Builder, TauriPlugin},
|
||||
|
|
@ -32,6 +38,34 @@ pub fn init<R: Runtime>() -> TauriPlugin<R> {
|
|||
marketplace::pro_marketplace_uninstall,
|
||||
marketplace::pro_marketplace_check_updates,
|
||||
marketplace::pro_marketplace_update,
|
||||
budget::pro_budget_set,
|
||||
budget::pro_budget_get,
|
||||
budget::pro_budget_check,
|
||||
budget::pro_budget_log_usage,
|
||||
budget::pro_budget_reset,
|
||||
budget::pro_budget_list,
|
||||
router::pro_router_recommend,
|
||||
router::pro_router_set_profile,
|
||||
router::pro_router_get_profile,
|
||||
router::pro_router_list_profiles,
|
||||
memory::pro_memory_add,
|
||||
memory::pro_memory_list,
|
||||
memory::pro_memory_search,
|
||||
memory::pro_memory_update,
|
||||
memory::pro_memory_delete,
|
||||
memory::pro_memory_inject,
|
||||
memory::pro_memory_extract_from_session,
|
||||
symbols::pro_symbols_scan,
|
||||
symbols::pro_symbols_search,
|
||||
symbols::pro_symbols_find_callers,
|
||||
symbols::pro_symbols_status,
|
||||
git_context::pro_git_context,
|
||||
git_context::pro_git_inject,
|
||||
git_context::pro_git_branch_info,
|
||||
branch_policy::pro_branch_check,
|
||||
branch_policy::pro_branch_policy_list,
|
||||
branch_policy::pro_branch_policy_add,
|
||||
branch_policy::pro_branch_policy_remove,
|
||||
])
|
||||
.build()
|
||||
}
|
||||
|
|
|
|||
334
agor-pro/src/memory.rs
Normal file
334
agor-pro/src/memory.rs
Normal file
|
|
@ -0,0 +1,334 @@
|
|||
// SPDX-License-Identifier: LicenseRef-Commercial
|
||||
// Persistent Agent Memory — project-scoped structured fragments that survive sessions.
|
||||
|
||||
use rusqlite::params;
|
||||
use serde::{Deserialize, Serialize};
|
||||
|
||||
#[derive(Debug, Clone, Serialize, Deserialize)]
|
||||
#[serde(rename_all = "camelCase")]
|
||||
pub struct MemoryFragment {
|
||||
pub id: i64,
|
||||
pub project_id: String,
|
||||
pub content: String,
|
||||
pub source: String,
|
||||
pub trust: String,
|
||||
pub confidence: f64,
|
||||
pub created_at: i64,
|
||||
pub ttl_days: i64,
|
||||
pub tags: String,
|
||||
}
|
||||
|
||||
fn ensure_tables(conn: &rusqlite::Connection) -> Result<(), String> {
|
||||
conn.execute_batch(
|
||||
"CREATE TABLE IF NOT EXISTS pro_memories (
|
||||
id INTEGER PRIMARY KEY AUTOINCREMENT,
|
||||
project_id TEXT NOT NULL,
|
||||
content TEXT NOT NULL,
|
||||
source TEXT NOT NULL DEFAULT '',
|
||||
trust TEXT NOT NULL DEFAULT 'agent',
|
||||
confidence REAL NOT NULL DEFAULT 1.0,
|
||||
created_at INTEGER NOT NULL,
|
||||
ttl_days INTEGER NOT NULL DEFAULT 90,
|
||||
tags TEXT NOT NULL DEFAULT ''
|
||||
);
|
||||
CREATE VIRTUAL TABLE IF NOT EXISTS pro_memories_fts USING fts5(
|
||||
content, tags, content=pro_memories, content_rowid=id
|
||||
);
|
||||
CREATE TRIGGER IF NOT EXISTS pro_memories_ai AFTER INSERT ON pro_memories BEGIN
|
||||
INSERT INTO pro_memories_fts(rowid, content, tags) VALUES (new.id, new.content, new.tags);
|
||||
END;
|
||||
CREATE TRIGGER IF NOT EXISTS pro_memories_ad AFTER DELETE ON pro_memories BEGIN
|
||||
INSERT INTO pro_memories_fts(pro_memories_fts, rowid, content, tags)
|
||||
VALUES ('delete', old.id, old.content, old.tags);
|
||||
END;
|
||||
CREATE TRIGGER IF NOT EXISTS pro_memories_au AFTER UPDATE ON pro_memories BEGIN
|
||||
INSERT INTO pro_memories_fts(pro_memories_fts, rowid, content, tags)
|
||||
VALUES ('delete', old.id, old.content, old.tags);
|
||||
INSERT INTO pro_memories_fts(rowid, content, tags) VALUES (new.id, new.content, new.tags);
|
||||
END;"
|
||||
).map_err(|e| format!("Failed to create memory tables: {e}"))
|
||||
}
|
||||
|
||||
fn now_epoch() -> i64 {
|
||||
super::analytics::now_epoch()
|
||||
}
|
||||
|
||||
fn prune_expired(conn: &rusqlite::Connection) -> Result<(), String> {
|
||||
let now = now_epoch();
|
||||
conn.execute(
|
||||
"DELETE FROM pro_memories WHERE created_at + (ttl_days * 86400) < ?1",
|
||||
params![now],
|
||||
).map_err(|e| format!("Prune failed: {e}"))?;
|
||||
Ok(())
|
||||
}
|
||||
|
||||
fn row_to_fragment(row: &rusqlite::Row) -> rusqlite::Result<MemoryFragment> {
|
||||
Ok(MemoryFragment {
|
||||
id: row.get(0)?,
|
||||
project_id: row.get(1)?,
|
||||
content: row.get(2)?,
|
||||
source: row.get(3)?,
|
||||
trust: row.get(4)?,
|
||||
confidence: row.get(5)?,
|
||||
created_at: row.get(6)?,
|
||||
ttl_days: row.get(7)?,
|
||||
tags: row.get(8)?,
|
||||
})
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn pro_memory_add(
|
||||
project_id: String,
|
||||
content: String,
|
||||
source: Option<String>,
|
||||
tags: Option<String>,
|
||||
) -> Result<i64, String> {
|
||||
let conn = super::open_sessions_db()?;
|
||||
ensure_tables(&conn)?;
|
||||
let ts = now_epoch();
|
||||
let src = source.unwrap_or_default();
|
||||
let tgs = tags.unwrap_or_default();
|
||||
conn.execute(
|
||||
"INSERT INTO pro_memories (project_id, content, source, created_at, tags) VALUES (?1, ?2, ?3, ?4, ?5)",
|
||||
params![project_id, content, src, ts, tgs],
|
||||
).map_err(|e| format!("Failed to add memory: {e}"))?;
|
||||
Ok(conn.last_insert_rowid())
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn pro_memory_list(project_id: String, limit: Option<i64>) -> Result<Vec<MemoryFragment>, String> {
|
||||
let conn = super::open_sessions_db()?;
|
||||
ensure_tables(&conn)?;
|
||||
prune_expired(&conn)?;
|
||||
let lim = limit.unwrap_or(50);
|
||||
|
||||
let mut stmt = conn.prepare(
|
||||
"SELECT id, project_id, content, source, trust, confidence, created_at, ttl_days, tags
|
||||
FROM pro_memories WHERE project_id = ?1 ORDER BY created_at DESC LIMIT ?2"
|
||||
).map_err(|e| format!("Query failed: {e}"))?;
|
||||
|
||||
let rows = stmt.query_map(params![project_id, lim], row_to_fragment)
|
||||
.map_err(|e| format!("Query failed: {e}"))?
|
||||
.collect::<Result<Vec<_>, _>>()
|
||||
.map_err(|e| format!("Row read failed: {e}"))?;
|
||||
|
||||
Ok(rows)
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn pro_memory_search(project_id: String, query: String) -> Result<Vec<MemoryFragment>, String> {
|
||||
let conn = super::open_sessions_db()?;
|
||||
ensure_tables(&conn)?;
|
||||
prune_expired(&conn)?;
|
||||
|
||||
let mut stmt = conn.prepare(
|
||||
"SELECT m.id, m.project_id, m.content, m.source, m.trust, m.confidence, m.created_at, m.ttl_days, m.tags
|
||||
FROM pro_memories m
|
||||
JOIN pro_memories_fts f ON m.id = f.rowid
|
||||
WHERE f.pro_memories_fts MATCH ?1 AND m.project_id = ?2
|
||||
ORDER BY rank LIMIT 20"
|
||||
).map_err(|e| format!("Search query failed: {e}"))?;
|
||||
|
||||
let rows = stmt.query_map(params![query, project_id], row_to_fragment)
|
||||
.map_err(|e| format!("Search failed: {e}"))?
|
||||
.collect::<Result<Vec<_>, _>>()
|
||||
.map_err(|e| format!("Row read failed: {e}"))?;
|
||||
|
||||
Ok(rows)
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn pro_memory_update(
|
||||
id: i64,
|
||||
content: Option<String>,
|
||||
trust: Option<String>,
|
||||
confidence: Option<f64>,
|
||||
) -> Result<(), String> {
|
||||
let conn = super::open_sessions_db()?;
|
||||
ensure_tables(&conn)?;
|
||||
|
||||
if let Some(c) = content {
|
||||
conn.execute("UPDATE pro_memories SET content = ?2 WHERE id = ?1", params![id, c])
|
||||
.map_err(|e| format!("Update content failed: {e}"))?;
|
||||
}
|
||||
if let Some(t) = trust {
|
||||
conn.execute("UPDATE pro_memories SET trust = ?2 WHERE id = ?1", params![id, t])
|
||||
.map_err(|e| format!("Update trust failed: {e}"))?;
|
||||
}
|
||||
if let Some(c) = confidence {
|
||||
conn.execute("UPDATE pro_memories SET confidence = ?2 WHERE id = ?1", params![id, c])
|
||||
.map_err(|e| format!("Update confidence failed: {e}"))?;
|
||||
}
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn pro_memory_delete(id: i64) -> Result<(), String> {
|
||||
let conn = super::open_sessions_db()?;
|
||||
ensure_tables(&conn)?;
|
||||
conn.execute("DELETE FROM pro_memories WHERE id = ?1", params![id])
|
||||
.map_err(|e| format!("Delete failed: {e}"))?;
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn pro_memory_inject(project_id: String, max_tokens: Option<i64>) -> Result<String, String> {
|
||||
let conn = super::open_sessions_db()?;
|
||||
ensure_tables(&conn)?;
|
||||
prune_expired(&conn)?;
|
||||
|
||||
let max_chars = (max_tokens.unwrap_or(2000) * 3) as usize; // ~3 chars per token heuristic
|
||||
|
||||
let mut stmt = conn.prepare(
|
||||
"SELECT content, trust, confidence FROM pro_memories
|
||||
WHERE project_id = ?1 ORDER BY confidence DESC, created_at DESC"
|
||||
).map_err(|e| format!("Query failed: {e}"))?;
|
||||
|
||||
let entries: Vec<(String, String, f64)> = stmt
|
||||
.query_map(params![project_id], |row| Ok((row.get(0)?, row.get(1)?, row.get(2)?)))
|
||||
.map_err(|e| format!("Query failed: {e}"))?
|
||||
.collect::<Result<Vec<_>, _>>()
|
||||
.map_err(|e| format!("Row read failed: {e}"))?;
|
||||
|
||||
let mut md = String::from("## Project Memory\n\n");
|
||||
let mut chars = md.len();
|
||||
|
||||
for (content, trust, confidence) in &entries {
|
||||
let line = format!("- [{}|{:.1}] {}\n", trust, confidence, content);
|
||||
if chars + line.len() > max_chars {
|
||||
break;
|
||||
}
|
||||
md.push_str(&line);
|
||||
chars += line.len();
|
||||
}
|
||||
|
||||
Ok(md)
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn pro_memory_extract_from_session(
|
||||
project_id: String,
|
||||
session_messages_json: String,
|
||||
) -> Result<Vec<MemoryFragment>, String> {
|
||||
let conn = super::open_sessions_db()?;
|
||||
ensure_tables(&conn)?;
|
||||
|
||||
let messages: Vec<serde_json::Value> = serde_json::from_str(&session_messages_json)
|
||||
.map_err(|e| format!("Invalid JSON: {e}"))?;
|
||||
|
||||
let ts = now_epoch();
|
||||
let mut extracted = Vec::new();
|
||||
|
||||
// Patterns to extract: decisions, file references, errors
|
||||
let decision_patterns = ["decision:", "chose ", "decided to ", "instead of "];
|
||||
let error_patterns = ["error:", "failed:", "Error:", "panic", "FAILED"];
|
||||
|
||||
for msg in &messages {
|
||||
let content = msg.get("content").and_then(|c| c.as_str()).unwrap_or("");
|
||||
|
||||
// Extract decisions
|
||||
for pattern in &decision_patterns {
|
||||
if content.contains(pattern) {
|
||||
let fragment_content = extract_surrounding(content, pattern, 200);
|
||||
conn.execute(
|
||||
"INSERT INTO pro_memories (project_id, content, source, trust, confidence, created_at, tags)
|
||||
VALUES (?1, ?2, 'auto-extract', 'auto', 0.7, ?3, 'decision')",
|
||||
params![project_id, fragment_content, ts],
|
||||
).map_err(|e| format!("Insert failed: {e}"))?;
|
||||
let id = conn.last_insert_rowid();
|
||||
extracted.push(MemoryFragment {
|
||||
id,
|
||||
project_id: project_id.clone(),
|
||||
content: fragment_content,
|
||||
source: "auto-extract".into(),
|
||||
trust: "auto".into(),
|
||||
confidence: 0.7,
|
||||
created_at: ts,
|
||||
ttl_days: 90,
|
||||
tags: "decision".into(),
|
||||
});
|
||||
break; // One extraction per message
|
||||
}
|
||||
}
|
||||
|
||||
// Extract errors
|
||||
for pattern in &error_patterns {
|
||||
if content.contains(pattern) {
|
||||
let fragment_content = extract_surrounding(content, pattern, 300);
|
||||
conn.execute(
|
||||
"INSERT INTO pro_memories (project_id, content, source, trust, confidence, created_at, tags)
|
||||
VALUES (?1, ?2, 'auto-extract', 'auto', 0.6, ?3, 'error')",
|
||||
params![project_id, fragment_content, ts],
|
||||
).map_err(|e| format!("Insert failed: {e}"))?;
|
||||
let id = conn.last_insert_rowid();
|
||||
extracted.push(MemoryFragment {
|
||||
id,
|
||||
project_id: project_id.clone(),
|
||||
content: fragment_content,
|
||||
source: "auto-extract".into(),
|
||||
trust: "auto".into(),
|
||||
confidence: 0.6,
|
||||
created_at: ts,
|
||||
ttl_days: 90,
|
||||
tags: "error".into(),
|
||||
});
|
||||
break;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
Ok(extracted)
|
||||
}
|
||||
|
||||
/// Extract surrounding text around a pattern match, up to max_chars.
|
||||
fn extract_surrounding(text: &str, pattern: &str, max_chars: usize) -> String {
|
||||
if let Some(pos) = text.find(pattern) {
|
||||
let start = pos.saturating_sub(50);
|
||||
let end = (pos + max_chars).min(text.len());
|
||||
// Ensure valid UTF-8 boundaries
|
||||
let start = text.floor_char_boundary(start);
|
||||
let end = text.ceil_char_boundary(end);
|
||||
text[start..end].to_string()
|
||||
} else {
|
||||
text.chars().take(max_chars).collect()
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
mod tests {
|
||||
use super::*;
|
||||
|
||||
#[test]
|
||||
fn test_memory_fragment_serializes_camel_case() {
|
||||
let f = MemoryFragment {
|
||||
id: 1,
|
||||
project_id: "proj1".into(),
|
||||
content: "We decided to use SQLite".into(),
|
||||
source: "session-abc".into(),
|
||||
trust: "agent".into(),
|
||||
confidence: 0.9,
|
||||
created_at: 1710000000,
|
||||
ttl_days: 90,
|
||||
tags: "decision,architecture".into(),
|
||||
};
|
||||
let json = serde_json::to_string(&f).unwrap();
|
||||
assert!(json.contains("projectId"));
|
||||
assert!(json.contains("createdAt"));
|
||||
assert!(json.contains("ttlDays"));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_memory_fragment_deserializes() {
|
||||
let json = r#"{"id":1,"projectId":"p","content":"test","source":"s","trust":"human","confidence":1.0,"createdAt":0,"ttlDays":30,"tags":"t"}"#;
|
||||
let f: MemoryFragment = serde_json::from_str(json).unwrap();
|
||||
assert_eq!(f.project_id, "p");
|
||||
assert_eq!(f.trust, "human");
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_extract_surrounding() {
|
||||
let text = "We chose SQLite instead of PostgreSQL for simplicity";
|
||||
let result = extract_surrounding(text, "chose ", 100);
|
||||
assert!(result.contains("chose SQLite"));
|
||||
}
|
||||
}
|
||||
194
agor-pro/src/router.rs
Normal file
194
agor-pro/src/router.rs
Normal file
|
|
@ -0,0 +1,194 @@
|
|||
// SPDX-License-Identifier: LicenseRef-Commercial
|
||||
// Smart Model Router — select optimal model based on task type and project config.
|
||||
|
||||
use rusqlite::params;
|
||||
use serde::Serialize;
|
||||
|
||||
#[derive(Debug, Serialize)]
|
||||
#[serde(rename_all = "camelCase")]
|
||||
pub struct ModelRecommendation {
|
||||
pub model: String,
|
||||
pub reason: String,
|
||||
pub estimated_cost_factor: f64,
|
||||
pub profile: String,
|
||||
}
|
||||
|
||||
#[derive(Debug, Clone, Serialize)]
|
||||
#[serde(rename_all = "camelCase")]
|
||||
pub struct RoutingProfile {
|
||||
pub name: String,
|
||||
pub description: String,
|
||||
pub rules: Vec<String>,
|
||||
}
|
||||
|
||||
fn ensure_tables(conn: &rusqlite::Connection) -> Result<(), String> {
|
||||
conn.execute_batch(
|
||||
"CREATE TABLE IF NOT EXISTS pro_router_profiles (
|
||||
project_id TEXT PRIMARY KEY,
|
||||
profile TEXT NOT NULL DEFAULT 'balanced'
|
||||
);"
|
||||
).map_err(|e| format!("Failed to create router tables: {e}"))
|
||||
}
|
||||
|
||||
fn get_profiles() -> Vec<RoutingProfile> {
|
||||
vec![
|
||||
RoutingProfile {
|
||||
name: "cost_saver".into(),
|
||||
description: "Minimize cost — use cheapest viable model".into(),
|
||||
rules: vec![
|
||||
"All roles use cheapest model".into(),
|
||||
"Only upgrade for prompts > 10000 chars".into(),
|
||||
],
|
||||
},
|
||||
RoutingProfile {
|
||||
name: "quality_first".into(),
|
||||
description: "Maximize quality — always use premium model".into(),
|
||||
rules: vec![
|
||||
"All roles use premium model".into(),
|
||||
"No downgrade regardless of prompt size".into(),
|
||||
],
|
||||
},
|
||||
RoutingProfile {
|
||||
name: "balanced".into(),
|
||||
description: "Match model to task — role and prompt size heuristic".into(),
|
||||
rules: vec![
|
||||
"Manager/Architect → premium model".into(),
|
||||
"Tester/Reviewer → mid-tier model".into(),
|
||||
"Short prompts (<2000 chars) → cheap model".into(),
|
||||
"Long prompts (>8000 chars) → premium model".into(),
|
||||
],
|
||||
},
|
||||
]
|
||||
}
|
||||
|
||||
fn select_model(profile: &str, role: &str, prompt_length: i64, provider: &str) -> (String, String, f64) {
|
||||
let (cheap, mid, premium) = match provider {
|
||||
"codex" => ("gpt-4.1-mini", "gpt-4.1", "gpt-5"),
|
||||
"ollama" => ("qwen3:8b", "qwen3:8b", "qwen3:32b"),
|
||||
_ => ("claude-haiku-4-5", "claude-sonnet-4-5", "claude-opus-4"),
|
||||
};
|
||||
|
||||
match profile {
|
||||
"cost_saver" => {
|
||||
if prompt_length > 10_000 {
|
||||
(mid.into(), "Long prompt upgrade in cost_saver profile".into(), 0.5)
|
||||
} else {
|
||||
(cheap.into(), "Cost saver: cheapest model".into(), 0.1)
|
||||
}
|
||||
}
|
||||
"quality_first" => {
|
||||
(premium.into(), "Quality first: premium model".into(), 1.0)
|
||||
}
|
||||
_ => {
|
||||
// Balanced: role + prompt heuristic
|
||||
match role {
|
||||
"manager" | "architect" => {
|
||||
(premium.into(), format!("Balanced: premium for {role} role"), 1.0)
|
||||
}
|
||||
"tester" | "reviewer" => {
|
||||
(mid.into(), format!("Balanced: mid-tier for {role} role"), 0.5)
|
||||
}
|
||||
_ => {
|
||||
if prompt_length < 2_000 {
|
||||
(cheap.into(), "Balanced: cheap for short prompt".into(), 0.1)
|
||||
} else if prompt_length > 8_000 {
|
||||
(premium.into(), "Balanced: premium for long prompt".into(), 1.0)
|
||||
} else {
|
||||
(mid.into(), "Balanced: mid-tier default".into(), 0.5)
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn pro_router_recommend(
|
||||
project_id: String,
|
||||
role: String,
|
||||
prompt_length: i64,
|
||||
provider: Option<String>,
|
||||
) -> Result<ModelRecommendation, String> {
|
||||
let conn = super::open_sessions_db()?;
|
||||
ensure_tables(&conn)?;
|
||||
|
||||
let profile = conn.prepare("SELECT profile FROM pro_router_profiles WHERE project_id = ?1")
|
||||
.map_err(|e| format!("Query failed: {e}"))?
|
||||
.query_row(params![project_id], |row| row.get::<_, String>(0))
|
||||
.unwrap_or_else(|_| "balanced".into());
|
||||
|
||||
let prov = provider.as_deref().unwrap_or("claude");
|
||||
let (model, reason, cost_factor) = select_model(&profile, &role, prompt_length, prov);
|
||||
|
||||
Ok(ModelRecommendation { model, reason, estimated_cost_factor: cost_factor, profile })
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn pro_router_set_profile(project_id: String, profile: String) -> Result<(), String> {
|
||||
let valid = ["cost_saver", "quality_first", "balanced"];
|
||||
if !valid.contains(&profile.as_str()) {
|
||||
return Err(format!("Invalid profile '{}'. Valid: {:?}", profile, valid));
|
||||
}
|
||||
let conn = super::open_sessions_db()?;
|
||||
ensure_tables(&conn)?;
|
||||
conn.execute(
|
||||
"INSERT INTO pro_router_profiles (project_id, profile) VALUES (?1, ?2)
|
||||
ON CONFLICT(project_id) DO UPDATE SET profile = ?2",
|
||||
params![project_id, profile],
|
||||
).map_err(|e| format!("Failed to set profile: {e}"))?;
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn pro_router_get_profile(project_id: String) -> Result<String, String> {
|
||||
let conn = super::open_sessions_db()?;
|
||||
ensure_tables(&conn)?;
|
||||
let profile = conn.prepare("SELECT profile FROM pro_router_profiles WHERE project_id = ?1")
|
||||
.map_err(|e| format!("Query failed: {e}"))?
|
||||
.query_row(params![project_id], |row| row.get::<_, String>(0))
|
||||
.unwrap_or_else(|_| "balanced".into());
|
||||
Ok(profile)
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn pro_router_list_profiles() -> Vec<RoutingProfile> {
|
||||
get_profiles()
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
mod tests {
|
||||
use super::*;
|
||||
|
||||
#[test]
|
||||
fn test_recommendation_serializes_camel_case() {
|
||||
let r = ModelRecommendation {
|
||||
model: "claude-sonnet-4-5".into(),
|
||||
reason: "test".into(),
|
||||
estimated_cost_factor: 0.5,
|
||||
profile: "balanced".into(),
|
||||
};
|
||||
let json = serde_json::to_string(&r).unwrap();
|
||||
assert!(json.contains("estimatedCostFactor"));
|
||||
assert!(json.contains("\"profile\":\"balanced\""));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_select_model_balanced_manager() {
|
||||
let (model, _, cost) = select_model("balanced", "manager", 5000, "claude");
|
||||
assert_eq!(model, "claude-opus-4");
|
||||
assert_eq!(cost, 1.0);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_select_model_cost_saver() {
|
||||
let (model, _, cost) = select_model("cost_saver", "worker", 1000, "claude");
|
||||
assert_eq!(model, "claude-haiku-4-5");
|
||||
assert!(cost < 0.2);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_select_model_codex_provider() {
|
||||
let (model, _, _) = select_model("quality_first", "manager", 5000, "codex");
|
||||
assert_eq!(model, "gpt-5");
|
||||
}
|
||||
}
|
||||
295
agor-pro/src/symbols.rs
Normal file
295
agor-pro/src/symbols.rs
Normal file
|
|
@ -0,0 +1,295 @@
|
|||
// SPDX-License-Identifier: LicenseRef-Commercial
|
||||
// Codebase Symbol Graph — stub implementation using regex parsing.
|
||||
// Full tree-sitter implementation deferred until tree-sitter dep is added.
|
||||
|
||||
use serde::Serialize;
|
||||
use std::collections::HashMap;
|
||||
use std::path::{Path, PathBuf};
|
||||
use std::sync::Mutex;
|
||||
|
||||
static SYMBOL_CACHE: std::sync::LazyLock<Mutex<HashMap<String, Vec<Symbol>>>> =
|
||||
std::sync::LazyLock::new(|| Mutex::new(HashMap::new()));
|
||||
|
||||
#[derive(Debug, Clone, Serialize)]
|
||||
#[serde(rename_all = "camelCase")]
|
||||
pub struct Symbol {
|
||||
pub name: String,
|
||||
pub kind: String,
|
||||
pub file_path: String,
|
||||
pub line_number: usize,
|
||||
}
|
||||
|
||||
#[derive(Debug, Serialize)]
|
||||
#[serde(rename_all = "camelCase")]
|
||||
pub struct CallerRef {
|
||||
pub file_path: String,
|
||||
pub line_number: usize,
|
||||
pub context: String,
|
||||
}
|
||||
|
||||
#[derive(Debug, Serialize)]
|
||||
#[serde(rename_all = "camelCase")]
|
||||
pub struct ScanResult {
|
||||
pub files_scanned: usize,
|
||||
pub symbols_found: usize,
|
||||
pub duration_ms: u64,
|
||||
}
|
||||
|
||||
#[derive(Debug, Serialize)]
|
||||
#[serde(rename_all = "camelCase")]
|
||||
pub struct IndexStatus {
|
||||
pub indexed: bool,
|
||||
pub symbols_count: usize,
|
||||
pub last_scan: Option<String>,
|
||||
}
|
||||
|
||||
/// Common directories to skip during scan.
|
||||
const SKIP_DIRS: &[&str] = &[
|
||||
".git", "node_modules", "target", "dist", "build", ".next",
|
||||
"__pycache__", ".venv", "venv", ".tox",
|
||||
];
|
||||
|
||||
/// Supported extensions for symbol extraction.
|
||||
const SUPPORTED_EXT: &[&str] = &["ts", "rs", "py", "js", "tsx", "jsx"];
|
||||
|
||||
fn should_skip(name: &str) -> bool {
|
||||
SKIP_DIRS.contains(&name)
|
||||
}
|
||||
|
||||
fn walk_files(dir: &Path, files: &mut Vec<PathBuf>) {
|
||||
let Ok(entries) = std::fs::read_dir(dir) else { return };
|
||||
for entry in entries.flatten() {
|
||||
let path = entry.path();
|
||||
if path.is_dir() {
|
||||
if let Some(name) = path.file_name().and_then(|n| n.to_str()) {
|
||||
if !should_skip(name) {
|
||||
walk_files(&path, files);
|
||||
}
|
||||
}
|
||||
} else if let Some(ext) = path.extension().and_then(|e| e.to_str()) {
|
||||
if SUPPORTED_EXT.contains(&ext) {
|
||||
files.push(path);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
fn extract_symbols_from_file(path: &Path) -> Vec<Symbol> {
|
||||
let Ok(content) = std::fs::read_to_string(path) else { return vec![] };
|
||||
let file_str = path.to_string_lossy().to_string();
|
||||
let ext = path.extension().and_then(|e| e.to_str()).unwrap_or("");
|
||||
let mut symbols = Vec::new();
|
||||
|
||||
for (line_idx, line) in content.lines().enumerate() {
|
||||
let trimmed = line.trim();
|
||||
match ext {
|
||||
"rs" => {
|
||||
if let Some(name) = extract_after(trimmed, "fn ") {
|
||||
symbols.push(Symbol { name, kind: "function".into(), file_path: file_str.clone(), line_number: line_idx + 1 });
|
||||
} else if let Some(name) = extract_after(trimmed, "struct ") {
|
||||
symbols.push(Symbol { name, kind: "struct".into(), file_path: file_str.clone(), line_number: line_idx + 1 });
|
||||
} else if let Some(name) = extract_after(trimmed, "enum ") {
|
||||
symbols.push(Symbol { name, kind: "type".into(), file_path: file_str.clone(), line_number: line_idx + 1 });
|
||||
} else if let Some(name) = extract_after(trimmed, "const ") {
|
||||
symbols.push(Symbol { name, kind: "const".into(), file_path: file_str.clone(), line_number: line_idx + 1 });
|
||||
} else if let Some(name) = extract_after(trimmed, "trait ") {
|
||||
symbols.push(Symbol { name, kind: "type".into(), file_path: file_str.clone(), line_number: line_idx + 1 });
|
||||
}
|
||||
}
|
||||
"ts" | "tsx" | "js" | "jsx" => {
|
||||
if let Some(name) = extract_after(trimmed, "function ") {
|
||||
symbols.push(Symbol { name, kind: "function".into(), file_path: file_str.clone(), line_number: line_idx + 1 });
|
||||
} else if let Some(name) = extract_after(trimmed, "class ") {
|
||||
symbols.push(Symbol { name, kind: "class".into(), file_path: file_str.clone(), line_number: line_idx + 1 });
|
||||
} else if let Some(name) = extract_ts_const_fn(trimmed) {
|
||||
symbols.push(Symbol { name, kind: "function".into(), file_path: file_str.clone(), line_number: line_idx + 1 });
|
||||
} else if let Some(name) = extract_after(trimmed, "interface ") {
|
||||
symbols.push(Symbol { name, kind: "type".into(), file_path: file_str.clone(), line_number: line_idx + 1 });
|
||||
} else if let Some(name) = extract_after(trimmed, "type ") {
|
||||
symbols.push(Symbol { name, kind: "type".into(), file_path: file_str.clone(), line_number: line_idx + 1 });
|
||||
}
|
||||
}
|
||||
"py" => {
|
||||
if let Some(name) = extract_after(trimmed, "def ") {
|
||||
symbols.push(Symbol { name, kind: "function".into(), file_path: file_str.clone(), line_number: line_idx + 1 });
|
||||
} else if let Some(name) = extract_after(trimmed, "class ") {
|
||||
symbols.push(Symbol { name, kind: "class".into(), file_path: file_str.clone(), line_number: line_idx + 1 });
|
||||
}
|
||||
}
|
||||
_ => {}
|
||||
}
|
||||
}
|
||||
|
||||
symbols
|
||||
}
|
||||
|
||||
/// Extract identifier after a keyword (e.g., "fn " -> function name).
|
||||
fn extract_after(line: &str, prefix: &str) -> Option<String> {
|
||||
if !line.starts_with(prefix) && !line.starts_with(&format!("pub {prefix}"))
|
||||
&& !line.starts_with(&format!("export {prefix}"))
|
||||
&& !line.starts_with(&format!("pub(crate) {prefix}"))
|
||||
&& !line.starts_with(&format!("async {prefix}"))
|
||||
&& !line.starts_with(&format!("pub async {prefix}"))
|
||||
&& !line.starts_with(&format!("export async {prefix}"))
|
||||
&& !line.starts_with(&format!("export default {prefix}"))
|
||||
{
|
||||
return None;
|
||||
}
|
||||
let after = line.find(prefix)? + prefix.len();
|
||||
let rest = &line[after..];
|
||||
let name: String = rest.chars()
|
||||
.take_while(|c| c.is_alphanumeric() || *c == '_')
|
||||
.collect();
|
||||
if name.is_empty() { None } else { Some(name) }
|
||||
}
|
||||
|
||||
/// Extract arrow function / const fn pattern: `const foo = (` or `export const foo = (`
|
||||
fn extract_ts_const_fn(line: &str) -> Option<String> {
|
||||
let stripped = line.strip_prefix("export ")
|
||||
.or(Some(line))?;
|
||||
let rest = stripped.strip_prefix("const ")?;
|
||||
let name: String = rest.chars()
|
||||
.take_while(|c| c.is_alphanumeric() || *c == '_')
|
||||
.collect();
|
||||
if name.is_empty() { return None; }
|
||||
// Check if it looks like a function assignment
|
||||
if rest.contains("= (") || rest.contains("= async (") || rest.contains("=> ") {
|
||||
Some(name)
|
||||
} else {
|
||||
None
|
||||
}
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn pro_symbols_scan(project_path: String) -> Result<ScanResult, String> {
|
||||
let start = std::time::Instant::now();
|
||||
let root = PathBuf::from(&project_path);
|
||||
if !root.is_dir() {
|
||||
return Err(format!("Not a directory: {project_path}"));
|
||||
}
|
||||
|
||||
let mut files = Vec::new();
|
||||
walk_files(&root, &mut files);
|
||||
|
||||
let mut all_symbols = Vec::new();
|
||||
for file in &files {
|
||||
all_symbols.extend(extract_symbols_from_file(file));
|
||||
}
|
||||
|
||||
let result = ScanResult {
|
||||
files_scanned: files.len(),
|
||||
symbols_found: all_symbols.len(),
|
||||
duration_ms: start.elapsed().as_millis() as u64,
|
||||
};
|
||||
|
||||
let mut cache = SYMBOL_CACHE.lock().map_err(|e| format!("Lock failed: {e}"))?;
|
||||
cache.insert(project_path, all_symbols);
|
||||
|
||||
Ok(result)
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn pro_symbols_search(project_path: String, query: String) -> Result<Vec<Symbol>, String> {
|
||||
let cache = SYMBOL_CACHE.lock().map_err(|e| format!("Lock failed: {e}"))?;
|
||||
let symbols = cache.get(&project_path).cloned().unwrap_or_default();
|
||||
let query_lower = query.to_lowercase();
|
||||
|
||||
let results: Vec<Symbol> = symbols.into_iter()
|
||||
.filter(|s| s.name.to_lowercase().contains(&query_lower))
|
||||
.take(50)
|
||||
.collect();
|
||||
|
||||
Ok(results)
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn pro_symbols_find_callers(project_path: String, symbol_name: String) -> Result<Vec<CallerRef>, String> {
|
||||
let root = PathBuf::from(&project_path);
|
||||
if !root.is_dir() {
|
||||
return Err(format!("Not a directory: {project_path}"));
|
||||
}
|
||||
|
||||
let mut files = Vec::new();
|
||||
walk_files(&root, &mut files);
|
||||
|
||||
let mut callers = Vec::new();
|
||||
for file in &files {
|
||||
let Ok(content) = std::fs::read_to_string(file) else { continue };
|
||||
for (idx, line) in content.lines().enumerate() {
|
||||
if line.contains(&symbol_name) {
|
||||
callers.push(CallerRef {
|
||||
file_path: file.to_string_lossy().to_string(),
|
||||
line_number: idx + 1,
|
||||
context: line.trim().to_string(),
|
||||
});
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Cap results
|
||||
callers.truncate(100);
|
||||
Ok(callers)
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn pro_symbols_status(project_path: String) -> Result<IndexStatus, String> {
|
||||
let cache = SYMBOL_CACHE.lock().map_err(|e| format!("Lock failed: {e}"))?;
|
||||
match cache.get(&project_path) {
|
||||
Some(symbols) => Ok(IndexStatus {
|
||||
indexed: true,
|
||||
symbols_count: symbols.len(),
|
||||
last_scan: None, // In-memory only, no timestamp tracking
|
||||
}),
|
||||
None => Ok(IndexStatus {
|
||||
indexed: false,
|
||||
symbols_count: 0,
|
||||
last_scan: None,
|
||||
}),
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
mod tests {
|
||||
use super::*;
|
||||
|
||||
#[test]
|
||||
fn test_symbol_serializes_camel_case() {
|
||||
let s = Symbol {
|
||||
name: "processEvent".into(),
|
||||
kind: "function".into(),
|
||||
file_path: "/src/lib.rs".into(),
|
||||
line_number: 42,
|
||||
};
|
||||
let json = serde_json::to_string(&s).unwrap();
|
||||
assert!(json.contains("filePath"));
|
||||
assert!(json.contains("lineNumber"));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_scan_result_serializes_camel_case() {
|
||||
let r = ScanResult {
|
||||
files_scanned: 10,
|
||||
symbols_found: 50,
|
||||
duration_ms: 123,
|
||||
};
|
||||
let json = serde_json::to_string(&r).unwrap();
|
||||
assert!(json.contains("filesScanned"));
|
||||
assert!(json.contains("symbolsFound"));
|
||||
assert!(json.contains("durationMs"));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_extract_after_rust_fn() {
|
||||
assert_eq!(extract_after("fn hello()", "fn "), Some("hello".into()));
|
||||
assert_eq!(extract_after("pub fn world()", "fn "), Some("world".into()));
|
||||
assert_eq!(extract_after("pub async fn go()", "fn "), Some("go".into()));
|
||||
assert_eq!(extract_after("let x = 5;", "fn "), None);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_extract_ts_const_fn() {
|
||||
assert_eq!(extract_ts_const_fn("const foo = (x: number) => x"), Some("foo".into()));
|
||||
assert_eq!(extract_ts_const_fn("export const bar = async ("), Some("bar".into()));
|
||||
assert_eq!(extract_ts_const_fn("const DATA = 42"), None);
|
||||
}
|
||||
}
|
||||
Loading…
Add table
Add a link
Reference in a new issue