refactor(core): add LLMClient facade trait; decouple TUI from Provider/MCP details
This commit is contained in:
@@ -7,7 +7,10 @@ use crate::consent::{ConsentManager, ConsentScope};
|
||||
use crate::tools::{Tool, WebScrapeTool, WebSearchTool};
|
||||
use crate::types::ModelInfo;
|
||||
use crate::types::{ChatResponse, Message, Role};
|
||||
use crate::{Error, LlmProvider, Result, mode::Mode, send_via_stream};
|
||||
use crate::{
|
||||
ChatStream, Error, LlmProvider, Result, facade::llm_client::LlmClient, mode::Mode,
|
||||
send_via_stream,
|
||||
};
|
||||
use anyhow::anyhow;
|
||||
use futures::{StreamExt, future::BoxFuture, stream};
|
||||
use reqwest::Client as HttpClient;
|
||||
@@ -564,3 +567,27 @@ impl LlmProvider for RemoteMcpClient {
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
#[async_trait::async_trait]
|
||||
impl LlmClient for RemoteMcpClient {
|
||||
async fn list_models(&self) -> Result<Vec<ModelInfo>> {
|
||||
<Self as LlmProvider>::list_models(self).await
|
||||
}
|
||||
|
||||
async fn send_chat(&self, request: crate::types::ChatRequest) -> Result<ChatResponse> {
|
||||
<Self as LlmProvider>::send_prompt(self, request).await
|
||||
}
|
||||
|
||||
async fn stream_chat(&self, request: crate::types::ChatRequest) -> Result<ChatStream> {
|
||||
let stream = <Self as LlmProvider>::stream_prompt(self, request).await?;
|
||||
Ok(Box::pin(stream))
|
||||
}
|
||||
|
||||
async fn list_tools(&self) -> Result<Vec<McpToolDescriptor>> {
|
||||
<Self as McpClient>::list_tools(self).await
|
||||
}
|
||||
|
||||
async fn call_tool(&self, call: McpToolCall) -> Result<McpToolResponse> {
|
||||
<Self as McpClient>::call_tool(self, call).await
|
||||
}
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user