diff --git a/apps/llm/function.py b/apps/llm/function.py index 760eb7ecb4a0fae511c70d209f98481637de17e2..96a2757241492cd5728d02bc52d5e653f6b7d603 100644 --- a/apps/llm/function.py +++ b/apps/llm/function.py @@ -6,6 +6,7 @@ import logging import re from textwrap import dedent from typing import Any +import httpx from jinja2 import BaseLoader from jinja2.sandbox import SandboxedEnvironment @@ -48,7 +49,7 @@ class FunctionLLM: self._params["timeout"] = 300 if self._config.backend == "ollama": import ollama - + if not self._config.api_key: self._client = ollama.AsyncClient(host=self._config.endpoint) else: @@ -64,11 +65,16 @@ class FunctionLLM: if not self._config.api_key: self._client = openai.AsyncOpenAI( - base_url=self._config.endpoint) + base_url=self._config.endpoint, + http_client=httpx.AsyncClient( + verify=False) # 关闭 openai 的 SSL 验证 + ) else: self._client = openai.AsyncOpenAI( base_url=self._config.endpoint, api_key=self._config.api_key, + http_client=httpx.AsyncClient( + verify=False) # 关闭 openai 的 SSL 验证 ) async def _call_openai( diff --git a/apps/llm/reasoning.py b/apps/llm/reasoning.py index 1944273bce90e5f7ede75b2eed7e18bd7e60bad1..d7c8c12ceedb187a8c98223fe1c885f238ad1107 100644 --- a/apps/llm/reasoning.py +++ b/apps/llm/reasoning.py @@ -4,7 +4,7 @@ import logging from collections.abc import AsyncGenerator from dataclasses import dataclass - +import httpx from openai import AsyncOpenAI from openai.types.chat import ChatCompletionChunk @@ -111,12 +111,14 @@ class ReasoningLLM: if not self._config.key: self._client = AsyncOpenAI( base_url=self._config.endpoint, + http_client=httpx.AsyncClient(verify=False) # 关闭 SSL 验证 ) return self._client = AsyncOpenAI( api_key=self._config.key, base_url=self._config.endpoint, + http_client=httpx.AsyncClient(verify=False) # 关闭 SSL 验证 ) @staticmethod