feat: ShouldPreserveThinkingSuffix (#2189)

This commit is contained in:
Seefs
2025-11-07 17:43:33 +08:00
committed by GitHub
parent 43ee7a98b4
commit e082268533
14 changed files with 138 additions and 19 deletions

View File

@@ -189,7 +189,9 @@ func RequestOpenAI2ClaudeMessage(c *gin.Context, textRequest dto.GeneralOpenAIRe
// https://docs.anthropic.com/en/docs/build-with-claude/extended-thinking#important-considerations-when-using-extended-thinking // https://docs.anthropic.com/en/docs/build-with-claude/extended-thinking#important-considerations-when-using-extended-thinking
claudeRequest.TopP = 0 claudeRequest.TopP = 0
claudeRequest.Temperature = common.GetPointer[float64](1.0) claudeRequest.Temperature = common.GetPointer[float64](1.0)
claudeRequest.Model = strings.TrimSuffix(textRequest.Model, "-thinking") if !model_setting.ShouldPreserveThinkingSuffix(textRequest.Model) {
claudeRequest.Model = strings.TrimSuffix(textRequest.Model, "-thinking")
}
} }
if textRequest.ReasoningEffort != "" { if textRequest.ReasoningEffort != "" {

View File

@@ -127,7 +127,8 @@ func (a *Adaptor) Init(info *relaycommon.RelayInfo) {
func (a *Adaptor) GetRequestURL(info *relaycommon.RelayInfo) (string, error) { func (a *Adaptor) GetRequestURL(info *relaycommon.RelayInfo) (string, error) {
if model_setting.GetGeminiSettings().ThinkingAdapterEnabled { if model_setting.GetGeminiSettings().ThinkingAdapterEnabled &&
!model_setting.ShouldPreserveThinkingSuffix(info.OriginModelName) {
// 新增逻辑:处理 -thinking-<budget> 格式 // 新增逻辑:处理 -thinking-<budget> 格式
if strings.Contains(info.UpstreamModelName, "-thinking-") { if strings.Contains(info.UpstreamModelName, "-thinking-") {
parts := strings.Split(info.UpstreamModelName, "-thinking-") parts := strings.Split(info.UpstreamModelName, "-thinking-")

View File

@@ -27,6 +27,7 @@ import (
"github.com/QuantumNous/new-api/relay/common_handler" "github.com/QuantumNous/new-api/relay/common_handler"
relayconstant "github.com/QuantumNous/new-api/relay/constant" relayconstant "github.com/QuantumNous/new-api/relay/constant"
"github.com/QuantumNous/new-api/service" "github.com/QuantumNous/new-api/service"
"github.com/QuantumNous/new-api/setting/model_setting"
"github.com/QuantumNous/new-api/types" "github.com/QuantumNous/new-api/types"
"github.com/gin-gonic/gin" "github.com/gin-gonic/gin"
@@ -224,7 +225,8 @@ func (a *Adaptor) ConvertOpenAIRequest(c *gin.Context, info *relaycommon.RelayIn
request.Usage = json.RawMessage(`{"include":true}`) request.Usage = json.RawMessage(`{"include":true}`)
} }
// 适配 OpenRouter 的 thinking 后缀 // 适配 OpenRouter 的 thinking 后缀
if strings.HasSuffix(info.UpstreamModelName, "-thinking") { if !model_setting.ShouldPreserveThinkingSuffix(info.OriginModelName) &&
strings.HasSuffix(info.UpstreamModelName, "-thinking") {
info.UpstreamModelName = strings.TrimSuffix(info.UpstreamModelName, "-thinking") info.UpstreamModelName = strings.TrimSuffix(info.UpstreamModelName, "-thinking")
request.Model = info.UpstreamModelName request.Model = info.UpstreamModelName
if len(request.Reasoning) == 0 { if len(request.Reasoning) == 0 {

View File

@@ -168,7 +168,8 @@ func (a *Adaptor) getRequestUrl(info *relaycommon.RelayInfo, modelName, suffix s
func (a *Adaptor) GetRequestURL(info *relaycommon.RelayInfo) (string, error) { func (a *Adaptor) GetRequestURL(info *relaycommon.RelayInfo) (string, error) {
suffix := "" suffix := ""
if a.RequestMode == RequestModeGemini { if a.RequestMode == RequestModeGemini {
if model_setting.GetGeminiSettings().ThinkingAdapterEnabled { if model_setting.GetGeminiSettings().ThinkingAdapterEnabled &&
!model_setting.ShouldPreserveThinkingSuffix(info.OriginModelName) {
// 新增逻辑:处理 -thinking-<budget> 格式 // 新增逻辑:处理 -thinking-<budget> 格式
if strings.Contains(info.UpstreamModelName, "-thinking-") { if strings.Contains(info.UpstreamModelName, "-thinking-") {
parts := strings.Split(info.UpstreamModelName, "-thinking-") parts := strings.Split(info.UpstreamModelName, "-thinking-")

View File

@@ -16,6 +16,7 @@ import (
"github.com/QuantumNous/new-api/relay/channel/openai" "github.com/QuantumNous/new-api/relay/channel/openai"
relaycommon "github.com/QuantumNous/new-api/relay/common" relaycommon "github.com/QuantumNous/new-api/relay/common"
"github.com/QuantumNous/new-api/relay/constant" "github.com/QuantumNous/new-api/relay/constant"
"github.com/QuantumNous/new-api/setting/model_setting"
"github.com/QuantumNous/new-api/types" "github.com/QuantumNous/new-api/types"
"github.com/gin-gonic/gin" "github.com/gin-gonic/gin"
@@ -291,7 +292,9 @@ func (a *Adaptor) ConvertOpenAIRequest(c *gin.Context, info *relaycommon.RelayIn
return nil, errors.New("request is nil") return nil, errors.New("request is nil")
} }
if strings.HasSuffix(info.UpstreamModelName, "-thinking") && strings.HasPrefix(info.UpstreamModelName, "deepseek") { if !model_setting.ShouldPreserveThinkingSuffix(info.OriginModelName) &&
strings.HasSuffix(info.UpstreamModelName, "-thinking") &&
strings.HasPrefix(info.UpstreamModelName, "deepseek") {
info.UpstreamModelName = strings.TrimSuffix(info.UpstreamModelName, "-thinking") info.UpstreamModelName = strings.TrimSuffix(info.UpstreamModelName, "-thinking")
request.Model = info.UpstreamModelName request.Model = info.UpstreamModelName
request.THINKING = json.RawMessage(`{"type": "enabled"}`) request.THINKING = json.RawMessage(`{"type": "enabled"}`)

View File

@@ -67,7 +67,9 @@ func ClaudeHelper(c *gin.Context, info *relaycommon.RelayInfo) (newAPIError *typ
request.TopP = 0 request.TopP = 0
request.Temperature = common.GetPointer[float64](1.0) request.Temperature = common.GetPointer[float64](1.0)
} }
request.Model = strings.TrimSuffix(request.Model, "-thinking") if !model_setting.ShouldPreserveThinkingSuffix(info.OriginModelName) {
request.Model = strings.TrimSuffix(request.Model, "-thinking")
}
info.UpstreamModelName = request.Model info.UpstreamModelName = request.Model
} }

View File

@@ -1,16 +1,23 @@
package model_setting package model_setting
import ( import (
"strings"
"github.com/QuantumNous/new-api/setting/config" "github.com/QuantumNous/new-api/setting/config"
) )
type GlobalSettings struct { type GlobalSettings struct {
PassThroughRequestEnabled bool `json:"pass_through_request_enabled"` PassThroughRequestEnabled bool `json:"pass_through_request_enabled"`
ThinkingModelBlacklist []string `json:"thinking_model_blacklist"`
} }
// 默认配置 // 默认配置
var defaultOpenaiSettings = GlobalSettings{ var defaultOpenaiSettings = GlobalSettings{
PassThroughRequestEnabled: false, PassThroughRequestEnabled: false,
ThinkingModelBlacklist: []string{
"moonshotai/kimi-k2-thinking",
"kimi-k2-thinking",
},
} }
// 全局实例 // 全局实例
@@ -24,3 +31,18 @@ func init() {
func GetGlobalSettings() *GlobalSettings { func GetGlobalSettings() *GlobalSettings {
return &globalSettings return &globalSettings
} }
// ShouldPreserveThinkingSuffix 判断模型是否配置为保留 thinking/-nothinking 后缀
func ShouldPreserveThinkingSuffix(modelName string) bool {
target := strings.TrimSpace(modelName)
if target == "" {
return false
}
for _, entry := range globalSettings.ThinkingModelBlacklist {
if strings.TrimSpace(entry) == target {
return true
}
}
return false
}

View File

@@ -37,6 +37,7 @@ const ModelSetting = () => {
'claude.default_max_tokens': '', 'claude.default_max_tokens': '',
'claude.thinking_adapter_budget_tokens_percentage': 0.8, 'claude.thinking_adapter_budget_tokens_percentage': 0.8,
'global.pass_through_request_enabled': false, 'global.pass_through_request_enabled': false,
'global.thinking_model_blacklist': '[]',
'general_setting.ping_interval_enabled': false, 'general_setting.ping_interval_enabled': false,
'general_setting.ping_interval_seconds': 60, 'general_setting.ping_interval_seconds': 60,
'gemini.thinking_adapter_enabled': false, 'gemini.thinking_adapter_enabled': false,
@@ -56,7 +57,8 @@ const ModelSetting = () => {
item.key === 'gemini.version_settings' || item.key === 'gemini.version_settings' ||
item.key === 'claude.model_headers_settings' || item.key === 'claude.model_headers_settings' ||
item.key === 'claude.default_max_tokens' || item.key === 'claude.default_max_tokens' ||
item.key === 'gemini.supported_imagine_models' item.key === 'gemini.supported_imagine_models' ||
item.key === 'global.thinking_model_blacklist'
) { ) {
if (item.value !== '') { if (item.value !== '') {
item.value = JSON.stringify(JSON.parse(item.value), null, 2); item.value = JSON.stringify(JSON.parse(item.value), null, 2);

View File

@@ -561,6 +561,9 @@
"启用绘图功能": "Enable drawing function", "启用绘图功能": "Enable drawing function",
"启用请求体透传功能": "Enable request body pass-through functionality", "启用请求体透传功能": "Enable request body pass-through functionality",
"启用请求透传": "Enable request pass-through", "启用请求透传": "Enable request pass-through",
"禁用思考处理的模型列表": "Models skipping thinking handling",
"列出的模型将不会自动添加或移除-thinking/-nothinking 后缀": "Models in this list will not automatically add or remove the -thinking/-nothinking suffix.",
"请输入JSON数组如 [\"model-a\",\"model-b\"]": "Enter a JSON array, e.g. [\"model-a\",\"model-b\"]",
"启用额度消费日志记录": "Enable quota consumption logging", "启用额度消费日志记录": "Enable quota consumption logging",
"启用验证": "Enable Authentication", "启用验证": "Enable Authentication",
"周": "week", "周": "week",

View File

@@ -564,6 +564,9 @@
"启用绘图功能": "Activer la fonction de dessin", "启用绘图功能": "Activer la fonction de dessin",
"启用请求体透传功能": "Activer la fonctionnalité de transmission du corps de la requête", "启用请求体透传功能": "Activer la fonctionnalité de transmission du corps de la requête",
"启用请求透传": "Activer la transmission de la requête", "启用请求透传": "Activer la transmission de la requête",
"禁用思考处理的模型列表": "Liste noire des modèles pour le traitement thinking",
"列出的模型将不会自动添加或移除-thinking/-nothinking 后缀": "Les modèles listés ici n'ajouteront ni ne retireront automatiquement le suffixe -thinking/-nothinking.",
"请输入JSON数组如 [\"model-a\",\"model-b\"]": "Saisissez un tableau JSON, par ex. [\"model-a\",\"model-b\"]",
"启用额度消费日志记录": "Activer la journalisation de la consommation de quota", "启用额度消费日志记录": "Activer la journalisation de la consommation de quota",
"启用验证": "Activer l'authentification", "启用验证": "Activer l'authentification",
"周": "semaine", "周": "semaine",

View File

@@ -561,6 +561,9 @@
"启用绘图功能": "画像生成機能を有効にする", "启用绘图功能": "画像生成機能を有効にする",
"启用请求体透传功能": "リクエストボディのパススルー機能を有効にします。", "启用请求体透传功能": "リクエストボディのパススルー機能を有効にします。",
"启用请求透传": "リクエストパススルーを有効にする", "启用请求透传": "リクエストパススルーを有効にする",
"禁用思考处理的模型列表": "Thinking処理を無効化するモデル一覧",
"列出的模型将不会自动添加或移除-thinking/-nothinking 后缀": "ここに含まれるモデルでは-thinking/-nothinkingサフィックスを自動的に追加・削除しません。",
"请输入JSON数组如 [\"model-a\",\"model-b\"]": "JSON配列を入力してください[\"model-a\",\"model-b\"]",
"启用额度消费日志记录": "クォータ消費のログ記録を有効にする", "启用额度消费日志记录": "クォータ消費のログ記録を有効にする",
"启用验证": "認証を有効にする", "启用验证": "認証を有効にする",
"周": "週", "周": "週",

View File

@@ -567,6 +567,9 @@
"启用绘图功能": "Включить функцию рисования", "启用绘图功能": "Включить функцию рисования",
"启用请求体透传功能": "Включить функцию прозрачной передачи тела запроса", "启用请求体透传功能": "Включить функцию прозрачной передачи тела запроса",
"启用请求透传": "Включить прозрачную передачу запросов", "启用请求透传": "Включить прозрачную передачу запросов",
"禁用思考处理的模型列表": "Список моделей без обработки thinking",
"列出的模型将不会自动添加或移除-thinking/-nothinking 后缀": "Для этих моделей суффиксы -thinking/-nothinking не будут добавляться или удаляться автоматически.",
"请输入JSON数组如 [\"model-a\",\"model-b\"]": "Введите JSON-массив, например [\"model-a\",\"model-b\"]",
"启用额度消费日志记录": "Включить журналирование потребления квоты", "启用额度消费日志记录": "Включить журналирование потребления квоты",
"启用验证": "Включить проверку", "启用验证": "Включить проверку",
"周": "Неделя", "周": "Неделя",

View File

@@ -558,6 +558,9 @@
"启用绘图功能": "启用绘图功能", "启用绘图功能": "启用绘图功能",
"启用请求体透传功能": "启用请求体透传功能", "启用请求体透传功能": "启用请求体透传功能",
"启用请求透传": "启用请求透传", "启用请求透传": "启用请求透传",
"禁用思考处理的模型列表": "禁用思考处理的模型列表",
"列出的模型将不会自动添加或移除-thinking/-nothinking 后缀": "列出的模型将不会自动添加或移除-thinking/-nothinking 后缀",
"请输入JSON数组如 [\"model-a\",\"model-b\"]": "请输入JSON数组如 [\"model-a\",\"model-b\"]",
"启用额度消费日志记录": "启用额度消费日志记录", "启用额度消费日志记录": "启用额度消费日志记录",
"启用验证": "启用验证", "启用验证": "启用验证",
"周": "周", "周": "周",

View File

@@ -29,23 +29,44 @@ import {
} from '../../../helpers'; } from '../../../helpers';
import { useTranslation } from 'react-i18next'; import { useTranslation } from 'react-i18next';
const thinkingExample = JSON.stringify(
['moonshotai/kimi-k2-thinking', 'kimi-k2-thinking'],
null,
2,
);
const defaultGlobalSettingInputs = {
'global.pass_through_request_enabled': false,
'global.thinking_model_blacklist': '[]',
'general_setting.ping_interval_enabled': false,
'general_setting.ping_interval_seconds': 60,
};
export default function SettingGlobalModel(props) { export default function SettingGlobalModel(props) {
const { t } = useTranslation(); const { t } = useTranslation();
const [loading, setLoading] = useState(false); const [loading, setLoading] = useState(false);
const [inputs, setInputs] = useState({ const [inputs, setInputs] = useState(defaultGlobalSettingInputs);
'global.pass_through_request_enabled': false,
'general_setting.ping_interval_enabled': false,
'general_setting.ping_interval_seconds': 60,
});
const refForm = useRef(); const refForm = useRef();
const [inputsRow, setInputsRow] = useState(inputs); const [inputsRow, setInputsRow] = useState(defaultGlobalSettingInputs);
const normalizeValueBeforeSave = (key, value) => {
if (key === 'global.thinking_model_blacklist') {
const text = typeof value === 'string' ? value.trim() : '';
return text === '' ? '[]' : value;
}
return value;
};
function onSubmit() { function onSubmit() {
const updateArray = compareObjects(inputs, inputsRow); const updateArray = compareObjects(inputs, inputsRow);
if (!updateArray.length) return showWarning(t('你似乎并没有修改什么')); if (!updateArray.length) return showWarning(t('你似乎并没有修改什么'));
const requestQueue = updateArray.map((item) => { const requestQueue = updateArray.map((item) => {
let value = String(inputs[item.key]); const normalizedValue = normalizeValueBeforeSave(
item.key,
inputs[item.key],
);
let value = String(normalizedValue);
return API.put('/api/option/', { return API.put('/api/option/', {
key: item.key, key: item.key,
@@ -74,14 +95,30 @@ export default function SettingGlobalModel(props) {
useEffect(() => { useEffect(() => {
const currentInputs = {}; const currentInputs = {};
for (let key in props.options) { for (const key of Object.keys(defaultGlobalSettingInputs)) {
if (Object.keys(inputs).includes(key)) { if (props.options[key] !== undefined) {
currentInputs[key] = props.options[key]; let value = props.options[key];
if (key === 'global.thinking_model_blacklist') {
try {
value =
value && String(value).trim() !== ''
? JSON.stringify(JSON.parse(value), null, 2)
: defaultGlobalSettingInputs[key];
} catch (error) {
value = defaultGlobalSettingInputs[key];
}
}
currentInputs[key] = value;
} else {
currentInputs[key] = defaultGlobalSettingInputs[key];
} }
} }
setInputs(currentInputs); setInputs(currentInputs);
setInputsRow(structuredClone(currentInputs)); setInputsRow(structuredClone(currentInputs));
refForm.current.setValues(currentInputs); if (refForm.current) {
refForm.current.setValues(currentInputs);
}
}, [props.options]); }, [props.options]);
return ( return (
@@ -110,6 +147,38 @@ export default function SettingGlobalModel(props) {
/> />
</Col> </Col>
</Row> </Row>
<Row>
<Col span={24}>
<Form.TextArea
label={t('禁用思考处理的模型列表')}
field={'global.thinking_model_blacklist'}
placeholder={
t('例如:') +
'\n' +
thinkingExample
}
rows={4}
rules={[
{
validator: (rule, value) => {
if (!value || value.trim() === '') return true;
return verifyJSON(value);
},
message: t('不是合法的 JSON 字符串'),
},
]}
extraText={t(
'列出的模型将不会自动添加或移除-thinking/-nothinking 后缀',
)}
onChange={(value) =>
setInputs({
...inputs,
'global.thinking_model_blacklist': value,
})
}
/>
</Col>
</Row>
<Form.Section text={t('连接保活设置')}> <Form.Section text={t('连接保活设置')}>
<Row style={{ marginTop: 10 }}> <Row style={{ marginTop: 10 }}>