mirror of
https://github.com/openclaw/openclaw.git
synced 2026-05-11 12:41:42 +00:00
feat(memory): add Ollama embedding provider (#26349)
Merged via /review-pr -> /prepare-pr -> /merge-pr.
Prepared head SHA: ac41386543
Co-authored-by: nico-hoff <43175972+nico-hoff@users.noreply.github.com>
Co-authored-by: gumadeiras <5599352+gumadeiras@users.noreply.github.com>
Reviewed-by: @gumadeiras
This commit is contained in:
@@ -6,7 +6,7 @@ const asConfig = (cfg: OpenClawConfig): OpenClawConfig => cfg;
|
||||
|
||||
describe("memory search config", () => {
|
||||
function configWithDefaultProvider(
|
||||
provider: "openai" | "local" | "gemini" | "mistral",
|
||||
provider: "openai" | "local" | "gemini" | "mistral" | "ollama",
|
||||
): OpenClawConfig {
|
||||
return asConfig({
|
||||
agents: {
|
||||
@@ -156,6 +156,13 @@ describe("memory search config", () => {
|
||||
expect(resolved?.model).toBe("mistral-embed");
|
||||
});
|
||||
|
||||
it("includes remote defaults and model default for ollama without overrides", () => {
|
||||
const cfg = configWithDefaultProvider("ollama");
|
||||
const resolved = resolveMemorySearchConfig(cfg, "main");
|
||||
expectDefaultRemoteBatch(resolved);
|
||||
expect(resolved?.model).toBe("nomic-embed-text");
|
||||
});
|
||||
|
||||
it("defaults session delta thresholds", () => {
|
||||
const cfg = asConfig({
|
||||
agents: {
|
||||
|
||||
Reference in New Issue
Block a user