fix: harden Ollama autodiscovery and warning behavior (#29201)

* agents: auto-discover Ollama models without API key

* tests: cover Ollama autodiscovery warning behavior
This commit is contained in:
Vincent Koc
2026-02-27 15:22:34 -08:00
committed by GitHub
parent d17c083803
commit fa5e71d1ae
2 changed files with 133 additions and 9 deletions

View File

@@ -0,0 +1,105 @@
import { mkdtempSync } from "node:fs";
import { tmpdir } from "node:os";
import { join } from "node:path";
import { afterEach, describe, expect, it, vi } from "vitest";
import { resolveImplicitProviders } from "./models-config.providers.js";
describe("Ollama auto-discovery", () => {
let originalVitest: string | undefined;
let originalNodeEnv: string | undefined;
let originalFetch: typeof globalThis.fetch;
afterEach(() => {
if (originalVitest !== undefined) {
process.env.VITEST = originalVitest;
} else {
delete process.env.VITEST;
}
if (originalNodeEnv !== undefined) {
process.env.NODE_ENV = originalNodeEnv;
} else {
delete process.env.NODE_ENV;
}
globalThis.fetch = originalFetch;
delete process.env.OLLAMA_API_KEY;
});
function setupDiscoveryEnv() {
originalVitest = process.env.VITEST;
originalNodeEnv = process.env.NODE_ENV;
delete process.env.VITEST;
delete process.env.NODE_ENV;
originalFetch = globalThis.fetch;
}
it("auto-registers ollama provider when models are discovered locally", async () => {
setupDiscoveryEnv();
globalThis.fetch = vi.fn().mockImplementation(async (url: string | URL) => {
if (String(url).includes("/api/tags")) {
return {
ok: true,
json: async () => ({
models: [{ name: "deepseek-r1:latest" }, { name: "llama3.3:latest" }],
}),
};
}
throw new Error(`Unexpected fetch: ${url}`);
}) as typeof fetch;
const agentDir = mkdtempSync(join(tmpdir(), "openclaw-test-"));
const providers = await resolveImplicitProviders({ agentDir });
expect(providers?.ollama).toBeDefined();
expect(providers?.ollama?.apiKey).toBe("ollama-local");
expect(providers?.ollama?.api).toBe("ollama");
expect(providers?.ollama?.baseUrl).toBe("http://127.0.0.1:11434");
expect(providers?.ollama?.models).toHaveLength(2);
expect(providers?.ollama?.models?.[0]?.id).toBe("deepseek-r1:latest");
expect(providers?.ollama?.models?.[0]?.reasoning).toBe(true);
expect(providers?.ollama?.models?.[1]?.reasoning).toBe(false);
});
it("does not warn when Ollama is unreachable and not explicitly configured", async () => {
setupDiscoveryEnv();
const warnSpy = vi.spyOn(console, "warn").mockImplementation(() => {});
globalThis.fetch = vi
.fn()
.mockRejectedValue(new Error("connect ECONNREFUSED 127.0.0.1:11434")) as typeof fetch;
const agentDir = mkdtempSync(join(tmpdir(), "openclaw-test-"));
const providers = await resolveImplicitProviders({ agentDir });
expect(providers?.ollama).toBeUndefined();
const ollamaWarnings = warnSpy.mock.calls.filter(
(args) => typeof args[0] === "string" && args[0].includes("Ollama"),
);
expect(ollamaWarnings).toHaveLength(0);
warnSpy.mockRestore();
});
it("warns when Ollama is unreachable and explicitly configured", async () => {
setupDiscoveryEnv();
const warnSpy = vi.spyOn(console, "warn").mockImplementation(() => {});
globalThis.fetch = vi
.fn()
.mockRejectedValue(new Error("connect ECONNREFUSED 127.0.0.1:11434")) as typeof fetch;
const agentDir = mkdtempSync(join(tmpdir(), "openclaw-test-"));
await resolveImplicitProviders({
agentDir,
explicitProviders: {
ollama: {
baseUrl: "http://127.0.0.1:11434/v1",
api: "openai-completions",
models: [],
},
},
});
const ollamaWarnings = warnSpy.mock.calls.filter(
(args) => typeof args[0] === "string" && args[0].includes("Ollama"),
);
expect(ollamaWarnings.length).toBeGreaterThan(0);
warnSpy.mockRestore();
});
});

View File

@@ -236,7 +236,10 @@ export function resolveOllamaApiBase(configuredBaseUrl?: string): string {
return trimmed.replace(/\/v1$/i, "");
}
async function discoverOllamaModels(baseUrl?: string): Promise<ModelDefinitionConfig[]> {
async function discoverOllamaModels(
baseUrl?: string,
opts?: { quiet?: boolean },
): Promise<ModelDefinitionConfig[]> {
// Skip Ollama discovery in test environments
if (process.env.VITEST || process.env.NODE_ENV === "test") {
return [];
@@ -247,7 +250,9 @@ async function discoverOllamaModels(baseUrl?: string): Promise<ModelDefinitionCo
signal: AbortSignal.timeout(5000),
});
if (!response.ok) {
log.warn(`Failed to discover Ollama models: ${response.status}`);
if (!opts?.quiet) {
log.warn(`Failed to discover Ollama models: ${response.status}`);
}
return [];
}
const data = (await response.json()) as OllamaTagsResponse;
@@ -270,7 +275,9 @@ async function discoverOllamaModels(baseUrl?: string): Promise<ModelDefinitionCo
};
});
} catch (error) {
log.warn(`Failed to discover Ollama models: ${String(error)}`);
if (!opts?.quiet) {
log.warn(`Failed to discover Ollama models: ${String(error)}`);
}
return [];
}
}
@@ -690,8 +697,11 @@ async function buildVeniceProvider(): Promise<ProviderConfig> {
};
}
async function buildOllamaProvider(configuredBaseUrl?: string): Promise<ProviderConfig> {
const models = await discoverOllamaModels(configuredBaseUrl);
async function buildOllamaProvider(
configuredBaseUrl?: string,
opts?: { quiet?: boolean },
): Promise<ProviderConfig> {
const models = await discoverOllamaModels(configuredBaseUrl, opts);
return {
baseUrl: resolveOllamaApiBase(configuredBaseUrl),
api: "ollama",
@@ -959,15 +969,24 @@ export async function resolveImplicitProviders(params: {
break;
}
// Ollama provider - only add if explicitly configured.
// Ollama provider - auto-discover if running locally, or add if explicitly configured.
// Use the user's configured baseUrl (from explicit providers) for model
// discovery so that remote / non-default Ollama instances are reachable.
const ollamaKey =
resolveEnvApiKeyVarName("ollama") ??
resolveApiKeyFromProfiles({ provider: "ollama", store: authStore });
if (ollamaKey) {
const ollamaBaseUrl = params.explicitProviders?.ollama?.baseUrl;
providers.ollama = { ...(await buildOllamaProvider(ollamaBaseUrl)), apiKey: ollamaKey };
const ollamaBaseUrl = params.explicitProviders?.ollama?.baseUrl;
const hasExplicitOllamaConfig = Boolean(params.explicitProviders?.ollama);
// Only suppress warnings for implicit local probing when user has not
// explicitly configured Ollama.
const ollamaProvider = await buildOllamaProvider(ollamaBaseUrl, {
quiet: !ollamaKey && !hasExplicitOllamaConfig,
});
if (ollamaProvider.models.length > 0 || ollamaKey) {
providers.ollama = {
...ollamaProvider,
apiKey: ollamaKey ?? "ollama-local",
};
}
// vLLM provider - OpenAI-compatible local server (opt-in via env/profile).