|
| 1 | +""" |
| 2 | +Unified CUA (Computer Use Agent) template. |
| 3 | +
|
| 4 | +Supports Anthropic, OpenAI, and Gemini providers with automatic fallback. |
| 5 | +Configure via environment variables: |
| 6 | + CUA_PROVIDER - Primary provider: anthropic, openai, or gemini (default: anthropic) |
| 7 | + CUA_FALLBACK_PROVIDERS - Comma-separated fallback order (e.g. "openai,gemini") |
| 8 | + ANTHROPIC_API_KEY - Required if using Anthropic |
| 9 | + OPENAI_API_KEY - Required if using OpenAI |
| 10 | + GOOGLE_API_KEY - Required if using Gemini |
| 11 | +""" |
| 12 | + |
| 13 | +import os |
| 14 | +from typing import NotRequired, Optional, TypedDict |
| 15 | + |
| 16 | +import kernel |
| 17 | +from session import KernelBrowserSession |
| 18 | +from tools import KernelExecutor |
| 19 | +from providers import run_with_fallback, ProviderConfig |
| 20 | + |
| 21 | +# Parse provider configuration |
| 22 | +PRIMARY_PROVIDER = os.getenv("CUA_PROVIDER", "anthropic") |
| 23 | +FALLBACK_PROVIDERS = [ |
| 24 | + p.strip() for p in os.getenv("CUA_FALLBACK_PROVIDERS", "").split(",") if p.strip() |
| 25 | +] |
| 26 | +PROVIDER_CHAIN = [PRIMARY_PROVIDER] + FALLBACK_PROVIDERS |
| 27 | + |
| 28 | +API_KEY_MAP = { |
| 29 | + "anthropic": "ANTHROPIC_API_KEY", |
| 30 | + "openai": "OPENAI_API_KEY", |
| 31 | + "gemini": "GOOGLE_API_KEY", |
| 32 | +} |
| 33 | + |
| 34 | +configured = [p for p in PROVIDER_CHAIN if os.getenv(API_KEY_MAP.get(p, ""))] |
| 35 | +if not configured: |
| 36 | + keys = [API_KEY_MAP.get(p, p) for p in PROVIDER_CHAIN] |
| 37 | + raise ValueError( |
| 38 | + f"No API keys found for configured providers {PROVIDER_CHAIN}. " |
| 39 | + f"Set at least one of: {', '.join(keys)}" |
| 40 | + ) |
| 41 | + |
| 42 | + |
| 43 | +class CuaInput(TypedDict): |
| 44 | + query: str |
| 45 | + provider: NotRequired[str] |
| 46 | + model: NotRequired[str] |
| 47 | + record_replay: NotRequired[bool] |
| 48 | + |
| 49 | + |
| 50 | +class CuaOutput(TypedDict): |
| 51 | + result: str |
| 52 | + provider: str |
| 53 | + replay_url: Optional[str] |
| 54 | + |
| 55 | + |
| 56 | +app = kernel.App("python-cua") |
| 57 | + |
| 58 | + |
| 59 | +@app.action("cua-task") |
| 60 | +async def cua_task( |
| 61 | + ctx: kernel.KernelContext, |
| 62 | + payload: CuaInput, |
| 63 | +) -> CuaOutput: |
| 64 | + if not payload or not payload.get("query"): |
| 65 | + raise ValueError("Query is required") |
| 66 | + |
| 67 | + # Allow per-request provider override |
| 68 | + if payload.get("provider"): |
| 69 | + provider_chain = [payload["provider"]] + [p for p in PROVIDER_CHAIN if p != payload["provider"]] |
| 70 | + else: |
| 71 | + provider_chain = PROVIDER_CHAIN |
| 72 | + |
| 73 | + record_replay = payload.get("record_replay", False) |
| 74 | + |
| 75 | + async with KernelBrowserSession( |
| 76 | + invocation_id=ctx.invocation_id, |
| 77 | + stealth=True, |
| 78 | + record_replay=record_replay, |
| 79 | + ) as session: |
| 80 | + print("Kernel browser live view url:", session.live_view_url) |
| 81 | + |
| 82 | + executor = KernelExecutor(session.kernel, session.session_id) |
| 83 | + |
| 84 | + result = await run_with_fallback( |
| 85 | + provider_chain, |
| 86 | + ProviderConfig( |
| 87 | + query=payload["query"], |
| 88 | + model=payload.get("model"), |
| 89 | + viewport_width=session.viewport_width, |
| 90 | + viewport_height=session.viewport_height, |
| 91 | + ), |
| 92 | + executor, |
| 93 | + ) |
| 94 | + |
| 95 | + return { |
| 96 | + "result": result.result, |
| 97 | + "provider": result.provider, |
| 98 | + "replay_url": session.replay_view_url, |
| 99 | + } |
0 commit comments