AxAIServiceOptions
type AxAIServiceOptions = object;
Defined in: https://github.com/ax-llm/ax/blob/71ea5064d766efdc031d375243a8e525911833e7/src/ax/ai/types.ts#L710
Runtime options for AI service requests.
These options control how requests are made to the AI service, including debugging, rate limiting, streaming, function calling, and extended thinking.
Example
const options: AxAIServiceOptions = {
stream: true,
thinkingTokenBudget: 'medium',
debug: true
};
await gen.forward(ai, values, options);
Properties
abortSignal?
optional abortSignal: AbortSignal;
Defined in: https://github.com/ax-llm/ax/blob/71ea5064d766efdc031d375243a8e525911833e7/src/ax/ai/types.ts#L750
AbortSignal for cancelling in-flight requests.
contextCache?
optional contextCache: AxContextCacheOptions;
Defined in: https://github.com/ax-llm/ax/blob/71ea5064d766efdc031d375243a8e525911833e7/src/ax/ai/types.ts#L866
Context caching options for large prompt prefixes.
When enabled, large prompt prefixes can be cached for cost savings and lower latency on subsequent requests.
Currently supported by: Google Gemini/Vertex AI
corsProxy?
optional corsProxy: string;
Defined in: https://github.com/ax-llm/ax/blob/71ea5064d766efdc031d375243a8e525911833e7/src/ax/ai/types.ts#L848
CORS proxy URL for browser environments.
When running in a browser, API calls may be blocked by CORS. Specify a proxy URL to route requests through.
Example
'https://cors-anywhere.herokuapp.com/'
customLabels?
optional customLabels: Record<string, string>;
Defined in: https://github.com/ax-llm/ax/blob/71ea5064d766efdc031d375243a8e525911833e7/src/ax/ai/types.ts#L886
Custom labels for OpenTelemetry metrics.
These labels are merged with axGlobals.customLabels (service-level
options override global settings).
Example
{ environment: 'production', feature: 'search' }
debug?
optional debug: boolean;
Defined in: https://github.com/ax-llm/ax/blob/71ea5064d766efdc031d375243a8e525911833e7/src/ax/ai/types.ts#L717
Enable debug logging for troubleshooting.
When true, logs detailed information about prompts, responses, and the generation pipeline. Useful for understanding AI behavior.
debugHideSystemPrompt?
optional debugHideSystemPrompt: boolean;
Defined in: https://github.com/ax-llm/ax/blob/71ea5064d766efdc031d375243a8e525911833e7/src/ax/ai/types.ts#L759
Hide system prompt in debug output (for cleaner logs).
examplesInSystem?
optional examplesInSystem: boolean;
Defined in: https://github.com/ax-llm/ax/blob/71ea5064d766efdc031d375243a8e525911833e7/src/ax/ai/types.ts#L876
Render examples/demos in the system prompt instead of as message pairs.
false(default) - Examples rendered as alternating user/assistant messagestrue- Examples embedded in system prompt (legacy behavior)
Message pair rendering generally produces better results.
excludeContentFromTrace?
optional excludeContentFromTrace: boolean;
Defined in: https://github.com/ax-llm/ax/blob/71ea5064d766efdc031d375243a8e525911833e7/src/ax/ai/types.ts#L747
Exclude message content from OpenTelemetry traces (for privacy).
fetch?
optional fetch: typeof fetch;
Defined in: https://github.com/ax-llm/ax/blob/71ea5064d766efdc031d375243a8e525911833e7/src/ax/ai/types.ts#L731
Custom fetch implementation (useful for proxies or custom HTTP handling).
functionCallMode?
optional functionCallMode: "auto" | "native" | "prompt";
Defined in: https://github.com/ax-llm/ax/blob/71ea5064d766efdc031d375243a8e525911833e7/src/ax/ai/types.ts#L783
How to handle function/tool calling.
'auto'- Let the provider decide the best approach (default)'native'- Use the provider’s native function calling API. Fails if the model doesn’t support it.'prompt'- Simulate function calling via prompt engineering. Works with any model but may be less reliable.
Default
'auto'
logger?
optional logger: AxLoggerFunction;
Defined in: https://github.com/ax-llm/ax/blob/71ea5064d766efdc031d375243a8e525911833e7/src/ax/ai/types.ts#L753
Custom logger function for debug output.
meter?
optional meter: Meter;
Defined in: https://github.com/ax-llm/ax/blob/71ea5064d766efdc031d375243a8e525911833e7/src/ax/ai/types.ts#L737
OpenTelemetry meter for metrics collection.
rateLimiter?
optional rateLimiter: AxRateLimiterFunction;
Defined in: https://github.com/ax-llm/ax/blob/71ea5064d766efdc031d375243a8e525911833e7/src/ax/ai/types.ts#L728
Custom rate limiter function to control request throughput.
retry?
optional retry: Partial<RetryConfig>;
Defined in: https://github.com/ax-llm/ax/blob/71ea5064d766efdc031d375243a8e525911833e7/src/ax/ai/types.ts#L856
Retry configuration for failed requests.
Controls automatic retry behavior for transient errors (rate limits, timeouts, server errors).
sessionId?
optional sessionId: string;
Defined in: https://github.com/ax-llm/ax/blob/71ea5064d766efdc031d375243a8e525911833e7/src/ax/ai/types.ts#L756
Session identifier for conversation tracking and memory isolation.
showThoughts?
optional showThoughts: boolean;
Defined in: https://github.com/ax-llm/ax/blob/71ea5064d766efdc031d375243a8e525911833e7/src/ax/ai/types.ts#L827
Include the model’s thinking/reasoning in the output.
When true and thinkingTokenBudget is set, the model’s internal reasoning
is included in the response. Useful for debugging and understanding AI behavior.
Default
false
stepIndex?
optional stepIndex: number;
Defined in: https://github.com/ax-llm/ax/blob/71ea5064d766efdc031d375243a8e525911833e7/src/ax/ai/types.ts#L838
Internal: Current step index for multi-step operations.
stream?
optional stream: boolean;
Defined in: https://github.com/ax-llm/ax/blob/71ea5064d766efdc031d375243a8e525911833e7/src/ax/ai/types.ts#L770
Enable streaming responses.
When true, the AI returns responses as a stream of chunks, enabling real-time display of generated text.
thinkingTokenBudget?
optional thinkingTokenBudget: "minimal" | "low" | "medium" | "high" | "highest" | "none";
Defined in: https://github.com/ax-llm/ax/blob/71ea5064d766efdc031d375243a8e525911833e7/src/ax/ai/types.ts#L811
Token budget for extended thinking (chain-of-thought reasoning).
Extended thinking allows models to “think through” complex problems before responding. Higher budgets allow deeper reasoning but cost more.
Approximate token allocations:
'none'- Disabled (default)'minimal'- ~1,000 tokens (~750 words of thinking)'low'- ~4,000 tokens'medium'- ~10,000 tokens'high'- ~20,000 tokens'highest'- ~32,000+ tokens (provider maximum)
Provider support:
- Anthropic Claude: Full support with
claude-sonnet-4and above - OpenAI: Supported with o1/o3 models (uses
reasoning_effort) - Google: Supported with Gemini 2.0 Flash Thinking
- DeepSeek: Supported with DeepSeek-R1
Example
// Enable medium thinking for complex reasoning
await gen.forward(ai, values, { thinkingTokenBudget: 'medium' });
timeout?
optional timeout: number;
Defined in: https://github.com/ax-llm/ax/blob/71ea5064d766efdc031d375243a8e525911833e7/src/ax/ai/types.ts#L744
Request timeout in milliseconds.
Default
300000 (5 minutes)
traceContext?
optional traceContext: Context;
Defined in: https://github.com/ax-llm/ax/blob/71ea5064d766efdc031d375243a8e525911833e7/src/ax/ai/types.ts#L762
OpenTelemetry trace context for distributed tracing.
tracer?
optional tracer: Tracer;
Defined in: https://github.com/ax-llm/ax/blob/71ea5064d766efdc031d375243a8e525911833e7/src/ax/ai/types.ts#L734
OpenTelemetry tracer for distributed tracing.
useExpensiveModel?
optional useExpensiveModel: "yes";
Defined in: https://github.com/ax-llm/ax/blob/71ea5064d766efdc031d375243a8e525911833e7/src/ax/ai/types.ts#L835
Hint to use a more capable (and expensive) model for complex tasks.
Some providers offer tiered models. Setting this to ‘yes’ requests the higher-capability tier when available.
verbose?
optional verbose: boolean;
Defined in: https://github.com/ax-llm/ax/blob/71ea5064d766efdc031d375243a8e525911833e7/src/ax/ai/types.ts#L725
Enable low-level HTTP request/response logging.
More verbose than debug. Shows raw HTTP traffic including headers.
Useful for debugging API issues.