Defines the callbacks fired by an LLMChain and LLM module.
A callback handler is a map that defines the specific callback event with a function to execute for that event.
Example
A sample configured callback handler that forwards received data to a specific LiveView.
live_view_pid = self()
my_handlers = %{
on_llm_new_delta: fn _chain, new_deltas -> send(live_view_pid, {:received_delta, new_deltas}) end,
on_message_processed: fn _chain, new_message -> send(live_view_pid, {:received_message, new_message}) end,
on_error_message_created: fn _chain, new_message -> send(live_view_pid, {:received_message, new_message}) end
}
model = SomeLLM.new!(%{...})
chain =
%{llm: model}
|> LLMChain.new!()
|> LLMChain.add_callback(my_handlers)
Summary
Types
The supported set of callbacks for an LLM module.
Executed when an LLMChain, in response to an error from the LLM, generates a new, automated response message intended to be returned to the LLM.
Executed when an LLMChain has completed processing a received assistant message. This fires when a message is complete either after assembling streaming deltas or when a full message is received when not streaming.
Executed when processing a received message errors or fails. The erroring message is included in the callback with the state of processing that was completed before erroring.
Executed when the chain failed multiple times used up the max_retry_count
resulting in the process aborting and returning an error.
Executed when a tool call is identified during streaming, before execution begins.
Executed when a single tool execution completes successfully.
Executed when a single tool execution fails.
Executed when the chain begins executing a tool call.
Executed when the chain uses one or more tools and the resulting ToolResults are generated as part of a tool response message.
Executed when an LLM is streaming a response and a new MessageDelta (or token) was received.
Executed when an LLM is not streaming and a full message was received.
Executed when an LLM (typically a service) responds with rate limiting information.
Executed when an LLM response is received through an HTTP response. The entire set of raw response headers can be received and processed.
Executed when an LLM response reports the token usage in a
LangChain.TokenUsage struct. The data returned depends on the LLM.
Types
@type chain_callback_handler() :: %{ optional(:on_llm_new_delta) => llm_new_delta(), optional(:on_llm_new_message) => llm_new_message(), optional(:on_llm_ratelimit_info) => llm_ratelimit_info(), optional(:on_llm_token_usage) => llm_token_usage(), optional(:on_llm_response_headers) => llm_response_headers(), optional(:on_message_processed) => chain_message_processed(), optional(:on_message_processing_error) => chain_message_processing_error(), optional(:on_error_message_created) => chain_error_message_created(), optional(:on_tool_call_identified) => chain_tool_call_identified(), optional(:on_tool_execution_started) => chain_tool_execution_started(), optional(:on_tool_execution_completed) => chain_tool_execution_completed(), optional(:on_tool_execution_failed) => chain_tool_execution_failed(), optional(:on_tool_response_created) => chain_tool_response_created(), optional(:on_retries_exceeded) => chain_retries_exceeded() }
The supported set of callbacks for an LLM module.
@type chain_error_message_created() :: (LangChain.Chains.LLMChain.t(), LangChain.Message.t() -> any())
Executed when an LLMChain, in response to an error from the LLM, generates a new, automated response message intended to be returned to the LLM.
@type chain_message_processed() :: (LangChain.Chains.LLMChain.t(), LangChain.Message.t() -> any())
Executed when an LLMChain has completed processing a received assistant message. This fires when a message is complete either after assembling streaming deltas or when a full message is received when not streaming.
This is the best way to be notified when a message is "done" and should be handled by the application.
The handler's return value is discarded.
@type chain_message_processing_error() :: (LangChain.Chains.LLMChain.t(), LangChain.Message.t() -> any())
Executed when processing a received message errors or fails. The erroring message is included in the callback with the state of processing that was completed before erroring.
The handler's return value is discarded.
@type chain_retries_exceeded() :: (LangChain.Chains.LLMChain.t() -> any())
Executed when the chain failed multiple times used up the max_retry_count
resulting in the process aborting and returning an error.
The handler's return value is discarded.
@type chain_tool_call_identified() :: (LangChain.Chains.LLMChain.t(), LangChain.Message.ToolCall.t(), LangChain.Function.t() -> any())
Executed when a tool call is identified during streaming, before execution begins.
This fires as soon as we have enough information to identify the tool (at minimum, the name field).
The tool call may be incomplete - call_id might not be available yet, and arguments may be partial.
This callback provides early notification for UI feedback like "Searching web..." while the LLM is still streaming the complete tool call.
Timing:
- Fires: As soon as tool name is detected in streaming deltas
- Before: Tool arguments are fully received
- Before: Tool execution begins
Arguments:
- First: LLMChain.t() - Current chain state
- Second: ToolCall.t() - Tool call struct (may be incomplete, but has name)
- Third: Function.t() - Function definition (includes display_text)
The handler's return value is discarded.
Example
callback_handler = %{
on_tool_call_identified: fn _chain, tool_call, func ->
IO.puts("Tool identified: #{func.display_text || tool_call.name}")
end
}
@type chain_tool_execution_completed() :: (LangChain.Chains.LLMChain.t(), LangChain.Message.ToolCall.t(), LangChain.Message.ToolResult.t() -> any())
Executed when a single tool execution completes successfully.
Fires after individual tool execution, before results are aggregated. Useful for showing per-tool success indicators.
- First argument: LLMChain.t()
- Second argument: ToolCall that was executed
- Third argument: ToolResult that was generated
The handler's return value is discarded.
@type chain_tool_execution_failed() :: (LangChain.Chains.LLMChain.t(), LangChain.Message.ToolCall.t(), term() -> any())
Executed when a single tool execution fails.
Fires when tool execution raises an exception or returns an error result.
- First argument: LLMChain.t()
- Second argument: ToolCall that failed
- Third argument: Error reason or exception
The handler's return value is discarded.
@type chain_tool_execution_started() :: (LangChain.Chains.LLMChain.t(), LangChain.Message.ToolCall.t(), LangChain.Function.t() -> any())
Executed when the chain begins executing a tool call.
This fires immediately before tool execution starts, allowing UIs to show real-time feedback like "Searching the web..." or "Creating file...".
- First argument: LLMChain.t()
- Second argument: ToolCall struct being executed
- Third argument: Function struct for the tool (includes display_text)
The handler's return value is discarded.
@type chain_tool_response_created() :: (LangChain.Chains.LLMChain.t(), LangChain.Message.t() -> any())
Executed when the chain uses one or more tools and the resulting ToolResults are generated as part of a tool response message.
The handler's return value is discarded.
@type llm_new_delta() :: (LangChain.Chains.LLMChain.t(), [LangChain.MessageDelta.t()] -> any())
Executed when an LLM is streaming a response and a new MessageDelta (or token) was received.
:indexis optionally present if the LLM supports sendingnversions of a response.
The return value is discarded.
@type llm_new_message() :: (LangChain.Chains.LLMChain.t(), LangChain.Message.t() -> any())
Executed when an LLM is not streaming and a full message was received.
The return value is discarded.
@type llm_ratelimit_info() :: (LangChain.Chains.LLMChain.t(), info :: %{required(String.t()) => any()} -> any())
Executed when an LLM (typically a service) responds with rate limiting information.
The specific rate limit information depends on the LLM. It returns a map with all the available information included.
The return value is discarded.
@type llm_response_headers() :: (LangChain.Chains.LLMChain.t(), response_headers :: map() -> any())
Executed when an LLM response is received through an HTTP response. The entire set of raw response headers can be received and processed.
The return value is discarded.
Example
A function declaration that matches the signature.
def handle_llm_response_headers(chain, response_headers) do
# This demonstrates how to send the response headers to a
# LiveView assuming the LiveView's pid was stored in the chain's
# custom_context.
send(chain.custom_context.live_view_pid, {:req_response_headers, response_headers})
IO.inspect(response_headers)
end
@type llm_token_usage() :: (LangChain.Chains.LLMChain.t(), LangChain.TokenUsage.t() -> any())
Executed when an LLM response reports the token usage in a
LangChain.TokenUsage struct. The data returned depends on the LLM.
The return value is discarded.