# `EXLA`
[🔗](https://github.com/elixir-nx/nx/blob/v0.11.0/exla/lib/exla.ex#L1)

[Google's XLA](https://www.tensorflow.org/xla/) (Accelerated Linear Algebra) compiler/backend for Nx.

It supports just-in-time (JIT) compilation to GPU (both CUDA and ROCm) and TPUs.

## XLA binaries

EXLA relies on the `XLA` package to provide the necessary XLA binaries.
Whenever possible it tries to download precompiled builds, but you may
need to build from source if there is no version matching your target
environment. For more details, including GPU/TPU support and requirements
see the `XLA` docs.

> #### Version requirements {: .info}
>
> For precise requirements, such as CUDA and cuDNN versions, see `XLA` docs.

## Configuration

EXLA ships with a backend to store tensors and run computations on.
Generally speaking, the backend is enabled globally in your `config/config.exs`
(or `config/ENV.exs`) with the following:

    import Config
    config :nx, :default_backend, EXLA.Backend

In a script/notebook, you would do:

    Mix.install([
      {:exla, "~> 0.2"}
    ])

    Nx.global_default_backend(EXLA.Backend)

From now on, all created tensors will be allocated directly on the given
`EXLA.Backend`. You can use functions such as `Nx.backend_transfer/2` to
explicitly transfer tensors.

EXLA will pick an available client to allocate and compute tensors, in this
order: `:cuda`, `:rocm`, `:tpu`, and `:host` (CPU). See the "Clients" section
below for more information.

To use GPUs/TPUs, you must also set the appropriate value for the
[`XLA_TARGET`](https://github.com/elixir-nx/xla#xla_target) environment
variable. If you have GPU/TPU enabled, we recommend setting the environment
variable for your machine altogether. For CUDA, setting
`ELIXIR_ERL_OPTIONS="+sssdio 128"` is also required on more complex operations
to increase CUDA's compiler stack size.

Note that setting the `EXLA.Backend` does not enable the EXLA compiler.
You must still pass the `compiler: EXLA` option to `Nx.Defn` functions
or call the functions in this module.

### Options

The options accepted by EXLA backend/compiler are:

  * `:client` - an atom representing the client to use. The default
    client is chosen on this order: `:cuda`, `:rocm`, `:tpu`, and `:host`.

  * `:device_id` - the default device id to run the computation
      on. Defaults to the `:default_device_id` on the client

  * `:precision` - control the tradeoff between speed and accuracy for
    array computations on accelerator backends (i.e. TPU and GPU).
    It must be one of:

    * `:default` - Fastest mode, but least accurate. Performs computations
      in bfloat16

    * `:high` - Slower but more accurate. Performs float32 computations in
      3 bfloat16 passes, or using tensorfloat32 where available

    * `:highest` - Slowest but most accurate. Performs computations in float32
      or float64 as applicable

## Clients

The `EXLA` library uses a client for compiling and executing code.
Those clients are typically bound to a platform, such as CPU or
GPU.

Those clients are singleton resources on Google's XLA library,
therefore they are treated as a singleton resource on this library
too. EXLA ships with runtime client configuration for each supported
platform:

    config :exla, :clients,
      cuda: [platform: :cuda],
      rocm: [platform: :rocm],
      tpu: [platform: :tpu],
      host: [platform: :host]

In a script/notebook, you can set those after `Mix.install/2`,
but before any tensor operation is performed:

    Application.put_env(:exla, :clients,
      cuda: [platform: :cuda],
      rocm: [platform: :rocm],
      tpu: [platform: :tpu],
      host: [platform: :host]
    )

You can provide your own list of clients, replacing the list above
or configuring each client as listed below. You can also specify
`:default_client` to set a particular client by default or
`:preferred_clients` to change the order of clients preference,
but those configurations are rarely set in practice.

> **Important!** you should avoid using multiple clients for the
> same platform. If you have multiple clients per platform, they
> can race each other and fight for resources, such as memory.
> Therefore, we recommend developers to stick with the default
> clients above.

### Client options

Each client configuration accepts the following options:

  * `:platform` - the platform the client runs on. It can be
    `:host` (CPU), `:cuda`, `:rocm`, or `:tpu`. Defaults to `:host`.

  * `:default_device_id` - the default device ID to run on.
    For example, if you have two GPUs, you can choose a different
    one as the default. Defaults to device 0 (the first device).

  * `:preallocate`- if the memory should be preallocated on
    GPU devices. Defaults to `true`.

  * `:memory_fraction` - how much memory of a GPU device to
    allocate. Defaults to `0.9`.

### Memory preallocation

XLA preallocates memory in GPU devices. This means that, if you are to
run multiple notebooks or multiple instances of your application, the
second, third, and so on instances won't be able to allocate memory.

You can disable this behaviour by setting `preallocate: false` on the
client configuration, as specified above. You may also use
`:memory_fraction` to control how much is preallocated.

### GPU Runtime Issues

GPU Executions run in dirty IO threads, which have a considerably smaller
stack size than regular scheduler threads. This may lead to problems with
certain CUDA or cuDNN versions, leading to segmentation fails. In a development
environment, it is suggested to set:

    ELIXIR_ERL_OPTIONS="+sssdio 128"

To increase the stack size of dirty IO threads from 40 kilowords to
128 kilowords. In a release, you can set this flag in your `vm.args`.

## Distribution

EXLA allows its tensors to be sent across nodes, as long as the parent
node (which effectively holds the tensor) keeps a reference to the
tensor while it is read by any other node it was sent to.

The result of `EXLA.compile/3` can also be shared across nodes.
On invocation, the underlying executable is automatically serialized
and sent to other nodes, without requiring a full recompilation,
as long as the same conditions as above apply.

## Docker considerations

EXLA should run fine on Docker with one important consideration:
you must not start the Erlang VM as the root process in Docker.
That's because when the Erlang VM runs as root, it has to manage
all child programs.

At the same time, Google's XLA shells out to child programs and
must retain control over how child programs terminate.

To address this, simply make sure you wrap the Erlang VM in
another process, such as a shell process. In other words, if you
are using releases, instead of this:

    CMD path/to/release start

do this:

    CMD sh -c "path/to/release start"

If you are using Mix inside your Docker containers, instead of this:

    CMD mix run

do this:

    CMD sh -c "mix run"

Alternatively, you can pass the `--init` flag to `docker run`,
so it runs an `init` inside the container that forwards signals
and reaps processes.

The `--init` flag uses the [`tini`](https://github.com/krallin/tini)
project, so for cases where the flag may not available (e.g.
kubernetes) you may want to install it.

## Telemetry events

EXLA executes a telemetry event every time a function is JIT-compiled.
The events are named `[:exla, :compilation]` and include the following
measurements, given in microseconds:

  * `:eval_time` - the time spent on turning the function into XLA
    computation
  * `:compile_time` - the time spent on compiling the XLA computation
    into an executable
  * `:total_time` - the sum of `:eval_time` and `:compile_time`

The metadata is:

  * `:key` - the compilation key for debugging

## Sharding

EXLA supports sharding by default through the `EXLA.shard_jit/3` function
and XLA's automatic SPMD partitioning. This means that you can write a function
and provide input sharding information along with sharded inputs, and the function
will automatically propagate and partition the computation across the devices
in a best-effort approach.

# `cached?`

Checks if the compilation of function with args is cached.

Note that hooks are part of the cache, and
therefore they must be included in the options.

## Examples

    iex> fun = fn a, b -> Nx.add(a, b) end
    iex> left = Nx.tensor(1, type: {:u, 8})
    iex> right = Nx.tensor([1, 2, 3], type: {:u, 16})
    iex> EXLA.jit(fun).(left, right)
    iex> EXLA.cached?(fun, [left, right])
    true
    iex> EXLA.cached?(fun, [left, Nx.tensor([1, 2, 3, 4], type: {:u, 16})])
    false

Compiled functions are also cached, unless cache is set to false:

    iex> fun = fn a, b -> Nx.subtract(a, b) end
    iex> left = Nx.tensor(1, type: {:u, 8})
    iex> right = Nx.tensor([1, 2, 3], type: {:u, 16})
    iex> EXLA.compile(fun, [left, right], cache: false)
    iex> EXLA.cached?(fun, [left, right])
    false
    iex> EXLA.compile(fun, [left, right])
    iex> EXLA.cached?(fun, [left, right])
    true

# `compile`

A shortcut for `Nx.Defn.compile/3` with the EXLA compiler.

    iex> fun = EXLA.compile(&Nx.add(&1, &1), [Nx.template({3}, {:s, 32})])
    iex> fun.(Nx.tensor([1, 2, 3]))
    #Nx.Tensor<
      s32[3]
      [2, 4, 6]
    >

The returned function can be sent across nodes, as long as the parent
node (which effectively holds the function) keeps a reference to the
function while it is invoked by any other node it was sent to. On
invocation, the underlying executable is automatically serialized
and sent to other nodes, without requiring a full recompilation.

See `jit/2` for supported options.

# `jit`

A shortcut for `Nx.Defn.jit/2` with the EXLA compiler.

    iex> EXLA.jit(&Nx.add(&1, &1)).(Nx.tensor([1, 2, 3]))
    #Nx.Tensor<
      s32[3]
      [2, 4, 6]
    >

Results are allocated on the `EXLA.Backend`. Note that the
`EXLA.Backend` is asynchronous: operations on its tensors
*may* return immediately, before the tensor data is available.
The backend will then block only when trying to read the data
or when passing it to another operation.

## Options

It accepts the same option as `Nx.Defn.jit/2` plus:

  * `:cache` - cache the results of compilation, defaults to `true`.
    You may disable it by setting it to `false`. You can also set it
    to a binary, representing a filesystem path to store the cache.
    EXLA will ensure the arguments and parameters across invocations
    have the same shape, but it is ultimately your responsibility
    to provide a unique cache path.

  * `:client` - an atom representing the client to use. The default
    client is chosen on this order: `:cuda`, `:rocm`, `:tpu`, and `:host`.

  * `:debug` - print compile and debugging information, defaults to `false`.

  * `:device_id` - the default device id to run the computation on.
    Defaults to the `:default_device_id` on the client

  * `:lazy_transfers` - when `:always`, it lazily transfers data to the device
    instead of upfront. This is useful to reduce memory allocation on GPU/TPU
    devices at the cost of increased latency. **It is recommended to only enable
    this if the input tensors are allocated on host and the computation is
    running on GPU/TPU with a limited amount of memory**

# `jit_apply`

A shortcut for `Nx.Defn.jit_apply/3` with the EXLA compiler.

    iex> EXLA.jit_apply(&Nx.add(&1, &1), [Nx.tensor([1, 2, 3])])
    #Nx.Tensor<
      s32[3]
      [2, 4, 6]
    >

See `jit/2` for supported options.

# `shard_jit`

A shortcut for `Nx.Defn.shard_jit/3` with the EXLA compiler.

## Example

    mesh = %Nx.Mesh{name: "mesh", shape: {2}}
    fun = EXLA.shard_jit(&Nx.add(&1, &1), mesh, input_shardings: [%{0 => [0]}])
    # Pass sharded inputs (one arglist per partition)
    fun.([[Nx.tensor([1, 2, 3])], [Nx.tensor([4, 5, 6])]])
    #=> #Nx.Tensor<
    #     s32[6]
    #     [2, 4, 6, 8, 10, 12]
    #   >

## Options

  * `:input_shardings` - a list of maps specifying how to shard each input tensor.
    Each map has tensor dimension (integer index or atom name) as keys and lists of
    mesh axis indices as values. Dimensions not specified are replicated.

    Examples:
      - `[%{0 => [0], 1 => [1]}]` - shard first input's dim 0 on mesh axis 0, dim 1 on mesh axis 1
      - `[%{0 => [0]}, %{1 => [1]}]` - first input sharded on dim 0, second input sharded on dim 1
      - `[%{0 => [0, 1]}]` - shard dim 0 across both mesh axes 0 and 1
      - `[%{}]` - fully replicated tensor

Also accepts the same options as `compile/3`.

# `to_mlir_module`

Takes in a function, the argument templates and the compilation
options and returns the textual representation of the MLIR module.

## Options

  * `:within_defn_compiler` - a boolean that indicates whether
    this function is being called from within a `defn` compiler.
    Defaults to `false`.

## Examples

    iex> fun = fn x, y -> Nx.add(Nx.sin(x), Nx.cos(y)) end
    iex> args = [1.0, 2.0]
    iex> %{mlir_module: mlir_module} = EXLA.to_mlir_module(fun, args)
    iex> mlir_module
    """
    module {
      func.func public @main(%arg0: tensor<f32>, %arg1: tensor<f32>) -> tensor<f32> {
        %0 = stablehlo.sine %arg0 : tensor<f32>
        %1 = stablehlo.cosine %arg1 : tensor<f32>
        %2 = stablehlo.add %0, %1 : tensor<f32>
        return %2 : tensor<f32>
      }
    }
    """

---

*Consult [api-reference.md](api-reference.md) for complete listing*
