View Source DSL: AshPostgres.DataLayer

A postgres data layer that leverages Ecto's postgres capabilities.

postgres

Postgres data layer configuration

Nested DSLs

Examples

postgres do
  repo MyApp.Repo
  table "organizations"
end

Options

NameTypeDefaultDocs
repomodule | (any, any -> any)The repo that will be used to fetch your data. See the AshPostgres.Repo documentation for more. Can also be a function that takes a resource and a type :read | :mutate and returns the repo
migrate?booleantrueWhether or not to include this resource in the generated migrations with mix ash.generate_migrations
storage_typeskeyword[]A keyword list of attribute names to the ecto type that should be used for that attribute. Only necessary if you need to override the defaults.
migration_typeskeyword[]A keyword list of attribute names to the ecto migration type that should be used for that attribute. Only necessary if you need to override the defaults.
migration_defaultskeyword[]A keyword list of attribute names to the ecto migration default that should be used for that attribute. The string you use will be placed verbatim in the migration. Use fragments like fragment(\\"now()\\"), or for nil, use \\"nil\\".
calculations_to_sqlkeywordA keyword list of calculations and their SQL representation. Used when creating unique indexes for identities over calculations
identity_wheres_to_sqlkeywordA keyword list of identity names and the SQL representation of their where clause. See AshPostgres.DataLayer.Info.identity_wheres_to_sql/1 for more details.
base_filter_sqlString.tA raw sql version of the base_filter, e.g representative = true. Required if trying to create a unique constraint on a resource with a base_filter
simple_join_first_aggregateslist(atom)[]A list of :first type aggregate names that can be joined to using a simple join. Use when you have a :first aggregate that uses a to-many relationship , but your filter statement ensures that there is only one result. Optimizes the generated query.
skip_unique_indexesatom | list(atom)[]Skip generating unique indexes when generating migrations
unique_index_nameslist({list(atom), String.t} | {list(atom), String.t, String.t})[]A list of unique index names that could raise errors that are not configured in identities, or an mfa to a function that takes a changeset and returns the list. In the format {[:affected, :keys], "name_of_constraint"} or {[:affected, :keys], "name_of_constraint", "custom error message"}
exclusion_constraint_namesany[]A list of exclusion constraint names that could raise errors. Must be in the format {:affected_key, "name_of_constraint"} or {:affected_key, "name_of_constraint", "custom error message"}
identity_index_namesany[]A keyword list of identity names to the unique index name that they should use when being managed by the migration generator.
foreign_key_nameslist({atom | String.t, String.t} | {atom | String.t, String.t, String.t})[]A list of foreign keys that could raise errors, or an mfa to a function that takes a changeset and returns a list. In the format: {:key, "name_of_constraint"} or {:key, "name_of_constraint", "custom error message"}
migration_ignore_attributeslist(atom)[]A list of attributes that will be ignored when generating migrations.
tableString.tThe table to store and read the resource from. If this is changed, the migration generator will not remove the old table.
schemaString.tThe schema that the table is located in. Schema-based multitenancy will supercede this option. If this is changed, the migration generator will not remove the old schema.
polymorphic?booleanfalseDeclares this resource as polymorphic. See the polymorphic resources guide for more.

postgres.custom_indexes

A section for configuring indexes to be created by the migration generator.

In general, prefer to use identities for simple unique constraints. This is a tool to allow for declaring more complex indexes.

Nested DSLs

Examples

custom_indexes do
  index [:column1, :column2], unique: true, where: "thing = TRUE"
end

postgres.custom_indexes.index

index fields

Add an index to be managed by the migration generator.

Examples

index ["column", "column2"], unique: true, where: "thing = TRUE"

Arguments

NameTypeDefaultDocs
fieldsatom | String.t | list(atom | String.t)The fields to include in the index.

Options

NameTypeDefaultDocs
error_fieldslist(atom)The fields to attach the error to.
nameString.tthe name of the index. Defaults to "#{table}_#{column}_index".
uniquebooleanfalseindicates whether the index should be unique.
concurrentlybooleanfalseindicates whether the index should be created/dropped concurrently.
usingString.tconfigures the index type.
prefixString.tspecify an optional prefix for the index.
whereString.tspecify conditions for a partial index.
includelist(String.t)specify fields for a covering index. This is not supported by all databases. For more information on PostgreSQL support, please read the official docs.
nulls_distinctbooleantruespecify whether null values should be considered distinct for a unique index. Requires PostgreSQL 15 or later
messageString.tA custom message to use for unique indexes that have been violated
all_tenants?booleanfalseWhether or not the index should factor in the multitenancy attribute or not.

Introspection

Target: AshPostgres.CustomIndex

postgres.custom_statements

A section for configuring custom statements to be added to migrations.

Changing custom statements may require manual intervention, because Ash can't determine what order they should run in (i.e if they depend on table structure that you've added, or vice versa). As such, any down statements we run for custom statements happen first, and any up statements happen last.

Additionally, when changing a custom statement, we must make some assumptions, i.e that we should migrate the old structure down using the previously configured down and recreate it.

This may not be desired, and so what you may end up doing is simply modifying the old migration and deleting whatever was generated by the migration generator. As always: read your migrations after generating them!

Nested DSLs

Examples

custom_statements do
  # the name is used to detect if you remove or modify the statement
  statement :pgweb_idx do
    up "CREATE INDEX pgweb_idx ON pgweb USING GIN (to_tsvector('english', title || ' ' || body));"
    down "DROP INDEX pgweb_idx;"
  end
end

postgres.custom_statements.statement

statement name

Add a custom statement for migrations.

Examples

statement :pgweb_idx do
  up "CREATE INDEX pgweb_idx ON pgweb USING GIN (to_tsvector('english', title || ' ' || body));"
  down "DROP INDEX pgweb_idx;"
end

Arguments

NameTypeDefaultDocs
nameatomThe name of the statement, must be unique within the resource

Options

NameTypeDefaultDocs
upString.tHow to create the structure of the statement
downString.tHow to tear down the structure of the statement
code?booleanfalseBy default, we place the strings inside of ecto migration's execute/1 function and assume they are sql. Use this option if you want to provide custom elixir code to be placed directly in the migrations

Introspection

Target: AshPostgres.Statement

postgres.manage_tenant

Configuration for the behavior of a resource that manages a tenant

Examples

manage_tenant do
  template ["organization_", :id]
  create? true
  update? false
end

Options

NameTypeDefaultDocs
templateString.t | atom | list(String.t | atom)A template that will cause the resource to create/manage the specified schema.
create?booleantrueWhether or not to automatically create a tenant when a record is created
update?booleantrueWhether or not to automatically update the tenant name if the record is udpated

postgres.references

A section for configuring the references (foreign keys) in resource migrations.

This section is only relevant if you are using the migration generator with this resource. Otherwise, it has no effect.

Nested DSLs

Examples

references do
  reference :post, on_delete: :delete, on_update: :update, name: "comments_to_posts_fkey"
end

Options

NameTypeDefaultDocs
polymorphic_on_delete:delete | :nilify | :nothing | :restrict | {:nilify, atom | list(atom)}For polymorphic resources, configures the on_delete behavior of the automatically generated foreign keys to source tables.
polymorphic_on_update:update | :nilify | :nothing | :restrictFor polymorphic resources, configures the on_update behavior of the automatically generated foreign keys to source tables.

postgres.references.reference

reference relationship

Configures the reference for a relationship in resource migrations.

Keep in mind that multiple relationships can theoretically involve the same destination and foreign keys. In those cases, you only need to configure the reference behavior for one of them. Any conflicts will result in an error, across this resource and any other resources that share a table with this one. For this reason, instead of adding a reference configuration for :nothing, its best to just leave the configuration out, as that is the default behavior if no relationship anywhere has configured the behavior of that reference.

Examples

reference :post, on_delete: :delete, on_update: :update, name: "comments_to_posts_fkey"

Arguments

NameTypeDefaultDocs
relationshipatomThe relationship to be configured

Options

NameTypeDefaultDocs
ignore?booleanIf set to true, no reference is created for the given relationship. This is useful if you need to define it in some custom way
on_delete:delete | :nilify | :nothing | :restrict | {:nilify, atom | list(atom)}What should happen to records of this resource when the referenced record of the destination resource is deleted.
on_update:update | :nilify | :nothing | :restrictWhat should happen to records of this resource when the referenced destination_attribute of the destination record is update.
deferrablefalse | true | :initiallyfalseWhether or not the constraint is deferrable. This only affects the migration generator.
nameString.tThe name of the foreign key to generate in the database. Defaults to <table>_<source_attribute>_fkey
match_withkeywordDefines additional keys to the foreign key in order to build a composite foreign key. The key should be the name of the source attribute (in the current resource), the value the name of the destination attribute.
match_type:simple | :partial | :fullselect if the match is :simple, :partial, or :full
index?booleanfalseWhether to create or not a corresponding index

Introspection

Target: AshPostgres.Reference

postgres.check_constraints

A section for configuring the check constraints for a given table.

This can be used to automatically create those check constraints, or just to provide message when they are raised

Nested DSLs

Examples

check_constraints do
  check_constraint :price, "price_must_be_positive", check: "price > 0", message: "price must be positive"
end

postgres.check_constraints.check_constraint

check_constraint attribute, name

Add a check constraint to be validated.

If a check constraint exists on the table but not in this section, and it produces an error, a runtime error will be raised.

Provide a list of attributes instead of a single attribute to add the message to multiple attributes.

By adding the check option, the migration generator will include it when generating migrations.

Examples

check_constraint :price, "price_must_be_positive", check: "price > 0", message: "price must be positive"

Arguments

NameTypeDefaultDocs
attributeanyThe attribute or list of attributes to which an error will be added if the check constraint fails
nameString.tThe name of the constraint

Options

NameTypeDefaultDocs
messageString.tThe message to be added if the check constraint fails
checkString.tThe contents of the check. If this is set, the migration generator will include it when generating migrations

Introspection

Target: AshPostgres.CheckConstraint