View Source ExOpenAI.Components.CreateModerationResponse (ex_openai.ex v1.8.0)
Schema representing a CreateModerationResponse within the OpenAI API
Summary
Types
@type t() :: %ExOpenAI.Components.CreateModerationResponse{ id: String.t(), model: String.t(), results: [ %{ categories: %{ harassment: boolean(), "harassment/threatening": boolean(), hate: boolean(), "hate/threatening": boolean(), illicit: boolean(), "illicit/violent": boolean(), "self-harm": boolean(), "self-harm/instructions": boolean(), "self-harm/intent": boolean(), sexual: boolean(), "sexual/minors": boolean(), violence: boolean(), "violence/graphic": boolean() }, category_applied_input_types: %{ harassment: [:text], "harassment/threatening": [:text], hate: [:text], "hate/threatening": [:text], illicit: [:text], "illicit/violent": [:text], "self-harm": [:image | :text], "self-harm/instructions": [:image | :text], "self-harm/intent": [:image | :text], sexual: [:image | :text], "sexual/minors": [:text], violence: [:image | :text], "violence/graphic": [:image | :text] }, category_scores: %{ harassment: float(), "harassment/threatening": float(), hate: float(), "hate/threatening": float(), illicit: float(), "illicit/violent": float(), "self-harm": float(), "self-harm/instructions": float(), "self-harm/intent": float(), sexual: float(), "sexual/minors": float(), violence: float(), "violence/graphic": float() }, flagged: boolean() } ] }