Package-level declarations
Types
Access to a resource was denied.
Specifies the set of parameters needed to perform aggregation in the aggregate transform.
An object representing a value allowed for a property.
A resource to be created or added already exists.
Specifies an optional value when connecting to the Redshift cluster.
Specifies an Amazon Redshift node.
Specifies an Amazon Redshift source.
Specifies an Amazon Redshift target.
A failed annotation.
Specifies a transform that maps data property keys in the data source to data property keys in the data target. You can rename keys, modify the data types for keys, and choose which keys to drop from the dataset.
Specifies a connector to an Amazon Athena data source.
A structure containing the Lake Formation audit context.
The authentication configuration for a connection returned by the DescribeConnectionType
API.
A structure containing the authentication configuration.
A structure containing the authentication configuration in the CreateConnection request.
The set of properties required for the the OAuth2 AUTHORIZATION_CODE
grant type workflow.
A list of errors that can occur when registering partition indexes for an existing table.
For supplying basic auth credentials when not providing a SecretArn
value.
Specifies a target that uses a Glue Data Catalog table.
Represents a table optimizer to retrieve in the BatchGetTableOptimizer
operation.
Contains details on one of the errors in the error list returned by the BatchGetTableOptimizer
operation.
Records an error that occurred when attempting to stop a specified job run.
Records a successful request to stop a specified JobRun
.
Contains details for one of the table optimizers returned by the BatchGetTableOptimizer
operation.
Contains information about a batch update partition error.
A structure that contains the values and structure used to update a partition.
Defines column statistics supported for bit sequence data values.
The details of a blueprint.
The details of a blueprint run.
Defines column statistics supported for Boolean data columns.
Specifies the supported authentication types returned by the DescribeConnectionType
API.
Specifies a Delta Lake data source that is registered in the Glue Data Catalog.
Specifies a table definition in the Glue Data Catalog.
Specifies a Hudi data source that is registered in the Glue Data Catalog.
A structure containing migration status information.
A structure that describes catalog properties.
Specifies an Apache Kafka data store in the Data Catalog.
Specifies a Kinesis data source in the Glue Data Catalog.
A structure that specifies data lake access properties and other custom properties.
Property attributes that include configuration properties for the catalog resource.
A policy that specifies update behavior for the crawler.
Specifies a data store in the Glue Data Catalog.
Specifies an Glue Data Catalog target.
Classifiers are triggered during a crawl task. A classifier checks whether a given file is in a format it can handle. If it is, the classifier creates a schema in the form of a StructType
object that matches that data format.
Specifies how Amazon CloudWatch data should be encrypted.
CodeGenConfigurationNode
enumerates all valid Node types. One and only one of its member variables can be populated.
Represents a directional edge in a directed acyclic graph (DAG).
Represents a node in a directed acyclic graph (DAG)
An argument or property of a node.
Encapsulates a column name that failed and the reason for failure.
A structure containing the column name and column importance score for a column.
A filter that uses both column-level and row-level filtering.
Represents the generated column-level statistics for a table or partition.
Contains the individual types of column statistics data. Only one data object should be set and indicated by the Type
attribute.
Encapsulates a ColumnStatistics
object that failed and the reason for failure.
An exception thrown when you try to stop a task run when there is no task running.
The object that shows the details of the column stats run.
An exception thrown when you try to start another job while running a column stats generation job.
The settings for a column statistics task.
An exception thrown when you try to stop a task run.
The configuration for a compaction optimizer. This configuration defines how data files in your table will be compacted to improve query performance and reduce storage costs.
A structure that contains compaction metrics for the optimizer run.
An object containing configuration for a compute environment (such as Spark, Python or Athena) returned by the DescribeConnectionType
API.
Two processes are trying to modify a resource simultaneously.
Too many jobs are being run concurrently.
A specified condition was not satisfied.
Condition expression defined in the Glue Studio data preparation recipe node.
Specifies the values that an admin sets for each job or session parameter configured in a Glue usage profile.
The CreatePartitions
API was called on a table that has indexes enabled.
The confusion matrix shows you what your transform is predicting accurately and what types of errors it is making.
Defines a connection to a data source.
A structure that is used to specify a connection to create or update.
The data structure used by the Data Catalog to encrypt the password as part of CreateConnection
or UpdateConnection
and store it in the ENCRYPTED_PASSWORD
field in the connection properties. You can enable catalog encryption or only password encryption.
Specifies the connections used by a job.
Brief information about a supported connection type returned by the ListConnectionTypes
API.
Represents a variant of a connection type in Glue Data Catalog. Connection type variants provide specific configurations and behaviors for different implementations of the same general connection type.
Specifies a source generated with standard connection options.
Specifies a target generated with standard connection options.
Contains the information for a run of a crawler.
Metrics for a specified crawler.
The details of a Crawler node present in the workflow.
The specified crawler is not running.
The operation cannot be performed because the crawler is already running.
The specified crawler is stopping.
Specifies data stores to crawl.
A list of fields, comparators and value that you can use to filter the crawler runs for a specified crawler.
Specifies a custom CSV classifier for CreateClassifier
to create.
A request to create a data quality ruleset.
Specifies a grok
classifier for CreateClassifier
to create.
The configuration parameters required to create a new Iceberg table in the Glue Data Catalog, including table properties and metadata specifications.
Specifies a JSON classifier for CreateClassifier
to create.
Request to create a new session.
Specifies an XML classifier for CreateClassifier
to create.
A classifier for custom CSV
content.
Specifies a transform that uses custom code you provide to perform the data transformation. The output is a collection of DynamicFrames.
An object representing a custom pattern for detecting sensitive data across the columns and rows of your structured data.
A structure that describes a target database for resource linking.
The structure used to create or update a database.
Contains configuration information for maintaining Data Catalog security.
Input properties to configure data lake access for your catalog resource in the Glue Data Catalog.
The output properties of the data lake access configuration for your catalog resource in the Glue Data Catalog.
The Lake Formation principal.
An Inclusion Annotation.
A summary of metrics showing the total counts of processed rows and rules, including their pass/fail statistics based on row-level results.
Describes the result of the evaluation of a data quality analyzer.
Specifies how Data Quality assets in your account should be encrypted.
Additional run options you can specify for an evaluation run.
Describes the data quality metric value according to the analysis of historical data.
Describes the observation generated after evaluating the rules and analyzers.
Describes a data quality result.
Describes a data quality result.
Criteria used to return data quality results.
Describes the result of a data quality rule recommendation run.
A filter for listing data quality recommendation runs.
Describes the result of the evaluation of a data quality rule.
Describes the result of a data quality ruleset evaluation run.
The filter criteria.
The criteria used to filter data quality rulesets.
Describes a data quality ruleset returned by GetDataQualityRuleset
.
An object representing an Glue table.
A data source (an Glue table) for which you want data quality results.
Defines column statistics supported for timestamp data columns.
Defines column statistics supported for fixed-point number data columns.
Contains a numeric value in decimal format.
Specifies a Delta data store to crawl one or more Delta tables.
A development endpoint where a developer can remotely debug extract, transform, and load (ETL) scripts.
Custom libraries to be loaded into a development endpoint.
Specifies the direct JDBC source connection.
Specifies an Apache Kafka data store.
Specifies a direct Amazon Kinesis data source.
A policy that specifies update behavior for the crawler.
Defines column statistics supported for floating-point number data columns.
Options to configure how your data quality evaluation results are published.
Options to configure how your job will stop if your data quality evaluation fails.
Specifies a transform that removes rows of repeating data from a data set.
Specifies a transform that chooses the data property keys that you want to drop.
Specifies a transform that removes columns from the dataset if all values in the column are 'null'. By default, Glue Studio will recognize null objects, but some values such as empty strings, strings that are "null", -1 integers or other placeholders such as zeros, are not automatically recognized as nulls.
Specifies the set of parameters needed to perform the dynamic transform.
Specifies a DynamoDB data source in the Glue Data Catalog.
Specifies an Amazon DynamoDB table to crawl.
Specifies the encryption-at-rest configuration for the Data Catalog.
Specifies an encryption configuration.
A specified entity does not exist
Contains details about an error.
An object containing error details.
Specifies your data quality evaluation criteria.
Specifies your data quality evaluation criteria.
Evaluation metrics provide an estimate of the quality of your machine learning transform.
Batch condition that must be met (specified number of events received or batch time window expired) before EventBridge event trigger fires.
A run attempt for a column statistics task run.
An execution property of a job.
Specifies configuration properties for an exporting labels task run.
A catalog that points to an entity outside the Glue Data Catalog.
A database that points to an entity outside the Glue Data Catalog.
A federated resource already exists.
A table that points to an entity outside the Glue Data Catalog.
A federation source failed.
A federation source failed, but the operation may be retried.
Specifies a transform that locates records in the dataset that have missing values and adds a new field with a value determined by imputation. The input data set is used to train the machine learning model that determines what the missing value should be.
Specifies a filter expression.
Represents a single entry in the list of values for a FilterExpression
.
The evaluation metrics for the find matches algorithm. The quality of your machine learning transform is measured by getting your transform to predict some matches and comparing the results to known matches from the same dataset. The quality metrics are based on a subset of your data, so they are not precise.
The parameters to configure the find matches transform.
Specifies configuration properties for a Find Matches task run.
Filters the connection definitions that are returned by the GetConnections
API operation.
The response for the data quality result.
The response for the Data Quality rule recommendation run.
Returns the data quality ruleset response.
An encryption operation failed.
Base class for all service related exceptions thrown by the Glue client
A structure for returning a resource policy.
Specifies a user-defined schema when a schema cannot be determined by Glue.
Specifies a single column in a Glue schema definition.
Specifies the data store in the governed Glue Data Catalog.
Specifies a data target that writes to Amazon S3 using the Glue Data Catalog.
A classifier that uses grok
patterns.
Specifies an Apache Hudi data source.
The configuration for an Iceberg compaction optimizer. This configuration defines parameters for optimizing the layout of data files in Iceberg tables.
Compaction metrics for Iceberg for the optimizer run.
A structure that defines an Apache Iceberg metadata table to create in the catalog.
The configuration for an Iceberg orphan file deletion optimizer.
Orphan file deletion metrics for Iceberg for the optimizer run.
Defines a single partition field within an Iceberg partition specification, including the source field, transformation function, partition name, and unique identifier.
Defines the partitioning specification for an Iceberg table, determining how table data will be organized and partitioned for optimal query performance.
The configuration for an Iceberg snapshot retention optimizer.
Snapshot retention metrics for Iceberg for the optimizer run.
Defines the schema structure for an Iceberg table, including field definitions, data types, and schema metadata.
Defines a single field within an Iceberg sort order specification, including the source field, transformation, sort direction, and null value ordering.
Defines the sort order specification for an Iceberg table, determining how data should be ordered within partitions to optimize query performance.
Defines a single field within an Iceberg table schema, including its identifier, name, data type, nullability, and documentation.
Defines a complete set of updates to be applied to an Iceberg table, including schema changes, partitioning modifications, sort order adjustments, location updates, and property changes.
Specifies an Apache Iceberg data source where Iceberg tables are stored in Amazon S3.
The same unique identifier was associated with two different records.
The blueprint is in an invalid state to perform a requested operation.
The session is in an invalid state to perform a requested operation.
The workflow is in an invalid state to perform a requested operation.
Specifies configuration properties for an importing labels task run.
A structure for an integration that writes data into a resource.
Describes a zero-ETL integration.
Properties associated with the integration.
The requested operation conflicts with another operation.
An error associated with a zero-ETL integration.
A filter that can be used when invoking a DescribeIntegrations
request.
The specified integration could not be found.
A structure that describes how data is partitioned on the target.
The data processed through your integration exceeded your quota.
An internal server error occurred.
An internal service error occurred.
The input provided was not valid.
The integration is in an invalid state.
An error that indicates your data is in an invalid state.
Additional connection options for the connector.
Specifies a connector to a JDBC data source.
Specifies a data target that writes to Amazon S3 in Apache Parquet columnar storage.
Specifies a JDBC data store to crawl.
Defines a point that a job can resume processing.
Specifies how job bookmark data should be encrypted.
Specifies code that runs when a job is run.
The details of a Job node present in the workflow.
Specifies a column to be joined.
A classifier for JSON
content.
Additional options for streaming.
A partition key pair consisting of a name and a type.
Additional options for the Amazon Kinesis streaming data source.
The KMS key specified is not accessible.
Specifies configuration properties for a labeling set generation task run.
Specifies Lake Formation configuration settings for the crawler.
When there are multiple versions of a blueprint and the latest version has some errors, this attribute indicates the last successful blueprint definition that is available with the service.
Status and error information about the most recent crawl.
Specifies data lineage configuration settings for the crawler.
Defines column statistics supported for integer data columns.
Defines a mapping.
A structure containing metadata information for a schema version.
A structure containing a key value pair for metadata.
Describes the metric based observation generated based on evaluated data quality metrics.
Specifies a Microsoft SQL server data source in the Glue Data Catalog.
Specifies a target that uses Microsoft SQL.
A structure for a machine learning transform.
The machine learning transform is not ready to run.
The encryption-at-rest settings of the transform that apply to accessing user data.
Specifies an Amazon DocumentDB or MongoDB data store to crawl.
Specifies a MySQL data source in the Glue Data Catalog.
Specifies a target that uses MySQL.
There is no applicable schedule.
Specifies configuration properties of a notification.
Represents whether certain values are recognized as null values for removal.
Represents a custom null value such as a zeros or other value being used as a null placeholder unique to the dataset.
The OAuth2 client app used for the connection.
The credentials used when the authentication type is OAuth2 authentication.
A structure containing properties for OAuth2 authentication.
A structure containing properties for OAuth2 in the CreateConnection request.
A structure representing an open format table.
The operation is not available in the region.
The operation timed out.
Specifies an Oracle data source in the Glue Data Catalog.
Specifies a target that uses Oracle SQL.
The configuration for an orphan file deletion optimizer.
A structure that contains orphan file deletion metrics for the optimizer run.
A structure containing other metadata for a schema version belonging to the same metadata key.
Contains information about a partition error.
A structure for a partition index.
A descriptor for a partition index in a table.
The structure used to create and update a partition.
Contains a list of values defining partitions.
The operation timed out.
The OAuth client app in GetConnection response.
Specifies a transform that identifies, removes or masks PII data.
Specifies a PostgresSQL data source in the Glue Data Catalog.
Specifies a target that uses Postgres SQL.
A job run that was used in the predicate of a conditional trigger that triggered this job run.
Permissions granted to a principal.
Specifies the job and session values that an admin configures in an Glue usage profile.
Defines a property predicate.
Left blank.
A structure used as a protocol between query engines and Lake Formation or Glue. Contains both a Lake Formation generated authorization identifier and information from the request's authorization context.
Actions defined in the Glue Studio data preparation recipe node.
A reference to a Glue DataBrew recipe.
A recipe step used in a Glue Studio data preparation recipe node.
When crawling an Amazon S3 data source after the first crawl is complete, specifies whether to crawl the entire dataset again or to crawl only folders that were added since the last crawler run. For more information, see Incremental Crawls in Glue in the developer guide.
Specifies an Amazon Redshift data store.
Specifies a target that uses Amazon Redshift.
A wrapper structure that may contain the registry name and Amazon Resource Name (ARN).
A structure containing the details for a registry.
Specifies a Relational database data source in the Glue Data Catalog.
Specifies a transform that renames a single data property key.
The resource could not be found.
A resource was not ready for a transaction.
A resource numerical limit was exceeded.
The URIs for function resources.
The configuration for a snapshot retention optimizer.
A structure that contains retention metrics for the optimizer run.
A run identifier.
Metrics for the optimizer run.
Specifies a Delta Lake data source that is registered in the Glue Data Catalog. The data source must be stored in Amazon S3.
Specifies a Hudi data source that is registered in the Glue Data Catalog. The Hudi data source must be stored in Amazon S3.
Specifies an Amazon S3 data store in the Glue Data Catalog.
Specifies a data target that writes to Amazon S3 using the Glue Data Catalog.
Specifies a command-separated value (CSV) data store stored in Amazon S3.
Specifies a target that writes to a Delta Lake data source in the Glue Data Catalog.
Specifies a target that writes to a Delta Lake data source in Amazon S3.
Specifies a Delta Lake data source stored in Amazon S3.
Specifies additional connection options for the Amazon S3 data store.
Specifies a data target that writes to Amazon S3.
Specifies how Amazon Simple Storage Service (Amazon S3) data should be encrypted.
Specifies an S3 Excel data source.
Specifies a data target that writes to Amazon S3 in Apache Parquet columnar storage.
Specifies a target that writes to a Hudi data source in the Glue Data Catalog.
Specifies a target that writes to a Hudi data source in Amazon S3.
Specifies a Hudi data source stored in Amazon S3.
Specifies a HyperDirect data target that writes to Amazon S3.
Specifies a target that writes to an Iceberg data source in Amazon S3.
Specifies a JSON data store stored in Amazon S3.
Specifies an Apache Parquet data store stored in Amazon S3.
Specifies additional connection options for the Amazon S3 data store.
The specified scheduler is not running.
The specified scheduler is already running.
The specified scheduler is transitioning.
A policy that specifies update and deletion behaviors for the crawler.
A key-value pair representing a column and data type that this transform can run against. The Schema
parameter of the MLTransform
may contain up to 100 of these structures.
An object that contains minimal details for a schema.
An object that references a schema stored in the Glue Schema Registry.
An object that contains the error details for an operation on a schema version.
An object containing the details about a schema version.
A structure containing the schema version information.
Specifies a security configuration.
Specifies a transform that chooses the data property keys that you want to keep.
Specifies a transform that chooses one DynamicFrame
from a collection of DynamicFrames
. The output is the selected DynamicFrame
The SessionCommand
that runs the job.
Specifies skewed values in a table. Skewed values are those that occur with very high frequency.
Specifies configuration for Snowflake nodes in Glue Studio.
Specifies a Snowflake data source.
Specifies a Snowflake target.
Specifies a field to sort by and a sort order.
The details for a source control configuration for a job, allowing synchronization of job artifacts to or from a remote repository.
The resource properties associated with the integration source.
Properties used by the source leg to process data from the source.
Specifies a connector to an Apache Spark data source.
Specifies a target that uses an Apache Spark connector.
Specifies a transform that splits data property keys into two DynamicFrames
. The output is a collection of DynamicFrames
: one with selected data property keys, and one with the remaining data property keys.
The request of the Data Quality rule recommendation request.
The batch condition that started the workflow run. Either the number of events in the batch size arrived, in which case the BatchSize member is non-zero, or the batch window expired, in which case the BatchWindow member is non-zero.
The code execution output in JSON format.
The code execution output in JSON format.
A Statistic Annotation.
The statistic model result.
Summary information about a statistic.
A structure containing information about an asynchronous change to a table.
Describes the physical storage of table data.
Specifies options related to data preview for viewing a sample of your data.
Defines column statistics supported for character sequence data values.
A structure specifying the dialect and dialect version used by the query engine.
An error record for table operations.
A structure that describes a target table for resource linking.
A structure used to define a table.
Contains details about an optimizer associated with a table.
Contains details on the configuration of a table optimizer. You pass this configuration when creating or updating a table optimizer.
Contains details for a table optimizer run.
An object that describes the VPC configuration for a table optimizer.
A structure containing information about the state of an asynchronous change to a table.
Specifies a version of a table.
An error record for table-version operations.
The resource properties associated with the integration target.
A structure that describes a target catalog for resource linking.
The target resource could not be found.
Properties used by the target leg to partition the data on the target.
The criteria that are used to filter the task runs for the machine learning transform.
The configuration properties for the task run.
The sorting criteria that are used to sort the list of task runs for the machine learning transform.
A structure that is used to specify testing a connection to a service.
The throttling threshhold was exceeded.
A timestamped inclusion annotation.
A timestamp filter.
Specifies the parameters in the config file of the dynamic transform.
The encryption-at-rest settings of the transform that apply to accessing user data. Machine learning transforms can access user data encrypted in Amazon S3 using KMS.
The criteria used to filter the machine learning transforms.
The algorithm-specific parameters that are associated with the machine learning transform.
The sorting criteria that are associated with the machine learning transform.
The details of a Trigger node present in the workflow.
A structure used to provide information used to update a trigger. This object updates the previous trigger definition by overwriting it completely.
A partition that contains unfiltered metadata.
Specifies a custom CSV classifier to be updated.
Specifies a grok classifier to update when passed to UpdateClassifier
.
Input parameters specific to updating Apache Iceberg tables in Glue Data Catalog, containing the update operations to be applied to an existing Iceberg table.
Contains the update operations to be applied to an existing Iceberg table in AWS Glue Data Catalog, defining the new state of the table metadata.
Specifies a JSON classifier to be updated.
Input parameters for updating open table format tables in GlueData Catalog, serving as a wrapper for format-specific update operations such as Apache Iceberg.
Specifies an XML classifier to be updated.
The options to configure an upsert operation when writing to a Redshift target .
Describes an Glue usage profile.
Represents the equivalent of a Hive user-defined function (UDF
) definition.
A structure used to create or update a user-defined function.
A value could not be validated.
There was a version conflict.
A structure containing details for representations.
A structure containing details for creating or updating an Glue view.
A structure that contains the dialect of the view, and the query that defines the view.
A structure containing details of a representation to update or create a Lake Formation view.
A structure that contains information for an analytical engine to validate a view, prior to persisting the view metadata. Used in the case of direct UpdateTable
or CreateTable
API calls.
A workflow graph represents the complete workflow containing all the Glue components present in the workflow and all the directed connections between them.
A workflow run is an execution of a workflow providing all the runtime information.
Workflow run statistics provides statistics about the workflow run.
A classifier for XML
content.