SpecialistOff.NET / Вопросы / Статьи / Фрагменты кода / Резюме / Метки / Помощь / Файлы
НазадМетки: airflow apache airflow
Airflow has limited support for Microsoft Azure: interfaces exist only for Azure Blob Storage and Azure Data Lake. Hook, Sensor and Operator for Blob Storage and Azure Data Lake Hook are in contrib section.
All classes communicate via the Window Azure Storage Blob protocol. Make sure that a Airflow connection of type wasb exists. Authorization can be done by supplying a login (=Storage account name) and password (=KEY), or login and SAS token in the extra field (see connection wasb_defaultfor an example).
airflow.contrib.hooks.wasb_hook.WasbHookInterface with Azure Blob Storage.
airflow.contrib.sensors.wasb_sensor.WasbBlobSensorChecks if a blob is present on Azure Blob storage.
airflow.contrib.operators.wasb_delete_blob_operator.WasbDeleteBlobOperatorDeletes blob(s) on Azure Blob Storage.
airflow.contrib.sensors.wasb_sensor.WasbPrefixSensorChecks if blobs matching a prefix are present on Azure Blob storage.
airflow.contrib.operators.file_to_wasb.FileToWasbOperatorUploads a local file to a container as a blob.
Cloud variant of a SMB file share. Make sure that a Airflow connection of type wasb exists. Authorization can be done by supplying a login (=Storage account name) and password (=Storage account key), or login and SAS token in the extra field (see connection wasb_default for an example).
airflow.contrib.hooks.azure_fileshare_hook.AzureFileShareHook:Interface with Azure File Share.
Airflow can be configured to read and write task logs in Azure Blob Storage. See Writing Logs to Azure Blob Storage.
AzureCosmosDBHook communicates via the Azure Cosmos library. Make sure that a Airflow connection of type azure_cosmos exists. Authorization can be done by supplying a login (=Endpoint uri), password (=secret key) and extra fields database_name and collection_name to specify the default database and collection to use (see connection azure_cosmos_default for an example).
airflow.contrib.hooks.azure_cosmos_hook.AzureCosmosDBHookInterface with Azure CosmosDB.
airflow.contrib.operators.azure_cosmos_operator.AzureCosmosInsertDocumentOperatorSimple operator to insert document into CosmosDB.
airflow.contrib.sensors.azure_cosmos_sensor.AzureCosmosDocumentSensorSimple sensor to detect document existence in CosmosDB.
AzureDataLakeHook communicates via a REST API compatible with WebHDFS. Make sure that a Airflow connection of type azure_data_lake exists. Authorization can be done by supplying a login (=Client ID), password (=Client Secret) and extra fields tenant (Tenant) and account_name (Account Name) (see connection azure_data_lake_default for an example).
airflow.contrib.hooks.azure_data_lake_hook.AzureDataLakeHookInterface with Azure Data Lake.
airflow.contrib.operators.adls_list_operator.AzureDataLakeStorageListOperatorLists the files located in a specified Azure Data Lake path.
airflow.contrib.operators.adls_to_gcs.AdlsToGoogleCloudStorageOperatorCopies files from an Azure Data Lake path to a Google Cloud Storage bucket.
Azure Container Instances provides a method to run a docker container without having to worry about managing infrastructure. The AzureContainerInstanceHook requires a service principal. The credentials for this principal can either be defined in the extra field key_path, as an environment variable named AZURE_AUTH_LOCATION, or by providing a login/password and tenantId in extras.
The AzureContainerRegistryHook requires a host/login/password to be defined in the connection.
airflow.contrib.hooks.azure_container_volume_hook.AzureContainerVolumeHookInterface with Azure Container Volumes
airflow.contrib.operators.azure_container_instances_operator.AzureContainerInstancesOperatorStart/Monitor a new ACI.
airflow.contrib.hooks.azure_container_instance_hook.AzureContainerInstanceHookWrapper around a single ACI.
airflow.contrib.hooks.azure_container_registry_hook.AzureContainerRegistryHookInterface with ACR
Airflow has extensive support for Amazon Web Services. But note that the Hooks, Sensors and Operators are in the contrib section.
airflow.contrib.hooks.emr_hook.EmrHookInterface with AWS EMR.
airflow.contrib.operators.emr_add_steps_operator.EmrAddStepsOperatorAdds steps to an existing EMR JobFlow.
airflow.contrib.operators.emr_create_job_flow_operator.EmrCreateJobFlowOperatorCreates an EMR JobFlow, reading the config from the EMR connection.
airflow.contrib.operators.emr_terminate_job_flow_operator.EmrTerminateJobFlowOperatorTerminates an EMR JobFlow.
airflow.hooks.S3_hook.S3HookInterface with AWS S3.
airflow.operators.s3_file_transform_operator.S3FileTransformOperatorCopies data from a source S3 location to a temporary location on the local filesystem.
airflow.contrib.operators.s3_list_operator.S3ListOperatorLists the files matching a key prefix from a S3 location.
airflow.contrib.operators.s3_to_gcs_operator.S3ToGoogleCloudStorageOperatorSyncs an S3 location with a Google Cloud Storage bucket.
airflow.contrib.operators.s3_to_gcs_transfer_operator.S3ToGoogleCloudStorageTransferOperatorSyncs an S3 bucket with a Google Cloud Storage bucket using the GCP Storage Transfer Service.
airflow.operators.s3_to_hive_operator.S3ToHiveTransferMoves data from S3 to Hive. The operator downloads a file from S3, stores the file locally before loading it into a Hive table.
airflow.contrib.operators.awsbatch_operator.AWSBatchOperatorExecute a task on AWS Batch Service.
airflow.contrib.sensors.aws_redshift_cluster_sensor.AwsRedshiftClusterSensorWaits for a Redshift cluster to reach a specific status.
airflow.contrib.hooks.redshift_hook.RedshiftHookInteract with AWS Redshift, using the boto3 library.
airflow.operators.redshift_to_s3_operator.RedshiftToS3TransferExecutes an unload command to S3 as CSV with or without headers.
airflow.operators.s3_to_redshift_operator.S3ToRedshiftTransferExecutes an copy command from S3 as CSV with or without headers.
airflow.contrib.operators.hive_to_dynamodb.HiveToDynamoDBTransferOperatorMoves data from Hive to DynamoDB.
airflow.contrib.hooks.aws_dynamodb_hook.AwsDynamoDBHookInterface with AWS DynamoDB.
airflow.contrib.hooks.aws_lambda_hook.AwsLambdaHookInterface with AWS Lambda.
airflow.contrib.hooks.aws_firehose_hook.AwsFirehoseHookInterface with AWS Kinesis Firehose.
airflow.contrib.hooks.sagemaker_hook.SageMakerHookInterface with Amazon SageMaker.
airflow.contrib.operators.sagemaker_training_operator.SageMakerTrainingOperatorCreate a SageMaker training job.
airflow.contrib.operators.sagemaker_tuning_operator.SageMakerTuningOperatorCreate a SageMaker tuning job.
airflow.contrib.operators.sagemaker_model_operator.SageMakerModelOperatorCreate a SageMaker model.
airflow.contrib.operators.sagemaker_transform_operator.SageMakerTransformOperatorCreate a SageMaker transform job.
airflow.contrib.operators.sagemaker_endpoint_config_operator.SageMakerEndpointConfigOperatorCreate a SageMaker endpoint config.
airflow.contrib.operators.sagemaker_endpoint_operator.SageMakerEndpointOperatorCreate a SageMaker endpoint.
Databricks has contributed an Airflow operator which enables submitting runs to the Databricks platform. Internally the operator talks to the api/2.0/jobs/runs/submit endpoint.
airflow.contrib.operators.databricks_operator.DatabricksSubmitRunOperatorSubmits a Spark job run to Databricks using the api/2.0/jobs/runs/submit API endpoint.
Airflow has extensive support for the Google Cloud Platform. But note that most Hooks and Operators are in the contrib section. Meaning that they have a beta status, meaning that they can have breaking changes between minor releases.
See the GCP connection type documentation to configure connections to GCP.
Airflow can be configured to read and write task logs in Google Cloud Storage. See Writing Logs to Google Cloud Storage.
All hooks is based on airflow.contrib.hooks.gcp_api_base_hook.GoogleCloudBaseHook.
airflow.contrib.operators.bigquery_check_operator.BigQueryCheckOperatorPerforms checks against a SQL query that will return a single row with different values.
airflow.contrib.operators.bigquery_check_operator.BigQueryIntervalCheckOperatorChecks that the values of metrics given as SQL expressions are within a certain tolerance of the ones from days_back before.
airflow.contrib.operators.bigquery_check_operator.BigQueryValueCheckOperatorPerforms a simple value check using SQL code.
airflow.contrib.operators.bigquery_get_data.BigQueryGetDataOperatorFetches the data from a BigQuery table and returns data in a python list
airflow.contrib.operators.bigquery_operator.BigQueryCreateEmptyDatasetOperatorCreates an empty BigQuery dataset.
airflow.contrib.operators.bigquery_operator.BigQueryCreateEmptyTableOperatorCreates a new, empty table in the specified BigQuery dataset optionally with schema.
airflow.contrib.operators.bigquery_operator.BigQueryCreateExternalTableOperatorCreates a new, external table in the dataset with the data in Google Cloud Storage.
airflow.contrib.operators.bigquery_operator.BigQueryDeleteDatasetOperatorDeletes an existing BigQuery dataset.
airflow.contrib.operators.bigquery_operator.BigQueryOperatorExecutes BigQuery SQL queries in a specific BigQuery database.
airflow.contrib.operators.bigquery_table_delete_operator.BigQueryTableDeleteOperatorDeletes an existing BigQuery table.
airflow.contrib.operators.bigquery_to_bigquery.BigQueryToBigQueryOperatorCopy a BigQuery table to another BigQuery table.
airflow.contrib.operators.bigquery_to_gcs.BigQueryToCloudStorageOperatorTransfers a BigQuery table to a Google Cloud Storage bucket
They also use airflow.contrib.hooks.bigquery_hook.BigQueryHook to communicate with Google Cloud Platform.
airflow.contrib.operators.gcp_spanner_operator.CloudSpannerInstanceDatabaseDeleteOperatordeletes an existing database from a Google Cloud Spanner instance or returns success if the database is missing.
airflow.contrib.operators.gcp_spanner_operator.CloudSpannerInstanceDatabaseDeployOperatorcreates a new database in a Google Cloud instance or returns success if the database already exists.
airflow.contrib.operators.gcp_spanner_operator.CloudSpannerInstanceDatabaseQueryOperatorexecutes an arbitrary DML query (INSERT, UPDATE, DELETE).
airflow.contrib.operators.gcp_spanner_operator.CloudSpannerInstanceDatabaseUpdateOperatorupdates the structure of a Google Cloud Spanner database.
airflow.contrib.operators.gcp_spanner_operator.CloudSpannerInstanceDeleteOperatordeletes a Google Cloud Spanner instance.
airflow.contrib.operators.gcp_spanner_operator.CloudSpannerInstanceDeployOperatorcreates a new Google Cloud Spanner instance, or if an instance with the same name exists, updates the instance.
They also use airflow.contrib.hooks.gcp_spanner_hook.CloudSpannerHook to communicate with Google Cloud Platform.
airflow.contrib.operators.gcp_sql_operator.CloudSqlInstanceCreateOperatorcreate a new Cloud SQL instance.
airflow.contrib.operators.gcp_sql_operator.CloudSqlInstanceDatabaseCreateOperatorcreates a new database inside a Cloud SQL instance.
airflow.contrib.operators.gcp_sql_operator.CloudSqlInstanceDatabaseDeleteOperatordeletes a database from a Cloud SQL instance.
airflow.contrib.operators.gcp_sql_operator.CloudSqlInstanceDatabasePatchOperatorupdates a database inside a Cloud SQL instance.
airflow.contrib.operators.gcp_sql_operator.CloudSqlInstanceDeleteOperatordelete a Cloud SQL instance.
airflow.contrib.operators.gcp_sql_operator.CloudSqlInstanceExportOperatorexports data from a Cloud SQL instance.
airflow.contrib.operators.gcp_sql_operator.CloudSqlInstanceImportOperatorimports data into a Cloud SQL instance.
airflow.contrib.operators.gcp_sql_operator.CloudSqlInstancePatchOperatorpatch a Cloud SQL instance.
airflow.contrib.operators.gcp_sql_operator.CloudSqlQueryOperatorrun query in a Cloud SQL instance.
They also use airflow.contrib.hooks.gcp_sql_hook.CloudSqlDatabaseHook and airflow.contrib.hooks.gcp_sql_hook.CloudSqlHook to communicate with Google Cloud Platform.
airflow.contrib.operators.gcp_bigtable_operator.BigtableClusterUpdateOperatorupdates the number of nodes in a Google Cloud Bigtable cluster.
airflow.contrib.operators.gcp_bigtable_operator.BigtableInstanceCreateOperatorcreates a Cloud Bigtable instance.
airflow.contrib.operators.gcp_bigtable_operator.BigtableInstanceDeleteOperatordeletes a Google Cloud Bigtable instance.
airflow.contrib.operators.gcp_bigtable_operator.BigtableTableCreateOperatorcreates a table in a Google Cloud Bigtable instance.
airflow.contrib.operators.gcp_bigtable_operator.BigtableTableDeleteOperatordeletes a table in a Google Cloud Bigtable instance.
airflow.contrib.operators.gcp_bigtable_operator.BigtableTableWaitForReplicationSensor(sensor) waits for a table to be fully replicated.
They also use airflow.contrib.hooks.gcp_bigtable_hook.BigtableHook to communicate with Google Cloud Platform.
airflow.contrib.operators.gcp_cloud_build_operator.CloudBuildCreateBuildOperatorStarts a build with the specified configuration.
They also use airflow.contrib.hooks.gcp_cloud_build_hook.CloudBuildHook to communicate with Google Cloud Platform.
airflow.contrib.operators.gcp_compute_operator.GceInstanceStartOperatorstart an existing Google Compute Engine instance.
airflow.contrib.operators.gcp_compute_operator.GceInstanceStopOperatorstop an existing Google Compute Engine instance.
airflow.contrib.operators.gcp_compute_operator.GceSetMachineTypeOperatorchange the machine type for a stopped instance.
airflow.contrib.operators.gcp_compute_operator.GceInstanceTemplateCopyOperatorcopy the Instance Template, applying specified changes.
airflow.contrib.operators.gcp_compute_operator.GceInstanceGroupManagerUpdateTemplateOperatorpatch the Instance Group Manager, replacing source Instance Template URL with the destination one.
The operators have the common base operator airflow.contrib.operators.gcp_compute_operator.GceBaseOperator
They also use airflow.contrib.hooks.gcp_compute_hook.GceHook to communicate with Google Cloud Platform.
airflow.contrib.operators.gcp_function_operator.GcfFunctionDeployOperatordeploy Google Cloud Function to Google Cloud Platform
airflow.contrib.operators.gcp_function_operator.GcfFunctionDeleteOperatordelete Google Cloud Function in Google Cloud Platform
They also use airflow.contrib.hooks.gcp_function_hook.GcfHook to communicate with Google Cloud Platform.
airflow.contrib.operators.dataflow_operator.DataFlowJavaOperatorlaunching Cloud Dataflow jobs written in Java.
airflow.contrib.operators.dataflow_operator.DataflowTemplateOperatorlaunching a templated Cloud DataFlow batch job.
airflow.contrib.operators.dataflow_operator.DataFlowPythonOperatorlaunching Cloud Dataflow jobs written in python.
They also use airflow.contrib.hooks.gcp_dataflow_hook.DataFlowHook to communicate with Google Cloud Platform.
airflow.contrib.operators.dataproc_operator.DataprocClusterCreateOperatorCreate a new cluster on Google Cloud Dataproc.
airflow.contrib.operators.dataproc_operator.DataprocClusterDeleteOperatorDelete a cluster on Google Cloud Dataproc.
airflow.contrib.operators.dataproc_operator.DataprocClusterScaleOperatorScale up or down a cluster on Google Cloud Dataproc.
airflow.contrib.operators.dataproc_operator.DataProcHadoopOperatorStart a Hadoop Job on a Cloud DataProc cluster.
airflow.contrib.operators.dataproc_operator.DataProcHiveOperatorStart a Hive query Job on a Cloud DataProc cluster.
airflow.contrib.operators.dataproc_operator.DataProcPigOperatorStart a Pig query Job on a Cloud DataProc cluster.
airflow.contrib.operators.dataproc_operator.DataProcPySparkOperatorStart a PySpark Job on a Cloud DataProc cluster.
airflow.contrib.operators.dataproc_operator.DataProcSparkOperatorStart a Spark Job on a Cloud DataProc cluster.
airflow.contrib.operators.dataproc_operator.DataProcSparkSqlOperatorStart a Spark SQL query Job on a Cloud DataProc cluster.
airflow.contrib.operators.dataproc_operator.DataprocWorkflowTemplateInstantiateInlineOperatorInstantiate a WorkflowTemplate Inline on Google Cloud Dataproc.
airflow.contrib.operators.dataproc_operator.DataprocWorkflowTemplateInstantiateOperatorInstantiate a WorkflowTemplate on Google Cloud Dataproc.
airflow.contrib.operators.datastore_export_operator.DatastoreExportOperatorExport entities from Google Cloud Datastore to Cloud Storage.
airflow.contrib.operators.datastore_import_operator.DatastoreImportOperatorImport entities from Cloud Storage to Google Cloud Datastore.
They also use airflow.contrib.hooks.datastore_hook.DatastoreHook to communicate with Google Cloud Platform.
airflow.contrib.operators.mlengine_operator.MLEngineBatchPredictionOperatorStart a Cloud ML Engine batch prediction job.
airflow.contrib.operators.mlengine_operator.MLEngineModelOperatorManages a Cloud ML Engine model.
airflow.contrib.operators.mlengine_operator.MLEngineTrainingOperatorStart a Cloud ML Engine training job.
airflow.contrib.operators.mlengine_operator.MLEngineVersionOperatorManages a Cloud ML Engine model version.
They also use airflow.contrib.hooks.gcp_mlengine_hook.MLEngineHook to communicate with Google Cloud Platform.
airflow.contrib.operators.file_to_gcs.FileToGoogleCloudStorageOperatorUploads a file to Google Cloud Storage.
airflow.contrib.operators.gcs_acl_operator.GoogleCloudStorageBucketCreateAclEntryOperatorCreates a new ACL entry on the specified bucket.
airflow.contrib.operators.gcs_acl_operator.GoogleCloudStorageObjectCreateAclEntryOperatorCreates a new ACL entry on the specified object.
airflow.contrib.operators.gcs_download_operator.GoogleCloudStorageDownloadOperatorDownloads a file from Google Cloud Storage.
airflow.contrib.operators.gcs_list_operator.GoogleCloudStorageListOperatorList all objects from the bucket with the give string prefix and delimiter in name.
airflow.contrib.operators.gcs_operator.GoogleCloudStorageCreateBucketOperatorCreates a new cloud storage bucket.
airflow.contrib.operators.gcs_to_bq.GoogleCloudStorageToBigQueryOperatorLoads files from Google cloud storage into BigQuery.
airflow.contrib.operators.gcs_to_gcs.GoogleCloudStorageToGoogleCloudStorageOperatorCopies objects from a bucket to another, with renaming if requested.
airflow.contrib.operators.mysql_to_gcs.MySqlToGoogleCloudStorageOperatorCopy data from any MySQL Database to Google cloud storage in JSON format.
airflow.contrib.operators.mssql_to_gcs.MsSqlToGoogleCloudStorageOperatorCopy data from any Microsoft SQL Server Database to Google Cloud Storage in JSON format.
airflow.contrib.sensors.gcs_sensor.GoogleCloudStorageObjectSensorChecks for the existence of a file in Google Cloud Storage.
airflow.contrib.sensors.gcs_sensor.GoogleCloudStorageObjectUpdatedSensorChecks if an object is updated in Google Cloud Storage.
airflow.contrib.sensors.gcs_sensor.GoogleCloudStoragePrefixSensorChecks for the existence of a objects at prefix in Google Cloud Storage.
airflow.contrib.sensors.gcs_sensor.GoogleCloudStorageUploadSessionCompleteSessionChecks for changes in the number of objects at prefix in Google Cloud Storage bucket and returns True if the inactivity period has passed with no increase in the number of objects for situations when many objects are being uploaded to a bucket with no formal success signal.
airflow.contrib.operators.gcs_delete_operator.GoogleCloudStorageDeleteOperatorDeletes objects from a Google Cloud Storage bucket.
They also use airflow.contrib.hooks.gcs_hook.GoogleCloudStorageHook to communicate with Google Cloud Platform.
airflow.contrib.operators.gcp_transfer_operator.GcpTransferServiceJobDeleteOperatorDeletes a transfer job.
airflow.contrib.operators.gcp_transfer_operator.GcpTransferServiceJobCreateOperatorCreates a transfer job.
airflow.contrib.operators.gcp_transfer_operator.GcpTransferServiceJobUpdateOperatorUpdates a transfer job.
airflow.contrib.operators.gcp_transfer_operator.GcpTransferServiceOperationCancelOperatorCancels a transfer operation.
airflow.contrib.operators.gcp_transfer_operator.GcpTransferServiceOperationGetOperatorGets a transfer operation.
airflow.contrib.operators.gcp_transfer_operator.GcpTransferServiceOperationPauseOperatorPauses a transfer operation
airflow.contrib.operators.gcp_transfer_operator.GcpTransferServiceOperationResumeOperatorResumes a transfer operation.
airflow.contrib.operators.gcp_transfer_operator.GcpTransferServiceOperationsListOperatorGets a list of transfer operations.
airflow.contrib.operators.gcp_transfer_operator.GoogleCloudStorageToGoogleCloudStorageTransferOperatorCopies objects from a Google Cloud Storage bucket to another bucket.
airflow.contrib.operators.gcp_transfer_operator.S3ToGoogleCloudStorageTransferOperatorSynchronizes an S3 bucket with a Google Cloud Storage bucket.
airflow.contrib.sensors.gcp_transfer_operator.GCPTransferServiceWaitForJobStatusSensorWaits for at least one operation belonging to the job to have the expected status.
They also use airflow.contrib.hooks.gcp_transfer_hook.GCPTransferServiceHook to communicate with Google Cloud Platform.
airflow.contrib.operators.gcp_vision_operator.CloudVisionAddProductToProductSetOperatorAdds a Product to the specified ProductSet.
airflow.contrib.operators.gcp_vision_operator.CloudVisionAnnotateImageOperatorRun image detection and annotation for an image.
airflow.contrib.operators.gcp_vision_operator.CloudVisionProductCreateOperatorCreates a new Product resource.
airflow.contrib.operators.gcp_vision_operator.CloudVisionProductDeleteOperatorPermanently deletes a product and its reference images.
airflow.contrib.operators.gcp_vision_operator.CloudVisionProductGetOperatorGets information associated with a Product.
airflow.contrib.operators.gcp_vision_operator.CloudVisionProductSetCreateOperatorCreates a new ProductSet resource.
airflow.contrib.operators.gcp_vision_operator.CloudVisionProductSetDeleteOperatorPermanently deletes a ProductSet.
airflow.contrib.operators.gcp_vision_operator.CloudVisionProductSetGetOperatorGets information associated with a ProductSet.
airflow.contrib.operators.gcp_vision_operator.CloudVisionProductSetUpdateOperatorMakes changes to a ProductSet resource.
airflow.contrib.operators.gcp_vision_operator.CloudVisionProductUpdateOperatorMakes changes to a Product resource.
airflow.contrib.operators.gcp_vision_operator.CloudVisionReferenceImageCreateOperatorCreates a new ReferenceImage resource.
airflow.contrib.operators.gcp_vision_operator.CloudVisionRemoveProductFromProductSetOperatorRemoves a Product from the specified ProductSet.
airflow.contrib.operators.gcp_vision_operator.CloudVisionAnnotateImageOperatorRun image detection and annotation for an image.
airflow.contrib.operators.gcp_vision_operator.CloudVisionDetectTextOperatorRun text detection for an image
airflow.contrib.operators.gcp_vision_operator.CloudVisionDetectDocumentTextOperatorRun document text detection for an image
airflow.contrib.operators.gcp_vision_operator.CloudVisionDetectImageLabelsOperatorRun image labels detection for an image
airflow.contrib.operators.gcp_vision_operator.CloudVisionDetectImageSafeSearchOperatorRun safe search detection for an image
They also use airflow.contrib.hooks.gcp_vision_hook.CloudVisionHook to communicate with Google Cloud Platform.
airflow.contrib.operators.gcp_text_to_speech_operator.GcpTextToSpeechSynthesizeOperatorSynthesizes input text into audio file and stores this file to GCS.
They also use airflow.contrib.hooks.gcp_text_to_speech_hook.GCPTextToSpeechHook to communicate with Google Cloud Platform.
airflow.contrib.operators.gcp_speech_to_text_operator.GcpSpeechToTextRecognizeSpeechOperatorRecognizes speech in audio input and returns text.
They also use airflow.contrib.hooks.gcp_speech_to_text_hook.GCPSpeechToTextHook to communicate with Google Cloud Platform.
airflow.contrib.operators.gcp_translate_speech_operator.GcpTranslateSpeechOperatorRecognizes speech in audio input and translates it.
airflow.contrib.hooks.gcp_speech_to_text_hook.GCPSpeechToTextHook andairflow.contrib.hooks.gcp_translate_hook.CloudTranslateHook to communicate with Google Cloud Platform.
airflow.contrib.operators.gcp_translate_operator.CloudTranslateTextOperatorTranslate a string or list of strings.
airflow.contrib.operators.gcp_video_intelligence_operator.CloudVideoIntelligenceDetectVideoLabelsOperatorPerforms video annotation, annotating video labels.
airflow.contrib.operators.gcp_video_intelligence_operator.CloudVideoIntelligenceDetectVideoExplicitContentOperatorPerforms video annotation, annotating explicit content.
airflow.contrib.operators.gcp_video_intelligence_operator.CloudVideoIntelligenceDetectVideoShotsOperatorPerforms video annotation, annotating video shots.
They also use airflow.contrib.hooks.gcp_video_intelligence_hook.CloudVideoIntelligenceHook to communicate with Google Cloud Platform.
airflow.contrib.operators.gcp_container_operator.GKEClusterCreateOperatorCreates a Kubernetes Cluster in Google Cloud Platform
airflow.contrib.operators.gcp_container_operator.GKEClusterDeleteOperatorDeletes a Kubernetes Cluster in Google Cloud Platform
airflow.contrib.operators.gcp_container_operator.GKEPodOperatorExecutes a task in a Kubernetes pod in the specified Google Kubernetes Engine cluster
They also use airflow.contrib.hooks.gcp_container_hook.GKEClusterHook to communicate with Google Cloud Platform.
airflow.contrib.operators.gcp_natural_language_operator.CloudLanguageAnalyzeEntitiesFinds named entities (currently proper names and common nouns) in the text along with entity types, salience, mentions for each entity, and other properties.
airflow.contrib.operators.gcp_natural_language_operator.CloudLanguageAnalyzeEntitySentimentFinds entities, similar to AnalyzeEntities in the text and analyzes sentiment associated with each entity and its mentions.
airflow.contrib.operators.gcp_natural_language_operator.CloudLanguageAnalyzeSentimentAnalyzes the sentiment of the provided text.
airflow.contrib.operators.gcp_natural_language_operator.CloudLanguageClassifyTextOperatorClassifies a document into categories.
They also use airflow.contrib.hooks.gcp_natural_language_operator.CloudNaturalLanguageHook to communicate with Google Cloud Platform.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPCancelDLPJobOperatorStarts asynchronous cancellation on a long-running DlpJob.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPCreateDeidentifyTemplateOperatorCreates a DeidentifyTemplate for re-using frequently used configuration for de-identifying content, images, and storage.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPCreateDLPJobOperatorCreates a new job to inspect storage or calculate risk metrics.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPCreateInspectTemplateOperatorCreates an InspectTemplate for re-using frequently used configuration for inspecting content, images, and storage.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPCreateJobTriggerOperatorCreates a job trigger to run DLP actions such as scanning storage for sensitive information on a set schedule.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPCreateStoredInfoTypeOperatorCreates a pre-built stored infoType to be used for inspection.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPDeidentifyContentOperatorDe-identifies potentially sensitive info from a ContentItem. This method has limits on input size and output size.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPDeleteDeidentifyTemplateOperatorDeletes a DeidentifyTemplate.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPDeleteDlpJobOperatorDeletes a long-running DlpJob. This method indicates that the client is no longer interested in the DlpJob result. The job will be cancelled if possible.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPDeleteInspectTemplateOperatorDeletes an InspectTemplate.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPDeleteJobTriggerOperatorDeletes a job trigger.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPDeleteStoredInfoTypeOperatorDeletes a stored infoType.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPGetDeidentifyTemplateOperatorGets a DeidentifyTemplate.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPGetDlpJobOperatorGets the latest state of a long-running DlpJob.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPGetInspectTemplateOperatorGets an InspectTemplate.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPGetJobTripperOperatorGets a job trigger.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPGetStoredInfoTypeOperatorGets a stored infoType.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPInspectContentOperatorFinds potentially sensitive info in content. This method has limits on input size, processing time, and output size.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPListDeidentifyTemplatesOperatorLists DeidentifyTemplates.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPListDlpJobsOperatorLists DlpJobs that match the specified filter in the request.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPListInfoTypesOperatorReturns a list of the sensitive information types that the DLP API supports.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPListInspectTemplatesOperatorLists InspectTemplates.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPListJobTriggersOperatorLists job triggers.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPListStoredInfoTypesOperatorLists stored infoTypes.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPRedactImageOperatorRedacts potentially sensitive info from an image. This method has limits on input size, processing time, and output size.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPReidentifyContentOperatorRe-identifies content that has been de-identified.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPUpdateDeidentifyTemplateOperatorUpdates the DeidentifyTemplate.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPUpdateInspectTemplateOperatorUpdates the InspectTemplate.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPUpdateJobTriggerOperatorUpdates a job trigger.
airflow.contrib.operators.gcp_dlp_operator.CloudDLPUpdateStoredInfoTypeOperatorUpdates the stored infoType by creating a new version.
They also use airflow.contrib.hooks.gcp_dlp_hook.CloudDLPHook to communicate with Google Cloud Platform.
Apache Airflow has a native operator and hooks to talk to Qubole, which lets you submit your big data jobs directly to Qubole from Apache Airflow.
airflow.contrib.operators.qubole_operator.QuboleOperatorExecute tasks (commands) on QDS (https://qubole.com).
airflow.contrib.sensors.qubole_sensor.QubolePartitionSensorWait for a Hive partition to show up in QHS (Qubole Hive Service) and check for its presence via QDS APIs
airflow.contrib.sensors.qubole_sensor.QuboleFileSensorWait for a file or folder to be present in cloud storage and check for its presence via QDS APIs
airflow.contrib.operators.qubole_check_operator.QuboleCheckOperatorPerforms checks against Qubole Commands. QuboleCheckOperator expects a command that will be executed on QDS.
airflow.contrib.operators.qubole_check_operator.QuboleValueCheckOperatorPerforms a simple value check using Qubole command. By default, each value on the first row of this Qubole command is compared with a pre-defined value