azure.mgmt.datalake.analytics.job.models module¶
-
class
azure.mgmt.datalake.analytics.job.models.
JobInnerError
(**kwargs)[source]¶ Bases:
msrest.serialization.Model
The Data Lake Analytics job error details.
Variables are only populated by the server, and will be ignored when sending a request.
- Variables
error_id (str) – The specific identifier for the type of error encountered in the job.
severity (str or SeverityTypes) – The severity level of the failure. Possible values include: ‘Warning’, ‘Error’, ‘Info’, ‘SevereWarning’, ‘Deprecated’, ‘UserWarning’
source (str) – The ultimate source of the failure (usually either SYSTEM or USER).
message (str) – The user friendly error message for the failure.
description (str) – The error message description.
details (str) – The details of the error message.
diagnostic_code (int) – The diagnostic error code.
component (str) – The component that failed.
resolution (str) – The recommended resolution for the failure, if any.
help_link (str) – The link to MSDN or Azure help for this type of error, if any.
internal_diagnostics (str) – The internal diagnostic stack trace if the user requesting the job error details has sufficient permissions it will be retrieved, otherwise it will be empty.
inner_error (JobInnerError) – The inner error of this specific job error message, if any.
-
class
azure.mgmt.datalake.analytics.job.models.
JobErrorDetails
(**kwargs)[source]¶ Bases:
msrest.serialization.Model
The Data Lake Analytics job error details.
Variables are only populated by the server, and will be ignored when sending a request.
- Variables
error_id (str) – The specific identifier for the type of error encountered in the job.
severity (str or SeverityTypes) – The severity level of the failure. Possible values include: ‘Warning’, ‘Error’, ‘Info’, ‘SevereWarning’, ‘Deprecated’, ‘UserWarning’
source (str) – The ultimate source of the failure (usually either SYSTEM or USER).
message (str) – The user friendly error message for the failure.
description (str) – The error message description.
details (str) – The details of the error message.
line_number (int) – The specific line number in the job where the error occured.
start_offset (int) – The start offset in the job where the error was found
end_offset (int) – The end offset in the job where the error was found.
resolution (str) – The recommended resolution for the failure, if any.
file_path (str) – The path to any supplemental error files, if any.
help_link (str) – The link to MSDN or Azure help for this type of error, if any.
internal_diagnostics (str) – The internal diagnostic stack trace if the user requesting the job error details has sufficient permissions it will be retrieved, otherwise it will be empty.
inner_error (JobInnerError) – The inner error of this specific job error message, if any.
-
class
azure.mgmt.datalake.analytics.job.models.
JobStateAuditRecord
(**kwargs)[source]¶ Bases:
msrest.serialization.Model
The Data Lake Analytics job state audit records for tracking the lifecycle of a job.
Variables are only populated by the server, and will be ignored when sending a request.
-
class
azure.mgmt.datalake.analytics.job.models.
JobProperties
(*, script: str, runtime_version: str = None, **kwargs)[source]¶ Bases:
msrest.serialization.Model
The common Data Lake Analytics job properties.
You probably want to use the sub-classes and not this class directly. Known sub-classes are: USqlJobProperties, HiveJobProperties, ScopeJobProperties
All required parameters must be populated in order to send to Azure.
-
class
azure.mgmt.datalake.analytics.job.models.
JobInformation
(*, name: str, type, properties, degree_of_parallelism: int = 1, priority: int = None, log_file_patterns=None, related=None, tags=None, **kwargs)[source]¶ Bases:
azure.mgmt.datalake.analytics.job.models.job_information_basic_py3.JobInformationBasic
The extended Data Lake Analytics job information properties returned when retrieving a specific job.
Variables are only populated by the server, and will be ignored when sending a request.
All required parameters must be populated in order to send to Azure.
- Variables
job_id (str) – The job’s unique identifier (a GUID).
submitter (str) – The user or account that submitted the job.
submit_time (datetime) – The time the job was submitted to the service.
start_time (datetime) – The start time of the job.
end_time (datetime) – The completion time of the job.
state (str or JobState) – The job state. When the job is in the Ended state, refer to Result and ErrorMessage for details. Possible values include: ‘Accepted’, ‘Compiling’, ‘Ended’, ‘New’, ‘Queued’, ‘Running’, ‘Scheduling’, ‘Starting’, ‘Paused’, ‘WaitingForCapacity’
result (str or JobResult) – The result of job execution or the current result of the running job. Possible values include: ‘None’, ‘Succeeded’, ‘Cancelled’, ‘Failed’
log_folder (str) – The log folder path to use in the following format: adl://<accountName>.azuredatalakestore.net/system/jobservice/jobs/Usql/2016/03/13/17/18/5fe51957-93bc-4de0-8ddc-c5a4753b068b/logs/.
error_message (list[JobErrorDetails]) – The error message details for the job, if the job failed.
state_audit_records (list[JobStateAuditRecord]) – The job state audit records, indicating when various operations have been performed on this job.
- Parameters
name (str) – Required. The friendly name of the job.
type (str or JobType) – Required. The job type of the current job (Hive, USql, or Scope (for internal use only)). Possible values include: ‘USql’, ‘Hive’, ‘Scope’
degree_of_parallelism (int) – The degree of parallelism used for this job. This must be greater than 0, if set to less than 0 it will default to 1. Default value: 1 .
priority (int) – The priority value for the current job. Lower numbers have a higher priority. By default, a job has a priority of 1000. This must be greater than 0.
log_file_patterns (list[str]) – The list of log file name patterns to find in the logFolder. ‘*’ is the only matching character allowed. Example format: jobExecution*.log or mylog.txt
related (JobRelationshipProperties) – The recurring job relationship information properties.
tags (dict[str, str]) – The key-value pairs used to add additional metadata to the job information. (Only for use internally with Scope job type.)
properties (JobProperties) – Required. The job specific properties.
-
class
azure.mgmt.datalake.analytics.job.models.
JobRelationshipProperties
(*, recurrence_id: str, pipeline_id: str = None, pipeline_name: str = None, pipeline_uri: str = None, run_id: str = None, recurrence_name: str = None, **kwargs)[source]¶ Bases:
msrest.serialization.Model
Job relationship information properties including pipeline information, correlation information, etc.
All required parameters must be populated in order to send to Azure.
- Parameters
pipeline_id (str) – The job relationship pipeline identifier (a GUID).
pipeline_name (str) – The friendly name of the job relationship pipeline, which does not need to be unique.
pipeline_uri (str) – The pipeline uri, unique, links to the originating service for this pipeline.
run_id (str) – The run identifier (a GUID), unique identifier of the iteration of this pipeline.
recurrence_id (str) – Required. The recurrence identifier (a GUID), unique per activity/script, regardless of iterations. This is something to link different occurrences of the same job together.
recurrence_name (str) – The recurrence name, user friendly name for the correlation between jobs.
-
class
azure.mgmt.datalake.analytics.job.models.
JobInformationBasic
(*, name: str, type, degree_of_parallelism: int = 1, priority: int = None, log_file_patterns=None, related=None, tags=None, **kwargs)[source]¶ Bases:
msrest.serialization.Model
The common Data Lake Analytics job information properties.
Variables are only populated by the server, and will be ignored when sending a request.
All required parameters must be populated in order to send to Azure.
- Variables
job_id (str) – The job’s unique identifier (a GUID).
submitter (str) – The user or account that submitted the job.
submit_time (datetime) – The time the job was submitted to the service.
start_time (datetime) – The start time of the job.
end_time (datetime) – The completion time of the job.
state (str or JobState) – The job state. When the job is in the Ended state, refer to Result and ErrorMessage for details. Possible values include: ‘Accepted’, ‘Compiling’, ‘Ended’, ‘New’, ‘Queued’, ‘Running’, ‘Scheduling’, ‘Starting’, ‘Paused’, ‘WaitingForCapacity’
result (str or JobResult) – The result of job execution or the current result of the running job. Possible values include: ‘None’, ‘Succeeded’, ‘Cancelled’, ‘Failed’
log_folder (str) – The log folder path to use in the following format: adl://<accountName>.azuredatalakestore.net/system/jobservice/jobs/Usql/2016/03/13/17/18/5fe51957-93bc-4de0-8ddc-c5a4753b068b/logs/.
- Parameters
name (str) – Required. The friendly name of the job.
type (str or JobType) – Required. The job type of the current job (Hive, USql, or Scope (for internal use only)). Possible values include: ‘USql’, ‘Hive’, ‘Scope’
degree_of_parallelism (int) – The degree of parallelism used for this job. This must be greater than 0, if set to less than 0 it will default to 1. Default value: 1 .
priority (int) – The priority value for the current job. Lower numbers have a higher priority. By default, a job has a priority of 1000. This must be greater than 0.
log_file_patterns (list[str]) – The list of log file name patterns to find in the logFolder. ‘*’ is the only matching character allowed. Example format: jobExecution*.log or mylog.txt
related (JobRelationshipProperties) – The recurring job relationship information properties.
tags (dict[str, str]) – The key-value pairs used to add additional metadata to the job information. (Only for use internally with Scope job type.)
-
class
azure.mgmt.datalake.analytics.job.models.
JobResource
(*, name: str = None, resource_path: str = None, type=None, **kwargs)[source]¶ Bases:
msrest.serialization.Model
The Data Lake Analytics job resources.
- Parameters
name (str) – The name of the resource.
resource_path (str) – The path to the resource.
type (str or JobResourceType) – The job resource type. Possible values include: ‘VertexResource’, ‘JobManagerResource’, ‘StatisticsResource’, ‘VertexResourceInUserFolder’, ‘JobManagerResourceInUserFolder’, ‘StatisticsResourceInUserFolder’
-
class
azure.mgmt.datalake.analytics.job.models.
JobStatisticsVertex
(**kwargs)[source]¶ Bases:
msrest.serialization.Model
The detailed information for a vertex.
Variables are only populated by the server, and will be ignored when sending a request.
- Variables
-
class
azure.mgmt.datalake.analytics.job.models.
ResourceUsageStatistics
(**kwargs)[source]¶ Bases:
msrest.serialization.Model
The statistics information for resource usage.
Variables are only populated by the server, and will be ignored when sending a request.
- Variables
average (float) – The average value.
minimum (long) – The minimum value.
maximum (long) – The maximum value.
-
class
azure.mgmt.datalake.analytics.job.models.
JobStatisticsVertexStage
(*, max_data_read_vertex=None, max_execution_time_vertex=None, max_peak_mem_usage_vertex=None, allocated_container_cpu_core_count=None, allocated_container_mem_size=None, used_vertex_cpu_core_count=None, used_vertex_peak_mem_size=None, **kwargs)[source]¶ Bases:
msrest.serialization.Model
The Data Lake Analytics job statistics vertex stage information.
Variables are only populated by the server, and will be ignored when sending a request.
- Variables
data_read (long) – The amount of data read, in bytes.
data_read_cross_pod (long) – The amount of data read across multiple pods, in bytes.
data_read_intra_pod (long) – The amount of data read in one pod, in bytes.
data_to_read (long) – The amount of data remaining to be read, in bytes.
data_written (long) – The amount of data written, in bytes.
duplicate_discard_count (int) – The number of duplicates that were discarded.
failed_count (int) – The number of failures that occured in this stage.
max_vertex_data_read (long) – The maximum amount of data read in a single vertex, in bytes.
min_vertex_data_read (long) – The minimum amount of data read in a single vertex, in bytes.
read_failure_count (int) – The number of read failures in this stage.
revocation_count (int) – The number of vertices that were revoked during this stage.
running_count (int) – The number of currently running vertices in this stage.
scheduled_count (int) – The number of currently scheduled vertices in this stage.
stage_name (str) – The name of this stage in job execution.
succeeded_count (int) – The number of vertices that succeeded in this stage.
temp_data_written (long) – The amount of temporary data written, in bytes.
total_count (int) – The total vertex count for this stage.
total_failed_time (timedelta) – The amount of time that failed vertices took up in this stage.
total_progress (int) – The current progress of this stage, as a percentage.
total_succeeded_time (timedelta) – The amount of time all successful vertices took in this stage.
total_peak_mem_usage (long) – The sum of the peak memory usage of all the vertices in the stage, in bytes.
total_execution_time (timedelta) – The sum of the total execution time of all the vertices in the stage.
estimated_vertex_cpu_core_count (int) – The estimated vertex CPU core count.
estimated_vertex_peak_cpu_core_count (int) – The estimated vertex peak CPU core count.
estimated_vertex_mem_size (long) – The estimated vertex memory size, in bytes.
- Parameters
max_data_read_vertex (JobStatisticsVertex) – the vertex with the maximum amount of data read.
max_execution_time_vertex (JobStatisticsVertex) – the vertex with the maximum execution time.
max_peak_mem_usage_vertex (JobStatisticsVertex) – the vertex with the maximum peak memory usage.
allocated_container_cpu_core_count (ResourceUsageStatistics) – The statistics information for the allocated container CPU core count.
allocated_container_mem_size (ResourceUsageStatistics) – The statistics information for the allocated container memory size.
used_vertex_cpu_core_count (ResourceUsageStatistics) – The statistics information for the used vertex CPU core count.
used_vertex_peak_mem_size (ResourceUsageStatistics) – The statistics information for the used vertex peak memory size.
-
class
azure.mgmt.datalake.analytics.job.models.
JobStatistics
(**kwargs)[source]¶ Bases:
msrest.serialization.Model
The Data Lake Analytics job execution statistics.
Variables are only populated by the server, and will be ignored when sending a request.
- Variables
last_update_time_utc (datetime) – The last update time for the statistics.
finalizing_time_utc (datetime) – The job finalizing start time.
stages (list[JobStatisticsVertexStage]) – The list of stages for the job.
-
class
azure.mgmt.datalake.analytics.job.models.
JobDataPath
(**kwargs)[source]¶ Bases:
msrest.serialization.Model
A Data Lake Analytics job data path item.
Variables are only populated by the server, and will be ignored when sending a request.
-
class
azure.mgmt.datalake.analytics.job.models.
Diagnostics
(**kwargs)[source]¶ Bases:
msrest.serialization.Model
Error diagnostic information for failed jobs.
Variables are only populated by the server, and will be ignored when sending a request.
- Variables
message (str) – The error message.
severity (str or SeverityTypes) – The severity of the error. Possible values include: ‘Warning’, ‘Error’, ‘Info’, ‘SevereWarning’, ‘Deprecated’, ‘UserWarning’
line_number (int) – The line number the error occured on.
column_number (int) – The column where the error occured.
start (int) – The starting index of the error.
end (int) – The ending index of the error.
-
class
azure.mgmt.datalake.analytics.job.models.
USqlJobProperties
(*, script: str, runtime_version: str = None, statistics=None, debug_data=None, **kwargs)[source]¶ Bases:
azure.mgmt.datalake.analytics.job.models.job_properties_py3.JobProperties
U-SQL job properties used when retrieving U-SQL jobs.
Variables are only populated by the server, and will be ignored when sending a request.
All required parameters must be populated in order to send to Azure.
- Parameters
runtime_version (str) – The runtime version of the Data Lake Analytics engine to use for the specific type of job being run.
script (str) – Required. The script to run. Please note that the maximum script size is 3 MB.
type (str) – Required. Constant filled by server.
statistics (JobStatistics) – The job specific statistics.
debug_data (JobDataPath) – The job specific debug data locations.
- Variables
resources (list[JobResource]) – The list of resources that are required by the job.
diagnostics (list[Diagnostics]) – The diagnostics for the job.
algebra_file_path (str) – The algebra file path after the job has completed.
total_compilation_time (timedelta) – The total time this job spent compiling. This value should not be set by the user and will be ignored if it is.
total_queued_time (timedelta) – The total time this job spent queued. This value should not be set by the user and will be ignored if it is.
total_running_time (timedelta) – The total time this job spent executing. This value should not be set by the user and will be ignored if it is.
total_paused_time (timedelta) – The total time this job spent paused. This value should not be set by the user and will be ignored if it is.
root_process_node_id (str) – The ID used to identify the job manager coordinating job execution. This value should not be set by the user and will be ignored if it is.
yarn_application_id (str) – The ID used to identify the yarn application executing the job. This value should not be set by the user and will be ignored if it is.
yarn_application_time_stamp (long) – The timestamp (in ticks) for the yarn application executing the job. This value should not be set by the user and will be ignored if it is.
compile_mode (str or CompileMode) – The specific compilation mode for the job used during execution. If this is not specified during submission, the server will determine the optimal compilation mode. Possible values include: ‘Semantic’, ‘Full’, ‘SingleBox’
-
class
azure.mgmt.datalake.analytics.job.models.
HiveJobProperties
(*, script: str, runtime_version: str = None, **kwargs)[source]¶ Bases:
azure.mgmt.datalake.analytics.job.models.job_properties_py3.JobProperties
Hive job properties used when retrieving Hive jobs.
Variables are only populated by the server, and will be ignored when sending a request.
All required parameters must be populated in order to send to Azure.
- Parameters
- Variables
logs_location (str) – The Hive logs location.
output_location (str) – The location of Hive job output files (both execution output and results).
statement_count (int) – The number of statements that will be run based on the script.
executed_statement_count (int) – The number of statements that have been run based on the script.
-
class
azure.mgmt.datalake.analytics.job.models.
ScopeJobResource
(*, name: str = None, path: str = None, **kwargs)[source]¶ Bases:
msrest.serialization.Model
The Scope job resources. (Only for use internally with Scope job type.).
-
class
azure.mgmt.datalake.analytics.job.models.
ScopeJobProperties
(*, script: str, runtime_version: str = None, notifier: str = None, **kwargs)[source]¶ Bases:
azure.mgmt.datalake.analytics.job.models.job_properties_py3.JobProperties
Scope job properties used when submitting and retrieving Scope jobs. (Only for use internally with Scope job type.).
Variables are only populated by the server, and will be ignored when sending a request.
All required parameters must be populated in order to send to Azure.
- Parameters
runtime_version (str) – The runtime version of the Data Lake Analytics engine to use for the specific type of job being run.
script (str) – Required. The script to run. Please note that the maximum script size is 3 MB.
type (str) – Required. Constant filled by server.
notifier (str) – The list of email addresses, separated by semi-colons, to notify when the job reaches a terminal state.
- Variables
resources (list[ScopeJobResource]) – The list of resources that are required by the job.
user_algebra_path (str) – The algebra file path after the job has completed.
total_compilation_time (timedelta) – The total time this job spent compiling. This value should not be set by the user and will be ignored if it is.
total_queued_time (timedelta) – The total time this job spent queued. This value should not be set by the user and will be ignored if it is.
total_running_time (timedelta) – The total time this job spent executing. This value should not be set by the user and will be ignored if it is.
total_paused_time (timedelta) – The total time this job spent paused. This value should not be set by the user and will be ignored if it is.
root_process_node_id (str) – The ID used to identify the job manager coordinating job execution. This value should not be set by the user and will be ignored if it is.
yarn_application_id (str) – The ID used to identify the yarn application executing the job. This value should not be set by the user and will be ignored if it is.
-
class
azure.mgmt.datalake.analytics.job.models.
JobPipelineRunInformation
(**kwargs)[source]¶ Bases:
msrest.serialization.Model
Run info for a specific job pipeline.
Variables are only populated by the server, and will be ignored when sending a request.
- Variables
run_id (str) – The run identifier of an instance of pipeline executions (a GUID).
last_submit_time (datetime) – The time this instance was last submitted.
-
class
azure.mgmt.datalake.analytics.job.models.
JobPipelineInformation
(**kwargs)[source]¶ Bases:
msrest.serialization.Model
Job Pipeline Information, showing the relationship of jobs and recurrences of those jobs in a pipeline.
Variables are only populated by the server, and will be ignored when sending a request.
- Variables
pipeline_id (str) – The job relationship pipeline identifier (a GUID).
pipeline_name (str) – The friendly name of the job relationship pipeline, which does not need to be unique.
pipeline_uri (str) – The pipeline uri, unique, links to the originating service for this pipeline.
num_jobs_failed (int) – The number of jobs in this pipeline that have failed.
num_jobs_canceled (int) – The number of jobs in this pipeline that have been canceled.
num_jobs_succeeded (int) – The number of jobs in this pipeline that have succeeded.
au_hours_failed (float) – The number of job execution hours that resulted in failed jobs.
au_hours_canceled (float) – The number of job execution hours that resulted in canceled jobs.
au_hours_succeeded (float) – The number of job execution hours that resulted in successful jobs.
last_submit_time (datetime) – The last time a job in this pipeline was submitted.
runs (list[JobPipelineRunInformation]) – The list of recurrence identifiers representing each run of this pipeline.
recurrences (list[str]) – The list of recurrence identifiers representing each run of this pipeline.
-
class
azure.mgmt.datalake.analytics.job.models.
JobRecurrenceInformation
(**kwargs)[source]¶ Bases:
msrest.serialization.Model
Recurrence job information for a specific recurrence.
Variables are only populated by the server, and will be ignored when sending a request.
- Variables
recurrence_id (str) – The recurrence identifier (a GUID), unique per activity/script, regardless of iterations. This is something to link different occurrences of the same job together.
recurrence_name (str) – The recurrence name, user friendly name for the correlation between jobs.
num_jobs_failed (int) – The number of jobs in this recurrence that have failed.
num_jobs_canceled (int) – The number of jobs in this recurrence that have been canceled.
num_jobs_succeeded (int) – The number of jobs in this recurrence that have succeeded.
au_hours_failed (float) – The number of job execution hours that resulted in failed jobs.
au_hours_canceled (float) – The number of job execution hours that resulted in canceled jobs.
au_hours_succeeded (float) – The number of job execution hours that resulted in successful jobs.
last_submit_time (datetime) – The last time a job in this recurrence was submitted.
-
class
azure.mgmt.datalake.analytics.job.models.
CreateJobProperties
(*, script: str, runtime_version: str = None, **kwargs)[source]¶ Bases:
msrest.serialization.Model
The common Data Lake Analytics job properties for job submission.
You probably want to use the sub-classes and not this class directly. Known sub-classes are: CreateUSqlJobProperties, CreateScopeJobProperties
All required parameters must be populated in order to send to Azure.
-
class
azure.mgmt.datalake.analytics.job.models.
BaseJobParameters
(*, type, properties, **kwargs)[source]¶ Bases:
msrest.serialization.Model
Data Lake Analytics Job Parameters base class for build and submit.
All required parameters must be populated in order to send to Azure.
- Parameters
type (str or JobType) – Required. The job type of the current job (Hive, USql, or Scope (for internal use only)). Possible values include: ‘USql’, ‘Hive’, ‘Scope’
properties (CreateJobProperties) – Required. The job specific properties.
-
class
azure.mgmt.datalake.analytics.job.models.
CreateJobParameters
(*, type, properties, name: str, degree_of_parallelism: int = 1, priority: int = None, log_file_patterns=None, related=None, **kwargs)[source]¶ Bases:
azure.mgmt.datalake.analytics.job.models.base_job_parameters_py3.BaseJobParameters
The parameters used to submit a new Data Lake Analytics job.
All required parameters must be populated in order to send to Azure.
- Parameters
type (str or JobType) – Required. The job type of the current job (Hive, USql, or Scope (for internal use only)). Possible values include: ‘USql’, ‘Hive’, ‘Scope’
properties (CreateJobProperties) – Required. The job specific properties.
name (str) – Required. The friendly name of the job to submit.
degree_of_parallelism (int) – The degree of parallelism to use for this job. This must be greater than 0, if set to less than 0 it will default to 1. Default value: 1 .
priority (int) – The priority value to use for the current job. Lower numbers have a higher priority. By default, a job has a priority of 1000. This must be greater than 0.
log_file_patterns (list[str]) – The list of log file name patterns to find in the logFolder. ‘*’ is the only matching character allowed. Example format: jobExecution*.log or mylog.txt
related (JobRelationshipProperties) – The recurring job relationship information properties.
-
class
azure.mgmt.datalake.analytics.job.models.
CreateScopeJobParameters
(*, type, properties, name: str, degree_of_parallelism: int = 1, priority: int = None, log_file_patterns=None, related=None, tags=None, **kwargs)[source]¶ Bases:
azure.mgmt.datalake.analytics.job.models.create_job_parameters_py3.CreateJobParameters
The parameters used to submit a new Data Lake Analytics Scope job. (Only for use internally with Scope job type.).
All required parameters must be populated in order to send to Azure.
- Parameters
type (str or JobType) – Required. The job type of the current job (Hive, USql, or Scope (for internal use only)). Possible values include: ‘USql’, ‘Hive’, ‘Scope’
properties (CreateJobProperties) – Required. The job specific properties.
name (str) – Required. The friendly name of the job to submit.
degree_of_parallelism (int) – The degree of parallelism to use for this job. This must be greater than 0, if set to less than 0 it will default to 1. Default value: 1 .
priority (int) – The priority value to use for the current job. Lower numbers have a higher priority. By default, a job has a priority of 1000. This must be greater than 0.
log_file_patterns (list[str]) – The list of log file name patterns to find in the logFolder. ‘*’ is the only matching character allowed. Example format: jobExecution*.log or mylog.txt
related (JobRelationshipProperties) – The recurring job relationship information properties.
tags (dict[str, str]) – The key-value pairs used to add additional metadata to the job information.
-
class
azure.mgmt.datalake.analytics.job.models.
CreateUSqlJobProperties
(*, script: str, runtime_version: str = None, compile_mode=None, **kwargs)[source]¶ Bases:
azure.mgmt.datalake.analytics.job.models.create_job_properties_py3.CreateJobProperties
U-SQL job properties used when submitting U-SQL jobs.
All required parameters must be populated in order to send to Azure.
- Parameters
runtime_version (str) – The runtime version of the Data Lake Analytics engine to use for the specific type of job being run.
script (str) – Required. The script to run. Please note that the maximum script size is 3 MB.
type (str) – Required. Constant filled by server.
compile_mode (str or CompileMode) – The specific compilation mode for the job used during execution. If this is not specified during submission, the server will determine the optimal compilation mode. Possible values include: ‘Semantic’, ‘Full’, ‘SingleBox’
-
class
azure.mgmt.datalake.analytics.job.models.
CreateScopeJobProperties
(*, script: str, runtime_version: str = None, resources=None, notifier: str = None, **kwargs)[source]¶ Bases:
azure.mgmt.datalake.analytics.job.models.create_job_properties_py3.CreateJobProperties
Scope job properties used when submitting Scope jobs. (Only for use internally with Scope job type.).
All required parameters must be populated in order to send to Azure.
- Parameters
runtime_version (str) – The runtime version of the Data Lake Analytics engine to use for the specific type of job being run.
script (str) – Required. The script to run. Please note that the maximum script size is 3 MB.
type (str) – Required. Constant filled by server.
resources (list[ScopeJobResource]) – The list of resources that are required by the job.
notifier (str) – The list of email addresses, separated by semi-colons, to notify when the job reaches a terminal state.
-
class
azure.mgmt.datalake.analytics.job.models.
BuildJobParameters
(*, type, properties, name: str = None, **kwargs)[source]¶ Bases:
azure.mgmt.datalake.analytics.job.models.base_job_parameters_py3.BaseJobParameters
The parameters used to build a new Data Lake Analytics job.
All required parameters must be populated in order to send to Azure.
- Parameters
type (str or JobType) – Required. The job type of the current job (Hive, USql, or Scope (for internal use only)). Possible values include: ‘USql’, ‘Hive’, ‘Scope’
properties (CreateJobProperties) – Required. The job specific properties.
name (str) – The friendly name of the job to build.
-
class
azure.mgmt.datalake.analytics.job.models.
UpdateJobParameters
(*, degree_of_parallelism: int = None, priority: int = None, tags=None, **kwargs)[source]¶ Bases:
msrest.serialization.Model
The parameters that can be used to update existing Data Lake Analytics job information properties. (Only for use internally with Scope job type.).
- Parameters
degree_of_parallelism (int) – The degree of parallelism used for this job. This must be greater than 0, if set to less than 0 it will default to 1.
priority (int) – The priority value for the current job. Lower numbers have a higher priority. By default, a job has a priority of 1000. This must be greater than 0.
tags (dict[str, str]) – The key-value pairs used to add additional metadata to the job information.
-
class
azure.mgmt.datalake.analytics.job.models.
JobInformationBasicPaged
(*args, **kwargs)[source]¶ Bases:
msrest.paging.Paged
A paging container for iterating over a list of
JobInformationBasic
objectBring async to Paging.
“async_command” is mandatory keyword argument for this mixin to work.
-
class
azure.mgmt.datalake.analytics.job.models.
JobPipelineInformationPaged
(*args, **kwargs)[source]¶ Bases:
msrest.paging.Paged
A paging container for iterating over a list of
JobPipelineInformation
objectBring async to Paging.
“async_command” is mandatory keyword argument for this mixin to work.
-
class
azure.mgmt.datalake.analytics.job.models.
JobRecurrenceInformationPaged
(*args, **kwargs)[source]¶ Bases:
msrest.paging.Paged
A paging container for iterating over a list of
JobRecurrenceInformation
objectBring async to Paging.
“async_command” is mandatory keyword argument for this mixin to work.
-
class
azure.mgmt.datalake.analytics.job.models.
SeverityTypes
[source]¶ -
An enumeration.
-
deprecated
= 'Deprecated'¶
-
error
= 'Error'¶
-
info
= 'Info'¶
-
severe_warning
= 'SevereWarning'¶
-
user_warning
= 'UserWarning'¶
-
warning
= 'Warning'¶
-
-
class
azure.mgmt.datalake.analytics.job.models.
JobType
[source]¶ -
An enumeration.
-
hive
= 'Hive'¶
-
scope
= 'Scope'¶
-
usql
= 'USql'¶
-
-
class
azure.mgmt.datalake.analytics.job.models.
JobState
[source]¶ -
An enumeration.
-
accepted
= 'Accepted'¶
-
compiling
= 'Compiling'¶
-
ended
= 'Ended'¶
-
new
= 'New'¶
-
paused
= 'Paused'¶
-
queued
= 'Queued'¶
-
running
= 'Running'¶
-
scheduling
= 'Scheduling'¶
-
starting
= 'Starting'¶
-
waiting_for_capacity
= 'WaitingForCapacity'¶
-
-
class
azure.mgmt.datalake.analytics.job.models.
JobResult
[source]¶ -
An enumeration.
-
cancelled
= 'Cancelled'¶
-
failed
= 'Failed'¶
-
none
= 'None'¶
-
succeeded
= 'Succeeded'¶
-
-
class
azure.mgmt.datalake.analytics.job.models.
JobResourceType
[source]¶ -
An enumeration.
-
job_manager_resource
= 'JobManagerResource'¶
-
job_manager_resource_in_user_folder
= 'JobManagerResourceInUserFolder'¶
-
statistics_resource
= 'StatisticsResource'¶
-
statistics_resource_in_user_folder
= 'StatisticsResourceInUserFolder'¶
-
vertex_resource
= 'VertexResource'¶
-
vertex_resource_in_user_folder
= 'VertexResourceInUserFolder'¶
-