azure.synapse.spark.models package¶
-
class
azure.synapse.spark.models.
SparkBatchJob
(*, id: int, livy_info: Optional[SparkBatchJobState] = None, name: Optional[str] = None, workspace_name: Optional[str] = None, spark_pool_name: Optional[str] = None, submitter_name: Optional[str] = None, submitter_id: Optional[str] = None, artifact_id: Optional[str] = None, job_type: Union[str, SparkJobType, None] = None, result: Union[str, SparkBatchJobResultType, None] = None, scheduler: Optional[SparkScheduler] = None, plugin: Optional[SparkServicePlugin] = None, errors: Optional[List[SparkServiceError]] = None, tags: Optional[Dict[str, str]] = None, app_id: Optional[str] = None, app_info: Optional[Dict[str, str]] = None, state: Optional[str] = None, log_lines: Optional[List[str]] = None, **kwargs)[source]¶ SparkBatchJob.
All required parameters must be populated in order to send to Azure.
- Parameters
livy_info (SparkBatchJobState) –
name (str) – The batch name.
workspace_name (str) – The workspace name.
spark_pool_name (str) – The Spark pool name.
submitter_name (str) – The submitter name.
submitter_id (str) – The submitter identifier.
artifact_id (str) – The artifact identifier.
job_type (str or SparkJobType) – The job type. Possible values include: “SparkBatch”, “SparkSession”.
result (str or SparkBatchJobResultType) – The Spark batch job result. Possible values include: “Uncertain”, “Succeeded”, “Failed”, “Cancelled”.
scheduler (SparkScheduler) – The scheduler information.
plugin (SparkServicePlugin) – The plugin information.
errors (list[SparkServiceError]) – The error information.
id (int) – Required. The session Id.
app_id (str) – The application id of this session.
state (str) – The batch state.
-
as_dict
(keep_readonly=True, key_transformer=<function attribute_transformer>, **kwargs)¶ Return a dict that can be JSONify using json.dump.
Advanced usage might optionaly use a callback as parameter:
Key is the attribute name used in Python. Attr_desc is a dict of metadata. Currently contains ‘type’ with the msrest type and ‘key’ with the RestAPI encoded key. Value is the current value in this object.
The string returned will be used to serialize the key. If the return type is a list, this is considered hierarchical result dict.
See the three examples in this file:
attribute_transformer
full_restapi_key_transformer
last_restapi_key_transformer
If you want XML serialization, you can pass the kwargs is_xml=True.
- Parameters
key_transformer (function) – A key transformer function.
- Returns
A dict JSON compatible object
- Return type
-
classmethod
deserialize
(data, content_type=None)¶ Parse a str using the RestAPI syntax and return a model.
-
classmethod
enable_additional_properties_sending
()¶
-
classmethod
from_dict
(data, key_extractors=None, content_type=None)¶ Parse a dict using given key extractor return a model.
By default consider key extractors (rest_key_case_insensitive_extractor, attribute_key_case_insensitive_extractor and last_rest_key_case_insensitive_extractor)
-
classmethod
is_xml_model
()¶
-
serialize
(keep_readonly=False, **kwargs)¶ Return the JSON that would be sent to azure from this model.
This is an alias to as_dict(full_restapi_key_transformer, keep_readonly=False).
If you want XML serialization, you can pass the kwargs is_xml=True.
-
class
azure.synapse.spark.models.
SparkBatchJobCollection
(*, from_property: int, total: int, sessions: Optional[List[SparkBatchJob]] = None, **kwargs)[source]¶ Response for batch list operation.
All required parameters must be populated in order to send to Azure.
- Parameters
from_property (int) – Required. The start index of fetched sessions.
total (int) – Required. Number of sessions fetched.
sessions (list[SparkBatchJob]) – Batch list.
-
as_dict
(keep_readonly=True, key_transformer=<function attribute_transformer>, **kwargs)¶ Return a dict that can be JSONify using json.dump.
Advanced usage might optionaly use a callback as parameter:
Key is the attribute name used in Python. Attr_desc is a dict of metadata. Currently contains ‘type’ with the msrest type and ‘key’ with the RestAPI encoded key. Value is the current value in this object.
The string returned will be used to serialize the key. If the return type is a list, this is considered hierarchical result dict.
See the three examples in this file:
attribute_transformer
full_restapi_key_transformer
last_restapi_key_transformer
If you want XML serialization, you can pass the kwargs is_xml=True.
- Parameters
key_transformer (function) – A key transformer function.
- Returns
A dict JSON compatible object
- Return type
-
classmethod
deserialize
(data, content_type=None)¶ Parse a str using the RestAPI syntax and return a model.
-
classmethod
enable_additional_properties_sending
()¶
-
classmethod
from_dict
(data, key_extractors=None, content_type=None)¶ Parse a dict using given key extractor return a model.
By default consider key extractors (rest_key_case_insensitive_extractor, attribute_key_case_insensitive_extractor and last_rest_key_case_insensitive_extractor)
-
classmethod
is_xml_model
()¶
-
serialize
(keep_readonly=False, **kwargs)¶ Return the JSON that would be sent to azure from this model.
This is an alias to as_dict(full_restapi_key_transformer, keep_readonly=False).
If you want XML serialization, you can pass the kwargs is_xml=True.
-
class
azure.synapse.spark.models.
SparkBatchJobOptions
(*, name: str, file: str, tags: Optional[Dict[str, str]] = None, artifact_id: Optional[str] = None, class_name: Optional[str] = None, arguments: Optional[List[str]] = None, jars: Optional[List[str]] = None, python_files: Optional[List[str]] = None, files: Optional[List[str]] = None, archives: Optional[List[str]] = None, configuration: Optional[Dict[str, str]] = None, driver_memory: Optional[str] = None, driver_cores: Optional[int] = None, executor_memory: Optional[str] = None, executor_cores: Optional[int] = None, executor_count: Optional[int] = None, **kwargs)[source]¶ SparkBatchJobOptions.
All required parameters must be populated in order to send to Azure.
- Parameters
-
as_dict
(keep_readonly=True, key_transformer=<function attribute_transformer>, **kwargs)¶ Return a dict that can be JSONify using json.dump.
Advanced usage might optionaly use a callback as parameter:
Key is the attribute name used in Python. Attr_desc is a dict of metadata. Currently contains ‘type’ with the msrest type and ‘key’ with the RestAPI encoded key. Value is the current value in this object.
The string returned will be used to serialize the key. If the return type is a list, this is considered hierarchical result dict.
See the three examples in this file:
attribute_transformer
full_restapi_key_transformer
last_restapi_key_transformer
If you want XML serialization, you can pass the kwargs is_xml=True.
- Parameters
key_transformer (function) – A key transformer function.
- Returns
A dict JSON compatible object
- Return type
-
classmethod
deserialize
(data, content_type=None)¶ Parse a str using the RestAPI syntax and return a model.
-
classmethod
enable_additional_properties_sending
()¶
-
classmethod
from_dict
(data, key_extractors=None, content_type=None)¶ Parse a dict using given key extractor return a model.
By default consider key extractors (rest_key_case_insensitive_extractor, attribute_key_case_insensitive_extractor and last_rest_key_case_insensitive_extractor)
-
classmethod
is_xml_model
()¶
-
serialize
(keep_readonly=False, **kwargs)¶ Return the JSON that would be sent to azure from this model.
This is an alias to as_dict(full_restapi_key_transformer, keep_readonly=False).
If you want XML serialization, you can pass the kwargs is_xml=True.
-
class
azure.synapse.spark.models.
SparkBatchJobState
(*, not_started_at: Optional[datetime.datetime] = None, starting_at: Optional[datetime.datetime] = None, running_at: Optional[datetime.datetime] = None, dead_at: Optional[datetime.datetime] = None, success_at: Optional[datetime.datetime] = None, terminated_at: Optional[datetime.datetime] = None, recovering_at: Optional[datetime.datetime] = None, current_state: Optional[str] = None, job_creation_request: Optional[SparkRequest] = None, **kwargs)[source]¶ SparkBatchJobState.
- Parameters
not_started_at (datetime) – the time that at which “not_started” livy state was first seen.
starting_at (datetime) – the time that at which “starting” livy state was first seen.
running_at (datetime) – the time that at which “running” livy state was first seen.
dead_at (datetime) – time that at which “dead” livy state was first seen.
success_at (datetime) – the time that at which “success” livy state was first seen.
terminated_at (datetime) – the time that at which “killed” livy state was first seen.
recovering_at (datetime) – the time that at which “recovering” livy state was first seen.
current_state (str) – the Spark job state.
job_creation_request (SparkRequest) –
-
as_dict
(keep_readonly=True, key_transformer=<function attribute_transformer>, **kwargs)¶ Return a dict that can be JSONify using json.dump.
Advanced usage might optionaly use a callback as parameter:
Key is the attribute name used in Python. Attr_desc is a dict of metadata. Currently contains ‘type’ with the msrest type and ‘key’ with the RestAPI encoded key. Value is the current value in this object.
The string returned will be used to serialize the key. If the return type is a list, this is considered hierarchical result dict.
See the three examples in this file:
attribute_transformer
full_restapi_key_transformer
last_restapi_key_transformer
If you want XML serialization, you can pass the kwargs is_xml=True.
- Parameters
key_transformer (function) – A key transformer function.
- Returns
A dict JSON compatible object
- Return type
-
classmethod
deserialize
(data, content_type=None)¶ Parse a str using the RestAPI syntax and return a model.
-
classmethod
enable_additional_properties_sending
()¶
-
classmethod
from_dict
(data, key_extractors=None, content_type=None)¶ Parse a dict using given key extractor return a model.
By default consider key extractors (rest_key_case_insensitive_extractor, attribute_key_case_insensitive_extractor and last_rest_key_case_insensitive_extractor)
-
classmethod
is_xml_model
()¶
-
serialize
(keep_readonly=False, **kwargs)¶ Return the JSON that would be sent to azure from this model.
This is an alias to as_dict(full_restapi_key_transformer, keep_readonly=False).
If you want XML serialization, you can pass the kwargs is_xml=True.
-
class
azure.synapse.spark.models.
SparkRequest
(*, name: Optional[str] = None, file: Optional[str] = None, class_name: Optional[str] = None, arguments: Optional[List[str]] = None, jars: Optional[List[str]] = None, python_files: Optional[List[str]] = None, files: Optional[List[str]] = None, archives: Optional[List[str]] = None, configuration: Optional[Dict[str, str]] = None, driver_memory: Optional[str] = None, driver_cores: Optional[int] = None, executor_memory: Optional[str] = None, executor_cores: Optional[int] = None, executor_count: Optional[int] = None, **kwargs)[source]¶ SparkRequest.
- Parameters
-
as_dict
(keep_readonly=True, key_transformer=<function attribute_transformer>, **kwargs)¶ Return a dict that can be JSONify using json.dump.
Advanced usage might optionaly use a callback as parameter:
Key is the attribute name used in Python. Attr_desc is a dict of metadata. Currently contains ‘type’ with the msrest type and ‘key’ with the RestAPI encoded key. Value is the current value in this object.
The string returned will be used to serialize the key. If the return type is a list, this is considered hierarchical result dict.
See the three examples in this file:
attribute_transformer
full_restapi_key_transformer
last_restapi_key_transformer
If you want XML serialization, you can pass the kwargs is_xml=True.
- Parameters
key_transformer (function) – A key transformer function.
- Returns
A dict JSON compatible object
- Return type
-
classmethod
deserialize
(data, content_type=None)¶ Parse a str using the RestAPI syntax and return a model.
-
classmethod
enable_additional_properties_sending
()¶
-
classmethod
from_dict
(data, key_extractors=None, content_type=None)¶ Parse a dict using given key extractor return a model.
By default consider key extractors (rest_key_case_insensitive_extractor, attribute_key_case_insensitive_extractor and last_rest_key_case_insensitive_extractor)
-
classmethod
is_xml_model
()¶
-
serialize
(keep_readonly=False, **kwargs)¶ Return the JSON that would be sent to azure from this model.
This is an alias to as_dict(full_restapi_key_transformer, keep_readonly=False).
If you want XML serialization, you can pass the kwargs is_xml=True.
-
class
azure.synapse.spark.models.
SparkScheduler
(*, submitted_at: Optional[datetime.datetime] = None, scheduled_at: Optional[datetime.datetime] = None, ended_at: Optional[datetime.datetime] = None, cancellation_requested_at: Optional[datetime.datetime] = None, current_state: Union[str, SchedulerCurrentState, None] = None, **kwargs)[source]¶ SparkScheduler.
- Parameters
-
as_dict
(keep_readonly=True, key_transformer=<function attribute_transformer>, **kwargs)¶ Return a dict that can be JSONify using json.dump.
Advanced usage might optionaly use a callback as parameter:
Key is the attribute name used in Python. Attr_desc is a dict of metadata. Currently contains ‘type’ with the msrest type and ‘key’ with the RestAPI encoded key. Value is the current value in this object.
The string returned will be used to serialize the key. If the return type is a list, this is considered hierarchical result dict.
See the three examples in this file:
attribute_transformer
full_restapi_key_transformer
last_restapi_key_transformer
If you want XML serialization, you can pass the kwargs is_xml=True.
- Parameters
key_transformer (function) – A key transformer function.
- Returns
A dict JSON compatible object
- Return type
-
classmethod
deserialize
(data, content_type=None)¶ Parse a str using the RestAPI syntax and return a model.
-
classmethod
enable_additional_properties_sending
()¶
-
classmethod
from_dict
(data, key_extractors=None, content_type=None)¶ Parse a dict using given key extractor return a model.
By default consider key extractors (rest_key_case_insensitive_extractor, attribute_key_case_insensitive_extractor and last_rest_key_case_insensitive_extractor)
-
classmethod
is_xml_model
()¶
-
serialize
(keep_readonly=False, **kwargs)¶ Return the JSON that would be sent to azure from this model.
This is an alias to as_dict(full_restapi_key_transformer, keep_readonly=False).
If you want XML serialization, you can pass the kwargs is_xml=True.
-
class
azure.synapse.spark.models.
SparkServiceError
(*, message: Optional[str] = None, error_code: Optional[str] = None, source: Union[str, SparkErrorSource, None] = None, **kwargs)[source]¶ SparkServiceError.
- Parameters
message (str) –
error_code (str) –
source (str or SparkErrorSource) – Possible values include: “System”, “User”, “Unknown”, “Dependency”.
-
as_dict
(keep_readonly=True, key_transformer=<function attribute_transformer>, **kwargs)¶ Return a dict that can be JSONify using json.dump.
Advanced usage might optionaly use a callback as parameter:
Key is the attribute name used in Python. Attr_desc is a dict of metadata. Currently contains ‘type’ with the msrest type and ‘key’ with the RestAPI encoded key. Value is the current value in this object.
The string returned will be used to serialize the key. If the return type is a list, this is considered hierarchical result dict.
See the three examples in this file:
attribute_transformer
full_restapi_key_transformer
last_restapi_key_transformer
If you want XML serialization, you can pass the kwargs is_xml=True.
- Parameters
key_transformer (function) – A key transformer function.
- Returns
A dict JSON compatible object
- Return type
-
classmethod
deserialize
(data, content_type=None)¶ Parse a str using the RestAPI syntax and return a model.
-
classmethod
enable_additional_properties_sending
()¶
-
classmethod
from_dict
(data, key_extractors=None, content_type=None)¶ Parse a dict using given key extractor return a model.
By default consider key extractors (rest_key_case_insensitive_extractor, attribute_key_case_insensitive_extractor and last_rest_key_case_insensitive_extractor)
-
classmethod
is_xml_model
()¶
-
serialize
(keep_readonly=False, **kwargs)¶ Return the JSON that would be sent to azure from this model.
This is an alias to as_dict(full_restapi_key_transformer, keep_readonly=False).
If you want XML serialization, you can pass the kwargs is_xml=True.
-
class
azure.synapse.spark.models.
SparkServicePlugin
(*, preparation_started_at: Optional[datetime.datetime] = None, resource_acquisition_started_at: Optional[datetime.datetime] = None, submission_started_at: Optional[datetime.datetime] = None, monitoring_started_at: Optional[datetime.datetime] = None, cleanup_started_at: Optional[datetime.datetime] = None, current_state: Union[str, PluginCurrentState, None] = None, **kwargs)[source]¶ SparkServicePlugin.
- Parameters
preparation_started_at (datetime) –
resource_acquisition_started_at (datetime) –
submission_started_at (datetime) –
monitoring_started_at (datetime) –
cleanup_started_at (datetime) –
current_state (str or PluginCurrentState) – Possible values include: “Preparation”, “ResourceAcquisition”, “Queued”, “Submission”, “Monitoring”, “Cleanup”, “Ended”.
-
as_dict
(keep_readonly=True, key_transformer=<function attribute_transformer>, **kwargs)¶ Return a dict that can be JSONify using json.dump.
Advanced usage might optionaly use a callback as parameter:
Key is the attribute name used in Python. Attr_desc is a dict of metadata. Currently contains ‘type’ with the msrest type and ‘key’ with the RestAPI encoded key. Value is the current value in this object.
The string returned will be used to serialize the key. If the return type is a list, this is considered hierarchical result dict.
See the three examples in this file:
attribute_transformer
full_restapi_key_transformer
last_restapi_key_transformer
If you want XML serialization, you can pass the kwargs is_xml=True.
- Parameters
key_transformer (function) – A key transformer function.
- Returns
A dict JSON compatible object
- Return type
-
classmethod
deserialize
(data, content_type=None)¶ Parse a str using the RestAPI syntax and return a model.
-
classmethod
enable_additional_properties_sending
()¶
-
classmethod
from_dict
(data, key_extractors=None, content_type=None)¶ Parse a dict using given key extractor return a model.
By default consider key extractors (rest_key_case_insensitive_extractor, attribute_key_case_insensitive_extractor and last_rest_key_case_insensitive_extractor)
-
classmethod
is_xml_model
()¶
-
serialize
(keep_readonly=False, **kwargs)¶ Return the JSON that would be sent to azure from this model.
This is an alias to as_dict(full_restapi_key_transformer, keep_readonly=False).
If you want XML serialization, you can pass the kwargs is_xml=True.
-
class
azure.synapse.spark.models.
SparkSession
(*, id: int, livy_info: Optional[SparkSessionState] = None, name: Optional[str] = None, workspace_name: Optional[str] = None, spark_pool_name: Optional[str] = None, submitter_name: Optional[str] = None, submitter_id: Optional[str] = None, artifact_id: Optional[str] = None, job_type: Union[str, SparkJobType, None] = None, result: Union[str, SparkSessionResultType, None] = None, scheduler: Optional[SparkScheduler] = None, plugin: Optional[SparkServicePlugin] = None, errors: Optional[List[SparkServiceError]] = None, tags: Optional[Dict[str, str]] = None, app_id: Optional[str] = None, app_info: Optional[Dict[str, str]] = None, state: Optional[str] = None, log_lines: Optional[List[str]] = None, **kwargs)[source]¶ SparkSession.
All required parameters must be populated in order to send to Azure.
- Parameters
livy_info (SparkSessionState) –
name (str) –
workspace_name (str) –
spark_pool_name (str) –
submitter_name (str) –
submitter_id (str) –
artifact_id (str) –
job_type (str or SparkJobType) – The job type. Possible values include: “SparkBatch”, “SparkSession”.
result (str or SparkSessionResultType) – Possible values include: “Uncertain”, “Succeeded”, “Failed”, “Cancelled”.
scheduler (SparkScheduler) –
plugin (SparkServicePlugin) –
errors (list[SparkServiceError]) –
tags (dict[str, str]) – A set of tags. Dictionary of
<string>
.id (int) – Required.
app_id (str) –
state (str) –
-
as_dict
(keep_readonly=True, key_transformer=<function attribute_transformer>, **kwargs)¶ Return a dict that can be JSONify using json.dump.
Advanced usage might optionaly use a callback as parameter:
Key is the attribute name used in Python. Attr_desc is a dict of metadata. Currently contains ‘type’ with the msrest type and ‘key’ with the RestAPI encoded key. Value is the current value in this object.
The string returned will be used to serialize the key. If the return type is a list, this is considered hierarchical result dict.
See the three examples in this file:
attribute_transformer
full_restapi_key_transformer
last_restapi_key_transformer
If you want XML serialization, you can pass the kwargs is_xml=True.
- Parameters
key_transformer (function) – A key transformer function.
- Returns
A dict JSON compatible object
- Return type
-
classmethod
deserialize
(data, content_type=None)¶ Parse a str using the RestAPI syntax and return a model.
-
classmethod
enable_additional_properties_sending
()¶
-
classmethod
from_dict
(data, key_extractors=None, content_type=None)¶ Parse a dict using given key extractor return a model.
By default consider key extractors (rest_key_case_insensitive_extractor, attribute_key_case_insensitive_extractor and last_rest_key_case_insensitive_extractor)
-
classmethod
is_xml_model
()¶
-
serialize
(keep_readonly=False, **kwargs)¶ Return the JSON that would be sent to azure from this model.
This is an alias to as_dict(full_restapi_key_transformer, keep_readonly=False).
If you want XML serialization, you can pass the kwargs is_xml=True.
-
class
azure.synapse.spark.models.
SparkSessionCollection
(*, from_property: int, total: int, sessions: Optional[List[SparkSession]] = None, **kwargs)[source]¶ SparkSessionCollection.
All required parameters must be populated in order to send to Azure.
- Parameters
from_property (int) – Required.
total (int) – Required.
sessions (list[SparkSession]) –
-
as_dict
(keep_readonly=True, key_transformer=<function attribute_transformer>, **kwargs)¶ Return a dict that can be JSONify using json.dump.
Advanced usage might optionaly use a callback as parameter:
Key is the attribute name used in Python. Attr_desc is a dict of metadata. Currently contains ‘type’ with the msrest type and ‘key’ with the RestAPI encoded key. Value is the current value in this object.
The string returned will be used to serialize the key. If the return type is a list, this is considered hierarchical result dict.
See the three examples in this file:
attribute_transformer
full_restapi_key_transformer
last_restapi_key_transformer
If you want XML serialization, you can pass the kwargs is_xml=True.
- Parameters
key_transformer (function) – A key transformer function.
- Returns
A dict JSON compatible object
- Return type
-
classmethod
deserialize
(data, content_type=None)¶ Parse a str using the RestAPI syntax and return a model.
-
classmethod
enable_additional_properties_sending
()¶
-
classmethod
from_dict
(data, key_extractors=None, content_type=None)¶ Parse a dict using given key extractor return a model.
By default consider key extractors (rest_key_case_insensitive_extractor, attribute_key_case_insensitive_extractor and last_rest_key_case_insensitive_extractor)
-
classmethod
is_xml_model
()¶
-
serialize
(keep_readonly=False, **kwargs)¶ Return the JSON that would be sent to azure from this model.
This is an alias to as_dict(full_restapi_key_transformer, keep_readonly=False).
If you want XML serialization, you can pass the kwargs is_xml=True.
-
class
azure.synapse.spark.models.
SparkSessionOptions
(*, name: str, tags: Optional[Dict[str, str]] = None, artifact_id: Optional[str] = None, file: Optional[str] = None, class_name: Optional[str] = None, arguments: Optional[List[str]] = None, jars: Optional[List[str]] = None, python_files: Optional[List[str]] = None, files: Optional[List[str]] = None, archives: Optional[List[str]] = None, configuration: Optional[Dict[str, str]] = None, driver_memory: Optional[str] = None, driver_cores: Optional[int] = None, executor_memory: Optional[str] = None, executor_cores: Optional[int] = None, executor_count: Optional[int] = None, **kwargs)[source]¶ SparkSessionOptions.
All required parameters must be populated in order to send to Azure.
- Parameters
-
as_dict
(keep_readonly=True, key_transformer=<function attribute_transformer>, **kwargs)¶ Return a dict that can be JSONify using json.dump.
Advanced usage might optionaly use a callback as parameter:
Key is the attribute name used in Python. Attr_desc is a dict of metadata. Currently contains ‘type’ with the msrest type and ‘key’ with the RestAPI encoded key. Value is the current value in this object.
The string returned will be used to serialize the key. If the return type is a list, this is considered hierarchical result dict.
See the three examples in this file:
attribute_transformer
full_restapi_key_transformer
last_restapi_key_transformer
If you want XML serialization, you can pass the kwargs is_xml=True.
- Parameters
key_transformer (function) – A key transformer function.
- Returns
A dict JSON compatible object
- Return type
-
classmethod
deserialize
(data, content_type=None)¶ Parse a str using the RestAPI syntax and return a model.
-
classmethod
enable_additional_properties_sending
()¶
-
classmethod
from_dict
(data, key_extractors=None, content_type=None)¶ Parse a dict using given key extractor return a model.
By default consider key extractors (rest_key_case_insensitive_extractor, attribute_key_case_insensitive_extractor and last_rest_key_case_insensitive_extractor)
-
classmethod
is_xml_model
()¶
-
serialize
(keep_readonly=False, **kwargs)¶ Return the JSON that would be sent to azure from this model.
This is an alias to as_dict(full_restapi_key_transformer, keep_readonly=False).
If you want XML serialization, you can pass the kwargs is_xml=True.
-
class
azure.synapse.spark.models.
SparkSessionState
(*, not_started_at: Optional[datetime.datetime] = None, starting_at: Optional[datetime.datetime] = None, idle_at: Optional[datetime.datetime] = None, dead_at: Optional[datetime.datetime] = None, shutting_down_at: Optional[datetime.datetime] = None, terminated_at: Optional[datetime.datetime] = None, recovering_at: Optional[datetime.datetime] = None, busy_at: Optional[datetime.datetime] = None, error_at: Optional[datetime.datetime] = None, current_state: Optional[str] = None, job_creation_request: Optional[SparkRequest] = None, **kwargs)[source]¶ SparkSessionState.
- Parameters
-
as_dict
(keep_readonly=True, key_transformer=<function attribute_transformer>, **kwargs)¶ Return a dict that can be JSONify using json.dump.
Advanced usage might optionaly use a callback as parameter:
Key is the attribute name used in Python. Attr_desc is a dict of metadata. Currently contains ‘type’ with the msrest type and ‘key’ with the RestAPI encoded key. Value is the current value in this object.
The string returned will be used to serialize the key. If the return type is a list, this is considered hierarchical result dict.
See the three examples in this file:
attribute_transformer
full_restapi_key_transformer
last_restapi_key_transformer
If you want XML serialization, you can pass the kwargs is_xml=True.
- Parameters
key_transformer (function) – A key transformer function.
- Returns
A dict JSON compatible object
- Return type
-
classmethod
deserialize
(data, content_type=None)¶ Parse a str using the RestAPI syntax and return a model.
-
classmethod
enable_additional_properties_sending
()¶
-
classmethod
from_dict
(data, key_extractors=None, content_type=None)¶ Parse a dict using given key extractor return a model.
By default consider key extractors (rest_key_case_insensitive_extractor, attribute_key_case_insensitive_extractor and last_rest_key_case_insensitive_extractor)
-
classmethod
is_xml_model
()¶
-
serialize
(keep_readonly=False, **kwargs)¶ Return the JSON that would be sent to azure from this model.
This is an alias to as_dict(full_restapi_key_transformer, keep_readonly=False).
If you want XML serialization, you can pass the kwargs is_xml=True.
-
class
azure.synapse.spark.models.
SparkStatement
(*, id: int, code: Optional[str] = None, state: Optional[str] = None, output: Optional[SparkStatementOutput] = None, **kwargs)[source]¶ SparkStatement.
All required parameters must be populated in order to send to Azure.
- Parameters
id (int) – Required.
code (str) –
state (str) –
output (SparkStatementOutput) –
-
as_dict
(keep_readonly=True, key_transformer=<function attribute_transformer>, **kwargs)¶ Return a dict that can be JSONify using json.dump.
Advanced usage might optionaly use a callback as parameter:
Key is the attribute name used in Python. Attr_desc is a dict of metadata. Currently contains ‘type’ with the msrest type and ‘key’ with the RestAPI encoded key. Value is the current value in this object.
The string returned will be used to serialize the key. If the return type is a list, this is considered hierarchical result dict.
See the three examples in this file:
attribute_transformer
full_restapi_key_transformer
last_restapi_key_transformer
If you want XML serialization, you can pass the kwargs is_xml=True.
- Parameters
key_transformer (function) – A key transformer function.
- Returns
A dict JSON compatible object
- Return type
-
classmethod
deserialize
(data, content_type=None)¶ Parse a str using the RestAPI syntax and return a model.
-
classmethod
enable_additional_properties_sending
()¶
-
classmethod
from_dict
(data, key_extractors=None, content_type=None)¶ Parse a dict using given key extractor return a model.
By default consider key extractors (rest_key_case_insensitive_extractor, attribute_key_case_insensitive_extractor and last_rest_key_case_insensitive_extractor)
-
classmethod
is_xml_model
()¶
-
serialize
(keep_readonly=False, **kwargs)¶ Return the JSON that would be sent to azure from this model.
This is an alias to as_dict(full_restapi_key_transformer, keep_readonly=False).
If you want XML serialization, you can pass the kwargs is_xml=True.
-
class
azure.synapse.spark.models.
SparkStatementCancellationResult
(*, msg: Optional[str] = None, **kwargs)[source]¶ SparkStatementCancellationResult.
- Parameters
msg (str) –
-
as_dict
(keep_readonly=True, key_transformer=<function attribute_transformer>, **kwargs)¶ Return a dict that can be JSONify using json.dump.
Advanced usage might optionaly use a callback as parameter:
Key is the attribute name used in Python. Attr_desc is a dict of metadata. Currently contains ‘type’ with the msrest type and ‘key’ with the RestAPI encoded key. Value is the current value in this object.
The string returned will be used to serialize the key. If the return type is a list, this is considered hierarchical result dict.
See the three examples in this file:
attribute_transformer
full_restapi_key_transformer
last_restapi_key_transformer
If you want XML serialization, you can pass the kwargs is_xml=True.
- Parameters
key_transformer (function) – A key transformer function.
- Returns
A dict JSON compatible object
- Return type
-
classmethod
deserialize
(data, content_type=None)¶ Parse a str using the RestAPI syntax and return a model.
-
classmethod
enable_additional_properties_sending
()¶
-
classmethod
from_dict
(data, key_extractors=None, content_type=None)¶ Parse a dict using given key extractor return a model.
By default consider key extractors (rest_key_case_insensitive_extractor, attribute_key_case_insensitive_extractor and last_rest_key_case_insensitive_extractor)
-
classmethod
is_xml_model
()¶
-
serialize
(keep_readonly=False, **kwargs)¶ Return the JSON that would be sent to azure from this model.
This is an alias to as_dict(full_restapi_key_transformer, keep_readonly=False).
If you want XML serialization, you can pass the kwargs is_xml=True.
-
class
azure.synapse.spark.models.
SparkStatementCollection
(*, total: int, statements: Optional[List[SparkStatement]] = None, **kwargs)[source]¶ SparkStatementCollection.
All required parameters must be populated in order to send to Azure.
- Parameters
total (int) – Required.
statements (list[SparkStatement]) –
-
as_dict
(keep_readonly=True, key_transformer=<function attribute_transformer>, **kwargs)¶ Return a dict that can be JSONify using json.dump.
Advanced usage might optionaly use a callback as parameter:
Key is the attribute name used in Python. Attr_desc is a dict of metadata. Currently contains ‘type’ with the msrest type and ‘key’ with the RestAPI encoded key. Value is the current value in this object.
The string returned will be used to serialize the key. If the return type is a list, this is considered hierarchical result dict.
See the three examples in this file:
attribute_transformer
full_restapi_key_transformer
last_restapi_key_transformer
If you want XML serialization, you can pass the kwargs is_xml=True.
- Parameters
key_transformer (function) – A key transformer function.
- Returns
A dict JSON compatible object
- Return type
-
classmethod
deserialize
(data, content_type=None)¶ Parse a str using the RestAPI syntax and return a model.
-
classmethod
enable_additional_properties_sending
()¶
-
classmethod
from_dict
(data, key_extractors=None, content_type=None)¶ Parse a dict using given key extractor return a model.
By default consider key extractors (rest_key_case_insensitive_extractor, attribute_key_case_insensitive_extractor and last_rest_key_case_insensitive_extractor)
-
classmethod
is_xml_model
()¶
-
serialize
(keep_readonly=False, **kwargs)¶ Return the JSON that would be sent to azure from this model.
This is an alias to as_dict(full_restapi_key_transformer, keep_readonly=False).
If you want XML serialization, you can pass the kwargs is_xml=True.
-
class
azure.synapse.spark.models.
SparkStatementOptions
(*, code: Optional[str] = None, kind: Union[str, SparkStatementLanguageType, None] = None, **kwargs)[source]¶ SparkStatementOptions.
- Parameters
code (str) –
kind (str or SparkStatementLanguageType) – Possible values include: “spark”, “pyspark”, “dotnetspark”, “sql”.
-
as_dict
(keep_readonly=True, key_transformer=<function attribute_transformer>, **kwargs)¶ Return a dict that can be JSONify using json.dump.
Advanced usage might optionaly use a callback as parameter:
Key is the attribute name used in Python. Attr_desc is a dict of metadata. Currently contains ‘type’ with the msrest type and ‘key’ with the RestAPI encoded key. Value is the current value in this object.
The string returned will be used to serialize the key. If the return type is a list, this is considered hierarchical result dict.
See the three examples in this file:
attribute_transformer
full_restapi_key_transformer
last_restapi_key_transformer
If you want XML serialization, you can pass the kwargs is_xml=True.
- Parameters
key_transformer (function) – A key transformer function.
- Returns
A dict JSON compatible object
- Return type
-
classmethod
deserialize
(data, content_type=None)¶ Parse a str using the RestAPI syntax and return a model.
-
classmethod
enable_additional_properties_sending
()¶
-
classmethod
from_dict
(data, key_extractors=None, content_type=None)¶ Parse a dict using given key extractor return a model.
By default consider key extractors (rest_key_case_insensitive_extractor, attribute_key_case_insensitive_extractor and last_rest_key_case_insensitive_extractor)
-
classmethod
is_xml_model
()¶
-
serialize
(keep_readonly=False, **kwargs)¶ Return the JSON that would be sent to azure from this model.
This is an alias to as_dict(full_restapi_key_transformer, keep_readonly=False).
If you want XML serialization, you can pass the kwargs is_xml=True.
-
class
azure.synapse.spark.models.
SparkStatementOutput
(*, execution_count: int, status: Optional[str] = None, data: object = None, error_name: Optional[str] = None, error_value: Optional[str] = None, traceback: Optional[List[str]] = None, **kwargs)[source]¶ SparkStatementOutput.
All required parameters must be populated in order to send to Azure.
- Parameters
-
as_dict
(keep_readonly=True, key_transformer=<function attribute_transformer>, **kwargs)¶ Return a dict that can be JSONify using json.dump.
Advanced usage might optionaly use a callback as parameter:
Key is the attribute name used in Python. Attr_desc is a dict of metadata. Currently contains ‘type’ with the msrest type and ‘key’ with the RestAPI encoded key. Value is the current value in this object.
The string returned will be used to serialize the key. If the return type is a list, this is considered hierarchical result dict.
See the three examples in this file:
attribute_transformer
full_restapi_key_transformer
last_restapi_key_transformer
If you want XML serialization, you can pass the kwargs is_xml=True.
- Parameters
key_transformer (function) – A key transformer function.
- Returns
A dict JSON compatible object
- Return type
-
classmethod
deserialize
(data, content_type=None)¶ Parse a str using the RestAPI syntax and return a model.
-
classmethod
enable_additional_properties_sending
()¶
-
classmethod
from_dict
(data, key_extractors=None, content_type=None)¶ Parse a dict using given key extractor return a model.
By default consider key extractors (rest_key_case_insensitive_extractor, attribute_key_case_insensitive_extractor and last_rest_key_case_insensitive_extractor)
-
classmethod
is_xml_model
()¶
-
serialize
(keep_readonly=False, **kwargs)¶ Return the JSON that would be sent to azure from this model.
This is an alias to as_dict(full_restapi_key_transformer, keep_readonly=False).
If you want XML serialization, you can pass the kwargs is_xml=True.
-
class
azure.synapse.spark.models.
PluginCurrentState
[source]¶ An enumeration.
-
cleanup
= 'Cleanup'¶
-
ended
= 'Ended'¶
-
monitoring
= 'Monitoring'¶
-
preparation
= 'Preparation'¶
-
queued
= 'Queued'¶
-
resource_acquisition
= 'ResourceAcquisition'¶
-
submission
= 'Submission'¶
-
-
class
azure.synapse.spark.models.
SchedulerCurrentState
[source]¶ An enumeration.
-
ended
= 'Ended'¶
-
queued
= 'Queued'¶
-
scheduled
= 'Scheduled'¶
-
-
class
azure.synapse.spark.models.
SparkBatchJobResultType
[source]¶ The Spark batch job result.
-
cancelled
= 'Cancelled'¶
-
failed
= 'Failed'¶
-
succeeded
= 'Succeeded'¶
-
uncertain
= 'Uncertain'¶
-
-
class
azure.synapse.spark.models.
SparkErrorSource
[source]¶ An enumeration.
-
dependency
= 'Dependency'¶
-
system
= 'System'¶
-
unknown
= 'Unknown'¶
-
user
= 'User'¶
-
-
class
azure.synapse.spark.models.
SparkJobType
[source]¶ The job type.
-
spark_batch
= 'SparkBatch'¶
-
spark_session
= 'SparkSession'¶
-