SynapseNotebookActivity Class

Execute Synapse notebook activity.

All required parameters must be populated in order to send to server.

Inheritance
azure.mgmt.datafactory.models._models_py3.ExecutionActivity
SynapseNotebookActivity

Constructor

SynapseNotebookActivity(*, name: str, notebook: _models.SynapseNotebookReference, additional_properties: Dict[str, MutableMapping[str, Any]] | None = None, description: str | None = None, state: str | _models.ActivityState | None = None, on_inactive_mark_as: str | _models.ActivityOnInactiveMarkAs | None = None, depends_on: List[_models.ActivityDependency] | None = None, user_properties: List[_models.UserProperty] | None = None, linked_service_name: _models.LinkedServiceReference | None = None, policy: _models.ActivityPolicy | None = None, spark_pool: _models.BigDataPoolParametrizationReference | None = None, parameters: Dict[str, _models.NotebookParameter] | None = None, executor_size: MutableMapping[str, Any] | None = None, conf: MutableMapping[str, Any] | None = None, driver_size: MutableMapping[str, Any] | None = None, num_executors: MutableMapping[str, Any] | None = None, configuration_type: str | _models.ConfigurationType | None = None, target_spark_configuration: _models.SparkConfigurationParametrizationReference | None = None, spark_config: Dict[str, MutableMapping[str, Any]] | None = None, **kwargs: Any)

Keyword-Only Parameters

Name Description
additional_properties
dict[str, <xref:JSON>]

Unmatched properties from the message are deserialized to this collection.

name
str

Activity name. Required.

description
str

Activity description.

state

Activity state. This is an optional property and if not provided, the state will be Active by default. Known values are: "Active" and "Inactive".

on_inactive_mark_as

Status result of the activity when the state is set to Inactive. This is an optional property and if not provided when the activity is inactive, the status will be Succeeded by default. Known values are: "Succeeded", "Failed", and "Skipped".

depends_on

Activity depends on condition.

user_properties

Activity user properties.

linked_service_name

Linked service reference.

policy

Activity policy.

notebook

Synapse notebook reference. Required.

spark_pool

The name of the big data pool which will be used to execute the notebook.

parameters

Notebook parameters.

executor_size
<xref:JSON>

Number of core and memory to be used for executors allocated in the specified Spark pool for the session, which will be used for overriding 'executorCores' and 'executorMemory' of the notebook you provide. Type: string (or Expression with resultType string).

conf
<xref:JSON>

Spark configuration properties, which will override the 'conf' of the notebook you provide.

driver_size
<xref:JSON>

Number of core and memory to be used for driver allocated in the specified Spark pool for the session, which will be used for overriding 'driverCores' and 'driverMemory' of the notebook you provide. Type: string (or Expression with resultType string).

num_executors
<xref:JSON>

Number of executors to launch for this session, which will override the 'numExecutors' of the notebook you provide. Type: integer (or Expression with resultType integer).

configuration_type

The type of the spark config. Known values are: "Default", "Customized", and "Artifact".

target_spark_configuration

The spark configuration of the spark job.

spark_config
dict[str, <xref:JSON>]

Spark configuration property.

Variables

Name Description
additional_properties
dict[str, <xref:JSON>]

Unmatched properties from the message are deserialized to this collection.

name
str

Activity name. Required.

type
str

Type of activity. Required.

description
str

Activity description.

state

Activity state. This is an optional property and if not provided, the state will be Active by default. Known values are: "Active" and "Inactive".

on_inactive_mark_as

Status result of the activity when the state is set to Inactive. This is an optional property and if not provided when the activity is inactive, the status will be Succeeded by default. Known values are: "Succeeded", "Failed", and "Skipped".

depends_on

Activity depends on condition.

user_properties

Activity user properties.

linked_service_name

Linked service reference.

policy

Activity policy.

notebook

Synapse notebook reference. Required.

spark_pool

The name of the big data pool which will be used to execute the notebook.

parameters

Notebook parameters.

executor_size
<xref:JSON>

Number of core and memory to be used for executors allocated in the specified Spark pool for the session, which will be used for overriding 'executorCores' and 'executorMemory' of the notebook you provide. Type: string (or Expression with resultType string).

conf
<xref:JSON>

Spark configuration properties, which will override the 'conf' of the notebook you provide.

driver_size
<xref:JSON>

Number of core and memory to be used for driver allocated in the specified Spark pool for the session, which will be used for overriding 'driverCores' and 'driverMemory' of the notebook you provide. Type: string (or Expression with resultType string).

num_executors
<xref:JSON>

Number of executors to launch for this session, which will override the 'numExecutors' of the notebook you provide. Type: integer (or Expression with resultType integer).

configuration_type

The type of the spark config. Known values are: "Default", "Customized", and "Artifact".

target_spark_configuration

The spark configuration of the spark job.

spark_config
dict[str, <xref:JSON>]

Spark configuration property.