You are viewing an outdated version of the documentation.

This documentation is for an older version (1.4.7) of Dagster. You can view the version of this page from our latest release below.

Source code for dagster_spark.ops

from dagster import (
    In,
    Nothing,
    Out,
    _check as check,
    op,
)

from .configs import define_spark_config


[docs]def create_spark_op( name, main_class, description=None, required_resource_keys=frozenset(["spark"]) ): check.str_param(name, "name") check.str_param(main_class, "main_class") check.opt_str_param(description, "description", "A parameterized Spark job.") check.set_param(required_resource_keys, "required_resource_keys") @op( name=name, description=description, config_schema=define_spark_config(), ins={"start": In(Nothing)}, out=Out(Nothing), tags={"kind": "spark", "main_class": main_class}, required_resource_keys=required_resource_keys, ) def spark_op(context): context.resources.spark.run_spark_job(context.op_config, main_class) return spark_op