Skip to content

Commit 8f23149

Browse files
committed
more
1 parent b3846eb commit 8f23149

23 files changed

+127
-76
lines changed

.github/workflows/build_and_test.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -511,7 +511,7 @@ jobs:
511511
- >-
512512
pyspark-core, pyspark-errors, pyspark-streaming, pyspark-logger
513513
- >-
514-
pyspark-mllib, pyspark-ml, pyspark-ml-connect
514+
pyspark-mllib, pyspark-ml, pyspark-ml-connect, pyspark-pipelines
515515
- >-
516516
pyspark-connect
517517
- >-

bin/spark-pipelines

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -30,4 +30,4 @@ fi
3030
export PYTHONPATH="${SPARK_HOME}/python/:$PYTHONPATH"
3131
export PYTHONPATH="${SPARK_HOME}/python/lib/py4j-0.10.9.9-src.zip:$PYTHONPATH"
3232

33-
${SPARK_HOME}/bin/spark-submit --conf spark.api.mode=connect "${SPARK_HOME}"/python/pyspark/sql/pipelines/cli.py "$@"
33+
$PYSPARK_PYTHON "${SPARK_HOME}"/python/pyspark/pipelines/cli.py "$@"

dev/sparktestsupport/modules.py

Lines changed: 12 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -556,10 +556,6 @@ def __hash__(self):
556556
"pyspark.sql.tests.pandas.test_pandas_udf_window",
557557
"pyspark.sql.tests.pandas.test_pandas_sqlmetrics",
558558
"pyspark.sql.tests.pandas.test_converter",
559-
"pyspark.sql.tests.pipelines.test_block_connect_access",
560-
"pyspark.sql.tests.pipelines.test_cli",
561-
"pyspark.sql.tests.pipelines.test_decorators",
562-
"pyspark.sql.tests.pipelines.test_graph_element_registry",
563559
"pyspark.sql.tests.test_python_datasource",
564560
"pyspark.sql.tests.test_python_streaming_datasource",
565561
"pyspark.sql.tests.test_readwriter",
@@ -1505,6 +1501,18 @@ def __hash__(self):
15051501
],
15061502
)
15071503

1504+
pyspark_pipelines = Module(
1505+
name="pyspark-pipelines",
1506+
dependencies=[],
1507+
source_file_regexes=["python/pyspark/pipelines"],
1508+
python_test_goals=[
1509+
"pyspark.pipelines.tests.test_block_connect_access",
1510+
"pyspark.pipelines.tests.test_cli",
1511+
"pyspark.pipelines.tests.test_decorators",
1512+
"pyspark.pipelines.tests.test_graph_element_registry",
1513+
"pyspark.pipelines.tests.test_init_cli",
1514+
],
1515+
)
15081516

15091517
sparkr = Module(
15101518
name="sparkr",

python/mypy.ini

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -117,6 +117,9 @@ ignore_errors = True
117117
[mypy-pyspark.pandas.tests.*]
118118
ignore_errors = True
119119

120+
[mypy-pyspark.pipelines.tests.*]
121+
ignore_errors = True
122+
120123
[mypy-pyspark.tests.*]
121124
ignore_errors = True
122125

@@ -191,6 +194,3 @@ ignore_missing_imports = True
191194
; Ignore errors for proto generated code
192195
[mypy-pyspark.sql.connect.proto.*, pyspark.sql.connect.proto, pyspark.sql.streaming.proto]
193196
ignore_errors = True
194-
195-
[mypy-pyspark.sql.pipelines.proto.*]
196-
ignore_errors = True

python/pyspark/sql/pipelines/__init__.py renamed to python/pyspark/pipelines/__init__.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@
1414
# See the License for the specific language governing permissions and
1515
# limitations under the License.
1616
#
17-
from pyspark.sql.pipelines.api import (
17+
from pyspark.pipelines.api import (
1818
append_flow,
1919
create_streaming_table,
2020
materialized_view,

python/pyspark/sql/pipelines/api.py renamed to python/pyspark/pipelines/api.py

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -17,13 +17,13 @@
1717
from typing import Callable, Dict, List, Optional, Union, overload
1818

1919
from pyspark.errors import PySparkTypeError
20-
from pyspark.sql.pipelines.graph_element_registry import get_active_graph_element_registry
21-
from pyspark.sql.pipelines.type_error_utils import validate_optional_list_of_str_arg
22-
from pyspark.sql.pipelines.flow import Flow, QueryFunction
23-
from pyspark.sql.pipelines.source_code_location import (
20+
from pyspark.pipelines.graph_element_registry import get_active_graph_element_registry
21+
from pyspark.pipelines.type_error_utils import validate_optional_list_of_str_arg
22+
from pyspark.pipelines.flow import Flow, QueryFunction
23+
from pyspark.pipelines.source_code_location import (
2424
get_caller_source_code_location,
2525
)
26-
from pyspark.sql.pipelines.dataset import (
26+
from pyspark.pipelines.dataset import (
2727
MaterializedView,
2828
StreamingTable,
2929
TemporaryView,

python/pyspark/sql/pipelines/cli.py renamed to python/pyspark/pipelines/cli.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -32,15 +32,15 @@
3232

3333
from pyspark.errors import PySparkException, PySparkTypeError
3434
from pyspark.sql import SparkSession
35-
from pyspark.sql.pipelines.graph_element_registry import (
35+
from pyspark.pipelines.graph_element_registry import (
3636
graph_element_registration_context,
3737
GraphElementRegistry,
3838
)
39-
from pyspark.sql.pipelines.init_cli import init
40-
from pyspark.sql.pipelines.spark_connect_graph_element_registry import (
39+
from pyspark.pipelines.init_cli import init
40+
from pyspark.pipelines.spark_connect_graph_element_registry import (
4141
SparkConnectGraphElementRegistry,
4242
)
43-
from pyspark.sql.pipelines.spark_connect_pipeline import (
43+
from pyspark.pipelines.spark_connect_pipeline import (
4444
create_dataflow_graph,
4545
start_run,
4646
handle_pipeline_events,

python/pyspark/sql/pipelines/dataset.py renamed to python/pyspark/pipelines/dataset.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@
1717
from dataclasses import dataclass
1818
from typing import Mapping, Optional, Sequence, Union
1919

20-
from pyspark.sql.pipelines.source_code_location import SourceCodeLocation
20+
from pyspark.pipelines.source_code_location import SourceCodeLocation
2121
from pyspark.sql.types import StructType
2222

2323

python/pyspark/sql/pipelines/flow.py renamed to python/pyspark/pipelines/flow.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@
1818
from typing import Callable, Dict
1919

2020
from pyspark.sql import DataFrame
21-
from pyspark.sql.pipelines.source_code_location import SourceCodeLocation
21+
from pyspark.pipelines.source_code_location import SourceCodeLocation
2222

2323
QueryFunction = Callable[[], DataFrame]
2424

0 commit comments

Comments
 (0)