4
4
from pyspark .sql import DataFrame
5
5
from pyspark .sql import functions as F
6
6
7
- from butterfree .clients import SparkClient
8
7
from butterfree .configs import environment
9
8
from butterfree .configs .db import MetastoreConfig
10
9
from butterfree .constants import DataType
@@ -75,11 +74,7 @@ def create_ymd(dataframe):
75
74
76
75
class TestFeatureSetPipeline :
77
76
def test_feature_set_pipeline (
78
- self ,
79
- mocked_df ,
80
- spark_session ,
81
- fixed_windows_output_feature_set_dataframe ,
82
- mocker ,
77
+ self , mocked_df , spark_session , fixed_windows_output_feature_set_dataframe ,
83
78
):
84
79
# arrange
85
80
table_reader_id = "a_source"
@@ -93,11 +88,6 @@ def test_feature_set_pipeline(
93
88
table_reader_table = table_reader_table ,
94
89
)
95
90
96
- spark_client = SparkClient ()
97
- spark_client .conn .conf .set (
98
- "spark.sql.sources.partitionOverwriteMode" , "dynamic"
99
- )
100
-
101
91
dbconfig = Mock ()
102
92
dbconfig .mode = "overwrite"
103
93
dbconfig .format_ = "parquet"
@@ -107,12 +97,6 @@ def test_feature_set_pipeline(
107
97
108
98
historical_writer = HistoricalFeatureStoreWriter (db_config = dbconfig )
109
99
110
- historical_writer .check_schema_hook = mocker .stub ("check_schema_hook" )
111
- historical_writer .check_schema_hook .run = mocker .stub ("run" )
112
- historical_writer .check_schema_hook .run .return_value = (
113
- fixed_windows_output_feature_set_dataframe
114
- )
115
-
116
100
# act
117
101
test_pipeline = FeatureSetPipeline (
118
102
source = Source (
@@ -187,7 +171,6 @@ def test_feature_set_pipeline_with_dates(
187
171
spark_session ,
188
172
fixed_windows_output_feature_set_date_dataframe ,
189
173
feature_set_pipeline ,
190
- mocker ,
191
174
):
192
175
# arrange
193
176
table_reader_table = "b_table"
@@ -211,7 +194,6 @@ def test_feature_set_pipeline_with_execution_date(
211
194
spark_session ,
212
195
fixed_windows_output_feature_set_date_dataframe ,
213
196
feature_set_pipeline ,
214
- mocker ,
215
197
):
216
198
# arrange
217
199
table_reader_table = "b_table"
@@ -233,7 +215,7 @@ def test_feature_set_pipeline_with_execution_date(
233
215
# assert
234
216
assert_dataframe_equality (df , target_df )
235
217
236
- def test_pipeline_with_hooks (self , spark_session , mocker ):
218
+ def test_pipeline_with_hooks (self , spark_session ):
237
219
# arrange
238
220
hook1 = AddHook (value = 1 )
239
221
0 commit comments