Skip to content

Commit 297a648

Browse files
authored
fix(datasets): Investigate ManagedTableDataset test failures (kedro-org#671)
* revert ignore databricks test and coverage * Update managed_table_dataset.py * attempt a fix * revert Signed-off-by: Sajid Alam <sajid_alam@mckinsey.com> * Update managed_table_dataset.py Signed-off-by: Sajid Alam <sajid_alam@mckinsey.com> * Update managed_table_dataset.py Signed-off-by: Sajid Alam <sajid_alam@mckinsey.com> * Update managed_table_dataset.py Signed-off-by: Sajid Alam <sajid_alam@mckinsey.com> * fix tests Signed-off-by: Sajid Alam <sajid_alam@mckinsey.com> * revert Signed-off-by: Sajid Alam <sajid_alam@mckinsey.com> * revert test Signed-off-by: Sajid Alam <sajid_alam@mckinsey.com> * add compatability Signed-off-by: Sajid Alam <sajid_alam@mckinsey.com> * Update test_managed_table_dataset.py Signed-off-by: Sajid Alam <sajid_alam@mckinsey.com> * Update managed_table_dataset.py Signed-off-by: Sajid Alam <sajid_alam@mckinsey.com> * Update test_managed_table_dataset.py Signed-off-by: Sajid Alam <sajid_alam@mckinsey.com> * Update test_managed_table_dataset.py Signed-off-by: Sajid Alam <sajid_alam@mckinsey.com> * debugging Signed-off-by: Sajid Alam <sajid_alam@mckinsey.com> * update tests and save method Signed-off-by: Sajid Alam <sajid_alam@mckinsey.com> * lint Signed-off-by: Sajid Alam <sajid_alam@mckinsey.com> * fix mypy Signed-off-by: Sajid Alam <sajid_alam@mckinsey.com> * test pyspark 3.4 and above Signed-off-by: Sajid Alam <sajid_alam@mckinsey.com> * Update pyproject.toml Signed-off-by: Sajid Alam <sajid_alam@mckinsey.com> * try DF.items in stead Signed-off-by: Sajid Alam <sajid_alam@mckinsey.com> * Update managed_table_dataset.py Signed-off-by: Sajid Alam <sajid_alam@mckinsey.com> * Update managed_table_dataset.py Signed-off-by: Sajid Alam <sajid_alam@mckinsey.com> --------- Signed-off-by: Sajid Alam <90610031+SajidAlamQB@users.noreply.github.com> Signed-off-by: Sajid Alam <sajid_alam@mckinsey.com>
1 parent ca46a05 commit 297a648

File tree

3 files changed

+3
-2
lines changed

3 files changed

+3
-2
lines changed

Makefile

+1-1
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@ test:
2727

2828
# Run test_tensorflow_model_dataset separately, because these tests are flaky when run as part of the full test-suite
2929
dataset-tests: dataset-doctests
30-
cd kedro-datasets && pytest tests --cov-config pyproject.toml --numprocesses 4 --dist loadfile --ignore tests/databricks --ignore tests/tensorflow
30+
cd kedro-datasets && pytest tests --cov-config pyproject.toml --numprocesses 4 --dist loadfile --ignore tests/tensorflow
3131
cd kedro-datasets && pytest tests/tensorflow/test_tensorflow_model_dataset.py --no-cov
3232

3333
extra_pytest_args-no-spark=--ignore kedro_datasets/databricks --ignore kedro_datasets/spark

kedro-datasets/kedro_datasets/databricks/managed_table_dataset.py

+1
Original file line numberDiff line numberDiff line change
@@ -22,6 +22,7 @@
2222
from kedro_datasets.spark.spark_dataset import _get_spark
2323

2424
logger = logging.getLogger(__name__)
25+
pd.DataFrame.iteritems = pd.DataFrame.items
2526

2627

2728
@dataclass(frozen=True)

kedro-datasets/pyproject.toml

+1-1
Original file line numberDiff line numberDiff line change
@@ -291,7 +291,7 @@ version = {attr = "kedro_datasets.__version__"}
291291
fail_under = 100
292292
show_missing = true
293293
# temporarily ignore kedro_datasets/__init__.py in coverage report
294-
omit = ["tests/*", "kedro_datasets/holoviews/*", "kedro_datasets/netcdf/*", "kedro_datasets/snowflake/*", "kedro_datasets/tensorflow/*", "kedro_datasets/__init__.py", "kedro_datasets/conftest.py", "kedro_datasets/databricks/*"]
294+
omit = ["tests/*", "kedro_datasets/holoviews/*", "kedro_datasets/netcdf/*", "kedro_datasets/snowflake/*", "kedro_datasets/tensorflow/*", "kedro_datasets/__init__.py", "kedro_datasets/conftest.py"]
295295
exclude_also = ["raise NotImplementedError", "if TYPE_CHECKING:"]
296296

297297
[tool.pytest.ini_options]

0 commit comments

Comments
 (0)