Skip to content

Commit

Permalink
Add merge mode test coverage for DML
Browse files Browse the repository at this point in the history
  • Loading branch information
Davis-Zhang-Onehouse committed Jan 9, 2025
1 parent 4efc241 commit 4c14c95
Show file tree
Hide file tree
Showing 3 changed files with 538 additions and 0 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -295,6 +295,19 @@ class HoodieSparkSqlTestBase extends FunSuite with BeforeAndAfterAll {
}
}

protected def withSparkSqlSessionConfig(configNameValues: (String, String)*)(f: => Unit): Unit = {
try {
configNameValues.foreach { case (configName, configValue) =>
spark.sql(s"set $configName=$configValue")
}
f
} finally {
configNameValues.foreach { case (configName, _) =>
spark.sql(s"reset $configName")
}
}
}

protected def withRecordType(recordTypes: Seq[HoodieRecordType] = Seq(HoodieRecordType.AVRO, HoodieRecordType.SPARK),
recordConfig: Map[HoodieRecordType, Map[String, String]] = Map.empty)(f: => Unit) {
// TODO HUDI-5264 Test parquet log with avro record in spark sql test
Expand Down
Original file line number Diff line number Diff line change
@@ -0,0 +1,246 @@
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/

package org.apache.spark.sql.hudi.dml

import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase

class TestMergeModeCommitTimeOrdering extends HoodieSparkSqlTestBase {

Seq("mor").foreach { tableType =>
// For COW commit time ordering does not work.
// Seq("cow", "mor").foreach { tableType =>
test(s"Test $tableType table with COMMIT_TIME_ORDERING merge mode") {
withSparkSqlSessionConfig("hoodie.merge.small.file.group.candidates.limit" -> "0",
"hoodie.record.merge.mode" -> "COMMIT_TIME_ORDERING") {
withRecordType()(withTempDir { tmp =>
val tableName = generateTableName
// Create table with COMMIT_TIME_ORDERING
spark.sql(
s"""
| create table $tableName (
| id int,
| name string,
| price double,
| ts long
| ) using hudi
| tblproperties (
| type = '$tableType',
| primaryKey = 'id',
| preCombineField = 'ts',
| hoodie.record.merge.mode = 'COMMIT_TIME_ORDERING'
| )
| location '${tmp.getCanonicalPath}'
""".stripMargin)

// Insert initial records with ts=100
spark.sql(
s"""
| insert into $tableName
| select 1 as id, 'A' as name, 10.0 as price, 100 as ts
| union all
| select 2, 'B', 20.0, 100
""".stripMargin)

// Verify inserting records with the same ts value are visible (COMMIT_TIME_ORDERING)
spark.sql(
s"""
| insert into $tableName
| select 1 as id, 'A_equal' as name, 60.0 as price, 100 as ts
| union all
| select 2, 'B_equal', 70.0, 100
""".stripMargin)

checkAnswer(s"select id, name, price, ts from $tableName order by id")(
Seq(1, "A_equal", 60.0, 100),
Seq(2, "B_equal", 70.0, 100)
)

// Verify updating records with the same ts value are visible (COMMIT_TIME_ORDERING)
spark.sql(
s"""
| update $tableName
| set price = 50.0, ts = 100
| where id = 1
""".stripMargin)

checkAnswer(s"select id, name, price, ts from $tableName order by id")(
Seq(1, "A_equal", 50.0, 100),
Seq(2, "B_equal", 70.0, 100)
)

// Verify inserting records with a lower ts value are visible (COMMIT_TIME_ORDERING)
spark.sql(
s"""
| insert into $tableName
| select 1 as id, 'A' as name, 30.0 as price, 99 as ts
| union all
| select 2, 'B', 40.0, 99
""".stripMargin)

checkAnswer(s"select id, name, price, ts from $tableName order by id")(
Seq(1, "A", 30.0, 99),
Seq(2, "B", 40.0, 99)
)

// Verify updating records with a lower ts value are visible (COMMIT_TIME_ORDERING)
spark.sql(
s"""
| update $tableName
| set price = 50.0, ts = 98
| where id = 1
""".stripMargin)

checkAnswer(s"select id, name, price, ts from $tableName order by id")(
Seq(1, "A", 50.0, 98),
Seq(2, "B", 40.0, 99)
)

// Verify inserting records with a higher ts value are visible (COMMIT_TIME_ORDERING)
spark.sql(
s"""
| insert into $tableName
| select 1 as id, 'A' as name, 30.0 as price, 101 as ts
| union all
| select 2, 'B', 40.0, 101
""".stripMargin)

// Verify records with ts=101 are visible
checkAnswer(s"select id, name, price, ts from $tableName order by id")(
Seq(1, "A", 30.0, 101),
Seq(2, "B", 40.0, 101)
)

// Verify updating records with a higher ts value are visible (COMMIT_TIME_ORDERING)
spark.sql(
s"""
| update $tableName
| set price = 50.0, ts = 102
| where id = 1
""".stripMargin)

// Verify final state after all operations
checkAnswer(s"select id, name, price, ts from $tableName order by id")(
Seq(1, "A", 50.0, 102),
Seq(2, "B", 40.0, 101)
)

// Delete record
spark.sql(s"delete from $tableName where id = 1")

// Verify deletion
checkAnswer(s"select id, name, price, ts from $tableName order by id")(
Seq(2, "B", 40.0, 101)
)
})
}
}

test(s"Test merge operations with COMMIT_TIME_ORDERING for $tableType table") {
withSparkSqlSessionConfig("hoodie.merge.small.file.group.candidates.limit" -> "0") {
withRecordType()(withTempDir { tmp =>
val tableName = generateTableName
// Create table with COMMIT_TIME_ORDERING
spark.sql(
s"""
| create table $tableName (
| id int,
| name string,
| price double,
| ts long
| ) using hudi
| tblproperties (
| type = '$tableType',
| primaryKey = 'id',
| preCombineField = 'ts',
| hoodie.record.merge.mode = 'COMMIT_TIME_ORDERING'
| )
| location '${tmp.getCanonicalPath}'
""".stripMargin)

// Insert initial records
spark.sql(
s"""
| insert into $tableName
| select 1 as id, 'A' as name, 10.0 as price, 100 as ts union all
| select 2, 'B', 20.0, 100 union all
| select 3, 'C', 30.0, 100 union all
| select 4, 'D', 40.0, 100 union all
| select 5, 'E', 50.0, 100 union all
| select 6, 'F', 60.0, 100
""".stripMargin)

// Merge operation - delete
spark.sql(
s"""
| merge into $tableName t
| using (
| select 1 as id, 'B2' as name, 25.0 as price, 101 as ts
| ) s
| on t.id = s.id
| when matched then delete
""".stripMargin)

// Merge operation - update with mixed ts values
spark.sql(
s"""
| merge into $tableName t
| using (
| select 4 as id, 'D2' as name, 45.0 as price, 101 as ts union all
| select 5, 'E2', 55.0, 99 as ts union all
| select 6, 'F2', 65.0, 100 as ts
| ) s
| on t.id = s.id
| when matched then update set *
""".stripMargin)

// Verify state after merges
checkAnswer(s"select id, name, price, ts from $tableName order by id")(
Seq(2, "B", 20.0, 100),
Seq(3, "C", 30.0, 100),
Seq(4, "D2", 45.0, 101),
Seq(5, "E2", 55.0, 99),
Seq(6, "F2", 65.0, 100)
)

// Insert new records through merge
spark.sql(
s"""
| merge into $tableName t
| using (
| select 7 as id, 'D2' as name, 45.0 as price, 100 as ts union all
| select 8, 'E2', 55.0, 100 as ts
| ) s
| on t.id = s.id
| when not matched then insert *
""".stripMargin)

// Verify final state
checkAnswer(s"select id, name, price, ts from $tableName order by id")(
Seq(2, "B", 20.0, 100),
Seq(3, "C", 30.0, 100),
Seq(4, "D2", 45.0, 101),
Seq(5, "E2", 55.0, 99),
Seq(6, "F2", 65.0, 100),
Seq(7, "D2", 45.0, 100),
Seq(8, "E2", 55.0, 100)
)
})
}
}
}
}
Loading

0 comments on commit 4c14c95

Please sign in to comment.