Skip to content

Commit 4b1f824

Browse files
authored
[GPU] Fix mvn unit test fails (openvinotoolkit#28372)
### Details: - *When shape flattening is required, change blocked format to plain format if reduction axis is not blocked axis.* ### Tickets: - *158306*
1 parent 0335554 commit 4b1f824

File tree

3 files changed

+52
-15
lines changed

3 files changed

+52
-15
lines changed

src/plugins/intel_gpu/src/graph/graph_optimizer/add_required_reorders.cpp

+36-12
Original file line numberDiff line numberDiff line change
@@ -246,18 +246,42 @@ void add_required_reorders::run(program& p) {
246246
}
247247

248248
// Remove padded-inputs in spatial axes not to use ref kernel which causes huge perf drop
249-
if (usr->is_type<mvn>() && usr->as<mvn>().input().is_padded_spatial()) {
250-
auto out_layout = usr->get_output_layout();
251-
// Check formats of implemented opt kernels without a spatial padding support
252-
if (out_layout.format == format::b_fs_yx_fsv16 || out_layout.format == format::b_fs_zyx_fsv16 ||
253-
out_layout.format == format::bs_fs_yx_bsv32_fsv16 || out_layout.format == format::bs_fs_yx_bsv32_fsv32) {
254-
auto& dep = usr->as<mvn>().input();
255-
cldnn::layout layout_wo_padding = dep.get_output_layout();
256-
layout_wo_padding.data_padding = cldnn::padding{};
257-
auto new_reorder = std::make_shared<reorder>(dep.id() + "_no_pad_reorder", dep.id(), layout_wo_padding);
258-
auto& new_reorder_node = p.get_or_create(new_reorder);
259-
p.add_intermediate(new_reorder_node, *usr, dep);
260-
new_reorder_node.recalc_output_layout(false);
249+
if (usr->is_type<mvn>()) {
250+
if (usr->as<mvn>().input().is_padded_spatial()) {
251+
auto out_layout = usr->get_output_layout();
252+
// Check formats of implemented opt kernels without a spatial padding support
253+
if (out_layout.format == format::b_fs_yx_fsv16 || out_layout.format == format::b_fs_zyx_fsv16 ||
254+
out_layout.format == format::bs_fs_yx_bsv32_fsv16 || out_layout.format == format::bs_fs_yx_bsv32_fsv32) {
255+
auto& dep = usr->as<mvn>().input();
256+
cldnn::layout layout_wo_padding = dep.get_output_layout();
257+
layout_wo_padding.data_padding = cldnn::padding{};
258+
auto new_reorder = std::make_shared<reorder>(dep.id() + "_no_pad_reorder", dep.id(), layout_wo_padding);
259+
auto& new_reorder_node = p.get_or_create(new_reorder);
260+
p.add_intermediate(new_reorder_node, *usr, dep);
261+
new_reorder_node.recalc_output_layout(false);
262+
}
263+
}
264+
265+
auto input_layout = usr->get_input_layout();
266+
auto input_pshape = input_layout.get_partial_shape();
267+
auto prim = usr->as<mvn>().get_primitive();
268+
269+
if (prim->requires_alignment(input_pshape)) {
270+
auto block_sizes = format::block_sizes(input_layout.format);
271+
auto axes = prim->reduction_axes;
272+
if (input_layout.is_dynamic() || block_sizes.size() > 1
273+
|| (block_sizes.size() == 1 &&
274+
input_pshape[block_sizes[0].first].get_length() % block_sizes[0].second != 0 &&
275+
std::count(axes.begin(), axes.end(), block_sizes[0].first) == 0)) {
276+
auto rank = input_pshape.size();
277+
input_layout.format = format::get_default_format(rank);
278+
auto& dep = usr->as<mvn>().input();
279+
auto new_reorder = std::make_shared<reorder>(dep.id() + "_to_plain", dep.id(), input_layout);
280+
auto& new_reorder_node = p.get_or_create(new_reorder);
281+
p.add_intermediate(new_reorder_node, *usr, dep);
282+
// Need to invalidate users because the output format of mvn follows input format.
283+
new_reorder_node.recalc_output_layout(true);
284+
}
261285
}
262286
}
263287

src/plugins/intel_gpu/src/graph/impls/ocl/mvn.cpp

+12-1
Original file line numberDiff line numberDiff line change
@@ -63,7 +63,18 @@ struct mvn_impl : typed_primitive_impl_ocl<mvn> {
6363
ov::PartialShape shape = ov::PartialShape::dynamic(new_rank);
6464

6565
auto& output_layout = updated_impl_params.output_layouts[0];
66+
6667
if (input_pshape.is_static()) {
68+
size_t flatten_axis = 0;
69+
// Change flatten axis if the format is single fsv.
70+
auto block_sizes = format::block_sizes(input_layout.format);
71+
if (block_sizes.size() == 1
72+
&& (input_pshape[block_sizes[0].first].get_length() % block_sizes[0].second == 0)
73+
&& (std::count(axes.begin(), axes.end(), block_sizes[0].first) == 0)
74+
&& block_sizes[0].first == 1) {
75+
flatten_axis = 1;
76+
}
77+
6778
for (size_t i = 0; i < new_rank; i++) {
6879
shape[i] = 1;
6980
}
@@ -72,7 +83,7 @@ struct mvn_impl : typed_primitive_impl_ocl<mvn> {
7283
// 1. normalized dimensions which are flattened and written to the last dim
7384
// 2. not normalized dims which are flattened and written to the first dim
7485
for (size_t i = 0; i < input_rank; i++) {
75-
shape[static_cast<int64_t>(i) < min ? 0 : (new_rank - 1)] *= input_pshape[i];
86+
shape[static_cast<int64_t>(i) < min ? flatten_axis : (new_rank - 1)] *= input_pshape[i];
7687
}
7788
}
7889

src/plugins/intel_gpu/tests/unit/fusions/mvn_fusion_test.cpp

+4-2
Original file line numberDiff line numberDiff line change
@@ -80,6 +80,7 @@ class MVNFusingTest : public ::BaseFusingTest<mvn_test_params> {
8080
#define CASE_MVN_I8_6 { 2, 16, 8, 8 }, { 1, 1, 1, 1 }, data_types::i8, format::b_fs_yx_fsv16, {1, 2, 3}, true, data_types::f32, format::bfyx
8181
#define CASE_MVN_I8_7 { 2, 16, 1, 8 }, { 1, 1, 8, 1 }, data_types::i8, format::b_fs_yx_fsv16, {1, 2, 3}, true, data_types::f32, format::bfyx
8282
#define CASE_MVN_I8_8 { 2, 16, 3, 8 }, { 1, 1, 3, 8 }, data_types::i8, format::b_fs_yx_fsv16, {3}, true, data_types::f32, format::bfyx
83+
#define CASE_MVN_I8_8_NA { 2, 15, 3, 8 }, { 1, 1, 3, 8 }, data_types::i8, format::b_fs_yx_fsv16, {3}, true, data_types::f32, format::bfyx
8384
#define CASE_MVN_3D_I8_1 { 1, 16, 8, 8, 8 }, { 1, 16, 8, 8, 8 }, data_types::i8, format::bfzyx, {2, 3, 4}, true, data_types::f32, format::bfzyx
8485
#define CASE_MVN_3D_I8_2 { 2, 16, 8, 8, 8 }, { 2, 16, 8, 8, 8 }, data_types::i8, format::bfzyx, {1, 2, 3, 4}, true, data_types::f32, format::bfzyx
8586
#define CASE_MVN_3D_I8_3 { 2, 16, 8, 8, 8 }, { 2, 1, 8, 8, 1 }, data_types::i8, format::bfzyx, {1, 2, 3, 4}, true, data_types::f32, format::bfzyx
@@ -170,7 +171,8 @@ INSTANTIATE_TEST_SUITE_P(fusings_gpu, mvn_scale_quantize_i8, ::testing::ValuesIn
170171
mvn_test_params{ CASE_MVN_I8_2, 2, 2, 4 },
171172
mvn_test_params{ CASE_MVN_I8_3, 2, 2, 4 },
172173
mvn_test_params{ CASE_MVN_I8_4, 2, 2, 4 },
173-
// mvn_test_params{ CASE_MVN_I8_8, 3, 3, 4 }, // TODO: It will be fix soon, test reference is wrong in new driver.
174+
mvn_test_params{ CASE_MVN_I8_8, 3, 3, 4 },
175+
mvn_test_params{ CASE_MVN_I8_8_NA, 3, 3, 4 },
174176
mvn_test_params{ CASE_MVN_3D_I8_1, 2, 2, 4 },
175177
mvn_test_params{ CASE_MVN_3D_I8_2, 2, 2, 4 },
176178
mvn_test_params{ CASE_MVN_U8_1, 2, 2, 4 },
@@ -221,7 +223,7 @@ INSTANTIATE_TEST_SUITE_P(fusings_gpu, mvn_scale_activation_eltwise_fp32_quantize
221223
mvn_test_params{ CASE_MVN_I8_5, 2, 4, 6 },
222224
mvn_test_params{ CASE_MVN_I8_6, 2, 4, 6 },
223225
mvn_test_params{ CASE_MVN_I8_7, 3, 4, 6 },
224-
// mvn_test_params{ CASE_MVN_I8_8, 3, 5, 6 }, // TODO: It will be fix soon, test reference is wrong in new driver.
226+
mvn_test_params{ CASE_MVN_I8_8, 3, 5, 6 },
225227
mvn_test_params{ CASE_MVN_3D_I8_1, 2, 4, 6 },
226228
mvn_test_params{ CASE_MVN_3D_I8_2, 2, 4, 6 },
227229
mvn_test_params{ CASE_MVN_3D_I8_3, 2, 4, 6 },

0 commit comments

Comments
 (0)