Skip to content

Commit 76a668b

Browse files
authored
[CPU] Limit f16 in Accuracy mode by hardware support (#25243)
### Details: - *Restrict usage of f16 in transformations pipeline if it's not supported by hardware for ACCURACY MODE* ### Tickets: - *145051*
1 parent 8c045e9 commit 76a668b

File tree

3 files changed

+16
-3
lines changed

3 files changed

+16
-3
lines changed

src/plugins/intel_cpu/src/nodes/conv.cpp

+2-1
Original file line numberDiff line numberDiff line change
@@ -531,8 +531,9 @@ void Convolution::getSupportedDescriptors() {
531531
auto dt = memory::data_type::f32;
532532

533533
// supported lower precisions: bf16, f16
534-
if (one_of(originalDT, memory::data_type::bf16, memory::data_type::f16))
534+
if (one_of(originalDT, memory::data_type::bf16, memory::data_type::f16) && hasHardwareSupport(originalPrec)) {
535535
dt = originalDT;
536+
}
536537

537538
// fallback to f32 on special case for performance reasons
538539
if (isDepthWise() && ndims == 5)

src/plugins/intel_cpu/src/transformations/transformation_pipeline.cpp

+2-1
Original file line numberDiff line numberDiff line change
@@ -356,7 +356,8 @@ void Transformations::PreLpt(const std::vector<ov::element::Type>& defaultPrecis
356356
// @todo should we always convert to f32 regardless of hardware support, as it is done for f16?
357357
if (!hasHardwareSupport(ov::element::bf16))
358358
map.insert({ov::element::bf16, ov::element::f32});
359-
if (!one_of(inferencePrecision, element::f16, element::undefined)) {
359+
// TODO: Remove 'hasHardwareSupport' when all nodes are able to handle f16 properly.
360+
if (!one_of(inferencePrecision, element::f16, element::undefined) || !hasHardwareSupport(element::f16)) {
360361
map.insert({ov::element::f16, ov::element::f32});
361362
}
362363
return map;

src/plugins/intel_cpu/tests/functional/custom/subgraph_tests/src/classes/undefined_et.cpp

+12-1
Original file line numberDiff line numberDiff line change
@@ -26,6 +26,7 @@
2626
// -----------
2727

2828
#include "custom/subgraph_tests/include/undefined_et.hpp"
29+
#include "utils/precision_support.h"
2930

3031
namespace ov {
3132
namespace test {
@@ -74,6 +75,12 @@ void UndefinedEtSubgraphTest::SetUp() {
7475
auto logical_not = std::make_shared<op::v1::LogicalNot>(cvt_f32);
7576

7677
function = std::make_shared<ov::Model>(OutputVector{logical_not->output(0)}, ParameterVector{param_0, param_1, param_2}, "UndefinedET");
78+
79+
// TODO: Need to remove when the hardware checking for f16 will be eliminated in the Transformations pipeline.
80+
if (m_data_et == element::f16 && !ov::intel_cpu::hasHardwareSupport(m_data_et)) {
81+
abs_threshold = 1.f;
82+
rel_threshold = 0.1f;
83+
}
7784
}
7885

7986
template<typename TD, typename TS>
@@ -146,14 +153,18 @@ TEST_P(UndefinedEtSubgraphTest, CompareWithRefs) {
146153

147154
size_t rnd_unfm_counter = 0lu;
148155
size_t logical_not_counter = 0lu;
156+
auto expected_dt = m_data_et;
157+
if (!ov::intel_cpu::hasHardwareSupport(expected_dt)) {
158+
expected_dt = element::f32;
159+
}
149160
for (const auto& node : compiledModel.get_runtime_model()->get_ops()) {
150161
auto rt_info = node->get_rt_info();
151162
auto it = rt_info.find(exec_model_info::LAYER_TYPE);
152163
ASSERT_NE(rt_info.end(), it);
153164
auto op_name = it->second.as<std::string>();
154165

155166
if (op_name == "RandomUniform") {
156-
ASSERT_EQ(node->get_output_element_type(0), m_data_et);
167+
ASSERT_EQ(node->get_output_element_type(0), expected_dt);
157168
rnd_unfm_counter++;
158169
}
159170
if (op_name == "Eltwise") {

0 commit comments

Comments
 (0)