@@ -579,7 +579,7 @@ std::vector<std::string> disabledTestPatterns() {
579
579
// TODO: Issue 92895
580
580
// on platforms which do not support AMX, we are disabling I8 input tests
581
581
retVector.emplace_back (R"( smoke_LPT/FakeQuantizeWithNotOptimalTransformation.CompareWithRefImpl.*CPU.*i8.*)" );
582
- if (!ov::with_cpu_x86_avx512_core_amx_bf16 () && !ov::with_cpu_x86_bfloat16 ()) {
582
+ if (( !ov::with_cpu_x86_avx512_core_amx_bf16 () && !ov::with_cpu_x86_bfloat16 ()) || ov::with_cpu_x86_avx2_vnni_2 ()) {
583
583
// ignored for not supported bf16 platforms
584
584
retVector.emplace_back (R"( .*smoke_Snippets_EnforcePrecision_bf16.*)" );
585
585
retVector.emplace_back (R"( .*smoke_Snippets_MHAWOTransposeEnforceBF16.*)" );
@@ -638,7 +638,7 @@ std::vector<std::string> disabledTestPatterns() {
638
638
retVector.emplace_back (R"( .*smoke_Deconv_(2|3)D_NSPC_INT8_AMX/DeconvolutionLayerCPUTest.*)" );
639
639
}
640
640
641
- if (ov::with_cpu_x86_avx512_core_fp16 () || ov::with_cpu_x86_avx2_vnni_2 ()) {
641
+ if (ov::with_cpu_x86_float16 ()) {
642
642
// Issue: 143852
643
643
retVector.emplace_back (R"( smoke_ConvertRangeSubgraphCPUTest/ConvertRangeSubgraphCPUTest\.CompareWithRefs.*Prc=f16.*)" );
644
644
retVector.emplace_back (R"( (smoke|nightly)_FC_3D_FP16/.*_Fused=Multiply\(PerChannel\).*)" );
@@ -655,6 +655,12 @@ std::vector<std::string> disabledTestPatterns() {
655
655
retVector.emplace_back (
656
656
R"( smoke_GroupConv_.*D_Gemm_BF16/GroupConvolutionLayerCPUTest.CompareWithRefs.*primitive=jit_gemm.*)" );
657
657
retVector.emplace_back (R"( smoke_.*MatMulLayerCPUTest.*INFERENCE_PRECISION_HINT=bf16.*_primitive=jit_gemm.*)" );
658
+ // Issue: 163147
659
+ retVector.emplace_back (
660
+ R"( smoke_CompareWithRefs_4D.*[Ff]using.*EltwiseLayerCPUTest\.CompareWithRefs.*INFERENCE_PRECISION_HINT=f16.*enforceSnippets=1.*)" );
661
+ // Issue: 163144
662
+ retVector.emplace_back (
663
+ R"( smoke_ScaledAttn_CPU/ScaledAttnLayerCPUTest.CompareWithRefs/netPRC=bf16.*_TS=\(2\.8\.16\.32\)_\(2\.8\.16\.32\)_\(2\.8\.16\.32\)_\(1\.8\.48\.32\)_\(1\.8\.48\.32\)_\(1\.8\.48\.32\)_\(16\.48\)_\(16\.1\)_\(1\.48\).*)" );
658
664
}
659
665
660
666
return retVector;
0 commit comments