Skip to content

Commit e567c9e

Browse files
[NPU] Switching the I/O identification convention to indices (openvinotoolkit#24248)
### Details: - Please see PR#10348 (compiler repository) for a detailed description and some extra validation. ### Tickets: - *CVS-142751*
1 parent 033a515 commit e567c9e

File tree

22 files changed

+1098
-902
lines changed

22 files changed

+1098
-902
lines changed
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,18 @@
1+
// Copyright (C) 2018-2024 Intel Corporation
2+
// SPDX-License-Identifier: Apache-2.0
3+
//
4+
5+
#include "execution_graph_tests/duplicate_inputs_outputs_names.hpp"
6+
7+
#include "common_test_utils/test_constants.hpp"
8+
9+
using namespace ExecutionGraphTests;
10+
11+
namespace {
12+
13+
INSTANTIATE_TEST_SUITE_P(smoke_duplicateInputsOutputsNames,
14+
ExecGraphDuplicateInputsOutputsNames,
15+
::testing::Values(ov::test::utils::DEVICE_CPU),
16+
ExecGraphDuplicateInputsOutputsNames::getTestCaseName);
17+
18+
} // namespace

src/plugins/intel_npu/src/al/include/intel_npu/al/icompiler.hpp

+92-32
Original file line numberDiff line numberDiff line change
@@ -10,6 +10,7 @@
1010
#include <memory>
1111
#include <set>
1212
#include <string>
13+
#include <string_view>
1314
#include <unordered_map>
1415
#include <unordered_set>
1516

@@ -22,48 +23,107 @@
2223
namespace intel_npu {
2324

2425
/**
25-
* @brief A helper structure used for storing the metadata found within the I/O nodes.
26-
* @details The "legacyName" attribute holds the name most commonly used as map key for multiple structures.
27-
* This value also corresponds to the identifier used by the OpenVINO 1.0 API.
28-
*
29-
* "originalShape" corresponds to the shape registered in the graph, while "transposedShape" holds the shape obtained
30-
* upon applying a transposition corresponding to the legacy layout value. Use the "transposedShape" one if not sure
31-
* which one you need.
26+
* @brief A helper structure used for storing metadata corresponding to one input/output entry.
3227
*/
33-
struct IONodeDescriptor {
34-
std::string legacyName;
35-
std::string currentNodeName;
28+
struct IODescriptor {
29+
/**
30+
* @brief The name of the input/output assigned by the compiler.
31+
* @details This value may differ from other name attributes:
32+
* - The compiler could have created additional inputs/outputs (e.g. for representing states). These are not
33+
* found in the original IR model.
34+
* - The compiler may append indices to names in the case where duplicate names are found.
35+
* @note The prefixes introduced by the compiler in order to differentiate the special cases (e.g. states and shape
36+
* tensors) were removed prior to initializing this field.
37+
*/
38+
std::string nameFromCompiler;
39+
40+
ov::element::Type precision;
41+
42+
ov::PartialShape shapeFromCompiler;
43+
44+
/**
45+
* @brief If set to "true", the current object describes a buffer which may be used for altering a state tensor.
46+
* @details This flag is set if the compiler prefixed the name using a "read value" prefix. The state input and
47+
* state output descriptors are also tied using the "relatedDescriptorIndex" attribute.
48+
*/
49+
bool isStateInput = false;
50+
51+
/**
52+
* @brief If set to "true", the current object describes a buffer which reflects the value of a state tensor.
53+
* @details This flag is set if the compiler prefixed the name using an "assign" prefix. The state input and
54+
* state output descriptors are also tied using the "relatedDescriptorIndex" attribute.
55+
*/
56+
bool isStateOutput = false;
57+
58+
/**
59+
* @brief If set to "true", the buffer of the tensor described here contains as value the shape of the referenced
60+
* tensor.
61+
* @details This flag is set if the compiler prefixed the name using a "shape" prefix.
62+
*
63+
* The referenced tensor bears the same name ("nameFromCompiler"), but its "isShapeTensor" value is set to
64+
* "false". The two descriptors are also tied using the "relatedDescriptorIndex" attribute.
65+
*/
66+
bool isShapeTensor = false;
67+
68+
/**
69+
* @brief Points towards a related descriptor.
70+
* @details The related descriptors are defined by (state input, state output) or (dynamic tensor, shape tensor)
71+
* pairs.
72+
*/
73+
std::optional<size_t> relatedDescriptorIndex;
74+
75+
/**
76+
* @brief The friendly name of the node extracted from the IR model.
77+
* @details In some cases, this field is required for constructing a dummy model which uses the same input/output
78+
* metadata as the original IR model.
79+
*
80+
* This field may be empty if the I/O entry is not found in the original IR model (i.e. the entry was added by the
81+
* compiler).
82+
*/
83+
std::string nodeFriendlyName;
84+
85+
/**
86+
* @brief The names of the output tensors extracted from the IR model.
87+
* @details In some cases, this field is required for constructing a dummy model which uses the same input/output
88+
* metadata as the original IR model.
89+
*
90+
* This field may be empty if the I/O entry is not found in the original IR model (i.e. the entry was added by the
91+
* compiler).
92+
*/
3693
std::unordered_set<std::string> outputTensorNames;
37-
ov::element::Type_t precision;
38-
ov::PartialShape originalShape;
39-
ov::PartialShape transposedShape;
40-
};
4194

42-
/**
43-
* @brief A helper map to represent descriptions for inputs and outputs
44-
* of a network
45-
*/
46-
using IONodeDescriptorMap = std::unordered_map<std::string, IONodeDescriptor>;
95+
/**
96+
* @brief The shape extracted from the IR model.
97+
* @details The values may differ from the ones found in "shapeFromCompiler" if batching is to be handled by the
98+
* plugin.
99+
*
100+
* This field may be empty if the I/O entry is not found in the original IR model (i.e. the entry was added
101+
* by the compiler).
102+
*/
103+
std::optional<ov::PartialShape> shapeFromIRModel = std::nullopt;
104+
};
47105

48106
struct NetworkMetadata final {
49107
std::string name;
50108

51-
std::vector<std::string> inputNames;
52-
std::vector<std::string> outputNames;
53-
std::vector<std::string> stateNames;
54-
std::vector<std::string> shapeNames;
109+
std::vector<IODescriptor> inputs;
110+
std::vector<IODescriptor> outputs;
111+
std::vector<IODescriptor> profilingOutputs;
55112

56-
IONodeDescriptorMap parameters;
57-
IONodeDescriptorMap results;
58-
IONodeDescriptorMap states;
59-
IONodeDescriptorMap shapes;
60-
IONodeDescriptorMap profilingOutputs;
113+
size_t numStreams = 1;
61114

62-
std::unordered_map<std::string, size_t> inputOrder;
63-
std::unordered_map<std::string, size_t> outputOrder;
115+
/**
116+
* @brief Binds the (state input, state output) and (dynamic tensor, shape tensor) pairs using the
117+
* "relatedDescriptorIndex" attribute.
118+
* @details For state inputs, the "relatedDescriptorIndex" value is set to the index of the output which bears the
119+
* same name. The reverse is also applied.
120+
*
121+
* For shape tensors, the lookup is performed in the same container (inputs or outputs). The value is once again set
122+
* to the index of the entry which bears the same name.
123+
*/
124+
void bindRelatedDescriptors();
64125

65-
int numStreams = 1;
66-
};
126+
}; // namespace intel_npu
67127

68128
/**
69129
* @struct NetworkDescription

src/plugins/intel_npu/src/al/include/sync_infer_request.hpp

+42-71
Original file line numberDiff line numberDiff line change
@@ -92,56 +92,32 @@ class SyncInferRequest : public ov::IInferRequest {
9292
*/
9393
void initialize_states();
9494

95+
protected:
9596
/**
96-
* @return The state tensors accessible by their names.
97-
*/
98-
std::unordered_map<std::string, std::shared_ptr<VariableState>>& get_variable_states() {
99-
return _variableStates;
100-
}
101-
102-
/**
103-
* @return The names used by the inputs in the order registered inside the model.
104-
*/
105-
std::vector<std::string> get_input_names() {
106-
return _metadata.inputNames;
107-
}
108-
109-
/**
110-
* @return The names used by the outputs in the order registered inside the model.
111-
*/
112-
std::vector<std::string> get_output_names() {
113-
return _metadata.outputNames;
114-
}
115-
116-
/**
117-
* @return The names used by the state variables in the order registered inside the model.
97+
* @see ov::ISyncInferRequest
11898
*/
119-
std::vector<std::string> get_state_names() {
120-
return _metadata.stateNames;
121-
}
99+
struct FoundPort {
100+
size_t idx;
101+
enum class Type { NOT_FOUND = 0, INPUT, OUTPUT } type;
122102

123-
/**
124-
* @return The names used by the shape variables in the order registered inside the model.
125-
*/
126-
std::vector<std::string> get_shape_names() {
127-
return _metadata.shapeNames;
128-
}
103+
bool found() {
104+
return type != Type::NOT_FOUND;
105+
}
106+
bool is_input() {
107+
return type == Type::INPUT;
108+
}
109+
bool is_output() {
110+
return !is_input();
111+
}
112+
};
129113

130114
/**
131-
* @return A map holding references towards all tensors used by the current inference request object.
115+
* @brief Finds input or output port
116+
* @return structure which contains index of Input/Output or report that port wasn't found
117+
* @see ov::ISyncInferRequest
132118
*/
133-
std::unordered_map<std::string, std::shared_ptr<ov::ITensor>>& get_all_tensors() {
134-
return _allTensors;
135-
}
119+
FoundPort find_port(const ov::Output<const ov::Node>& port) const;
136120

137-
/**
138-
* @return A map holding references towards all shapes tensors used by the current inference request object.
139-
*/
140-
std::unordered_map<std::string, std::shared_ptr<ov::ITensor>>& get_shapes_tensors() {
141-
return _shapesTensors;
142-
}
143-
144-
protected:
145121
/**
146122
* @brief Basic checks for input/output tensor
147123
*
@@ -163,45 +139,40 @@ class SyncInferRequest : public ov::IInferRequest {
163139
virtual void check_network_precision(const ov::element::Type_t precision) const = 0;
164140

165141
/**
166-
* @brief Indicates a kind of provided tensor. Marks special tensors, used for internal implementation
167-
*/
168-
enum class TensorType { InputOrOutput, Shape, State };
169-
170-
/**
171-
* @brief Allocates a tensor on host and stores the reference inside the "_allTensors" attribute. If a buffer
172-
* address is provided, then the tensor is built upon it and no additional data buffer is allocated.
173-
* @param tensorName The name by which the tensor shall be identified
142+
* @brief Allocates a tensor on host and stores the reference inside multiple attributes.
174143
* @param descriptor Tensor's metadata
175-
* @param isState If true, the tensor shall also be stored inside the state variables map. In this case, adding the
176-
* tensor to this structure would be required in order to correctly answer the state queries.
144+
* @param index The index which the allocated tensor shall use.
145+
* @param isInput Determines the containers in which the newly allocated tensors will be stored.
177146
* @param allocator If provided, the tensor uses the custom allocator instead of using the default one.
147+
* @param batchSize If provided, the value of the shape on the 0th axis is overriden with this value.
148+
* @return Pointer towards the allocated tensor
178149
*/
179-
void allocate_tensor(std::string tensorName,
180-
const IONodeDescriptor& descriptor,
181-
TensorType tensorType = TensorType::InputOrOutput,
182-
const ov::Allocator& allocator = {}) const;
183-
184-
// Mutable to return reference to ov::Tensor
185-
mutable std::unordered_map<std::string, std::shared_ptr<ov::ITensor>> _allTensors;
186-
mutable std::unordered_map<std::string, std::shared_ptr<ov::ITensor>> _shapesTensors;
187-
// A copy of each tensor is needed to maintain the original L0 memory allocation in case the user provides another
188-
// memory area for the tensor.
189-
mutable std::unordered_map<std::string, std::shared_ptr<ov::ITensor>> _copyAllTensors;
190-
191-
mutable std::unordered_map<std::string, std::shared_ptr<VariableState>> _variableStates;
150+
std::shared_ptr<ov::ITensor> allocate_tensor(const IODescriptor& descriptor,
151+
const size_t index,
152+
const bool isInput,
153+
const ov::Allocator& allocator = {},
154+
const std::optional<std::size_t> batchSize = std::nullopt) const;
192155

193156
// This is intel_npu::ICompiledModel pointer, but need to use OV base class because
194157
// ov::IInferRequest::get_compiled_model returns a refernce to shared_ptr!
195158
std::shared_ptr<const ov::ICompiledModel> _compiledModel;
196159

197160
NetworkMetadata _metadata;
198161

199-
// Stored in order to avoid additional processing when launching inferences
200-
std::vector<std::string> _inputAndStateInputNames;
201-
std::vector<std::string> _outputAndStateOutputNames;
162+
mutable std::vector<std::shared_ptr<ov::ITensor>> _userInputTensors;
163+
mutable std::vector<std::shared_ptr<ov::ITensor>> _userOutputTensors;
202164

203-
std::unordered_map<std::string, std::string> _nodeNameToLegacyName;
204-
std::unordered_map<std::string, std::string> _legacyNameToNodeName;
165+
mutable std::vector<ov::SoPtr<ov::IVariableState>> _variableStates;
166+
167+
/**
168+
* @see ov::ISyncInferRequest
169+
*/
170+
mutable std::unordered_map<size_t, FoundPort> _cachedPorts;
171+
172+
/**
173+
* @see ov::ISyncInferRequest
174+
*/
175+
mutable std::mutex _cacheMutex;
205176
};
206177

207178
} // namespace intel_npu
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,67 @@
1+
// Copyright (C) 2018-2024 Intel Corporation
2+
// SPDX-License-Identifier: Apache-2.0
3+
//
4+
5+
#include "intel_npu/al/icompiler.hpp"
6+
7+
namespace intel_npu {
8+
9+
void NetworkMetadata::bindRelatedDescriptors() {
10+
size_t ioIndex = 0;
11+
12+
for (IODescriptor& input : inputs) {
13+
if (input.relatedDescriptorIndex.has_value()) {
14+
++ioIndex;
15+
continue;
16+
}
17+
18+
if (input.isStateInput) {
19+
const auto relatedDescriptorIterator =
20+
std::find_if(outputs.begin(), outputs.end(), [&](const IODescriptor& output) {
21+
return output.isStateOutput && (output.nameFromCompiler == input.nameFromCompiler);
22+
});
23+
24+
if (relatedDescriptorIterator != outputs.end()) {
25+
input.relatedDescriptorIndex = std::distance(outputs.begin(), relatedDescriptorIterator);
26+
outputs.at(*input.relatedDescriptorIndex).relatedDescriptorIndex = ioIndex;
27+
}
28+
} else if (input.isShapeTensor) {
29+
const auto relatedDescriptorIterator =
30+
std::find_if(inputs.begin(), inputs.end(), [&](const IODescriptor& candidate) {
31+
return !candidate.isShapeTensor && (candidate.nameFromCompiler == input.nameFromCompiler);
32+
});
33+
34+
if (relatedDescriptorIterator != inputs.end()) {
35+
input.relatedDescriptorIndex = std::distance(inputs.begin(), relatedDescriptorIterator);
36+
inputs.at(*input.relatedDescriptorIndex).relatedDescriptorIndex = ioIndex;
37+
}
38+
}
39+
40+
++ioIndex;
41+
}
42+
43+
ioIndex = 0;
44+
45+
for (IODescriptor& output : outputs) {
46+
if (output.relatedDescriptorIndex.has_value()) {
47+
++ioIndex;
48+
continue;
49+
}
50+
51+
if (output.isShapeTensor) {
52+
const auto relatedDescriptorIterator =
53+
std::find_if(outputs.begin(), outputs.end(), [&](const IODescriptor& candidate) {
54+
return !candidate.isShapeTensor && (candidate.nameFromCompiler == output.nameFromCompiler);
55+
});
56+
57+
if (relatedDescriptorIterator != outputs.end()) {
58+
output.relatedDescriptorIndex = std::distance(outputs.begin(), relatedDescriptorIterator);
59+
outputs.at(*output.relatedDescriptorIndex).relatedDescriptorIndex = ioIndex;
60+
}
61+
}
62+
63+
++ioIndex;
64+
}
65+
}
66+
67+
} // namespace intel_npu

0 commit comments

Comments
 (0)