@@ -92,56 +92,32 @@ class SyncInferRequest : public ov::IInferRequest {
92
92
*/
93
93
void initialize_states ();
94
94
95
+ protected:
95
96
/* *
96
- * @return The state tensors accessible by their names.
97
- */
98
- std::unordered_map<std::string, std::shared_ptr<VariableState>>& get_variable_states () {
99
- return _variableStates;
100
- }
101
-
102
- /* *
103
- * @return The names used by the inputs in the order registered inside the model.
104
- */
105
- std::vector<std::string> get_input_names () {
106
- return _metadata.inputNames ;
107
- }
108
-
109
- /* *
110
- * @return The names used by the outputs in the order registered inside the model.
111
- */
112
- std::vector<std::string> get_output_names () {
113
- return _metadata.outputNames ;
114
- }
115
-
116
- /* *
117
- * @return The names used by the state variables in the order registered inside the model.
97
+ * @see ov::ISyncInferRequest
118
98
*/
119
- std::vector<std::string> get_state_names () {
120
- return _metadata. stateNames ;
121
- }
99
+ struct FoundPort {
100
+ size_t idx ;
101
+ enum class Type { NOT_FOUND = 0 , INPUT, OUTPUT } type;
122
102
123
- /* *
124
- * @return The names used by the shape variables in the order registered inside the model.
125
- */
126
- std::vector<std::string> get_shape_names () {
127
- return _metadata.shapeNames ;
128
- }
103
+ bool found () {
104
+ return type != Type::NOT_FOUND;
105
+ }
106
+ bool is_input () {
107
+ return type == Type::INPUT;
108
+ }
109
+ bool is_output () {
110
+ return !is_input ();
111
+ }
112
+ };
129
113
130
114
/* *
131
- * @return A map holding references towards all tensors used by the current inference request object.
115
+ * @brief Finds input or output port
116
+ * @return structure which contains index of Input/Output or report that port wasn't found
117
+ * @see ov::ISyncInferRequest
132
118
*/
133
- std::unordered_map<std::string, std::shared_ptr<ov::ITensor>>& get_all_tensors () {
134
- return _allTensors;
135
- }
119
+ FoundPort find_port (const ov::Output<const ov::Node>& port) const ;
136
120
137
- /* *
138
- * @return A map holding references towards all shapes tensors used by the current inference request object.
139
- */
140
- std::unordered_map<std::string, std::shared_ptr<ov::ITensor>>& get_shapes_tensors () {
141
- return _shapesTensors;
142
- }
143
-
144
- protected:
145
121
/* *
146
122
* @brief Basic checks for input/output tensor
147
123
*
@@ -163,45 +139,40 @@ class SyncInferRequest : public ov::IInferRequest {
163
139
virtual void check_network_precision (const ov::element::Type_t precision) const = 0;
164
140
165
141
/* *
166
- * @brief Indicates a kind of provided tensor. Marks special tensors, used for internal implementation
167
- */
168
- enum class TensorType { InputOrOutput, Shape, State };
169
-
170
- /* *
171
- * @brief Allocates a tensor on host and stores the reference inside the "_allTensors" attribute. If a buffer
172
- * address is provided, then the tensor is built upon it and no additional data buffer is allocated.
173
- * @param tensorName The name by which the tensor shall be identified
142
+ * @brief Allocates a tensor on host and stores the reference inside multiple attributes.
174
143
* @param descriptor Tensor's metadata
175
- * @param isState If true, the tensor shall also be stored inside the state variables map. In this case, adding the
176
- * tensor to this structure would be required in order to correctly answer the state queries .
144
+ * @param index The index which the allocated tensor shall use.
145
+ * @param isInput Determines the containers in which the newly allocated tensors will be stored .
177
146
* @param allocator If provided, the tensor uses the custom allocator instead of using the default one.
147
+ * @param batchSize If provided, the value of the shape on the 0th axis is overriden with this value.
148
+ * @return Pointer towards the allocated tensor
178
149
*/
179
- void allocate_tensor (std::string tensorName,
180
- const IONodeDescriptor& descriptor,
181
- TensorType tensorType = TensorType::InputOrOutput,
182
- const ov::Allocator& allocator = {}) const ;
183
-
184
- // Mutable to return reference to ov::Tensor
185
- mutable std::unordered_map<std::string, std::shared_ptr<ov::ITensor>> _allTensors;
186
- mutable std::unordered_map<std::string, std::shared_ptr<ov::ITensor>> _shapesTensors;
187
- // A copy of each tensor is needed to maintain the original L0 memory allocation in case the user provides another
188
- // memory area for the tensor.
189
- mutable std::unordered_map<std::string, std::shared_ptr<ov::ITensor>> _copyAllTensors;
190
-
191
- mutable std::unordered_map<std::string, std::shared_ptr<VariableState>> _variableStates;
150
+ std::shared_ptr<ov::ITensor> allocate_tensor (const IODescriptor& descriptor,
151
+ const size_t index,
152
+ const bool isInput,
153
+ const ov::Allocator& allocator = {},
154
+ const std::optional<std::size_t > batchSize = std::nullopt) const ;
192
155
193
156
// This is intel_npu::ICompiledModel pointer, but need to use OV base class because
194
157
// ov::IInferRequest::get_compiled_model returns a refernce to shared_ptr!
195
158
std::shared_ptr<const ov::ICompiledModel> _compiledModel;
196
159
197
160
NetworkMetadata _metadata;
198
161
199
- // Stored in order to avoid additional processing when launching inferences
200
- std::vector<std::string> _inputAndStateInputNames;
201
- std::vector<std::string> _outputAndStateOutputNames;
162
+ mutable std::vector<std::shared_ptr<ov::ITensor>> _userInputTensors;
163
+ mutable std::vector<std::shared_ptr<ov::ITensor>> _userOutputTensors;
202
164
203
- std::unordered_map<std::string, std::string> _nodeNameToLegacyName;
204
- std::unordered_map<std::string, std::string> _legacyNameToNodeName;
165
+ mutable std::vector<ov::SoPtr<ov::IVariableState>> _variableStates;
166
+
167
+ /* *
168
+ * @see ov::ISyncInferRequest
169
+ */
170
+ mutable std::unordered_map<size_t , FoundPort> _cachedPorts;
171
+
172
+ /* *
173
+ * @see ov::ISyncInferRequest
174
+ */
175
+ mutable std::mutex _cacheMutex;
205
176
};
206
177
207
178
} // namespace intel_npu
0 commit comments