File tree 1 file changed +9
-5
lines changed
1 file changed +9
-5
lines changed Original file line number Diff line number Diff line change @@ -62,12 +62,16 @@ options:
62
62
| - Available in both Python and C++.
63
63
| - Allows client applications in any programming language that supports REST or gRPC.
64
64
65
- Deploy deep learning models remotely, using
66
65
:doc: `OpenVINO™ Model Server <model-server/ovms_what_is_openvino_model_server >`
67
- - a high-performance solution, developed in C++ for scalability and optimized for
68
- Intel architectures. The deployment is straightforward, as you simply connect
69
- your application via gRPC or REST endpoints to a server, where OpenVINO's logic for
70
- inference tasks is applied.
66
+ provides a set of REST API endpoints dedicated to generative use cases. The endpoints
67
+ simplify writing AI applications, ensure scalability, and provide state-of-the-art
68
+ performance optimizations. They include OpenAI API for:
69
+ `text generation <https://openvino-doc.iotg.sclab.intel.com/seba-test-8/model-server/ovms_docs_rest_api_chat.html >`__,
70
+ `embeddings <https://openvino-doc.iotg.sclab.intel.com/seba-test-8/model-server/ovms_docs_rest_api_embeddings.html `__,
71
+ and `reranking <https://openvino-doc.iotg.sclab.intel.com/seba-test-8/model-server/ovms_docs_rest_api_rerank.html `__.
72
+ The model server supports deployments as containers or binary applications on Linux and Windows with CPU or GPU acceleration.
73
+ See the :doc: `demos <model-server/ovms_docs_demos >`.
74
+
71
75
72
76
73
77
The advantages of using OpenVINO for generative model deployment:
You can’t perform that action at this time.
0 commit comments