Skip to content

Commit 32efdaf

Browse files
authoredFeb 26, 2025··
[DOCS] Moving OVMS to top menu level (#29050)
1 parent 2d14fa9 commit 32efdaf

File tree

9 files changed

+32
-21
lines changed

9 files changed

+32
-21
lines changed
 

‎docs/CMakeLists.txt

+1-1
Original file line numberDiff line numberDiff line change
@@ -84,7 +84,7 @@ function(build_docs)
8484
list(APPEND commands COMMAND ${Python3_EXECUTABLE} ${FILE_HELPER_SCRIPT}
8585
--filetype=md
8686
--input_dir=${OVMS_DOCS_DIR}
87-
--output_dir=${SPHINX_SOURCE_DIR}/openvino-workflow/model-server
87+
--output_dir=${SPHINX_SOURCE_DIR}/model-server
8888
--exclude_dir=${SPHINX_SOURCE_DIR})
8989
list(APPEND commands COMMAND ${CMAKE_COMMAND} -E cmake_echo_color --green "FINISHED preprocessing OVMS")
9090
endif()

‎docs/articles_en/about-openvino/openvino-ecosystem/openvino-project.rst

+1-1
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,7 @@ and TensorFlow models during training.
3838

3939
| **OpenVINO Model Server**
4040
| :bdg-link-dark:`GitHub <https://github.com/openvinotoolkit/model_server>`
41-
:bdg-link-success:`User Guide <https://docs.openvino.ai/2025/openvino-workflow/model-server/ovms_what_is_openvino_model_server.html>`
41+
:bdg-link-success:`User Guide <https://docs.openvino.ai/2025/model-server/ovms_what_is_openvino_model_server.html>`
4242
4343
A high-performance system that can be used to access the host models via request to the model
4444
server.

‎docs/articles_en/about-openvino/openvino-ecosystem/openvino-project/openvino-security-add-on.rst

+1-1
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@ In this release, one person performs the role of both the Model Developer and th
1717
Overview
1818
########
1919

20-
The OpenVINO™ Security Add-on works with the :doc:`OpenVINO™ Model Server <../../../openvino-workflow/model-server/ovms_what_is_openvino_model_server>` on Intel® architecture. Together, the OpenVINO™ Security Add-on and the OpenVINO™ Model Server provide a way for Model Developers and Independent Software Vendors to use secure packaging and secure model execution to enable access control to the OpenVINO™ models, and for model Users to run inference within assigned limits.
20+
The OpenVINO™ Security Add-on works with the :doc:`OpenVINO™ Model Server <../../../../model-server/ovms_what_is_openvino_model_server>` on Intel® architecture. Together, the OpenVINO™ Security Add-on and the OpenVINO™ Model Server provide a way for Model Developers and Independent Software Vendors to use secure packaging and secure model execution to enable access control to the OpenVINO™ models, and for model Users to run inference within assigned limits.
2121

2222
The OpenVINO™ Security Add-on consists of three components that run in Kernel-based Virtual Machines (KVMs). These components provide a way to run security-sensitive operations in an isolated environment. A brief description of the three components are as follows. Click each triangled line for more information about each.
2323

‎docs/articles_en/about-openvino/performance-benchmarks.rst

+1-1
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@ Performance Benchmarks
1818

1919
This page presents benchmark results for the
2020
`Intel® Distribution of OpenVINO™ toolkit <https://software.intel.com/content/www/us/en/develop/tools/openvino-toolkit.html>`__
21-
and :doc:`OpenVINO Model Server <../openvino-workflow/model-server/ovms_what_is_openvino_model_server>`, for a representative
21+
and :doc:`OpenVINO Model Server <../../model-server/ovms_what_is_openvino_model_server>`, for a representative
2222
selection of public neural networks and Intel® devices. The results may help you decide which
2323
hardware to use in your applications or plan AI workload for the hardware you have already
2424
implemented in your solutions. Click the buttons below to see the chosen benchmark data.

‎docs/articles_en/openvino-workflow-generative.rst

+19-8
Original file line numberDiff line numberDiff line change
@@ -55,14 +55,22 @@ options:
5555
as well as conversion on the fly. For integration with the final product it may offer
5656
lower performance, though.
5757

58-
.. tab-item:: Base OpenVINO (not recommended)
58+
.. tab-item:: OpenVINO™ Model Server
5959

60-
Note that the base version of OpenVINO may also be used to run generative AI. Although it may
61-
offer a simpler environment, with fewer dependencies, it has significant limitations and a more
62-
demanding implementation process.
60+
| - Easy and quick deployment of models to edge or cloud.
61+
| - Includes endpoints for serving generative AI models.
62+
| - Available in both Python and C++.
63+
| - Allows client applications in any programming language that supports REST or gRPC.
6364
64-
To learn more, refer to the article for the 2024.6 OpenVINO version:
65-
`Generative AI with Base OpenVINO <https://docs.openvino.ai/2024/learn-openvino/llm_inference_guide/llm-inference-native-ov.html>`__
65+
:doc:`OpenVINO™ Model Server <model-server/ovms_what_is_openvino_model_server>`
66+
provides a set of REST API endpoints dedicated to generative use cases. The endpoints
67+
simplify writing AI applications, ensure scalability, and provide state-of-the-art
68+
performance optimizations. They include OpenAI API for:
69+
`text generation <https://openvino-doc.iotg.sclab.intel.com/seba-test-8/model-server/ovms_docs_rest_api_chat.html>`__,
70+
`embeddings <https://openvino-doc.iotg.sclab.intel.com/seba-test-8/model-server/ovms_docs_rest_api_embeddings.html>`__,
71+
and `reranking <https://openvino-doc.iotg.sclab.intel.com/seba-test-8/model-server/ovms_docs_rest_api_rerank.html>`__.
72+
The model server supports deployments as containers or binary applications on Linux and Windows with CPU or GPU acceleration.
73+
See the :doc:`demos <model-server/ovms_docs_demos>`.
6674

6775

6876

@@ -94,10 +102,13 @@ The advantages of using OpenVINO for generative model deployment:
94102
better performance than Python-based runtimes.
95103
96104

105+
You can run Generative AI models, using native OpenVINO API, although it is not recommended.
106+
If you want to learn how to do it, refer to
107+
`the 24.6 documentation <https://docs.openvino.ai/2024/learn-openvino/llm_inference_guide/llm-inference-native-ov.html>`__.
108+
109+
97110
Proceed to guides on:
98111

99112
* :doc:`OpenVINO GenAI <./openvino-workflow-generative/inference-with-genai>`
100113
* :doc:`Hugging Face and Optimum Intel <./openvino-workflow-generative/inference-with-optimum-intel>`
101114
* `Generative AI with Base OpenVINO <https://docs.openvino.ai/2024/learn-openvino/llm_inference_guide/llm-inference-native-ov.html>`__
102-
103-

‎docs/articles_en/openvino-workflow.rst

+4-5
Original file line numberDiff line numberDiff line change
@@ -13,8 +13,7 @@ OpenVINO Workflow
1313
Model Preparation <openvino-workflow/model-preparation>
1414
openvino-workflow/model-optimization
1515
Running Inference <openvino-workflow/running-inference>
16-
Deployment on a Local System <openvino-workflow/deployment-locally>
17-
Deployment on a Model Server <openvino-workflow/model-server/ovms_what_is_openvino_model_server>
16+
Deployment on a Local System <openvino-workflow/deployment-locally>
1817
openvino-workflow/torch-compile
1918

2019

@@ -86,11 +85,11 @@ OpenVINO uses the following functions for reading, converting, and saving models
8685
and the quickest way of running a deep learning model.
8786
8887
| :doc:`Deployment Option 1. Using OpenVINO Runtime <openvino-workflow/deployment-locally>`
89-
| Deploy a model locally, reading the file directly from your application and utilizing about-openvino/additional-resources available to the system.
88+
| Deploy a model locally, reading the file directly from your application and utilizing resources available to the system.
9089
| Deployment on a local system uses the steps described in the section on running inference.
9190
92-
| :doc:`Deployment Option 2. Using Model Server <openvino-workflow/model-server/ovms_what_is_openvino_model_server>`
93-
| Deploy a model remotely, connecting your application to an inference server and utilizing external about-openvino/additional-resources, with no impact on the app's performance.
91+
| :doc:`Deployment Option 2. Using Model Server <../model-server/ovms_what_is_openvino_model_server>`
92+
| Deploy a model remotely, connecting your application to an inference server and utilizing external resources, with no impact on the app's performance.
9493
| Deployment on OpenVINO Model Server is quick and does not require any additional steps described in the section on running inference.
9594
9695
| :doc:`Deployment Option 3. Using torch.compile for PyTorch 2.0 <openvino-workflow/torch-compile>`

‎docs/articles_en/openvino-workflow/running-inference/stateful-models.rst

+1-1
Original file line numberDiff line numberDiff line change
@@ -140,4 +140,4 @@ sequences.
140140
You can find more examples demonstrating how to work with states in other articles:
141141

142142
* `LLaVA-NeXT Multimodal Chatbot notebook <../../notebooks/llava-next-multimodal-chatbot-with-output.html>`__
143-
* :doc:`Serving Stateful Models with OpenVINO Model Server <../../openvino-workflow/model-server/ovms_docs_stateful_models>`
143+
* :doc:`Serving Stateful Models with OpenVINO Model Server <../../model-server/ovms_docs_stateful_models>`

‎docs/sphinx_setup/_static/selector-tool/assets/selector-C_bHXmvJ.js

+1-1
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

‎docs/sphinx_setup/index.rst

+3-2
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,7 @@ hardware and environments, on-premises and on-device, in the browser or in the c
3838
<li id="ov-homepage-slide3" class="splide__slide">
3939
<p class="ov-homepage-slide-title">Improved model serving</p>
4040
<p class="ov-homepage-slide-subtitle">OpenVINO Model Server has improved parallel inference!</p>
41-
<a class="ov-homepage-banner-btn" href="https://docs.openvino.ai/2025/openvino-workflow/model-server/ovms_what_is_openvino_model_server.html">Learn more</a>
41+
<a class="ov-homepage-banner-btn" href="https://docs.openvino.ai/2025/model-server/ovms_what_is_openvino_model_server.html">Learn more</a>
4242
</li>
4343
<li id="ov-homepage-slide4" class="splide__slide">
4444
<p class="ov-homepage-slide-title">OpenVINO via PyTorch 2.0 torch.compile()</p>
@@ -124,7 +124,7 @@ Places to Begin
124124

125125
Cloud-ready deployments for microservice applications.
126126

127-
.. button-link:: openvino-workflow/model-server/ovms_what_is_openvino_model_server.html
127+
.. button-link:: model-server/ovms_what_is_openvino_model_server.html
128128
:color: primary
129129
:outline:
130130

@@ -195,5 +195,6 @@ Key Features
195195
GET STARTED <get-started>
196196
HOW TO USE - MAIN WORKFLOW <openvino-workflow>
197197
HOW TO USE - GENERATIVE AI WORKFLOW <openvino-workflow-generative>
198+
HOW TO USE - MODEL SERVING <model-server/ovms_what_is_openvino_model_server>
198199
REFERENCE DOCUMENTATION <documentation>
199200
ABOUT OPENVINO <about-openvino>

0 commit comments

Comments
 (0)
Please sign in to comment.