Skip to content

Commit dca5ddc

Browse files
[DOCS] relnotes 24.3 first pass fix (#25799)
1 parent d47a579 commit dca5ddc

File tree

2 files changed

+29
-26
lines changed

2 files changed

+29
-26
lines changed

docs/articles_en/about-openvino/release-notes-openvino.rst

+17-19
Original file line numberDiff line numberDiff line change
@@ -1,9 +1,10 @@
1+
OpenVINO Release Notes
2+
=============================
3+
14
.. meta::
25
:description: See what has changed in OpenVINO with the latest release, as well as all
36
previous releases in this year's cycle.
47

5-
OpenVINO Release Notes
6-
=============================
78

89
.. toctree::
910
:maxdepth: 1
@@ -14,7 +15,7 @@ OpenVINO Release Notes
1415

1516

1617

17-
2024.3 - 30 July 2024
18+
2024.3 - 31 July 2024
1819
#############################
1920

2021
:doc:`System Requirements <./release-notes-openvino/system-requirements>` | :doc:`Release policy <./release-notes-openvino/release-policy>` | :doc:`Installation Guides <./../get-started/install-openvino>`
@@ -23,21 +24,21 @@ OpenVINO Release Notes
2324
What's new
2425
+++++++++++++++++++++++++++++
2526

26-
More Gen AI coverage and framework integrations to minimize code changes.
27+
* More Gen AI coverage and framework integrations to minimize code changes.
2728

28-
* OpenVINO pre-optimized models are now available in Hugging Face making it easier for developers
29-
to get started with these models.
29+
* OpenVINO pre-optimized models are now available in Hugging Face making it easier for developers
30+
to get started with these models.
3031

31-
Broader Large Language Model (LLM) support and more model compression techniques.
32+
* Broader Large Language Model (LLM) support and more model compression techniques.
3233

33-
* Significant improvement in LLM performance on Intel built-in and discrete GPUs with the addition
34-
of dynamic quantization, Multi-Head Attention (MHA), and OneDNN enhancements.
34+
* Significant improvement in LLM performance on Intel discrete GPUs with the addition of
35+
Multi-Head Attention (MHA) and OneDNN enhancements.
3536

36-
More portability and performance to run AI at the edge, in the cloud, or locally.
37+
* More portability and performance to run AI at the edge, in the cloud, or locally.
3738

38-
* Improved CPU performance when serving LLMs with the inclusion of vLLM and continuous batching
39-
in the OpenVINO Model Server (OVMS). vLLM is an easy-to-use open-source library that supports
40-
efficient LLM inferencing and model serving.
39+
* Improved CPU performance when serving LLMs with the inclusion of vLLM and continuous batching
40+
in the OpenVINO Model Server (OVMS). vLLM is an easy-to-use open-source library that supports
41+
efficient LLM inferencing and model serving.
4142

4243

4344

@@ -59,7 +60,7 @@ Common
5960

6061
* Increasing support for models like YoloV10 or PixArt-XL-2, thanks to enabling Squeeze and
6162
Concat layers.
62-
* Performance of precision conversion fp16/bf16 -> fp32.
63+
* Performance of precision conversion FP16/BF16 -> FP32.
6364

6465

6566

@@ -97,9 +98,6 @@ GPU Device Plugin
9798

9899
* LLMs and Stable Diffusion on discrete GPUs, due to latency decrease, through optimizations
99100
such as Multi-Head Attention (MHA) and oneDNN improvements.
100-
* First token latency of LLMs for large input cases on Core Ultra integrated GPU. It can be
101-
further improved with dynamic quantization enabled with an application
102-
`interface <https://docs.openvino.ai/2024/api/c_cpp_api/group__ov__dev__exec__model.html#_CPPv4N2ov4hint31dynamic_quantization_group_sizeE>`__.
103101
* Whisper models on discrete GPU.
104102

105103

@@ -191,7 +189,7 @@ Neural Network Compression Framework
191189
Act->MatMul and Act->MUltiply->MatMul to cover the Phi family models.
192190
* The representation of symmetrically quantized weights has been updated to a signed data type
193191
with no zero point. This allows NPU to support compressed LLMs with the symmetric mode.
194-
* bf16 models in Post-Training Quantization are now supported; nncf.quantize().
192+
* BF16 models in Post-Training Quantization are now supported; nncf.quantize().
195193
* `Activation Sparsity <https://arxiv.org/abs/2310.17157>`__ (Contextual Sparsity) algorithm in
196194
the Weight Compression method is now supported (preview), speeding up LLM inference.
197195
The algorithm is enabled by setting the ``target_sparsity_by_scope`` option in
@@ -431,7 +429,7 @@ Previous 2024 releases
431429
compression of LLMs. Enabled by `gptq=True`` in nncf.compress_weights().
432430
* Scale Estimation algorithm for more accurate 4-bit compressed LLMs. Enabled by
433431
`scale_estimation=True`` in nncf.compress_weights().
434-
* Added support for models with bf16 weights in nncf.compress_weights().
432+
* Added support for models with BF16 weights in nncf.compress_weights().
435433
* nncf.quantize() method is now the recommended path for quantization initialization of
436434
PyTorch models in Quantization-Aware Training. See example for more details.
437435
* compressed_model.nncf.get_config() and nncf.torch.load_from_config() API have been added to

docs/sphinx_setup/index.rst

+12-7
Original file line numberDiff line numberDiff line change
@@ -5,11 +5,11 @@ OpenVINO 2024.3
55
.. meta::
66
:google-site-verification: _YqumYQ98cmXUTwtzM_0WIIadtDc6r_TMYGbmGgNvrk
77

8-
**OpenVINO is an open-source toolkit** for optimizing and deploying deep learning models from cloud
9-
to edge. It accelerates deep learning inference across various use cases, such as generative AI, video,
10-
audio, and language with models from popular frameworks like PyTorch, TensorFlow, ONNX, and more.
11-
Convert and optimize models, and deploy across a mix of Intel® hardware and environments, on-premises
12-
and on-device, in the browser or in the cloud.
8+
**OpenVINO is an open-source toolkit** for optimizing and deploying deep learning models from
9+
cloud to edge. It accelerates deep learning inference across various use cases, such as
10+
generative AI, video, audio, and language with models from popular frameworks like PyTorch,
11+
TensorFlow, ONNX, and more. Convert and optimize models, and deploy across a mix of Intel®
12+
hardware and environments, on-premises and on-device, in the browser or in the cloud.
1313

1414
Check out the `OpenVINO Cheat Sheet. <https://docs.openvino.ai/2024/_static/download/OpenVINO_Quick_Start_Guide.pdf>`__
1515

@@ -26,16 +26,21 @@ Check out the `OpenVINO Cheat Sheet. <https://docs.openvino.ai/2024/_static/down
2626
<div class="splide__track">
2727
<ul class="splide__list">
2828
<li id="ov-homepage-slide1" class="splide__slide">
29+
<p class="ov-homepage-slide-title">OpenVINO models on Hugging Face!</p>
30+
<p class="ov-homepage-slide-subtitle">Get pre-optimized OpenVINO models, no need to convert!</p>
31+
<a class="ov-homepage-banner-btn" href="https://huggingface.co/OpenVINO">Visit Hugging Face</a>
32+
</li>
33+
<li id="ov-homepage-slide2" class="splide__slide">
2934
<p class="ov-homepage-slide-title">New Generative AI API</p>
3035
<p class="ov-homepage-slide-subtitle">Generate text with LLMs in only a few lines of code!</p>
3136
<a class="ov-homepage-banner-btn" href="https://docs.openvino.ai/2024/learn-openvino/llm_inference_guide/genai-guide.html">Check out our guide</a>
3237
</li>
33-
<li id="ov-homepage-slide2" class="splide__slide">
38+
<li id="ov-homepage-slide3" class="splide__slide">
3439
<p class="ov-homepage-slide-title">Improved model serving</p>
3540
<p class="ov-homepage-slide-subtitle">OpenVINO Model Server has improved parallel inferencing!</p>
3641
<a class="ov-homepage-banner-btn" href="https://docs.openvino.ai/2024/ovms_what_is_openvino_model_server.html">Learn more</a>
3742
</li>
38-
<li id="ov-homepage-slide3" class="splide__slide">
43+
<li id="ov-homepage-slide4" class="splide__slide">
3944
<p class="ov-homepage-slide-title">OpenVINO via PyTorch 2.0 torch.compile()</p>
4045
<p class="ov-homepage-slide-subtitle">Use OpenVINO directly in PyTorch-native applications!</p>
4146
<a class="ov-homepage-banner-btn" href="https://docs.openvino.ai/2024/openvino-workflow/torch-compile.html">Learn more</a>

0 commit comments

Comments
 (0)