diff --git a/docs/articles_en/learn-openvino/llm_inference_guide/genai-guide.rst b/docs/articles_en/learn-openvino/llm_inference_guide/genai-guide.rst index 06ac3642df7334..2d47f9deacc24c 100644 --- a/docs/articles_en/learn-openvino/llm_inference_guide/genai-guide.rst +++ b/docs/articles_en/learn-openvino/llm_inference_guide/genai-guide.rst @@ -317,7 +317,8 @@ OpenVINO GenAI Flavor includes the following API: * visibility - controls the visibility of the GenAI library. -Learn more about API in the `GenAI repository `__. + +Learn more in the `GenAI API reference `__. Additional Resources #################### diff --git a/docs/articles_en/openvino-workflow/model-optimization-guide/weight-compression.rst b/docs/articles_en/openvino-workflow/model-optimization-guide/weight-compression.rst index da4f34b8806aea..2d5fc0f1b98d46 100644 --- a/docs/articles_en/openvino-workflow/model-optimization-guide/weight-compression.rst +++ b/docs/articles_en/openvino-workflow/model-optimization-guide/weight-compression.rst @@ -7,10 +7,6 @@ Weight compression is a technique for enhancing the efficiency of models, especially those with large memory requirements. This method reduces the model's memory footprint, a crucial factor for Large Language Models (LLMs). -Weight compression is a technique for enhancing the efficiency of models, -especially those with large memory requirements. This method reduces the model's -memory footprint, a crucial factor for Large Language Models (LLMs). - Unlike full model quantization, where weights and activations are quantized, weight compression in `Neural Network Compression Framework (NNCF) `__ only targets the model's weights. This approach allows the activations to remain as