|
| 1 | +Key Features |
| 2 | +============== |
| 3 | + |
| 4 | +Easy Integration |
| 5 | +######################### |
| 6 | + |
| 7 | +| :doc:`Support for multiple frameworks <../openvino-workflow/model-preparation/convert-model-to-ir>` |
| 8 | +| Use deep learning models from PyTorch, TensorFlow, TensorFlow Lite, PaddlePaddle, and ONNX |
| 9 | + directly or convert them to the optimized OpenVINO IR format for improved performance. |
| 10 | +
|
| 11 | +| :doc:`Close integration with PyTorch <../openvino-workflow/torch-compile>` |
| 12 | +| For PyTorch-based applications, specify OpenVINO as a backend using |
| 13 | + :doc:`torch.compile <../openvino-workflow/torch-compile>` to improve model inference. Apply |
| 14 | + OpenVINO optimizations to your PyTorch models directly with a single line of code. |
| 15 | +
|
| 16 | +| :doc:`GenAI Out Of The Box <../learn-openvino/llm_inference_guide/genai-guide>` |
| 17 | +| With the genAI flavor of OpenVINO, you can run generative AI with just a couple lines of code. |
| 18 | + Check out the GenAI guide for instructions on how to do it. |
| 19 | +
|
| 20 | +| `Python / C++ / C / NodeJS APIs <https://docs.openvino.ai/2024/api/api_reference.html>`__ |
| 21 | +| OpenVINO offers the C++ API as a complete set of available methods. For less resource-critical |
| 22 | + solutions, the Python API provides almost full coverage, while C and NodeJS ones are limited |
| 23 | + to the methods most basic for their typical environments. The NodeJS API, is still in its |
| 24 | + early and active development. |
| 25 | +
|
| 26 | +| :doc:`Open source and easy to extend <../about-openvino/contributing>` |
| 27 | +| If you need a particular feature or inference accelerator to be supported, you are free to file |
| 28 | + a feature request or develop new components specific to your projects yourself. As open source, |
| 29 | + OpenVINO may be used and modified freely. See the extensibility guide for more information on |
| 30 | + how to adapt it to your needs. |
| 31 | +
|
| 32 | +Deployment |
| 33 | +######################### |
| 34 | + |
| 35 | +| :doc:`Local or remote <../openvino-workflow>` |
| 36 | +| Integrate the OpenVINO runtime directly with your application to run inference locally or use |
| 37 | + `OpenVINO Model Server <https://github.com/openvinotoolkit/model_server>`__ to shift the inference |
| 38 | + workload to a remote system, a separate server or a Kubernetes environment. For serving, |
| 39 | + OpenVINO is also integrated with `vLLM <https://docs.vllm.ai/en/stable/getting_started/openvino-installation.html>`__ |
| 40 | + and `Triton <https://github.com/triton-inference-server/openvino_backend>`__ services. |
| 41 | +
|
| 42 | +| :doc:`Scalable and portable <release-notes-openvino/system-requirements>` |
| 43 | +| Write an application once, deploy it anywhere, always making the most out of your hardware setup. |
| 44 | + The automatic device selection mode gives you the ultimate deployment flexibility on all major |
| 45 | + operating systems. Check out system requirements. |
| 46 | +
|
| 47 | +| **Light-weight** |
| 48 | +| Designed with minimal external dependencies, OpenVINO does not bloat your application |
| 49 | + and simplifies installation and dependency management. The custom compilation for your specific |
| 50 | + model(s) may further reduce the final binary size. |
| 51 | +
|
| 52 | +Performance |
| 53 | +######################### |
| 54 | + |
| 55 | +| :doc:`Model Optimization <../openvino-workflow/model-optimization>` |
| 56 | +| Optimize your deep learning models with NNCF, using various training-time and post-training |
| 57 | + compression methods, such as pruning, sparsity, quantization, and weight compression. Make |
| 58 | + your models take less space, run faster, and use less resources. |
| 59 | +
|
| 60 | +| :doc:`Top performance <../about-openvino/performance-benchmarks>` |
| 61 | +| OpenVINO is optimized to work with Intel hardware, delivering confirmed high performance for |
| 62 | + hundreds of models. Explore OpenVINO Performance Benchmarks to discover the optimal hardware |
| 63 | + configurations and plan your AI deployment based on verified data. |
| 64 | +
|
| 65 | +| :doc:`Enhanced App Start-Up Time <../openvino-workflow/running-inference/optimize-inference>` |
| 66 | +| If you need your application to launch immediately, OpenVINO will reduce first-inference latency, |
| 67 | + running inference on CPU until a more suited device is ready to take over. Once a model |
| 68 | + is compiled for inference, it is also cached, improving the start-up time even more. |
| 69 | +
|
0 commit comments