Overview

Request 1194790 accepted

- Update to 2024.3.0
- Summary of major features and improvements  
* More Gen AI coverage and framework integrations to minimize
code changes
+ OpenVINO pre-optimized models are now available in Hugging
Face making it easier for developers to get started with
these models.
* Broader Large Language Model (LLM) support and more model
compression techniques.
+ Significant improvement in LLM performance on Intel
discrete GPUs with the addition of Multi-Head Attention
(MHA) and OneDNN enhancements.
* More portability and performance to run AI at the edge, in the
cloud, or locally.
+ Improved CPU performance when serving LLMs with the
inclusion of vLLM and continuous batching in the OpenVINO
Model Server (OVMS). vLLM is an easy-to-use open-source
library that supports efficient LLM inferencing and model
serving.
+ Ubuntu 24.04 long-term support (LTS), 64-bit (Kernel 6.8+)
(preview support)
- Support Change and Deprecation Notices
* Using deprecated features and components is not advised.
They are available to enable a smooth transition to new
solutions and will be discontinued in the future. To keep
using discontinued features, you will have to revert to the
last LTS OpenVINO version supporting them. For more details,
refer to the OpenVINO Legacy Features and Components page.
* Discontinued in 2024.0:
+ Runtime components:
- Intel® Gaussian & Neural Accelerator (Intel® GNA)..Consider
using the Neural Processing Unit (NPU) for low-powered
systems like Intel® Core™ Ultra or 14th generation
and beyond.
- OpenVINO C++/C/Python 1.0 APIs (see 2023.3 API transition
guide for reference).
- All ONNX Frontend legacy API (known as ONNX_IMPORTER_API)
- 'PerfomanceMode.UNDEFINED' property as part of the OpenVINO
Python API
+ Tools:
- Deployment Manager. See installation and deployment guides
for current distribution options.
- Accuracy Checker.
- Post-Training Optimization Tool (POT). Neural Network
Compression Framework (NNCF) should be used instead.
- A Git patch for NNCF integration with huggingface/
transformers. The recommended approach is to use
huggingface/optimum-intel for applying NNCF optimization
on top of models from Hugging Face.
- Support for Apache MXNet, Caffe, and Kaldi model formats.
Conversion to ONNX may be used as a solution.
* Deprecated and to be removed in the future:
+ The OpenVINO™ Development Tools package (pip install
openvino-dev) will be removed from installation options
and distribution channels beginning with OpenVINO 2025.0.
+ Model Optimizer will be discontinued with OpenVINO 2025.0.
Consider using the new conversion methods instead. For
more details, see the model conversion transition guide.
+ OpenVINO property Affinity API will be discontinued with
OpenVINO 2025.0. It will be replaced with CPU binding
configurations (ov::hint::enable_cpu_pinning).
+ OpenVINO Model Server components:
- “auto shape” and “auto batch size” (reshaping a model
in runtime) will be removed in the future. OpenVINO’s
dynamic shape models are recommended instead.
+ A number of notebooks have been deprecated. For an
up-to-date listing of available notebooks, refer to
the OpenVINO™ Notebook index (openvinotoolkit.github.io).

Loading...

Alessandro de Oliveira Faria's avatar

@badshah400 help me?


Atri Bhattacharya's avatar

Request you to trust the process and give science:machinelearning project reviewers time to review. Thanks for your continued contributions and may you keep it coming going forward.


Alessandro de Oliveira Faria's avatar

@badshah400 Thank you and sorry for the anxiety.

Request History
Alessandro de Oliveira Faria's avatar

cabelo created request

- Update to 2024.3.0
- Summary of major features and improvements  
* More Gen AI coverage and framework integrations to minimize
code changes
+ OpenVINO pre-optimized models are now available in Hugging
Face making it easier for developers to get started with
these models.
* Broader Large Language Model (LLM) support and more model
compression techniques.
+ Significant improvement in LLM performance on Intel
discrete GPUs with the addition of Multi-Head Attention
(MHA) and OneDNN enhancements.
* More portability and performance to run AI at the edge, in the
cloud, or locally.
+ Improved CPU performance when serving LLMs with the
inclusion of vLLM and continuous batching in the OpenVINO
Model Server (OVMS). vLLM is an easy-to-use open-source
library that supports efficient LLM inferencing and model
serving.
+ Ubuntu 24.04 long-term support (LTS), 64-bit (Kernel 6.8+)
(preview support)
- Support Change and Deprecation Notices
* Using deprecated features and components is not advised.
They are available to enable a smooth transition to new
solutions and will be discontinued in the future. To keep
using discontinued features, you will have to revert to the
last LTS OpenVINO version supporting them. For more details,
refer to the OpenVINO Legacy Features and Components page.
* Discontinued in 2024.0:
+ Runtime components:
- Intel® Gaussian & Neural Accelerator (Intel® GNA)..Consider
using the Neural Processing Unit (NPU) for low-powered
systems like Intel® Core™ Ultra or 14th generation
and beyond.
- OpenVINO C++/C/Python 1.0 APIs (see 2023.3 API transition
guide for reference).
- All ONNX Frontend legacy API (known as ONNX_IMPORTER_API)
- 'PerfomanceMode.UNDEFINED' property as part of the OpenVINO
Python API
+ Tools:
- Deployment Manager. See installation and deployment guides
for current distribution options.
- Accuracy Checker.
- Post-Training Optimization Tool (POT). Neural Network
Compression Framework (NNCF) should be used instead.
- A Git patch for NNCF integration with huggingface/
transformers. The recommended approach is to use
huggingface/optimum-intel for applying NNCF optimization
on top of models from Hugging Face.
- Support for Apache MXNet, Caffe, and Kaldi model formats.
Conversion to ONNX may be used as a solution.
* Deprecated and to be removed in the future:
+ The OpenVINO™ Development Tools package (pip install
openvino-dev) will be removed from installation options
and distribution channels beginning with OpenVINO 2025.0.
+ Model Optimizer will be discontinued with OpenVINO 2025.0.
Consider using the new conversion methods instead. For
more details, see the model conversion transition guide.
+ OpenVINO property Affinity API will be discontinued with
OpenVINO 2025.0. It will be replaced with CPU binding
configurations (ov::hint::enable_cpu_pinning).
+ OpenVINO Model Server components:
- “auto shape” and “auto batch size” (reshaping a model
in runtime) will be removed in the future. OpenVINO’s
dynamic shape models are recommended instead.
+ A number of notebooks have been deprecated. For an
up-to-date listing of available notebooks, refer to
the OpenVINO™ Notebook index (openvinotoolkit.github.io).


Christian Goll's avatar

mslacken accepted request

Accepting it to the devel project, but please remove the ubuntu line from the changes file so that this one can go into factory

openSUSE Build Service is sponsored by