Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[DOCS] release notes 24.6 pass-1 #28057

Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
59 changes: 59 additions & 0 deletions docs/articles_en/about-openvino/release-notes-openvino.rst
Original file line number Diff line number Diff line change
Expand Up @@ -23,6 +23,65 @@ OpenVINO Release Notes



What's new
+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++

* OpenVINO 2024.6 LTS release includes updates for enhanced stability and improved LLM performance.
* Introduced support for Intel® Arc™ B-Series Graphics (formerly known as Battlemage)
* Memory optimizations implemented to improve the inference time memory and LLM performance on NPUs.
* Improved LLM performance with GenAI API optimizations and bug fixes.



OpenVINO™ Runtime
+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++

CPU Device Plugin
-----------------------------------------------------------------------------------------------

* KV cache now uses asymmetric U8 as the default precision, reducing memory stress for LLMs and
increasing their performance. This option can be controlled by model meta data.
* Quality and accuracy has been improved for selected models with several bug fixes.

GPU Device Plugin
-----------------------------------------------------------------------------------------------

* Device memory copy optimizations have been introduced for inference with **Intel® Arc™ B-Series
Graphics** (formerly known as Battlemage). Since it does not utilize L2 cache for copying memory
between the device and host, a dedicated `copy` operation is used, if inputs or results are
not expected in the device memory.
* ChatGLM4 inference on GPU has been optimized.

NPU Device Plugin
-----------------------------------------------------------------------------------------------

* LLM performance and inference time has been improved with memory optimizations.





OpenVINO.GenAI
+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++

* The encrypted_model_causal_lm sample is now available, showing how to decrypt a model.




Other Changes and Known Issues
+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++

Jupyter Notebooks
-----------------------------

* `Visual-language assistant with GLM-Edge-V and OpenVINO <https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/glm-edge-v/glm-edge-v.ipynb>`__
* `Local AI and OpenVINO <https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/localai/localai.ipynb>`__
* `Multimodal understanding and generation with Janus and OpenVINO <https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/janus-multimodal-generation/janus-multimodal-generation.ipynb>`__







Expand Down
Loading