Skip to content
This repository has been archived by the owner on Oct 25, 2024. It is now read-only.

update ipex api #1650

Draft
wants to merge 2 commits into
base: main
Choose a base branch
from
Draft

update ipex api #1650

wants to merge 2 commits into from

Conversation

intellinjun
Copy link
Collaborator

Type of Change

feature or bug fix or documentation or others
API changed or not
yes

Description

update ipex api
detail description
image

Signed-off-by: intellinjun <[email protected]>
Copy link

github-actions bot commented Jul 3, 2024

⛈️ Required checks status: Has failure 🔴

Warning
If you do not have the access to re-run the CI-Summary bot, please contact VincyZhang for help. If you push a new commit, all of the workflow will be re-triggered.

Groups summary

🟢 Format Scan Tests workflow
Check ID Status Error details
format-scan (pylint) success
format-scan (bandit) success
format-scan (cloc) success
format-scan (cpplint) success

These checks are required after the changes to intel_extension_for_transformers/neural_chat/models/model_utils.py.

🔴 NeuralChat Unit Test
Check ID Status Error details
neuralchat-unit-test-baseline cancelled 🚫
neuralchat-unit-test-PR-test failure download
Generate-NeuralChat-Report skipped

These checks are required after the changes to intel_extension_for_transformers/neural_chat/models/model_utils.py.

🟡 Chat Bot Test workflow
Check ID Status Error details
call-inference-llama-2-7b-chat-hf / inference test no_status
call-inference-mpt-7b-chat / inference test no_status

These checks are required after the changes to intel_extension_for_transformers/neural_chat/models/model_utils.py.


Thank you for your contribution! 💜

Note
This comment is automatically generated and will be updates every 180 seconds within the next 6 hours. If you have any other questions, contact VincyZhang or XuehaoSun for help.

@a32543254
Copy link
Contributor

@changwangss could you help review this pr?

model.eval(),
dtype=torch_dtype,
inplace=True,
level="O1",
auto_kernel_selection=True,
)
except AssertionError:
model = intel_ipex.optimize(
Copy link
Collaborator

@changwangss changwangss Jul 3, 2024

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Could you also change it (Line 849)? as I know, ipex.llm.optimize is recommanded to run bf16 LLMs inference by IPEX.

Copy link
Contributor

@a32543254 a32543254 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

Signed-off-by: intellinjun <[email protected]>
@a32543254 a32543254 marked this pull request as draft July 5, 2024 03:15
@a32543254
Copy link
Contributor

convert to draft since the intel_ipex.llm.optimize will make the model cannot run.

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants