From 73966e693cfa6d8acd57182d86beb5be7ce95ec9 Mon Sep 17 00:00:00 2001 From: Nuno Campos Date: Tue, 9 Jul 2024 20:01:51 +0100 Subject: [PATCH] openai: Create msg chunk in bg thread (#24032) Thank you for contributing to LangChain! - [ ] **PR title**: "package: description" - Where "package" is whichever of langchain, community, core, experimental, etc. is being modified. Use "docs: ..." for purely docs changes, "templates: ..." for template changes, "infra: ..." for CI changes. - Example: "community: add foobar LLM" - [ ] **PR message**: ***Delete this entire checklist*** and replace with - **Description:** a description of the change - **Issue:** the issue # it fixes, if applicable - **Dependencies:** any dependencies required for this change - **Twitter handle:** if your PR gets announced, and you'd like a mention, we'll gladly shout you out! - [ ] **Add tests and docs**: If you're adding a new integration, please include 1. a test for the integration, preferably unit tests that do not rely on network access, 2. an example notebook showing its use. It lives in `docs/docs/integrations` directory. - [ ] **Lint and test**: Run `make format`, `make lint` and `make test` from the root of the package(s) you've modified. See contribution guidelines for more: https://python.langchain.com/docs/contributing/ Additional guidelines: - Make sure optional dependencies are imported within a function. - Please do not add dependencies to pyproject.toml files (even optional ones) unless they are required for unit tests. - Most PRs should not touch more than one package. - Changes should be backwards compatible. - If you are adding something to community, do not re-import it in langchain. If no one reviews your PR within a few days, please @-mention one of baskaryan, efriis, eyurtsev, ccurme, vbarda, hwchase17. --- .../openai/langchain_openai/chat_models/base.py | 10 +++++++--- 1 file changed, 7 insertions(+), 3 deletions(-) diff --git a/libs/partners/openai/langchain_openai/chat_models/base.py b/libs/partners/openai/langchain_openai/chat_models/base.py index 7788efced211c..db5f2eec5a9e9 100644 --- a/libs/partners/openai/langchain_openai/chat_models/base.py +++ b/libs/partners/openai/langchain_openai/chat_models/base.py @@ -74,6 +74,7 @@ from langchain_core.outputs import ChatGeneration, ChatGenerationChunk, ChatResult from langchain_core.pydantic_v1 import BaseModel, Field, SecretStr, root_validator from langchain_core.runnables import Runnable, RunnableMap, RunnablePassthrough +from langchain_core.runnables.config import run_in_executor from langchain_core.tools import BaseTool from langchain_core.utils import ( convert_to_secret_str, @@ -632,8 +633,11 @@ async def _astream( choice = chunk["choices"][0] if choice["delta"] is None: continue - message_chunk = _convert_delta_to_message_chunk( - choice["delta"], default_chunk_class + message_chunk = await run_in_executor( + None, + _convert_delta_to_message_chunk, + choice["delta"], + default_chunk_class, ) generation_info = {} if finish_reason := choice.get("finish_reason"): @@ -672,7 +676,7 @@ async def _agenerate( return await agenerate_from_stream(stream_iter) payload = self._get_request_payload(messages, stop=stop, **kwargs) response = await self.async_client.create(**payload) - return self._create_chat_result(response) + return await run_in_executor(None, self._create_chat_result, response) @property def _identifying_params(self) -> Dict[str, Any]: