Skip to content

Commit

Permalink
.
Browse files Browse the repository at this point in the history
  • Loading branch information
Timothee Lacroix committed Dec 11, 2023
1 parent 12e50d1 commit 768f36f
Show file tree
Hide file tree
Showing 3 changed files with 4 additions and 4 deletions.
2 changes: 1 addition & 1 deletion src/mistralai/async_client.py
Original file line number Diff line number Diff line change
Expand Up @@ -223,7 +223,7 @@ async def chat(
max_tokens: Optional[int] = None,
top_p: Optional[float] = None,
random_seed: Optional[int] = None,
safe_mode: bool = True,
safe_mode: bool = False,
) -> ChatCompletionResponse:
""" A asynchronous chat endpoint that returns a single response.
Expand Down
4 changes: 2 additions & 2 deletions src/mistralai/client.py
Original file line number Diff line number Diff line change
Expand Up @@ -106,7 +106,7 @@ def chat(
max_tokens: Optional[int] = None,
top_p: Optional[float] = None,
random_seed: Optional[int] = None,
safe_mode: bool = True,
safe_mode: bool = False,
) -> ChatCompletionResponse:
""" A chat endpoint that returns a single response.
Expand Down Expand Up @@ -149,7 +149,7 @@ def chat_stream(
max_tokens: Optional[int] = None,
top_p: Optional[float] = None,
random_seed: Optional[int] = None,
safe_mode: bool = True,
safe_mode: bool = False,
) -> Iterable[ChatCompletionStreamResponse]:
""" A chat endpoint that streams responses.
Expand Down
2 changes: 1 addition & 1 deletion src/mistralai/client_base.py
Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,7 @@ def _make_chat_request(
top_p: Optional[float] = None,
random_seed: Optional[int] = None,
stream: Optional[bool] = None,
safe_mode: Optional[bool] = True,
safe_mode: Optional[bool] = False,
) -> Dict[str, Any]:
request_data: Dict[str, Any] = {
"model": model,
Expand Down

0 comments on commit 768f36f

Please sign in to comment.