Skip to content

Commit

Permalink
Merge pull request #41 from MicrosoftDocs/main
Browse files Browse the repository at this point in the history
8/29/2024 AM Publish
  • Loading branch information
Taojunshen authored Aug 29, 2024
2 parents 946f97f + 14e9159 commit fd34f15
Show file tree
Hide file tree
Showing 29 changed files with 77 additions and 543 deletions.
4 changes: 4 additions & 0 deletions .gitignore
Original file line number Diff line number Diff line change
@@ -1,3 +1,7 @@
# Operating System files
.DS_Store
Thumbs.db

log/
obj/
_site/
Expand Down
2 changes: 1 addition & 1 deletion articles/ai-services/openai/how-to/deployment-types.md
Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,7 @@ Azure OpenAI offers three types of deployments. These provide a varied level of

| **Offering** | **Global-Batch** | **Global-Standard** | **Standard** | **Provisioned** |
|---|:---|:---|:---|:---|
| **Best suited for** | Offline scoring <br><br> Workloads that are not latency sensitive and can be completed in hours.<br><br> For use cases that do not have data processing residency requirements.| Recommended starting place for customers. <br><br>Global-Standard will have the higher default quota and larger number of models available than Standard. <br><br> For production applications that do not have data processing residency requirements. | For customers with data residency requirements. Optimized for low to medium volume. | Real-time scoring for large consistent volume. Includes the highest commitments and limits.|
| **Best suited for** | Offline scoring <br><br> Workloads that are not latency sensitive and can be completed in hours.<br><br> For use cases that do not have data processing residency requirements.| Recommended starting place for customers. <br><br>Global-Standard will have the higher default quota and larger number of models available than Standard. | For customers with data residency requirements. Optimized for low to medium volume. | Real-time scoring for large consistent volume. Includes the highest commitments and limits.|
| **How it works** | Offline processing via files |Traffic may be routed anywhere in the world | | |
| **Getting started** | [Global-Batch](./batch.md) | [Model deployment](./create-resource.md) | [Model deployment](./create-resource.md) | [Provisioned onboarding](./provisioned-throughput-onboarding.md) |
| **Cost** | [Least expensive option](https://azure.microsoft.com/pricing/details/cognitive-services/openai-service/) <br> 50% less cost compared to Global Standard prices. Access to all new models with larger quota allocations. | [Global deployment pricing](https://azure.microsoft.com/pricing/details/cognitive-services/openai-service/) | [Regional pricing](https://azure.microsoft.com/pricing/details/cognitive-services/openai-service/) | May experience cost savings for consistent usage |
Expand Down
5 changes: 4 additions & 1 deletion articles/ai-studio/how-to/costs-plan-manage.md
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,10 @@ author: Blackmist

[!INCLUDE [Feature preview](~/reusable-content/ce-skilling/azure/includes/ai-studio/includes/feature-preview.md)]

This article describes how you plan for and manage costs for Azure AI Studio. First, you use the Azure pricing calculator to help plan for Azure AI Studio costs before you add any resources for the service to estimate costs. Next, as you add Azure resources, review the estimated costs.
This article describes how you plan for and manage costs for Azure AI Studio. First, you use the Azure pricing calculator to help plan for Azure AI Studio costs before you add any resources for the service to estimate costs. Next, as you add Azure resources, review the estimated costs.

> [!TIP]
> Azure AI Studio does not have a specific page in the Azure pricing calculator. Azure AI Studio is composed of several other Azure services, some of which are optional. This article provides information on using the pricing calculator to estimate costs for these services.
You use Azure AI services in Azure AI Studio. Costs for Azure AI services are only a portion of the monthly costs in your Azure bill. You're billed for all Azure services and resources used in your Azure subscription, including the third-party services.

Expand Down
8 changes: 5 additions & 3 deletions articles/ai-studio/how-to/deploy-models-cohere-command.md
Original file line number Diff line number Diff line change
Expand Up @@ -231,7 +231,7 @@ print_stream(result)
Explore other parameters that you can specify in the inference client. For a full list of all the supported parameters and their corresponding documentation, see [Azure AI Model Inference API reference](https://aka.ms/azureai/modelinference).

```python
from azure.ai.inference.models import ChatCompletionsResponseFormat
from azure.ai.inference.models import ChatCompletionsResponseFormatText

response = client.complete(
messages=[
Expand All @@ -244,7 +244,7 @@ response = client.complete(
stop=["<|endoftext|>"],
temperature=0,
top_p=1,
response_format={ "type": ChatCompletionsResponseFormat.TEXT },
response_format=ChatCompletionsResponseFormatText(),
)
```

Expand All @@ -256,13 +256,15 @@ Cohere Command chat models can create JSON outputs. Set `response_format` to `js


```python
from azure.ai.inference.models import ChatCompletionsResponseFormatJSON

response = client.complete(
messages=[
SystemMessage(content="You are a helpful assistant that always generate responses in JSON format, using."
" the following format: { ""answer"": ""response"" }."),
UserMessage(content="How many languages are in the world?"),
],
response_format={ "type": ChatCompletionsResponseFormat.JSON_OBJECT }
response_format=ChatCompletionsResponseFormatJSON()
)
```

Expand Down
8 changes: 4 additions & 4 deletions articles/ai-studio/how-to/deploy-models-cohere-embed.md
Original file line number Diff line number Diff line change
Expand Up @@ -617,12 +617,12 @@ Cohere Embed V3 models can optimize the embeddings based on its use case.

| Description | Language | Sample |
|-------------------------------------------|-------------------|-----------------------------------------------------------------|
| Web requests | Bash | [Command-R](https://aka.ms/samples/cohere-command-r/webrequests) - [Command-R+](https://aka.ms/samples/cohere-command-r-plus/webrequests) |
| Web requests | Bash | [cohere-embed.ipynb](https://aka.ms/samples/embed-v3/webrequests) |
| Azure AI Inference package for JavaScript | JavaScript | [Link](https://aka.ms/azsdk/azure-ai-inference/javascript/samples) |
| Azure AI Inference package for Python | Python | [Link](https://aka.ms/azsdk/azure-ai-inference/python/samples) |
| OpenAI SDK (experimental) | Python | [Link](https://aka.ms/samples/cohere-command/openaisdk) |
| LangChain | Python | [Link](https://aka.ms/samples/cohere/langchain) |
| Cohere SDK | Python | [Link](https://aka.ms/samples/cohere-python-sdk) |
| OpenAI SDK (experimental) | Python | [Link](https://aka.ms/samples/cohere-embed/openaisdk) |
| LangChain | Python | [Link](https://aka.ms/samples/cohere-embed/langchain) |
| Cohere SDK | Python | [Link](https://aka.ms/samples/cohere-embed/cohere-python-sdk) |
| LiteLLM SDK | Python | [Link](https://github.com/Azure/azureml-examples/blob/main/sdk/python/foundation-models/cohere/litellm.ipynb) |

#### Retrieval Augmented Generation (RAG) and tool use samples
Expand Down
4 changes: 2 additions & 2 deletions articles/ai-studio/how-to/deploy-models-jais.md
Original file line number Diff line number Diff line change
Expand Up @@ -201,7 +201,7 @@ print_stream(result)
Explore other parameters that you can specify in the inference client. For a full list of all the supported parameters and their corresponding documentation, see [Azure AI Model Inference API reference](https://aka.ms/azureai/modelinference).

```python
from azure.ai.inference.models import ChatCompletionsResponseFormat
from azure.ai.inference.models import ChatCompletionsResponseFormatText

response = client.complete(
messages=[
Expand All @@ -214,7 +214,7 @@ response = client.complete(
stop=["<|endoftext|>"],
temperature=0,
top_p=1,
response_format={ "type": ChatCompletionsResponseFormat.TEXT },
response_format=ChatCompletionsResponseFormatText(),
)
```

Expand Down
4 changes: 2 additions & 2 deletions articles/ai-studio/how-to/deploy-models-llama.md
Original file line number Diff line number Diff line change
Expand Up @@ -255,7 +255,7 @@ print_stream(result)
Explore other parameters that you can specify in the inference client. For a full list of all the supported parameters and their corresponding documentation, see [Azure AI Model Inference API reference](https://aka.ms/azureai/modelinference).

```python
from azure.ai.inference.models import ChatCompletionsResponseFormat
from azure.ai.inference.models import ChatCompletionsResponseFormatText

response = client.complete(
messages=[
Expand All @@ -268,7 +268,7 @@ response = client.complete(
stop=["<|endoftext|>"],
temperature=0,
top_p=1,
response_format={ "type": ChatCompletionsResponseFormat.TEXT },
response_format=ChatCompletionsResponseFormatText(),
)
```

Expand Down
8 changes: 5 additions & 3 deletions articles/ai-studio/how-to/deploy-models-mistral-nemo.md
Original file line number Diff line number Diff line change
Expand Up @@ -209,7 +209,7 @@ print_stream(result)
Explore other parameters that you can specify in the inference client. For a full list of all the supported parameters and their corresponding documentation, see [Azure AI Model Inference API reference](https://aka.ms/azureai/modelinference).

```python
from azure.ai.inference.models import ChatCompletionsResponseFormat
from azure.ai.inference.models import ChatCompletionsResponseFormatText

response = client.complete(
messages=[
Expand All @@ -222,7 +222,7 @@ response = client.complete(
stop=["<|endoftext|>"],
temperature=0,
top_p=1,
response_format={ "type": ChatCompletionsResponseFormat.TEXT },
response_format=ChatCompletionsResponseFormatText(),
)
```

Expand All @@ -234,13 +234,15 @@ Mistral Nemo chat model can create JSON outputs. Set `response_format` to `json_


```python
from azure.ai.inference.models import ChatCompletionsResponseFormatJSON

response = client.complete(
messages=[
SystemMessage(content="You are a helpful assistant that always generate responses in JSON format, using."
" the following format: { ""answer"": ""response"" }."),
UserMessage(content="How many languages are in the world?"),
],
response_format={ "type": ChatCompletionsResponseFormat.JSON_OBJECT }
response_format=ChatCompletionsResponseFormatJSON()
)
```

Expand Down
4 changes: 2 additions & 2 deletions articles/ai-studio/how-to/deploy-models-mistral-open.md
Original file line number Diff line number Diff line change
Expand Up @@ -257,7 +257,7 @@ print_stream(result)
Explore other parameters that you can specify in the inference client. For a full list of all the supported parameters and their corresponding documentation, see [Azure AI Model Inference API reference](https://aka.ms/azureai/modelinference).

```python
from azure.ai.inference.models import ChatCompletionsResponseFormat
from azure.ai.inference.models import ChatCompletionsResponseFormatText

response = client.complete(
messages=[
Expand All @@ -270,7 +270,7 @@ response = client.complete(
stop=["<|endoftext|>"],
temperature=0,
top_p=1,
response_format={ "type": ChatCompletionsResponseFormat.TEXT },
response_format=ChatCompletionsResponseFormatText(),
)
```

Expand Down
8 changes: 5 additions & 3 deletions articles/ai-studio/how-to/deploy-models-mistral.md
Original file line number Diff line number Diff line change
Expand Up @@ -239,7 +239,7 @@ print_stream(result)
Explore other parameters that you can specify in the inference client. For a full list of all the supported parameters and their corresponding documentation, see [Azure AI Model Inference API reference](https://aka.ms/azureai/modelinference).

```python
from azure.ai.inference.models import ChatCompletionsResponseFormat
from azure.ai.inference.models import ChatCompletionsResponseFormatText

response = client.complete(
messages=[
Expand All @@ -252,7 +252,7 @@ response = client.complete(
stop=["<|endoftext|>"],
temperature=0,
top_p=1,
response_format={ "type": ChatCompletionsResponseFormat.TEXT },
response_format=ChatCompletionsResponseFormatText(),
)
```

Expand All @@ -264,13 +264,15 @@ Mistral premium chat models can create JSON outputs. Set `response_format` to `j


```python
from azure.ai.inference.models import ChatCompletionsResponseFormatJSON

response = client.complete(
messages=[
SystemMessage(content="You are a helpful assistant that always generate responses in JSON format, using."
" the following format: { ""answer"": ""response"" }."),
UserMessage(content="How many languages are in the world?"),
],
response_format={ "type": ChatCompletionsResponseFormat.JSON_OBJECT }
response_format=ChatCompletionsResponseFormatJSON()
)
```

Expand Down
4 changes: 2 additions & 2 deletions articles/ai-studio/how-to/deploy-models-phi-3-5-moe.md
Original file line number Diff line number Diff line change
Expand Up @@ -219,7 +219,7 @@ print_stream(result)
Explore other parameters that you can specify in the inference client. For a full list of all the supported parameters and their corresponding documentation, see [Azure AI Model Inference API reference](https://aka.ms/azureai/modelinference).

```python
from azure.ai.inference.models import ChatCompletionsResponseFormat
from azure.ai.inference.models import ChatCompletionsResponseFormatText

response = client.complete(
messages=[
Expand All @@ -232,7 +232,7 @@ response = client.complete(
stop=["<|endoftext|>"],
temperature=0,
top_p=1,
response_format={ "type": ChatCompletionsResponseFormat.TEXT },
response_format=ChatCompletionsResponseFormatText(),
)
```

Expand Down
4 changes: 2 additions & 2 deletions articles/ai-studio/how-to/deploy-models-phi-3-5-vision.md
Original file line number Diff line number Diff line change
Expand Up @@ -215,7 +215,7 @@ print_stream(result)
Explore other parameters that you can specify in the inference client. For a full list of all the supported parameters and their corresponding documentation, see [Azure AI Model Inference API reference](https://aka.ms/azureai/modelinference).

```python
from azure.ai.inference.models import ChatCompletionsResponseFormat
from azure.ai.inference.models import ChatCompletionsResponseFormatText

response = client.complete(
messages=[
Expand All @@ -228,7 +228,7 @@ response = client.complete(
stop=["<|endoftext|>"],
temperature=0,
top_p=1,
response_format={ "type": ChatCompletionsResponseFormat.TEXT },
response_format=ChatCompletionsResponseFormatText(),
)
```

Expand Down
4 changes: 2 additions & 2 deletions articles/ai-studio/how-to/deploy-models-phi-3-vision.md
Original file line number Diff line number Diff line change
Expand Up @@ -215,7 +215,7 @@ print_stream(result)
Explore other parameters that you can specify in the inference client. For a full list of all the supported parameters and their corresponding documentation, see [Azure AI Model Inference API reference](https://aka.ms/azureai/modelinference).

```python
from azure.ai.inference.models import ChatCompletionsResponseFormat
from azure.ai.inference.models import ChatCompletionsResponseFormatText

response = client.complete(
messages=[
Expand All @@ -228,7 +228,7 @@ response = client.complete(
stop=["<|endoftext|>"],
temperature=0,
top_p=1,
response_format={ "type": ChatCompletionsResponseFormat.TEXT },
response_format=ChatCompletionsResponseFormatText(),
)
```

Expand Down
4 changes: 2 additions & 2 deletions articles/ai-studio/how-to/deploy-models-phi-3.md
Original file line number Diff line number Diff line change
Expand Up @@ -256,7 +256,7 @@ print_stream(result)
Explore other parameters that you can specify in the inference client. For a full list of all the supported parameters and their corresponding documentation, see [Azure AI Model Inference API reference](https://aka.ms/azureai/modelinference).

```python
from azure.ai.inference.models import ChatCompletionsResponseFormat
from azure.ai.inference.models import ChatCompletionsResponseFormatText

response = client.complete(
messages=[
Expand All @@ -269,7 +269,7 @@ response = client.complete(
stop=["<|endoftext|>"],
temperature=0,
top_p=1,
response_format={ "type": ChatCompletionsResponseFormat.TEXT },
response_format=ChatCompletionsResponseFormatText(),
)
```

Expand Down
1 change: 1 addition & 0 deletions articles/ai-studio/index.yml
Original file line number Diff line number Diff line change
Expand Up @@ -8,6 +8,7 @@ metadata:
ms.service: azure-ai-studio
ms.custom:
- build-2024
- copilot-learning-hub
ms.topic: landing-page
ms.reviewer: eur
ms.author: eur
Expand Down
4 changes: 2 additions & 2 deletions articles/ai-studio/reference/reference-model-inference-api.md
Original file line number Diff line number Diff line change
Expand Up @@ -318,7 +318,7 @@ The following example shows the response for a chat completion request indicatin

```python
import json
from azure.ai.inference.models import SystemMessage, UserMessage, ChatCompletionsResponseFormat
from azure.ai.inference.models import SystemMessage, UserMessage, ChatCompletionsResponseFormatJSON
from azure.core.exceptions import HttpResponseError

try:
Expand All @@ -327,7 +327,7 @@ try:
SystemMessage(content="You are a helpful assistant."),
UserMessage(content="How many languages are in the world?"),
],
response_format={ "type": ChatCompletionsResponseFormat.JSON_OBJECT }
response_format=ChatCompletionsResponseFormatJSON()
)
except HttpResponseError as ex:
if ex.status_code == 422:
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -2972,8 +2972,13 @@
},
{
"source_path_from_root": "/articles/machine-learning/v-fake/how-to-cicd-data-ingestion.md",
"redirect_url": "/azure/machine-learning/how-to-cicd-data-ingestion",
"redirect_document_id": true
"redirect_url": "/azure/machine-learning/how-to-devops-machine-learning",
"redirect_document_id": false
},
{
"source_path_from_root": "/articles/machine-learning/v1/how-to-cicd-data-ingestion.md",
"redirect_url": "/azure/machine-learning/how-to-devops-machine-learning",
"redirect_document_id": false
},
{
"source_path_from_root": "/articles/machine-learning/how-to-debug-pipelines-application-insights.md",
Expand Down
4 changes: 1 addition & 3 deletions articles/machine-learning/component-reference/add-columns.md
Original file line number Diff line number Diff line change
Expand Up @@ -18,8 +18,6 @@ This article describes a component in Azure Machine Learning designer.

Use this component to concatenate two datasets. You combine all columns from the two datasets that you specify as inputs to create a single dataset. If you need to concatenate more than two datasets, use several instances of **Add Columns**.



## How to configure Add Columns
1. Add the **Add Columns** component to your pipeline.

Expand All @@ -42,4 +40,4 @@ If there are two columns with the same name in the input datasets, a numeric suf

## Next steps

See the [set of components available](component-reference.md) to Azure Machine Learning.
See the [set of components available](component-reference.md) to Azure Machine Learning.
4 changes: 2 additions & 2 deletions articles/machine-learning/how-to-deploy-models-phi-3-5-moe.md
Original file line number Diff line number Diff line change
Expand Up @@ -219,7 +219,7 @@ print_stream(result)
Explore other parameters that you can specify in the inference client. For a full list of all the supported parameters and their corresponding documentation, see [Azure AI Model Inference API reference](reference-model-inference-api.md).

```python
from azure.ai.inference.models import ChatCompletionsResponseFormat
from azure.ai.inference.models import ChatCompletionsResponseFormatText

response = client.complete(
messages=[
Expand All @@ -232,7 +232,7 @@ response = client.complete(
stop=["<|endoftext|>"],
temperature=0,
top_p=1,
response_format={ "type": ChatCompletionsResponseFormat.TEXT },
response_format=ChatCompletionsResponseFormatText(),
)
```

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -214,7 +214,7 @@ print_stream(result)
Explore other parameters that you can specify in the inference client. For a full list of all the supported parameters and their corresponding documentation, see [Azure AI Model Inference API reference](reference-model-inference-api.md).

```python
from azure.ai.inference.models import ChatCompletionsResponseFormat
from azure.ai.inference.models import ChatCompletionsResponseFormatText

response = client.complete(
messages=[
Expand All @@ -227,7 +227,7 @@ response = client.complete(
stop=["<|endoftext|>"],
temperature=0,
top_p=1,
response_format={ "type": ChatCompletionsResponseFormat.TEXT },
response_format=ChatCompletionsResponseFormatText(),
)
```

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -213,7 +213,7 @@ print_stream(result)
Explore other parameters that you can specify in the inference client. For a full list of all the supported parameters and their corresponding documentation, see [Azure AI Model Inference API reference](reference-model-inference-api.md).

```python
from azure.ai.inference.models import ChatCompletionsResponseFormat
from azure.ai.inference.models import ChatCompletionsResponseFormatText

response = client.complete(
messages=[
Expand All @@ -226,7 +226,7 @@ response = client.complete(
stop=["<|endoftext|>"],
temperature=0,
top_p=1,
response_format={ "type": ChatCompletionsResponseFormat.TEXT },
response_format=ChatCompletionsResponseFormatText(),
)
```

Expand Down
Loading

0 comments on commit fd34f15

Please sign in to comment.