diff --git a/.docusaurus/docusaurus-plugin-content-docs/default/site-docs-04-part-1-labs-4-prompt-engineering-techniques-md-140.json b/.docusaurus/docusaurus-plugin-content-docs/default/site-docs-04-part-1-labs-4-prompt-engineering-techniques-md-140.json index 6b314170..90487e63 100644 --- a/.docusaurus/docusaurus-plugin-content-docs/default/site-docs-04-part-1-labs-4-prompt-engineering-techniques-md-140.json +++ b/.docusaurus/docusaurus-plugin-content-docs/default/site-docs-04-part-1-labs-4-prompt-engineering-techniques-md-140.json @@ -18,7 +18,7 @@ "permalink": "/Part-1-labs/Conversation-history" }, "next": { - "title": "System Message", - "permalink": "/Part-2-labs/System-Message" + "title": "Image Generation", + "permalink": "/Part-1-labs/Image Generation" } } \ No newline at end of file diff --git a/.docusaurus/docusaurus-plugin-content-docs/default/site-docs-04-part-1-labs-5-image-generation-md-3ed.json b/.docusaurus/docusaurus-plugin-content-docs/default/site-docs-04-part-1-labs-5-image-generation-md-3ed.json new file mode 100644 index 00000000..9f1b50dc --- /dev/null +++ b/.docusaurus/docusaurus-plugin-content-docs/default/site-docs-04-part-1-labs-5-image-generation-md-3ed.json @@ -0,0 +1,24 @@ +{ + "id": "Part-1-labs/Image Generation", + "title": "Image Generation", + "description": "Welcome to part 2 of this workshop, where we will interact with a text-to-image model, DALL-E 3. Start by clicking on the Images playground, and selecting dall-e-3 from the deployments drop-down menu.", + "source": "@site/docs/04-Part-1-labs/5-Image Generation.md", + "sourceDirName": "04-Part-1-labs", + "slug": "/Part-1-labs/Image Generation", + "permalink": "/Part-1-labs/Image Generation", + "draft": false, + "unlisted": false, + "tags": [], + "version": "current", + "sidebarPosition": 5, + "frontMatter": {}, + "sidebar": "tutorialSidebar", + "previous": { + "title": "Prompt engineering techniques", + "permalink": "/Part-1-labs/Prompt-engineering-techniques" + }, + "next": { + "title": "Multimodal", + "permalink": "/Part-1-labs/Multimodal" + } +} \ No newline at end of file diff --git a/.docusaurus/docusaurus-plugin-content-docs/default/site-docs-04-part-1-labs-6-multimodal-md-3f5.json b/.docusaurus/docusaurus-plugin-content-docs/default/site-docs-04-part-1-labs-6-multimodal-md-3f5.json new file mode 100644 index 00000000..de673212 --- /dev/null +++ b/.docusaurus/docusaurus-plugin-content-docs/default/site-docs-04-part-1-labs-6-multimodal-md-3f5.json @@ -0,0 +1,24 @@ +{ + "id": "Part-1-labs/Multimodal", + "title": "Multimodal", + "description": "By now we have interacted with LLMs using a single modality: inputting text and receiving either text or images. However, multimodal interfaces are becoming increasingly popular, as they allow users to interact with models using multiple modalities, such as text, images, and speech, thus facilitating human-computer interactions. In this section, we will explore how to use multimodal interfaces to interact with GPT-4o.", + "source": "@site/docs/04-Part-1-labs/6-Multimodal.md", + "sourceDirName": "04-Part-1-labs", + "slug": "/Part-1-labs/Multimodal", + "permalink": "/Part-1-labs/Multimodal", + "draft": false, + "unlisted": false, + "tags": [], + "version": "current", + "sidebarPosition": 6, + "frontMatter": {}, + "sidebar": "tutorialSidebar", + "previous": { + "title": "Image Generation", + "permalink": "/Part-1-labs/Image Generation" + }, + "next": { + "title": "System Message", + "permalink": "/Part-2-labs/System-Message" + } +} \ No newline at end of file diff --git a/.docusaurus/docusaurus-plugin-content-docs/default/site-docs-05-part-2-labs-3-system-message-md-0b5.json b/.docusaurus/docusaurus-plugin-content-docs/default/site-docs-05-part-2-labs-3-system-message-md-0b5.json index 67de73c2..19350aca 100644 --- a/.docusaurus/docusaurus-plugin-content-docs/default/site-docs-05-part-2-labs-3-system-message-md-0b5.json +++ b/.docusaurus/docusaurus-plugin-content-docs/default/site-docs-05-part-2-labs-3-system-message-md-0b5.json @@ -14,8 +14,8 @@ "frontMatter": {}, "sidebar": "tutorialSidebar", "previous": { - "title": "Prompt engineering techniques", - "permalink": "/Part-1-labs/Prompt-engineering-techniques" + "title": "Multimodal", + "permalink": "/Part-1-labs/Multimodal" }, "next": { "title": "Add knowledge", diff --git a/.docusaurus/docusaurus-plugin-content-docs/default/site-docs-06-part-3-labs-01-image-generation-md-668.json b/.docusaurus/docusaurus-plugin-content-docs/default/site-docs-06-part-3-labs-01-image-generation-md-668.json new file mode 100644 index 00000000..6fe1eb4b --- /dev/null +++ b/.docusaurus/docusaurus-plugin-content-docs/default/site-docs-06-part-3-labs-01-image-generation-md-668.json @@ -0,0 +1,24 @@ +{ + "id": "Part-3-labs/Image Generation", + "title": "Image Generation", + "description": "Welcome to part 2 of this workshop, where we will interact with a text-to-image model, DALL-E 3. Start by clicking on the Images playground, and selecting dall-e-3 from the deployments drop-down menu.", + "source": "@site/docs/06-Part-3-labs/01-Image Generation.md", + "sourceDirName": "06-Part-3-labs", + "slug": "/Part-3-labs/Image Generation", + "permalink": "/Part-3-labs/Image Generation", + "draft": false, + "unlisted": false, + "tags": [], + "version": "current", + "sidebarPosition": 1, + "frontMatter": {}, + "sidebar": "tutorialSidebar", + "previous": { + "title": "Function Calling", + "permalink": "/Part-2-labs/Function-Calling" + }, + "next": { + "title": "Multimodal", + "permalink": "/Part-3-labs/Multimodal" + } +} \ No newline at end of file diff --git a/.docusaurus/docusaurus-plugin-content-docs/default/site-docs-06-part-3-labs-02-multimodal-md-0ef.json b/.docusaurus/docusaurus-plugin-content-docs/default/site-docs-06-part-3-labs-02-multimodal-md-0ef.json new file mode 100644 index 00000000..901d30fa --- /dev/null +++ b/.docusaurus/docusaurus-plugin-content-docs/default/site-docs-06-part-3-labs-02-multimodal-md-0ef.json @@ -0,0 +1,24 @@ +{ + "id": "Part-3-labs/Multimodal", + "title": "Multimodal", + "description": "By now we have interacted with LLMs using a single modality: inputting text and receiving either text or images. However, multimodal interfaces are becoming increasingly popular, as they allow users to interact with models using multiple modalities, such as text, images, and speech, thus facilitating human-computer interactions. In this section, we will explore how to use multimodal interfaces to interact with GPT-4o.", + "source": "@site/docs/06-Part-3-labs/02-Multimodal.md", + "sourceDirName": "06-Part-3-labs", + "slug": "/Part-3-labs/Multimodal", + "permalink": "/Part-3-labs/Multimodal", + "draft": false, + "unlisted": false, + "tags": [], + "version": "current", + "sidebarPosition": 2, + "frontMatter": {}, + "sidebar": "tutorialSidebar", + "previous": { + "title": "Function Calling", + "permalink": "/Part-2-labs/Function-Calling" + }, + "next": { + "title": "AI Models & Deployments", + "permalink": "/ai-models" + } +} \ No newline at end of file diff --git a/.docusaurus/docusaurus-plugin-content-docs/default/site-docs-06-part-3-labs-5-image-generation-md-a53.json b/.docusaurus/docusaurus-plugin-content-docs/default/site-docs-06-part-3-labs-5-image-generation-md-a53.json new file mode 100644 index 00000000..b75b6bc4 --- /dev/null +++ b/.docusaurus/docusaurus-plugin-content-docs/default/site-docs-06-part-3-labs-5-image-generation-md-a53.json @@ -0,0 +1,24 @@ +{ + "id": "Part-3-labs/Image Generation", + "title": "Image Generation", + "description": "Welcome to part 2 of this workshop, where we will interact with a text-to-image model, DALL-E 3. Start by clicking on the Images playground, and selecting dall-e-3 from the deployments drop-down menu.", + "source": "@site/docs/06-Part-3-labs/5-Image Generation.md", + "sourceDirName": "06-Part-3-labs", + "slug": "/Part-3-labs/Image Generation", + "permalink": "/Part-3-labs/Image Generation", + "draft": false, + "unlisted": false, + "tags": [], + "version": "current", + "sidebarPosition": 5, + "frontMatter": {}, + "sidebar": "tutorialSidebar", + "previous": { + "title": "Multimodal", + "permalink": "/Part-3-labs/Multimodal" + }, + "next": { + "title": "AI Models & Deployments", + "permalink": "/ai-models" + } +} \ No newline at end of file diff --git a/.docusaurus/docusaurus-plugin-content-docs/default/site-docs-06-part-3-labs-6-multimodal-md-368.json b/.docusaurus/docusaurus-plugin-content-docs/default/site-docs-06-part-3-labs-6-multimodal-md-368.json new file mode 100644 index 00000000..7fa0ce4b --- /dev/null +++ b/.docusaurus/docusaurus-plugin-content-docs/default/site-docs-06-part-3-labs-6-multimodal-md-368.json @@ -0,0 +1,24 @@ +{ + "id": "Part-3-labs/Multimodal", + "title": "Multimodal", + "description": "By now we have interacted with LLMs using a single modality: inputting text and receiving either text or images. However, multimodal interfaces are becoming increasingly popular, as they allow users to interact with models using multiple modalities, such as text, images, and speech, thus facilitating human-computer interactions. In this section, we will explore how to use multimodal interfaces to interact with GPT-4o.", + "source": "@site/docs/06-Part-3-labs/6-Multimodal.md", + "sourceDirName": "06-Part-3-labs", + "slug": "/Part-3-labs/Multimodal", + "permalink": "/Part-3-labs/Multimodal", + "draft": false, + "unlisted": false, + "tags": [], + "version": "current", + "sidebarPosition": 6, + "frontMatter": {}, + "sidebar": "tutorialSidebar", + "previous": { + "title": "Function Calling", + "permalink": "/Part-2-labs/Function-Calling" + }, + "next": { + "title": "AI Models & Deployments", + "permalink": "/ai-models" + } +} \ No newline at end of file diff --git a/.docusaurus/docusaurus-plugin-content-docs/default/version-current-metadata-prop-751.json b/.docusaurus/docusaurus-plugin-content-docs/default/version-current-metadata-prop-751.json index a5f89549..e3c53c4e 100644 --- a/.docusaurus/docusaurus-plugin-content-docs/default/version-current-metadata-prop-751.json +++ b/.docusaurus/docusaurus-plugin-content-docs/default/version-current-metadata-prop-751.json @@ -50,6 +50,20 @@ "href": "/Part-1-labs/Prompt-engineering-techniques", "docId": "Part-1-labs/Prompt-engineering-techniques", "unlisted": false + }, + { + "type": "link", + "label": "Image Generation", + "href": "/Part-1-labs/Image Generation", + "docId": "Part-1-labs/Image Generation", + "unlisted": false + }, + { + "type": "link", + "label": "Multimodal", + "href": "/Part-1-labs/Multimodal", + "docId": "Part-1-labs/Multimodal", + "unlisted": false } ] }, @@ -172,6 +186,18 @@ "description": "Consumer conversational AI services like ChatGPT and Bing Chat use a trick to make the AI agent seem to remember the context of the conversation. The trick is that the foundation model is given the whole chat history at each turn, not just the latest prompt, but the user does not see this. An AI model cannot learn and has no memory of previous interactions if the user leaves and comes back but the application is using prompt engineering to add this 'memory'", "sidebar": "tutorialSidebar" }, + "Part-1-labs/Image Generation": { + "id": "Part-1-labs/Image Generation", + "title": "Image Generation", + "description": "Welcome to part 2 of this workshop, where we will interact with a text-to-image model, DALL-E 3. Start by clicking on the Images playground, and selecting dall-e-3 from the deployments drop-down menu.", + "sidebar": "tutorialSidebar" + }, + "Part-1-labs/Multimodal": { + "id": "Part-1-labs/Multimodal", + "title": "Multimodal", + "description": "By now we have interacted with LLMs using a single modality: inputting text and receiving either text or images. However, multimodal interfaces are becoming increasingly popular, as they allow users to interact with models using multiple modalities, such as text, images, and speech, thus facilitating human-computer interactions. In this section, we will explore how to use multimodal interfaces to interact with GPT-4o.", + "sidebar": "tutorialSidebar" + }, "Part-1-labs/Prompt-engineering-techniques": { "id": "Part-1-labs/Prompt-engineering-techniques", "title": "Prompt engineering techniques", diff --git a/.docusaurus/docusaurus-plugin-debug/default/docusaurus-debug-all-content-673.json b/.docusaurus/docusaurus-plugin-debug/default/docusaurus-debug-all-content-673.json index 4c448d2d..78518d7e 100644 --- a/.docusaurus/docusaurus-plugin-debug/default/docusaurus-debug-all-content-673.json +++ b/.docusaurus/docusaurus-plugin-debug/default/docusaurus-debug-all-content-673.json @@ -200,6 +200,54 @@ "permalink": "/Part-1-labs/Prompt-engineering-techniques" } }, + { + "id": "Part-1-labs/Image Generation", + "title": "Image Generation", + "description": "Welcome to part 2 of this workshop, where we will interact with a text-to-image model, DALL-E 3. Start by clicking on the Images playground, and selecting dall-e-3 from the deployments drop-down menu.", + "source": "@site/docs/04-Part-1-labs/5-Image Generation.md", + "sourceDirName": "04-Part-1-labs", + "slug": "/Part-1-labs/Image Generation", + "permalink": "/Part-1-labs/Image Generation", + "draft": false, + "unlisted": false, + "tags": [], + "version": "current", + "sidebarPosition": 5, + "frontMatter": {}, + "sidebar": "tutorialSidebar", + "previous": { + "title": "Prompt engineering techniques", + "permalink": "/Part-1-labs/Prompt-engineering-techniques" + }, + "next": { + "title": "Multimodal", + "permalink": "/Part-1-labs/Multimodal" + } + }, + { + "id": "Part-1-labs/Multimodal", + "title": "Multimodal", + "description": "By now we have interacted with LLMs using a single modality: inputting text and receiving either text or images. However, multimodal interfaces are becoming increasingly popular, as they allow users to interact with models using multiple modalities, such as text, images, and speech, thus facilitating human-computer interactions. In this section, we will explore how to use multimodal interfaces to interact with GPT-4o.", + "source": "@site/docs/04-Part-1-labs/6-Multimodal.md", + "sourceDirName": "04-Part-1-labs", + "slug": "/Part-1-labs/Multimodal", + "permalink": "/Part-1-labs/Multimodal", + "draft": false, + "unlisted": false, + "tags": [], + "version": "current", + "sidebarPosition": 6, + "frontMatter": {}, + "sidebar": "tutorialSidebar", + "previous": { + "title": "Image Generation", + "permalink": "/Part-1-labs/Image Generation" + }, + "next": { + "title": "System Message", + "permalink": "/Part-2-labs/System-Message" + } + }, { "id": "Part-1-labs/Prompt-engineering-techniques", "title": "Prompt engineering techniques", @@ -220,8 +268,8 @@ "permalink": "/Part-1-labs/Conversation-history" }, "next": { - "title": "System Message", - "permalink": "/Part-2-labs/System-Message" + "title": "Image Generation", + "permalink": "/Part-1-labs/Image Generation" } }, { @@ -288,8 +336,8 @@ "frontMatter": {}, "sidebar": "tutorialSidebar", "previous": { - "title": "Prompt engineering techniques", - "permalink": "/Part-1-labs/Prompt-engineering-techniques" + "title": "Multimodal", + "permalink": "/Part-1-labs/Multimodal" }, "next": { "title": "Add knowledge", @@ -372,6 +420,14 @@ { "type": "doc", "id": "Part-1-labs/Prompt-engineering-techniques" + }, + { + "type": "doc", + "id": "Part-1-labs/Image Generation" + }, + { + "type": "doc", + "id": "Part-1-labs/Multimodal" } ] }, diff --git a/.docusaurus/globalData.json b/.docusaurus/globalData.json index 99ab9adc..84e3c886 100644 --- a/.docusaurus/globalData.json +++ b/.docusaurus/globalData.json @@ -45,6 +45,16 @@ "path": "/Part-1-labs/Conversation-history", "sidebar": "tutorialSidebar" }, + { + "id": "Part-1-labs/Image Generation", + "path": "/Part-1-labs/Image Generation", + "sidebar": "tutorialSidebar" + }, + { + "id": "Part-1-labs/Multimodal", + "path": "/Part-1-labs/Multimodal", + "sidebar": "tutorialSidebar" + }, { "id": "Part-1-labs/Prompt-engineering-techniques", "path": "/Part-1-labs/Prompt-engineering-techniques", diff --git a/.docusaurus/registry.js b/.docusaurus/registry.js index 9b3c8228..7c8a6e51 100644 --- a/.docusaurus/registry.js +++ b/.docusaurus/registry.js @@ -10,22 +10,24 @@ export default { "__comp---theme-doc-version-roota-7-b-5de": [() => import(/* webpackChunkName: "__comp---theme-doc-version-roota-7-b-5de" */ "@theme/DocVersionRoot"), "@theme/DocVersionRoot", require.resolveWeak("@theme/DocVersionRoot")], "__comp---theme-docs-root-5-e-9-0b6": [() => import(/* webpackChunkName: "__comp---theme-docs-root-5-e-9-0b6" */ "@theme/DocsRoot"), "@theme/DocsRoot", require.resolveWeak("@theme/DocsRoot")], "__comp---theme-mdx-page-1-f-3-b90": [() => import(/* webpackChunkName: "__comp---theme-mdx-page-1-f-3-b90" */ "@theme/MDXPage"), "@theme/MDXPage", require.resolveWeak("@theme/MDXPage")], - "allContent---workshop-interact-with-open-ai-models-docusaurus-debug-content-246-957": [() => import(/* webpackChunkName: "allContent---workshop-interact-with-open-ai-models-docusaurus-debug-content-246-957" */ "~debug/default/docusaurus-debug-all-content-673.json"), "~debug/default/docusaurus-debug-all-content-673.json", require.resolveWeak("~debug/default/docusaurus-debug-all-content-673.json")], - "content---workshop-interact-with-open-ai-models-ai-models-3-a-6-901": [() => import(/* webpackChunkName: "content---workshop-interact-with-open-ai-models-ai-models-3-a-6-901" */ "@site/docs/06-concepts/1-Explore-Models.md"), "@site/docs/06-concepts/1-Explore-Models.md", require.resolveWeak("@site/docs/06-concepts/1-Explore-Models.md")], - "content---workshop-interact-with-open-ai-models-at-home-ebf-782": [() => import(/* webpackChunkName: "content---workshop-interact-with-open-ai-models-at-home-ebf-782" */ "@site/docs/06-concepts/3-At-home.md"), "@site/docs/06-concepts/3-At-home.md", require.resolveWeak("@site/docs/06-concepts/3-At-home.md")], - "content---workshop-interact-with-open-ai-models-bea-90e": [() => import(/* webpackChunkName: "content---workshop-interact-with-open-ai-models-bea-90e" */ "@site/docs/00-Welcome.md"), "@site/docs/00-Welcome.md", require.resolveWeak("@site/docs/00-Welcome.md")], - "content---workshop-interact-with-open-ai-models-llms-3-ba-6c4": [() => import(/* webpackChunkName: "content---workshop-interact-with-open-ai-models-llms-3-ba-6c4" */ "@site/docs/06-concepts/2-Understand-LLMs.md"), "@site/docs/06-concepts/2-Understand-LLMs.md", require.resolveWeak("@site/docs/06-concepts/2-Understand-LLMs.md")], - "content---workshop-interact-with-open-ai-models-markdown-page-393-78e": [() => import(/* webpackChunkName: "content---workshop-interact-with-open-ai-models-markdown-page-393-78e" */ "@site/src/pages/markdown-page.md"), "@site/src/pages/markdown-page.md", require.resolveWeak("@site/src/pages/markdown-page.md")], - "content---workshop-interact-with-open-ai-models-part-1-labs-basic-prompting-e-9-c-b3a": [() => import(/* webpackChunkName: "content---workshop-interact-with-open-ai-models-part-1-labs-basic-prompting-e-9-c-b3a" */ "@site/docs/04-Part-1-labs/1-Basic-Prompting.md"), "@site/docs/04-Part-1-labs/1-Basic-Prompting.md", require.resolveWeak("@site/docs/04-Part-1-labs/1-Basic-Prompting.md")], - "content---workshop-interact-with-open-ai-models-part-1-labs-conversation-history-733-a37": [() => import(/* webpackChunkName: "content---workshop-interact-with-open-ai-models-part-1-labs-conversation-history-733-a37" */ "@site/docs/04-Part-1-labs/2-Conversation-history.md"), "@site/docs/04-Part-1-labs/2-Conversation-history.md", require.resolveWeak("@site/docs/04-Part-1-labs/2-Conversation-history.md")], - "content---workshop-interact-with-open-ai-models-part-1-labs-prompt-engineering-techniques-140-8cd": [() => import(/* webpackChunkName: "content---workshop-interact-with-open-ai-models-part-1-labs-prompt-engineering-techniques-140-8cd" */ "@site/docs/04-Part-1-labs/4-Prompt-engineering-techniques.md"), "@site/docs/04-Part-1-labs/4-Prompt-engineering-techniques.md", require.resolveWeak("@site/docs/04-Part-1-labs/4-Prompt-engineering-techniques.md")], - "content---workshop-interact-with-open-ai-models-part-2-labs-add-knowledge-90-c-9ac": [() => import(/* webpackChunkName: "content---workshop-interact-with-open-ai-models-part-2-labs-add-knowledge-90-c-9ac" */ "@site/docs/05-Part-2-labs/5-Add-Knowledge.md"), "@site/docs/05-Part-2-labs/5-Add-Knowledge.md", require.resolveWeak("@site/docs/05-Part-2-labs/5-Add-Knowledge.md")], - "content---workshop-interact-with-open-ai-models-part-2-labs-function-calling-89-c-055": [() => import(/* webpackChunkName: "content---workshop-interact-with-open-ai-models-part-2-labs-function-calling-89-c-055" */ "@site/docs/05-Part-2-labs/6-Function-Calling.md"), "@site/docs/05-Part-2-labs/6-Function-Calling.md", require.resolveWeak("@site/docs/05-Part-2-labs/6-Function-Calling.md")], - "content---workshop-interact-with-open-ai-models-part-2-labs-system-message-0-b-5-69d": [() => import(/* webpackChunkName: "content---workshop-interact-with-open-ai-models-part-2-labs-system-message-0-b-5-69d" */ "@site/docs/05-Part-2-labs/3-System-Message.md"), "@site/docs/05-Part-2-labs/3-System-Message.md", require.resolveWeak("@site/docs/05-Part-2-labs/3-System-Message.md")], - "content---workshop-interact-with-open-ai-models-setup-9-f-0-3c5": [() => import(/* webpackChunkName: "content---workshop-interact-with-open-ai-models-setup-9-f-0-3c5" */ "@site/docs/01-Get-Started.md"), "@site/docs/01-Get-Started.md", require.resolveWeak("@site/docs/01-Get-Started.md")], - "content---workshop-interact-with-open-ai-models-summary-7-cc-80a": [() => import(/* webpackChunkName: "content---workshop-interact-with-open-ai-models-summary-7-cc-80a" */ "@site/docs/100-Summary.md"), "@site/docs/100-Summary.md", require.resolveWeak("@site/docs/100-Summary.md")], - "content---workshop-interact-with-open-ai-models-tokenization-1-b-8-2bd": [() => import(/* webpackChunkName: "content---workshop-interact-with-open-ai-models-tokenization-1-b-8-2bd" */ "@site/docs/06-concepts/3-Understand-Tokens.md"), "@site/docs/06-concepts/3-Understand-Tokens.md", require.resolveWeak("@site/docs/06-concepts/3-Understand-Tokens.md")], - "plugin---workshop-interact-with-open-ai-models-c-6-e-150": [() => import(/* webpackChunkName: "plugin---workshop-interact-with-open-ai-models-c-6-e-150" */ "/workspaces/AzureOpenAIService-Workshop/.docusaurus/docusaurus-plugin-content-docs/default/plugin-route-context-module-100.json"), "/workspaces/AzureOpenAIService-Workshop/.docusaurus/docusaurus-plugin-content-docs/default/plugin-route-context-module-100.json", require.resolveWeak("/workspaces/AzureOpenAIService-Workshop/.docusaurus/docusaurus-plugin-content-docs/default/plugin-route-context-module-100.json")], - "plugin---workshop-interact-with-open-ai-models-docusaurus-debug-d-63-78e": [() => import(/* webpackChunkName: "plugin---workshop-interact-with-open-ai-models-docusaurus-debug-d-63-78e" */ "/workspaces/AzureOpenAIService-Workshop/.docusaurus/docusaurus-plugin-debug/default/plugin-route-context-module-100.json"), "/workspaces/AzureOpenAIService-Workshop/.docusaurus/docusaurus-plugin-debug/default/plugin-route-context-module-100.json", require.resolveWeak("/workspaces/AzureOpenAIService-Workshop/.docusaurus/docusaurus-plugin-debug/default/plugin-route-context-module-100.json")], - "plugin---workshop-interact-with-open-ai-models-markdown-page-7-af-dfb": [() => import(/* webpackChunkName: "plugin---workshop-interact-with-open-ai-models-markdown-page-7-af-dfb" */ "/workspaces/AzureOpenAIService-Workshop/.docusaurus/docusaurus-plugin-content-pages/default/plugin-route-context-module-100.json"), "/workspaces/AzureOpenAIService-Workshop/.docusaurus/docusaurus-plugin-content-pages/default/plugin-route-context-module-100.json", require.resolveWeak("/workspaces/AzureOpenAIService-Workshop/.docusaurus/docusaurus-plugin-content-pages/default/plugin-route-context-module-100.json")], - "version---workshop-interact-with-open-ai-models-935-a61": [() => import(/* webpackChunkName: "version---workshop-interact-with-open-ai-models-935-a61" */ "~docs/default/version-current-metadata-prop-751.json"), "~docs/default/version-current-metadata-prop-751.json", require.resolveWeak("~docs/default/version-current-metadata-prop-751.json")],}; + "allContent---docusaurus-debug-content-246-d88": [() => import(/* webpackChunkName: "allContent---docusaurus-debug-content-246-d88" */ "~debug/default/docusaurus-debug-all-content-673.json"), "~debug/default/docusaurus-debug-all-content-673.json", require.resolveWeak("~debug/default/docusaurus-debug-all-content-673.json")], + "content---ai-models-3-a-6-1ba": [() => import(/* webpackChunkName: "content---ai-models-3-a-6-1ba" */ "@site/docs/06-concepts/1-Explore-Models.md"), "@site/docs/06-concepts/1-Explore-Models.md", require.resolveWeak("@site/docs/06-concepts/1-Explore-Models.md")], + "content---at-home-ebf-186": [() => import(/* webpackChunkName: "content---at-home-ebf-186" */ "@site/docs/06-concepts/3-At-home.md"), "@site/docs/06-concepts/3-At-home.md", require.resolveWeak("@site/docs/06-concepts/3-At-home.md")], + "content---bea-ef4": [() => import(/* webpackChunkName: "content---bea-ef4" */ "@site/docs/00-Welcome.md"), "@site/docs/00-Welcome.md", require.resolveWeak("@site/docs/00-Welcome.md")], + "content---llms-3-ba-cf2": [() => import(/* webpackChunkName: "content---llms-3-ba-cf2" */ "@site/docs/06-concepts/2-Understand-LLMs.md"), "@site/docs/06-concepts/2-Understand-LLMs.md", require.resolveWeak("@site/docs/06-concepts/2-Understand-LLMs.md")], + "content---markdown-page-393-038": [() => import(/* webpackChunkName: "content---markdown-page-393-038" */ "@site/src/pages/markdown-page.md"), "@site/src/pages/markdown-page.md", require.resolveWeak("@site/src/pages/markdown-page.md")], + "content---part-1-labs-basic-prompting-e-9-c-989": [() => import(/* webpackChunkName: "content---part-1-labs-basic-prompting-e-9-c-989" */ "@site/docs/04-Part-1-labs/1-Basic-Prompting.md"), "@site/docs/04-Part-1-labs/1-Basic-Prompting.md", require.resolveWeak("@site/docs/04-Part-1-labs/1-Basic-Prompting.md")], + "content---part-1-labs-conversation-history-733-d23": [() => import(/* webpackChunkName: "content---part-1-labs-conversation-history-733-d23" */ "@site/docs/04-Part-1-labs/2-Conversation-history.md"), "@site/docs/04-Part-1-labs/2-Conversation-history.md", require.resolveWeak("@site/docs/04-Part-1-labs/2-Conversation-history.md")], + "content---part-1-labs-image-generation-3-ed-08c": [() => import(/* webpackChunkName: "content---part-1-labs-image-generation-3-ed-08c" */ "@site/docs/04-Part-1-labs/5-Image Generation.md"), "@site/docs/04-Part-1-labs/5-Image Generation.md", require.resolveWeak("@site/docs/04-Part-1-labs/5-Image Generation.md")], + "content---part-1-labs-multimodal-3-f-5-49d": [() => import(/* webpackChunkName: "content---part-1-labs-multimodal-3-f-5-49d" */ "@site/docs/04-Part-1-labs/6-Multimodal.md"), "@site/docs/04-Part-1-labs/6-Multimodal.md", require.resolveWeak("@site/docs/04-Part-1-labs/6-Multimodal.md")], + "content---part-1-labs-prompt-engineering-techniques-140-67b": [() => import(/* webpackChunkName: "content---part-1-labs-prompt-engineering-techniques-140-67b" */ "@site/docs/04-Part-1-labs/4-Prompt-engineering-techniques.md"), "@site/docs/04-Part-1-labs/4-Prompt-engineering-techniques.md", require.resolveWeak("@site/docs/04-Part-1-labs/4-Prompt-engineering-techniques.md")], + "content---part-2-labs-add-knowledge-90-c-d3f": [() => import(/* webpackChunkName: "content---part-2-labs-add-knowledge-90-c-d3f" */ "@site/docs/05-Part-2-labs/5-Add-Knowledge.md"), "@site/docs/05-Part-2-labs/5-Add-Knowledge.md", require.resolveWeak("@site/docs/05-Part-2-labs/5-Add-Knowledge.md")], + "content---part-2-labs-function-calling-89-c-033": [() => import(/* webpackChunkName: "content---part-2-labs-function-calling-89-c-033" */ "@site/docs/05-Part-2-labs/6-Function-Calling.md"), "@site/docs/05-Part-2-labs/6-Function-Calling.md", require.resolveWeak("@site/docs/05-Part-2-labs/6-Function-Calling.md")], + "content---part-2-labs-system-message-0-b-5-995": [() => import(/* webpackChunkName: "content---part-2-labs-system-message-0-b-5-995" */ "@site/docs/05-Part-2-labs/3-System-Message.md"), "@site/docs/05-Part-2-labs/3-System-Message.md", require.resolveWeak("@site/docs/05-Part-2-labs/3-System-Message.md")], + "content---setup-9-f-0-e20": [() => import(/* webpackChunkName: "content---setup-9-f-0-e20" */ "@site/docs/01-Get-Started.md"), "@site/docs/01-Get-Started.md", require.resolveWeak("@site/docs/01-Get-Started.md")], + "content---summary-7-cc-457": [() => import(/* webpackChunkName: "content---summary-7-cc-457" */ "@site/docs/100-Summary.md"), "@site/docs/100-Summary.md", require.resolveWeak("@site/docs/100-Summary.md")], + "content---tokenization-1-b-8-e54": [() => import(/* webpackChunkName: "content---tokenization-1-b-8-e54" */ "@site/docs/06-concepts/3-Understand-Tokens.md"), "@site/docs/06-concepts/3-Understand-Tokens.md", require.resolveWeak("@site/docs/06-concepts/3-Understand-Tokens.md")], + "plugin---c-6-e-368": [() => import(/* webpackChunkName: "plugin---c-6-e-368" */ "/workspaces/AzureOpenAIService-Workshop/.docusaurus/docusaurus-plugin-content-docs/default/plugin-route-context-module-100.json"), "/workspaces/AzureOpenAIService-Workshop/.docusaurus/docusaurus-plugin-content-docs/default/plugin-route-context-module-100.json", require.resolveWeak("/workspaces/AzureOpenAIService-Workshop/.docusaurus/docusaurus-plugin-content-docs/default/plugin-route-context-module-100.json")], + "plugin---docusaurus-debug-d-63-ef3": [() => import(/* webpackChunkName: "plugin---docusaurus-debug-d-63-ef3" */ "/workspaces/AzureOpenAIService-Workshop/.docusaurus/docusaurus-plugin-debug/default/plugin-route-context-module-100.json"), "/workspaces/AzureOpenAIService-Workshop/.docusaurus/docusaurus-plugin-debug/default/plugin-route-context-module-100.json", require.resolveWeak("/workspaces/AzureOpenAIService-Workshop/.docusaurus/docusaurus-plugin-debug/default/plugin-route-context-module-100.json")], + "plugin---markdown-page-7-af-a57": [() => import(/* webpackChunkName: "plugin---markdown-page-7-af-a57" */ "/workspaces/AzureOpenAIService-Workshop/.docusaurus/docusaurus-plugin-content-pages/default/plugin-route-context-module-100.json"), "/workspaces/AzureOpenAIService-Workshop/.docusaurus/docusaurus-plugin-content-pages/default/plugin-route-context-module-100.json", require.resolveWeak("/workspaces/AzureOpenAIService-Workshop/.docusaurus/docusaurus-plugin-content-pages/default/plugin-route-context-module-100.json")], + "version---935-fcf": [() => import(/* webpackChunkName: "version---935-fcf" */ "~docs/default/version-current-metadata-prop-751.json"), "~docs/default/version-current-metadata-prop-751.json", require.resolveWeak("~docs/default/version-current-metadata-prop-751.json")],}; diff --git a/.docusaurus/routes.js b/.docusaurus/routes.js index 54038681..bb4849da 100644 --- a/.docusaurus/routes.js +++ b/.docusaurus/routes.js @@ -44,15 +44,15 @@ export default [ }, { path: '/', - component: ComponentCreator('/', '01d'), + component: ComponentCreator('/', '657'), routes: [ { path: '/', - component: ComponentCreator('/', '9a3'), + component: ComponentCreator('/', '49c'), routes: [ { path: '/', - component: ComponentCreator('/', 'b89'), + component: ComponentCreator('/', '34b'), routes: [ { path: '/ai-models/', @@ -84,6 +84,18 @@ export default [ exact: true, sidebar: "tutorialSidebar" }, + { + path: '/Part-1-labs/Image Generation/', + component: ComponentCreator('/Part-1-labs/Image Generation/', '602'), + exact: true, + sidebar: "tutorialSidebar" + }, + { + path: '/Part-1-labs/Multimodal/', + component: ComponentCreator('/Part-1-labs/Multimodal/', 'd4c'), + exact: true, + sidebar: "tutorialSidebar" + }, { path: '/Part-1-labs/Prompt-engineering-techniques/', component: ComponentCreator('/Part-1-labs/Prompt-engineering-techniques/', 'b9f'), diff --git a/.docusaurus/routesChunkNames.json b/.docusaurus/routesChunkNames.json index b8c2ba12..963cbedb 100644 --- a/.docusaurus/routesChunkNames.json +++ b/.docusaurus/routesChunkNames.json @@ -2,116 +2,124 @@ "/__docusaurus/debug/-0c4": { "__comp": "__comp---theme-debug-config-23-a-2ff", "__context": { - "plugin": "plugin---workshop-interact-with-open-ai-models-docusaurus-debug-d-63-78e" + "plugin": "plugin---docusaurus-debug-d-63-ef3" } }, "/__docusaurus/debug/config/-254": { "__comp": "__comp---theme-debug-config-23-a-2ff", "__context": { - "plugin": "plugin---workshop-interact-with-open-ai-models-docusaurus-debug-d-63-78e" + "plugin": "plugin---docusaurus-debug-d-63-ef3" } }, "/__docusaurus/debug/content/-091": { "__comp": "__comp---theme-debug-contentba-8-ce7", "__context": { - "plugin": "plugin---workshop-interact-with-open-ai-models-docusaurus-debug-d-63-78e" + "plugin": "plugin---docusaurus-debug-d-63-ef3" }, - "allContent": "allContent---workshop-interact-with-open-ai-models-docusaurus-debug-content-246-957" + "allContent": "allContent---docusaurus-debug-content-246-d88" }, "/__docusaurus/debug/globalData/-540": { "__comp": "__comp---theme-debug-global-dataede-0fa", "__context": { - "plugin": "plugin---workshop-interact-with-open-ai-models-docusaurus-debug-d-63-78e" + "plugin": "plugin---docusaurus-debug-d-63-ef3" } }, "/__docusaurus/debug/metadata/-3ff": { "__comp": "__comp---theme-debug-site-metadata-68-e-3d4", "__context": { - "plugin": "plugin---workshop-interact-with-open-ai-models-docusaurus-debug-d-63-78e" + "plugin": "plugin---docusaurus-debug-d-63-ef3" } }, "/__docusaurus/debug/registry/-044": { "__comp": "__comp---theme-debug-registry-679-501", "__context": { - "plugin": "plugin---workshop-interact-with-open-ai-models-docusaurus-debug-d-63-78e" + "plugin": "plugin---docusaurus-debug-d-63-ef3" } }, "/__docusaurus/debug/routes/-3fa": { "__comp": "__comp---theme-debug-routes-946-699", "__context": { - "plugin": "plugin---workshop-interact-with-open-ai-models-docusaurus-debug-d-63-78e" + "plugin": "plugin---docusaurus-debug-d-63-ef3" } }, "/markdown-page/-d0d": { "__comp": "__comp---theme-mdx-page-1-f-3-b90", "__context": { - "plugin": "plugin---workshop-interact-with-open-ai-models-markdown-page-7-af-dfb" + "plugin": "plugin---markdown-page-7-af-a57" }, - "content": "content---workshop-interact-with-open-ai-models-markdown-page-393-78e" + "content": "content---markdown-page-393-038" }, - "/-01d": { + "/-657": { "__comp": "__comp---theme-docs-root-5-e-9-0b6", "__context": { - "plugin": "plugin---workshop-interact-with-open-ai-models-c-6-e-150" + "plugin": "plugin---c-6-e-368" } }, - "/-9a3": { + "/-49c": { "__comp": "__comp---theme-doc-version-roota-7-b-5de", - "version": "version---workshop-interact-with-open-ai-models-935-a61" + "version": "version---935-fcf" }, - "/-b89": { + "/-34b": { "__comp": "__comp---theme-doc-roota-94-67a" }, "/ai-models/-90a": { "__comp": "__comp---theme-doc-item-178-a40", - "content": "content---workshop-interact-with-open-ai-models-ai-models-3-a-6-901" + "content": "content---ai-models-3-a-6-1ba" }, "/at-home/-c59": { "__comp": "__comp---theme-doc-item-178-a40", - "content": "content---workshop-interact-with-open-ai-models-at-home-ebf-782" + "content": "content---at-home-ebf-186" }, "/llms/-a1f": { "__comp": "__comp---theme-doc-item-178-a40", - "content": "content---workshop-interact-with-open-ai-models-llms-3-ba-6c4" + "content": "content---llms-3-ba-cf2" }, "/Part-1-labs/Basic-Prompting/-4ea": { "__comp": "__comp---theme-doc-item-178-a40", - "content": "content---workshop-interact-with-open-ai-models-part-1-labs-basic-prompting-e-9-c-b3a" + "content": "content---part-1-labs-basic-prompting-e-9-c-989" }, "/Part-1-labs/Conversation-history/-270": { "__comp": "__comp---theme-doc-item-178-a40", - "content": "content---workshop-interact-with-open-ai-models-part-1-labs-conversation-history-733-a37" + "content": "content---part-1-labs-conversation-history-733-d23" + }, + "/Part-1-labs/Image Generation/-602": { + "__comp": "__comp---theme-doc-item-178-a40", + "content": "content---part-1-labs-image-generation-3-ed-08c" + }, + "/Part-1-labs/Multimodal/-d4c": { + "__comp": "__comp---theme-doc-item-178-a40", + "content": "content---part-1-labs-multimodal-3-f-5-49d" }, "/Part-1-labs/Prompt-engineering-techniques/-b9f": { "__comp": "__comp---theme-doc-item-178-a40", - "content": "content---workshop-interact-with-open-ai-models-part-1-labs-prompt-engineering-techniques-140-8cd" + "content": "content---part-1-labs-prompt-engineering-techniques-140-67b" }, "/Part-2-labs/Add-Knowledge/-7a5": { "__comp": "__comp---theme-doc-item-178-a40", - "content": "content---workshop-interact-with-open-ai-models-part-2-labs-add-knowledge-90-c-9ac" + "content": "content---part-2-labs-add-knowledge-90-c-d3f" }, "/Part-2-labs/Function-Calling/-57b": { "__comp": "__comp---theme-doc-item-178-a40", - "content": "content---workshop-interact-with-open-ai-models-part-2-labs-function-calling-89-c-055" + "content": "content---part-2-labs-function-calling-89-c-033" }, "/Part-2-labs/System-Message/-129": { "__comp": "__comp---theme-doc-item-178-a40", - "content": "content---workshop-interact-with-open-ai-models-part-2-labs-system-message-0-b-5-69d" + "content": "content---part-2-labs-system-message-0-b-5-995" }, "/setup/-e1d": { "__comp": "__comp---theme-doc-item-178-a40", - "content": "content---workshop-interact-with-open-ai-models-setup-9-f-0-3c5" + "content": "content---setup-9-f-0-e20" }, "/summary/-0c6": { "__comp": "__comp---theme-doc-item-178-a40", - "content": "content---workshop-interact-with-open-ai-models-summary-7-cc-80a" + "content": "content---summary-7-cc-457" }, "/tokenization/-1ed": { "__comp": "__comp---theme-doc-item-178-a40", - "content": "content---workshop-interact-with-open-ai-models-tokenization-1-b-8-2bd" + "content": "content---tokenization-1-b-8-e54" }, "/-a80": { "__comp": "__comp---theme-doc-item-178-a40", - "content": "content---workshop-interact-with-open-ai-models-bea-90e" + "content": "content---bea-ef4" } } \ No newline at end of file diff --git a/.gitignore b/.gitignore new file mode 100644 index 00000000..4d54a081 --- /dev/null +++ b/.gitignore @@ -0,0 +1,9 @@ +node_modules/ +node_modules/.cache/webpack/client-development-en/0.pack +node_modules/.cache/webpack/client-development-en/1.pack +node_modules/.cache/webpack/client-development-en/2.pack +node_modules/.cache/webpack/client-development-en/3.pack +node_modules/.cache/webpack/client-development-en/4.pack +node_modules/.cache/webpack/client-development-en/5.pack +node_modules/.cache/webpack/client-development-en/index.pack +node_modules/.cache/webpack/client-development-en/index.pack.old diff --git a/docs/00-Welcome.md b/docs/00-Welcome.md index 95a5924b..09ab143e 100644 --- a/docs/00-Welcome.md +++ b/docs/00-Welcome.md @@ -20,12 +20,11 @@ By the end of this workshop, you should be able to: The lab is organized into sections, taking you from "understanding" (core concepts) to "doing" (hands-on exercises). Prioritize completing the exercises within the allotted time. You can then any remaining session time for open-ended exploration to build your own intuition for prompt engineering. -1. [Setup](/setup). Activate your access to the custom workshop Playground. +1. [Get started](/setup). Activate your access to the custom workshop Playground. 2. [Part 1](/Part-1-labs/Basic-Prompting/). Learn the basics of interacting with the GPT Models . 3. [Part 2](/Part-2-labs/System-Message/). Learn how to guide the models to ​create better answers. 4. [Concepts & resources](/Workshop-Interact-with-OpenAI-models/ai-models/). 5. [Learnings and Resources](/summary). Recap what you learned & identify next steps. - ## Learner Feedback -We would love to hear your feedback on this lab. Please visit the [GitHub repository for this lab](https://github.com/microsoft/Workshop-Interact-with-OpenAI-models/), and use the Issues tab for bugs or suggestions or create a pull-request with your suggestions. +We would love to hear your feedback on this lab. Please visit the [GitHub repository for this lab](https://github.com/GlobalAICommunity/AzureOpenAIService-Workshop), and use the Issues tab for bugs or suggestions or create a pull-request with your suggestions. diff --git a/docs/01-Get-Started.md b/docs/01-Get-Started.md index 44436a68..c84254d6 100644 --- a/docs/01-Get-Started.md +++ b/docs/01-Get-Started.md @@ -14,15 +14,14 @@ have the oppertunity to play with the prompts and learn how to interact with Ope ## Authenticate -Your workshop instructor will provide you with an **event key** that you can use to get your personal **access key** +Your workshop instructor will provide you with an **access key**. -Setup is now a quick 3-step process. +Setup is now a quick 2-step process. -1. 🗝 | **Get your event-key** - from your workshop organizers at the start of the event. -2. 📬 | **Get your access-key** - enter event key in [**this form**](https://globalai.community/workshop) to receive your access-key via email. -3. 💬 | **Use the access-key** - visit the [**Global AI Playground**](https://playground.globalai.community), enter access-key at top right, and click `Authorize`. +1. 🗝 | **Get your access-key** - from your workshop organizers at the start of the event. +3. 💬 | **Use the access-key** - visit the [**Azure AI Proxy Playground**](https://mango-bush-0a9e12903.5.azurestaticapps.net/), enter access-key at top right, and click `Authorize`. -![AI Tour Playground](./images/aitour-playground-chat.png) +![AI Tour Playground](./images/new/playground.png) ## Explore @@ -30,12 +29,12 @@ On successful login, the playground application should update to show a `Logout` Let's take a minute to familiarize ourselves with the different regions of this app. -![](./images/aitour-playground-regions.png) +![](./images/new/playground_2.png) - Region 1️⃣ | Once authorized, shows the event profile. - Region 2️⃣ | User prompt - used to enter text questions for AI - Region 3️⃣ | Conversation - interleaved user questions & assistant responses -- Region 4️⃣ | Configuration - tunable parameters, execution stats +- Region 4️⃣ | Configuration - model selection, tunable parameters, execution stats - Region 5️⃣ | System message - sets assistant persona, base context - Region 6️⃣ | Function calling - save custom functions to set context - Region 7️⃣ | Image generation - access the DALL-E model playground diff --git a/docs/04-Part-1-labs/1-Basic-Prompting.md b/docs/04-Part-1-labs/1-Basic-Prompting.md index 9e254f11..f5f5ff7d 100644 --- a/docs/04-Part-1-labs/1-Basic-Prompting.md +++ b/docs/04-Part-1-labs/1-Basic-Prompting.md @@ -62,7 +62,7 @@ Natural language generative AI models have a number of limitations: Here are some example prompts that demonstrate these weaknesses: ```text title="Enter in the user prompt:" -When did Queen Elizabeth II die? +Which country won the most medals during the 2024 Summer Olympics? ``` In this case, the model is limited by training data, which is current only up to June 2021. diff --git a/docs/06-Part-3-labs/01-Image Generation.md b/docs/04-Part-1-labs/5-Image Generation.md similarity index 93% rename from docs/06-Part-3-labs/01-Image Generation.md rename to docs/04-Part-1-labs/5-Image Generation.md index e5e9ec61..df93d9c1 100644 --- a/docs/06-Part-3-labs/01-Image Generation.md +++ b/docs/04-Part-1-labs/5-Image Generation.md @@ -4,8 +4,10 @@ Welcome to part 2 of this workshop, where we will interact with a text-to-image ## Creating our first images -> [!TIP] -> What is DALL-E 3? DALL-E 3 is a neural network based model that can generate graphical data from **natural language input**. Put more simply, you can provide DALL-E 3 with a **description** and it can generate an appropriate image. +:::tip What is DALL-E 3? +DALL-E 3 is a neural network based model that can generate graphical data from **natural language input**. Put more simply, you can provide DALL-E 3 with a **description** and it can generate an appropriate image. +::: + Let's begin with generating an image by entering a basic prompt in the description box, then clicking on generate: diff --git a/docs/04-Part-1-labs/6-Multimodal.md b/docs/04-Part-1-labs/6-Multimodal.md new file mode 100644 index 00000000..0fd66403 --- /dev/null +++ b/docs/04-Part-1-labs/6-Multimodal.md @@ -0,0 +1,74 @@ +# Multimodal + +By now we have interacted with LLMs using a single modality: inputting text and receiving either text or images. However, multimodal interfaces are becoming increasingly popular, as they allow users to interact with models using multiple modalities, such as text, images, and speech, thus facilitating human-computer interactions. In this section, we will explore how to use multimodal interfaces to interact with GPT-4o. + + +:::tip +**GPT-4o** is a multimodal model that incorporates both natural language processing and visual understanding. It is able to process as input a combination of text and images, and generate an output which is relevant to both modalities. +::: + +## Best Practices + +- **Contextual specificity**: Adding context to the scenario at hand gives the model a better understanding of an appropriate output. This level of specificity aids in focusing on relevant aspects and avoiding extraneous details.​ + +- **Task-oriented prompts**: Focusing on a specific task helps the model to develop the output while taking that perspective into consideration.​ + +- **Define output format**: Clearly mention the desired format for the output, such as markdown, JSON, HTML, etc. You can also suggest a specific structure, length, or specific attributes about the response.​ + +- **Handling refusals**: When the model indicates an inability to perform a task, refining the prompt can be an effective solution. More specific prompts can guide the model towards a clearer understanding and better execution of the task. Some tips to keep in mind:​ + - Request explanations for generated responses to enhance transparency in the model's output​ + - If using a single-image prompt, place the image before the text​ + - Ask the model to describe the image in detail first and complete your specific task from the description​ + +- **Prompt Tuning**: Try prompt tuning techniques that we explored for text generation scenarios such as:​ + - Breaking down requests (e.g. chain of thoughts)​ + - Adding examples (e.g. few-shot learning)​ + +## Excerise 1 + +1. In Project playground section select **Chat.** +2. In the chat text box, click on the attachment icon to upload a local image. + +3. [Download the image](../images/mm_1_Rental_Listing_s.jpg) + +4. Once you have uploaded the file, try these prompts to start interacting with the image: + + +```text title="Enter in the user prompt:" +What is on the image? +``` + +```text title="Enter in the user prompt:" +What is the house made of? +``` + +```text title="Enter in the user prompt:" +Does the house have a swimming pool? +``` + +## Excerise 2 + +1. [Download the image](../images/mm_2_TriangleCalculation_s.jpg) + +```text title="Enter in the user prompt:" +Answer the questions on the image. +``` + +## Excerise 3 + +1. [Download the images](../images/mm_3.zip) + +```text title="Enter in the user prompt:" + Create a tagline and short description for this rental home advertisement. + - The first picture is from the home + - The other pictures are from sights nearby + - In the description use the features of the house and make the ad more compelling with the sights. + - Do not talk about features not visible in the images. + - If you have information about the location of the images, use that information in the description +``` + + +## Next Steps + +Congratulations! You have now completed the 3rd part of the lab and you learnt how to interact with multimodal models. + diff --git a/docs/images/mm_1_Rental_Listing_s.jpg b/docs/images/mm_1_Rental_Listing_s.jpg new file mode 100644 index 00000000..ccc230f1 Binary files /dev/null and b/docs/images/mm_1_Rental_Listing_s.jpg differ diff --git a/docs/images/mm_2_TriangleCalculation_s.jpg b/docs/images/mm_2_TriangleCalculation_s.jpg new file mode 100644 index 00000000..6f18d7ff Binary files /dev/null and b/docs/images/mm_2_TriangleCalculation_s.jpg differ diff --git a/docs/images/mm_3.zip b/docs/images/mm_3.zip new file mode 100644 index 00000000..96ac9c85 Binary files /dev/null and b/docs/images/mm_3.zip differ diff --git a/docs/images/new/playground.png b/docs/images/new/playground.png new file mode 100644 index 00000000..2a1aeb7c Binary files /dev/null and b/docs/images/new/playground.png differ diff --git a/docs/images/new/playground_2.png b/docs/images/new/playground_2.png new file mode 100644 index 00000000..93a6df93 Binary files /dev/null and b/docs/images/new/playground_2.png differ diff --git a/node_modules/.cache/webpack/client-development-en/0.pack b/node_modules/.cache/webpack/client-development-en/0.pack index f3772630..64730c17 100644 Binary files a/node_modules/.cache/webpack/client-development-en/0.pack and b/node_modules/.cache/webpack/client-development-en/0.pack differ diff --git a/node_modules/.cache/webpack/client-development-en/1.pack b/node_modules/.cache/webpack/client-development-en/1.pack index 5d556cf2..fccaac0a 100644 Binary files a/node_modules/.cache/webpack/client-development-en/1.pack and b/node_modules/.cache/webpack/client-development-en/1.pack differ diff --git a/node_modules/.cache/webpack/client-development-en/2.pack b/node_modules/.cache/webpack/client-development-en/2.pack index 61eaba22..fe131c55 100644 Binary files a/node_modules/.cache/webpack/client-development-en/2.pack and b/node_modules/.cache/webpack/client-development-en/2.pack differ diff --git a/node_modules/.cache/webpack/client-development-en/3.pack b/node_modules/.cache/webpack/client-development-en/3.pack index e53a3f7d..82db3c2c 100644 Binary files a/node_modules/.cache/webpack/client-development-en/3.pack and b/node_modules/.cache/webpack/client-development-en/3.pack differ diff --git a/node_modules/.cache/webpack/client-development-en/4.pack b/node_modules/.cache/webpack/client-development-en/4.pack index f58d319c..c6e76d90 100644 Binary files a/node_modules/.cache/webpack/client-development-en/4.pack and b/node_modules/.cache/webpack/client-development-en/4.pack differ diff --git a/node_modules/.cache/webpack/client-development-en/5.pack b/node_modules/.cache/webpack/client-development-en/5.pack index e1fea7d3..66365c92 100644 Binary files a/node_modules/.cache/webpack/client-development-en/5.pack and b/node_modules/.cache/webpack/client-development-en/5.pack differ diff --git a/node_modules/.cache/webpack/client-development-en/index.pack b/node_modules/.cache/webpack/client-development-en/index.pack index 5d13874c..5f3b08f8 100644 Binary files a/node_modules/.cache/webpack/client-development-en/index.pack and b/node_modules/.cache/webpack/client-development-en/index.pack differ diff --git a/node_modules/.cache/webpack/client-development-en/index.pack.old b/node_modules/.cache/webpack/client-development-en/index.pack.old index c7dddeda..52820789 100644 Binary files a/node_modules/.cache/webpack/client-development-en/index.pack.old and b/node_modules/.cache/webpack/client-development-en/index.pack.old differ