AI-102 Language, Speech, and Conversation

Keep LUIS, Language service, speech, translation, QnA, and chatbot-style workloads together.

Exams
AI-102
Questions
141
Comments
1398

1. AI-102 Topic 3 Question 79

Sequence
1
Discussion ID
143420
Source URL
https://www.examtopics.com/discussions/microsoft/view/143420-exam-ai-102-topic-3-question-79-discussion/
Posted By
krzkrzkra
Posted At
July 6, 2024, 3:16 p.m.

Question

You are building an Azure AI Language Understanding solution.

You discover that many intents have similar utterances containing airport names or airport codes.

You need to minimize the number of utterances used to train the model.

Which type of custom entity should you use?

  • A. Pattern.any
  • B. machine-learning
  • C. regular expression
  • D. list

Suggested Answer

D

Answer Description Click to expand


Community Answer Votes

Comments 18 comments Click to expand

Comment 1

ID: 1719813 User: az4002024 Badges: - Relative Date: 4 days, 9 hours ago Absolute Date: Sun 08 Mar 2026 05:42 Selected Answer: A Upvotes: 1

Since airports can appear in many forms and combinations (Example variations:

JFK

John F Kennedy Airport

New York JFK

JFK Airport), Pattern.any looks better. List is used for Small, fixed vocabulary.

Comment 2

ID: 1698910 User: Ysfrangieh Badges: - Relative Date: 3 months ago Absolute Date: Thu 11 Dec 2025 18:03 Selected Answer: A Upvotes: 1

List entities ARE great for airports
BUT the question specifically says "similar utterances" - this is Pattern.any territory

Comment 3

ID: 1698506 User: Madhusmita Badges: - Relative Date: 3 months ago Absolute Date: Wed 10 Dec 2025 10:11 Selected Answer: D Upvotes: 1

A list entity is perfect for airport names and codes because they represent a finite, well-defined set of values with known synonyms. This approach dramatically reduces the number of training utterances needed by allowing you to generalize utterance patterns and handle all airport variations through the predefined list mappings rather than requiring separate training examples for each airport.

Comment 4

ID: 1628494 User: harshad883 Badges: - Relative Date: 3 months, 2 weeks ago Absolute Date: Wed 26 Nov 2025 11:16 Selected Answer: D Upvotes: 2

List entities allow you to define a set of known values (like airport names or codes) and their synonyms.
When an utterance contains any of these values, LUIS will recognize them without needing to add separate utterances for each airport.
This minimizes the number of training utterances because you don’t have to create variations for every airport name or code.

Comment 5

ID: 1606641 User: waqy Badges: - Relative Date: 6 months, 1 week ago Absolute Date: Sat 06 Sep 2025 15:21 Selected Answer: D Upvotes: 1

Not A (Pattern.any) because that entity just captures free text placeholders without restricting values.
You’d still need many utterances to cover variations.

With D (list), you define all airport names/codes once, and the model generalizes across intents—minimizing training utterances

Comment 6

ID: 1578802 User: StelSen Badges: - Relative Date: 8 months, 3 weeks ago Absolute Date: Thu 19 Jun 2025 05:06 Selected Answer: A Upvotes: 2

Pattern.any entities are ideal when you want to reduce the number of utterances needed for training.

They act as placeholders in patterns (like {fromAirport} or {toAirport}), and can match any word or phrase, such as airport names or codes.

This is especially useful when you have many similar utterances that differ only by variable data like city names, airport codes, etc.

Comment 7

ID: 1564903 User: marcellov Badges: - Relative Date: 10 months, 2 weeks ago Absolute Date: Wed 30 Apr 2025 01:41 Selected Answer: A Upvotes: 3

To minimize the number of utterances needed for training when many intents share similar utterances with variable airport names or codes, you should use the Pattern.any custom entity. This entity type allows you to define patterns with placeholders for variable parts (like airport names/codes), reducing the need to enumerate every possible utterance variation.

Suppose you have intents like:
BookFlight
CheckFlightStatus

And users might say things like:
"Book a flight to JFK"
"Book a flight to London Heathrow"
"Check the status of flight DL123"
"Check the status of flight to SFO"

The variable parts are airport names, airport codes, or flight numbers.

You define a Pattern.any entity called, for example, @destination or @flightCode.

Intent: BookFlight
Pattern:
"Book a flight to {@destination}"
Intent: CheckFlightStatus
Pattern:
"Check the status of flight {@flightCode}"
"Check the status of flight to {@destination}"

Comment 8

ID: 1333873 User: pabsinaz Badges: - Relative Date: 1 year, 2 months ago Absolute Date: Mon 30 Dec 2024 03:36 Selected Answer: D Upvotes: 1

Either A or D, both works. A less development effort, D a better solution. However, the requirement is not talking about minimizing development efforts.

Comment 9

ID: 1333871 User: pabsinaz Badges: - Relative Date: 1 year, 2 months ago Absolute Date: Mon 30 Dec 2024 03:35 Selected Answer: D Upvotes: 1

Either A or D, both works. A less development effort, D a better solution. Howeve,r the requirements it is not talking to minimize development effort.

Comment 10

ID: 1328128 User: MASANASA Badges: - Relative Date: 1 year, 2 months ago Absolute Date: Tue 17 Dec 2024 21:10 Selected Answer: D Upvotes: 2

pattern.any https://learn.microsoft.com/ja-jp/azure/ai-services/luis/reference-entity-pattern-any?tabs=V2
list https://learn.microsoft.com/en-us/azure/ai-services/language-service/conversational-language-understanding/concepts/entity-components

Comment 11

ID: 1323751 User: Andriki Badges: - Relative Date: 1 year, 3 months ago Absolute Date: Sun 08 Dec 2024 21:03 Selected Answer: D Upvotes: 2

List: fixed, close set of related words

Comment 12

ID: 1319278 User: nastolgia Badges: - Relative Date: 1 year, 3 months ago Absolute Date: Thu 28 Nov 2024 15:06 Selected Answer: D Upvotes: 1

It shoul be LIST

Comment 13

ID: 1311835 User: pmd30 Badges: - Relative Date: 1 year, 3 months ago Absolute Date: Thu 14 Nov 2024 11:36 Selected Answer: - Upvotes: 1

D. List
https://learn.microsoft.com/en-us/azure/ai-services/luis/reference-entity-list?tabs=V2
Suppose the app has a list, named Cities, allowing for variations of city names including city of airport (Sea-tac), airport code (SEA), postal zip code (98101), and phone area code (206).

List item Item synonyms
Seattle sea-tac, sea, 98101, 206, +1
Paris cdg, roissy, ory, 75001, 1, +33
book 2 tickets to paris

In the previous utterance, the word paris is mapped to the paris item as part of the Cities list entity. The list entity matches both the item's normalized name as well as the item synonyms.

Comment 14

ID: 1308555 User: Alan_CA Badges: - Relative Date: 1 year, 4 months ago Absolute Date: Thu 07 Nov 2024 22:19 Selected Answer: D Upvotes: 1

LIST entity

Comment 15

ID: 1289997 User: RajuTS Badges: - Relative Date: 1 year, 5 months ago Absolute Date: Fri 27 Sep 2024 13:22 Selected Answer: - Upvotes: 1

USing a list entity allows you to define a set of values and their synonyms which will help minimize the number of utterances needed to train the model.

Hence the answer must be:
D) List

Comment 16

ID: 1282160 User: mrg998 Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Wed 11 Sep 2024 15:48 Selected Answer: A Upvotes: 1

Pattern, you can use this to summarise multiple utterances into intents

Comment 17

ID: 1281197 User: famco Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Mon 09 Sep 2024 20:46 Selected Answer: - Upvotes: 1

"Patterns are designed to improve accuracy when multiple utterances are very similar. A pattern allows you to gain more accuracy for an intent without providing several more utterances."
So, it has to be pattern. The only one in the list is Pattern.Any.
Looks like the MIcrosoft guy just read this line and made the question and randomly chose a pattern type

Comment 17.1

ID: 1281200 User: famco Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Mon 09 Sep 2024 20:52 Selected Answer: - Upvotes: 1

regular-expression entity is given as an option to trick those who actually read the example for that using flight code. If you did not read much, then you are safe in Microsoft world

2. AI-102 Topic 3 Question 43

Sequence
2
Discussion ID
108161
Source URL
https://www.examtopics.com/discussions/microsoft/view/108161-exam-ai-102-topic-3-question-43-discussion/
Posted By
hcgkzsf
Posted At
May 2, 2023, 2:20 a.m.

Question

You are building a retail kiosk system that will use a custom neural voice.

You acquire audio samples and consent from the voice talent.

You need to create a voice talent profile.

What should you upload to the profile?

  • A. a .zip file that contains 10-second .wav files and the associated transcripts as .txt files
  • B. a five-minute .flac audio file and the associated transcript as a .txt file
  • C. a .wav or .mp3 file of the voice talent consenting to the creation of a synthetic version of their voice
  • D. a five-minute .wav or .mp3 file of the voice talent describing the kiosk system

Suggested Answer

C

Answer Description Click to expand


Community Answer Votes

Comments 19 comments Click to expand

Comment 1

ID: 1085028 User: tdctdc Badges: Highly Voted Relative Date: 2 years, 3 months ago Absolute Date: Fri 01 Dec 2023 08:16 Selected Answer: C Upvotes: 7

The question is about the profile, not data.

Comment 2

ID: 1719807 User: az4002024 Badges: Most Recent Relative Date: 4 days, 11 hours ago Absolute Date: Sun 08 Mar 2026 04:14 Selected Answer: C Upvotes: 1

A voice talent profile requires proof of consent from the person whose voice will be used. This ensures compliance with responsible AI policies.

Comment 3

ID: 1362550 User: gyaansastra Badges: - Relative Date: 1 year ago Absolute Date: Thu 27 Feb 2025 15:18 Selected Answer: A Upvotes: 1

A. a .zip file that contains 10-second .wav files and the associated transcripts as .txt files

Explanation:
This option aligns with the typical requirements for creating a custom neural voice. The system needs short, segmented audio samples along with their corresponding transcripts to accurately train the model and generate the synthetic voice.

By providing multiple short audio files and their transcripts, you ensure that the voice model can learn the nuances of the voice talent's speech and generate a high-quality custom voice.

Comment 4

ID: 1336172 User: VitaliiKurishko Badges: - Relative Date: 1 year, 2 months ago Absolute Date: Fri 03 Jan 2025 23:24 Selected Answer: A Upvotes: 1

https://learn.microsoft.com/en-us/azure/ai-services/speech-service/record-custom-voice-samples

Comment 5

ID: 1290106 User: fawzi008 Badges: - Relative Date: 1 year, 5 months ago Absolute Date: Fri 27 Sep 2024 17:23 Selected Answer: A Upvotes: 1

uploaded testing data must be A collection (.zip) of either .wave +.txt (if samples and transcripts are available)or .wave/.mp3 (if samples only available)
in both cases must be .zip

Comment 5.1

ID: 1292913 User: fawzi008 Badges: - Relative Date: 1 year, 5 months ago Absolute Date: Thu 03 Oct 2024 21:28 Selected Answer: - Upvotes: 1

wrong answer
as per https://learn.microsoft.com/en-us/azure/ai-services/speech-service/professional-voice-create-consent?pivots=speech-studio
file should be .wave or .mp3

Comment 6

ID: 1282155 User: mrg998 Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Wed 11 Sep 2024 15:40 Selected Answer: C Upvotes: 1

anwser is C, see here - https://learn.microsoft.com/en-us/azure/ai-services/speech-service/professional-voice-create-consent?pivots=speech-studio

Comment 7

ID: 1265816 User: anto69 Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Wed 14 Aug 2024 16:06 Selected Answer: A Upvotes: 1

Both me and ChatGPT agree with A

Comment 7.1

ID: 1700407 User: d335823 Badges: - Relative Date: 2 months, 3 weeks ago Absolute Date: Fri 19 Dec 2025 01:47 Selected Answer: - Upvotes: 1

According to ChatGPT, it's C

Comment 8

ID: 1260537 User: anto69 Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Sun 04 Aug 2024 08:02 Selected Answer: A Upvotes: 1

a .zip file that contains 10-second .wav files and the associated transcripts as .txt files

Comment 9

ID: 1235193 User: HaraTadahisa Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 08:24 Selected Answer: C Upvotes: 2

I say this answer is C.

Comment 10

ID: 1229220 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 12 Jun 2024 16:15 Selected Answer: C Upvotes: 3

C is answer.

Comment 11

ID: 1191185 User: Murtuza Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Sun 07 Apr 2024 22:31 Selected Answer: C Upvotes: 2

C is the correct answer

Comment 12

ID: 1141381 User: Tactable Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Mon 05 Feb 2024 20:19 Selected Answer: C Upvotes: 2

https://learn.microsoft.com/en-us/azure/ai-services/speech-service/how-to-custom-voice-talent

Comment 13

ID: 1139197 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 03 Feb 2024 12:48 Selected Answer: C Upvotes: 3

Based on the Azure AI documentation, the correct option for creating a voice talent profile for a custom neural voice is:

C. a .wav or .mp3 file of the voice talent consenting to the creation of a synthetic version of their voice.

This is because the documentation specifies the need for a recording of the voice talent's consent statement, acknowledging the use of their voice recordings by a specified company to create and use a synthetic version of their voice

Comment 14

ID: 1084398 User: RupRizal Badges: - Relative Date: 2 years, 3 months ago Absolute Date: Thu 30 Nov 2023 14:26 Selected Answer: - Upvotes: 2

C is the correct answer as only wav & mp3 formats are allowed. Zip is not allowed.

Comment 15

ID: 1047598 User: katrang Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Thu 19 Oct 2023 08:21 Selected Answer: C Upvotes: 1

I initially thought A, but now think it is C
https://learn.microsoft.com/en-us/azure/ai-services/speech-service/how-to-custom-voice-talent

Comment 16

ID: 1023947 User: AnonymousJhb Badges: - Relative Date: 2 years, 5 months ago Absolute Date: Tue 03 Oct 2023 14:54 Selected Answer: C Upvotes: 3

zip is not allowed. only wav & mp3.
And made synthetic. =C
https://learn.microsoft.com/en-us/azure/ai-services/speech-service/how-to-custom-voice-talent
follow the pictures very carefully.

Comment 17

ID: 996762 User: M25 Badges: - Relative Date: 2 years, 6 months ago Absolute Date: Sat 02 Sep 2023 11:23 Selected Answer: C Upvotes: 1

In Q#13 (How should you upload the samples?), the selected answer as option B (for speech samples to be used in training) corresponds to this Q#43’s option A.

Q#43 (What should you upload to the profile?) though, refers specifically to the voice talent’s profile, that should contain an audio sample of the given consent, and hereto “Only .wav and .mp3 files are accepted” (at the bottom of the screenshot provided in the link).
https://learn.microsoft.com/en-us/azure/ai-services/speech-service/how-to-custom-voice-talent#add-voice-talent

3. AI-102 Topic 11 Question 1

Sequence
8
Discussion ID
60817
Source URL
https://www.examtopics.com/discussions/microsoft/view/60817-exam-ai-102-topic-11-question-1-discussion/
Posted By
SuperPetey
Posted At
Aug. 27, 2021, 9:41 a.m.

Question

HOTSPOT -
You are planning the product creation project.
You need to build the REST endpoint to create the multilingual product descriptions.
How should you complete the URI? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Hot Area:
image

Suggested Answer

image
Answer Description Click to expand

Box 1: api.cognitive.microsofttranslator.com
Translator 3.0: Translate. Send a POST request to:
https://api.cognitive.microsofttranslator.com/translate?api-version=3.0

Box 2: /translate -
Reference:
https://docs.microsoft.com/en-us/azure/cognitive-services/translator/reference/v3-0-translate

Comments 14 comments Click to expand

Comment 1

ID: 432855 User: SuperPetey Badges: Highly Voted Relative Date: 4 years, 6 months ago Absolute Date: Fri 27 Aug 2021 09:41 Selected Answer: - Upvotes: 91

first drop-down should instead be: api-nam.cognitive.microsofttranslator.com this is because the case study specifically states under Business Requirements "Data storage and processing must occur in datacenters located in the United States."

see reference documentation for base urls per geo region:

https://docs.microsoft.com/en-us/azure/cognitive-services/translator/reference/v3-0-reference

Comment 1.1

ID: 648176 User: ninjia Badges: - Relative Date: 3 years, 6 months ago Absolute Date: Wed 17 Aug 2022 19:44 Selected Answer: - Upvotes: 1

Agreed

Comment 1.2

ID: 1133648 User: josebernabeo Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 27 Jan 2024 21:50 Selected Answer: - Upvotes: 3

Where do you read that? This is the question I see:
"HOTSPOT -
You are planning the product creation project.
You need to build the REST endpoint to create the multilingual product descriptions.
How should you complete the URI? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Hot Area:"

Comment 1.2.1

ID: 1330732 User: kennynelcon Badges: - Relative Date: 1 year, 2 months ago Absolute Date: Mon 23 Dec 2024 09:39 Selected Answer: - Upvotes: 2

Check description

Governance and Security Requirements
Wide World Importers identifies the following governance and security requirements:
Data storage and processing must occur in datacenters located in the United States.
Azure Cognitive Services must be inaccessible directly from the internet.

Comment 2

ID: 1287418 User: famco Badges: Most Recent Relative Date: 1 year, 5 months ago Absolute Date: Sat 21 Sep 2024 19:39 Selected Answer: - Upvotes: 1

From that wall of text who has the eye sight to pick "united states".

Comment 2.1

ID: 1715825 User: phvogel Badges: - Relative Date: 3 weeks ago Absolute Date: Wed 18 Feb 2026 19:35 Selected Answer: - Upvotes: 1

I read athe question before I read the case study and then look for what I need to answer the question. Once I saw "nam", I went looking for "United States"

Comment 3

ID: 1230533 User: NagaoShingo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 14 Jun 2024 16:37 Selected Answer: - Upvotes: 2

1. api-nam.cognitive.microsofttranslator.com
2. /translate

Comment 4

ID: 1214420 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Mon 20 May 2024 16:33 Selected Answer: - Upvotes: 1

Is this question still available on May 21, 2024?

Comment 5

ID: 1145914 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 10 Feb 2024 05:01 Selected Answer: - Upvotes: 1

1. api-nam.cognitive.microsofttranslator.com
2. /translate

Comment 6

ID: 1102017 User: Gvalli Badges: - Relative Date: 2 years, 2 months ago Absolute Date: Thu 21 Dec 2023 00:16 Selected Answer: - Upvotes: 1

A modified version of this was in the exam today.

Comment 7

ID: 918332 User: chimex1 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Thu 08 Jun 2023 15:58 Selected Answer: - Upvotes: 2

Does the exam for AI 102 includes labs or its multiple choice only.

Comment 8

ID: 634005 User: Eltooth Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Wed 20 Jul 2022 13:20 Selected Answer: - Upvotes: 3

api-nam
translate

Comment 9

ID: 461526 User: gs23mi Badges: - Relative Date: 4 years, 5 months ago Absolute Date: Wed 13 Oct 2021 14:01 Selected Answer: - Upvotes: 4

in the "existing environment" there is
An Azure Cognitive Services resource named wwics
so it should be "resource name".cognitiveservices.azure.com : wwics.cognitiveservices.azure.com/translator/
Ref. https://docs.microsoft.com/en-us/azure/cognitive-services/translator/reference/v3-0-reference

Comment 9.1

ID: 465140 User: vominhtri854 Badges: - Relative Date: 4 years, 4 months ago Absolute Date: Wed 20 Oct 2021 14:06 Selected Answer: - Upvotes: 1

but this for Customers with a resource located in Switzerland North or Switzerland West

4. AI-102 Topic 3 Question 72

Sequence
10
Discussion ID
135048
Source URL
https://www.examtopics.com/discussions/microsoft/view/135048-exam-ai-102-topic-3-question-72-discussion/
Posted By
warrior1234
Posted At
March 2, 2024, 2:10 p.m.

Question

You have an Azure AI service model named Model1 that identifies the intent of text input.

You develop a Python app named App1.

You need to configure App1 to use Model1.

Which package should you add to App1?

  • A. azure-cognitiveservices-language-textanalytics
  • B. azure-ai-language-conversations
  • C. azure-mgmt-cognitiveservices
  • D. azure-cognitiveservices-speech

Suggested Answer

B

Answer Description Click to expand


Community Answer Votes

Comments 19 comments Click to expand

Comment 1

ID: 1164118 User: warrior1234 Badges: Highly Voted Relative Date: 2 years ago Absolute Date: Sat 02 Mar 2024 14:10 Selected Answer: - Upvotes: 16

The correct package for working with Azure AI service for text intent identification, like Model1, is:

A. azure-cognitiveservices-language-textanalytics

Therefore, you should add the azure-cognitiveservices-language-textanalytics package to App1 for configuring it to use Model1. This package provides functionalities for working with the Language API in Azure Cognitive Services, which includes text analytics capabilities, such as identifying language and sentiment, and can be used for processing text input to determine intent.

Comment 2

ID: 1176321 User: [Removed] Badges: Highly Voted Relative Date: 1 year, 11 months ago Absolute Date: Mon 18 Mar 2024 08:08 Selected Answer: - Upvotes: 5

answer is B.
"Conversation App: It's used in extracting intents and entities in conversations"
- https://learn.microsoft.com/ko-kr/python/api/overview/azure/ai-language-conversations-readme?view=azure-python

"The (Text Analytics) API can be used to analyze unstructured text for tasks such as sentiment analysis, key phrase extraction and language detection."
https://learn.microsoft.com/en-us/python/api/azure-cognitiveservices-language-textanalytics/azure.cognitiveservices.language.textanalytics.textanalyticsclient?view=azure-python-previous

Comment 2.1

ID: 1188990 User: anto69 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Thu 04 Apr 2024 02:20 Selected Answer: - Upvotes: 2

Agree 100%

Comment 3

ID: 1714669 User: phvogel Badges: Most Recent Relative Date: 3 weeks, 3 days ago Absolute Date: Sun 15 Feb 2026 18:57 Selected Answer: B Upvotes: 1

The textanalytics package for Python lists its capabilities, which don't include intent (https://learn.microsoft.com/en-us/python/api/overview/azure/ai-textanalytics-readme?view=azure-python); on the other hand, the converstationapp for python does reference intent (https://learn.microsoft.com/en-us/python/api/overview/azure/ai-language-conversations-readme?view=azure-python).

Comment 4

ID: 1333947 User: pabsinaz Badges: - Relative Date: 1 year, 2 months ago Absolute Date: Mon 30 Dec 2024 08:12 Selected Answer: B Upvotes: 4

Option A, azure-cognitiveservices-language-textanalytics, is focused on text analytics capabilities such as sentiment analysis, key phrase extraction, and language detection. While it provides valuable tools for text analysis, it doesn't specifically cater to intent recognition within conversational contexts, which is what you're looking for in Model1.

On the other hand, Option B, azure-ai-language-conversations, is designed specifically for handling conversational AI tasks, including intent recognition. This makes it the most appropriate package for configuring App1 to use Model1 for identifying the intent of text input.

Comment 5

ID: 1318017 User: payer4120 Badges: - Relative Date: 1 year, 3 months ago Absolute Date: Tue 26 Nov 2024 11:33 Selected Answer: A Upvotes: 2

A is correct

Comment 6

ID: 1283954 User: famco Badges: - Relative Date: 1 year, 5 months ago Absolute Date: Sun 15 Sep 2024 09:24 Selected Answer: - Upvotes: 1

What an e v i l question from Microsoft. It depends on the text.

Comment 6.1

ID: 1283955 User: famco Badges: - Relative Date: 1 year, 5 months ago Absolute Date: Sun 15 Sep 2024 09:29 Selected Answer: - Upvotes: 1

The answer they expect is conversational language understanding. But asking the name of the library instead of a service, just an additional step to trip people,
CLU is all about what you do with the intent. If you are going to map it to an action or decision it is CLU. But for example intent can be used in the context of opinion mining. But do not over think these questions are from Microsoft employees.

Comment 7

ID: 1265833 User: anto69 Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Wed 14 Aug 2024 17:08 Selected Answer: B Upvotes: 1

Should be B:
this package is specifically designed for conversational language understanding, which includes intent recognition.

Comment 8

ID: 1247896 User: krzkrzkra Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sun 14 Jul 2024 19:56 Selected Answer: B Upvotes: 1

Selected Answer: B

Comment 9

ID: 1235179 User: HaraTadahisa Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 08:18 Selected Answer: B Upvotes: 2

I say this answer is B.

Comment 10

ID: 1229624 User: mon2002 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 07:02 Selected Answer: - Upvotes: 1

azure-cognitiveservices-speech.

Comment 11

ID: 1229206 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 12 Jun 2024 16:08 Selected Answer: B Upvotes: 1

B is answer.

Comment 12

ID: 1220359 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Tue 28 May 2024 16:39 Selected Answer: B Upvotes: 1

B is right answer.

Comment 13

ID: 1215751 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 22 May 2024 15:23 Selected Answer: D Upvotes: 1

When identifying text input intent, it is usually aggregated to cognitiveservices. not just Microsoft, but mega cloud products and solutions have too many similar names and should be aggregated.

Comment 14

ID: 1191510 User: Murtuza Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Mon 08 Apr 2024 12:28 Selected Answer: B Upvotes: 1

B is correct

Comment 15

ID: 1184535 User: chandiochan Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Thu 28 Mar 2024 04:32 Selected Answer: B Upvotes: 1

Must be B

Comment 16

ID: 1170084 User: GHill1982 Badges: - Relative Date: 2 years ago Absolute Date: Sun 10 Mar 2024 07:47 Selected Answer: B Upvotes: 1

B. azure-ai-language-conversations
This package is part of the Azure AI SDK for Python and contains functionality for building applications that can understand user intent using language models.

Comment 17

ID: 1164348 User: Harry300 Badges: - Relative Date: 2 years ago Absolute Date: Sat 02 Mar 2024 20:51 Selected Answer: B Upvotes: 2

Should be B.
Source: https://pypi.org/project/azure-ai-language-conversations/
"Conversation App: It's used in extracting intents and entities in conversations"

5. AI-102 Topic 3 Question 74

Sequence
11
Discussion ID
135056
Source URL
https://www.examtopics.com/discussions/microsoft/view/135056-exam-ai-102-topic-3-question-74-discussion/
Posted By
Delta64
Posted At
March 2, 2024, 4:55 p.m.

Question

DRAG DROP -

You plan to implement an Azure AI Search resource that will use custom skill based on sentiment analysis.

You need to create a custom model and configure Azure AI Search use the model.

Which five actions should you perform in sequence? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.

image

Suggested Answer

image
Answer Description Click to expand


Comments 20 comments Click to expand

Comment 1

ID: 1231057 User: fba825b Badges: Highly Voted Relative Date: 1 year, 8 months ago Absolute Date: Sat 15 Jun 2024 18:24 Selected Answer: - Upvotes: 30

1. Create an Azure Machine Learning workspace.
2. Create and train the model in the Azure Machine Learning studio.
3. Create an endpoint for the model.
4. Connect the custom skill to the endpoint.
5. Rerun the indexer to enrich the index.

Comment 1.1

ID: 1332319 User: TimZ1010 Badges: - Relative Date: 1 year, 2 months ago Absolute Date: Fri 27 Dec 2024 09:54 Selected Answer: - Upvotes: 2

1- Provision an Azure AI Services resource and obtain the endpoint.
2- Connect the custom skill to the endpoint
3- Rerun the indexer to enrich the index.
4- Create an Azure Machine Learning workspace.
5- Create and train the model in the Azure Machine Learning studio.

Comment 2

ID: 1181723 User: 3a0b61c Badges: Highly Voted Relative Date: 1 year, 11 months ago Absolute Date: Sun 24 Mar 2024 15:50 Selected Answer: - Upvotes: 8

https://learn.microsoft.com/en-us/training/modules/build-azure-machine-learn-custom-skill-for-azure-cognitive-search/03-enrich-search-index-use-model
1.Create an Azure Machine Learning workspace.
2.Create and train the model in the Azure Machine Learning studio.
3.Create an endpoint for the model.
4.Provision an Azure AI Services resource and obtain the endpoint.
5.Connect the custom skill to the endpoint.

Comment 2.1

ID: 1206310 User: MonicaKarim Badges: - Relative Date: 1 year, 10 months ago Absolute Date: Sat 04 May 2024 02:18 Selected Answer: - Upvotes: 5

in the link you provided, last step is to update your existing cognitive search that means is to create the search resource first:
1- Provision an Azure AI Services resource and obtain the endpoint.
2- Create an Azure Machine Learning workspace.
3- Create and train the model in the Azure Machine Learning studio.
4- Create an endpoint for the model.
6- Connect the custom skill to the endpoint.

Comment 2.1.1

ID: 1286351 User: mrg998 Badges: - Relative Date: 1 year, 5 months ago Absolute Date: Thu 19 Sep 2024 15:36 Selected Answer: - Upvotes: 5

that link literally says "rerun your indexer to enrich your index with the AML model." as the last step. The correct sequence from that link is:
1. Create an Azure Machine Learning workspace.
2. Create and train the model in the Azure Machine Learning studio.
3. Create an endpoint for the model.
4. Connect the custom skill to the endpoint.
5. Rerun the indexer to enrich the index

Comment 3

ID: 1714674 User: phvogel Badges: Most Recent Relative Date: 3 weeks, 3 days ago Absolute Date: Sun 15 Feb 2026 19:12 Selected Answer: - Upvotes: 1

According to the exercise at https://microsoftlearning.github.io/mslearn-knowledge-mining/Instructions/Exercises/02-search-skills.html#create-azure-resources, you have to both create the Search resource and run the indexer to incorporate the skill -- the two highest voted answers at this point are split on which step belongs in the answer. I suspect that the question author is assuming the search is already created and that you are adding the custom skill to it. That means the expected answer is:

Create an Azure Machine Learning workspace.
2. Create and train the model in the Azure Machine Learning studio.
3. Create an endpoint for the model.
4. Connect the custom skill to the endpoint.
5. Rerun the indexer to enrich the index.

Comment 4

ID: 1363017 User: gyaansastra Badges: - Relative Date: 1 year ago Absolute Date: Fri 28 Feb 2025 14:14 Selected Answer: - Upvotes: 2

To implement an Azure AI Search resource that uses a custom skill based on sentiment analysis, follow these actions in sequence:

Create an Azure Machine Learning workspace: Set up the environment where you can create and manage your machine learning experiments and models.

Create and train the model in the Azure Machine Learning studio: Develop and train your custom sentiment analysis model using the resources in your Azure Machine Learning workspace.

Create an endpoint for the model: Deploy your trained model and create an endpoint to access the model via REST API.

Connect the custom skill to the endpoint: Configure Azure AI Search to use your custom skill by connecting it to the model's endpoint.

Rerun the indexer to enrich the index: Update the search index by rerunning the indexer, which will now use the custom skill to analyze sentiment and enrich the data in the search index.

Comment 5

ID: 1266151 User: anto69 Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Thu 15 Aug 2024 04:52 Selected Answer: - Upvotes: 1

ChatGPT:
1- Provision an Azure AI Services resource and obtain the endpoint.
2- Create an Azure Machine Learning workspace.
3- Create and train the model in the Azure Machine Learning studio.
4- Create an endpoint for the model.
6- Connect the custom skill to the endpoint

Comment 5.1

ID: 1390360 User: Mattt Badges: - Relative Date: 12 months ago Absolute Date: Thu 13 Mar 2025 17:06 Selected Answer: - Upvotes: 1

ChatGPT 4.5

1. Create an Azure Machine Learning workspace.
2. Create and train the model in the Azure Machine Learning studio.
3. Create an endpoint for the model.
4. Connect the custom skill to the endpoint.
5. Rerun the indexer to enrich the index.

Comment 6

ID: 1248520 User: krzkrzkra Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Mon 15 Jul 2024 20:18 Selected Answer: - Upvotes: 1

1. Create an Azure Machine Learning workspace.
2. Create and train the model in the Azure Machine Learning studio.
3. Create an endpoint for the model.
4. Connect the custom skill to the endpoint.
5. Rerun the indexer to enrich the index.

Comment 7

ID: 1235782 User: PeteColag Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sun 23 Jun 2024 12:33 Selected Answer: - Upvotes: 4

Based on the detailed steps provided in the Microsoft Learn documentation, the sequence of actions indeed doesn't require provisioning a separate Azure AI Services resource. Instead, the custom skill is directly connected to the Azure Machine Learning model endpoint.

https://learn.microsoft.com/en-us/training/modules/build-azure-machine-learn-custom-skill-for-azure-cognitive-search/03-enrich-search-index-use-model

Comment 8

ID: 1235357 User: reiwanotora Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 13:25 Selected Answer: - Upvotes: 1

1. Create an Azure Machine Learning workspace.
2. Create and train the model in the Azure Machine Learning studio.
3. Create an endpoint for the model.
4. Connect the custom skill to the endpoint.
5. Rerun the indexer to enrich the index.

Comment 9

ID: 1229625 User: mon2002 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 07:06 Selected Answer: - Upvotes: 2

Create an Azure Machine Learning workspace.
Create and train the model in the Azure Machine Learning studio.
Create an endpoint for the model.
Provision an Azure AI Services resource and obtain the endpoint.
Connect the custom skill the endpoint.

Comment 10

ID: 1222683 User: monniq Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sat 01 Jun 2024 16:07 Selected Answer: - Upvotes: 3

Create an Azure Machine Learning workspace.
Create and train the model in the Azure Machine Learning studio.
Create an endpoint for the model.
Connect the custom skill to the endpoint.
Rerun the indexer to enrich the index.

Comment 11

ID: 1215749 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 22 May 2024 15:20 Selected Answer: - Upvotes: 3

Create workspace
Provision
Connect
Create studio
Rerun

Comment 12

ID: 1178668 User: Murtuza Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Wed 20 Mar 2024 21:17 Selected Answer: - Upvotes: 1

Create an Azure Machine Learning workspace.
Create and train the model in the Azure Machine Learning studio.
Create an endpoint for the model.
Provision an Azure AI Services resource and obtain the endpoint.
Connect the custom skill to the endpoint.

Comment 13

ID: 1177428 User: rober13 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Tue 19 Mar 2024 15:34 Selected Answer: - Upvotes: 1

Create an endpoint for the model.
Create and train the model in the Azure Machine Learning studio.
Create an Azure Machine Learning workspace.
Connect the custom skill to the endpoint.

so, with the options:
Create an Azure Machine Learning
Create and train
Provision an Azure
Create an endpoint
Connect

Comment 14

ID: 1174030 User: chandiochan Badges: - Relative Date: 1 year, 12 months ago Absolute Date: Fri 15 Mar 2024 05:05 Selected Answer: - Upvotes: 2

Create an Azure Machine Learning workspace.
Create and train the model in the Azure Machine Learning studio.
Create an endpoint for the model.
Provision an Azure AI Services resource and obtain the endpoint.
Connect the custom skill to the endpoint.

Comment 15

ID: 1164195 User: Delta64 Badges: - Relative Date: 2 years ago Absolute Date: Sat 02 Mar 2024 16:55 Selected Answer: - Upvotes: 1

According to GPT4:

1. Create an Azure Machine Learning workspace.
2. Create and train the model in the Azure Machine Learning studio.
3. Provision an Azure AI Services resource and obtain the endpoint.
4. Create an endpoint for the model.
Connect the custom skill to the endpoint.
Rerun the indexer to enrich the index.

Comment 15.1

ID: 1164197 User: Delta64 Badges: - Relative Date: 2 years ago Absolute Date: Sat 02 Mar 2024 16:59 Selected Answer: - Upvotes: 1

My bad, this is wrong information.

6. AI-102 Topic 3 Question 60

Sequence
13
Discussion ID
134937
Source URL
https://www.examtopics.com/discussions/microsoft/view/134937-exam-ai-102-topic-3-question-60-discussion/
Posted By
audlindr
Posted At
Feb. 29, 2024, 7:57 p.m.

Question

HOTSPOT
-

You are building a chatbot.

You need to use the Content Moderator API to identify aggressive and sexually explicit language.

Which three settings should you configure? To answer, select the appropriate settings in the answer area.

NOTE: Each correct selection is worth one point.

image

Suggested Answer

image
Answer Description Click to expand


Comments 9 comments Click to expand

Comment 1

ID: 1162938 User: audlindr Badges: Highly Voted Relative Date: 2 years ago Absolute Date: Thu 29 Feb 2024 19:57 Selected Answer: - Upvotes: 26

The answers should be
Resource Name
classify
ocp-Apim-Subscription-Key

autocorrect has nothing to do with the question

Comment 1.1

ID: 1162939 User: audlindr Badges: - Relative Date: 2 years ago Absolute Date: Thu 29 Feb 2024 19:57 Selected Answer: - Upvotes: 3

https://westus.dev.cognitive.microsoft.com/docs/services/57cf753a3f9b070c105bd2c1/operations/57cf753a3f9b070868a1f66f

Comment 1.1.1

ID: 1190705 User: Training Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Sun 07 Apr 2024 03:30 Selected Answer: - Upvotes: 3

https://learn.microsoft.com/en-us/training/modules/classify-and-moderate-text-with-azure-content-moderator/4-exercise-use-the-api-console

Comment 2

ID: 1714659 User: phvogel Badges: Most Recent Relative Date: 3 weeks, 3 days ago Absolute Date: Sun 15 Feb 2026 18:31 Selected Answer: - Upvotes: 1

Except, of course, you need to set the Resource name and subscription key even if you're not doing moderation so it's really only classify that has to be set to enable moderation.

Comment 3

ID: 1356216 User: sukantadey Badges: - Relative Date: 1 year ago Absolute Date: Thu 13 Feb 2025 18:08 Selected Answer: - Upvotes: 1

Resource Name
classify
ocp-Apim-Subscription-Key

Comment 4

ID: 1247892 User: krzkrzkra Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sun 14 Jul 2024 19:53 Selected Answer: - Upvotes: 4

NagaoShingo 1 month, 1 week ago
1. Resource name
2. classify
3. Ocp-Api-Suscription-Key

Comment 5

ID: 1231293 User: PeteColag Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sun 16 Jun 2024 11:58 Selected Answer: - Upvotes: 3

Resource Name: Provide the name of your Content Moderator resource.
Classify: Set to true to enable classification of content.
Ocp-Apim-Subscription-Key: Provide your subscription key for authentication.

All other fields are not necessary to set.

Comment 6

ID: 1225496 User: NagaoShingo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 06 Jun 2024 14:42 Selected Answer: - Upvotes: 3

1. Resource name
2. classify
3. Ocp-Api-Suscription-Key

Comment 7

ID: 1215770 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 22 May 2024 15:39 Selected Answer: - Upvotes: 2

Resource name
classify
Ocp-Api-Suscription-Key

7. AI-102 Topic 15 Question 3

Sequence
14
Discussion ID
77001
Source URL
https://www.examtopics.com/discussions/microsoft/view/77001-exam-ai-102-topic-15-question-3-discussion/
Posted By
sdokmak
Posted At
June 22, 2022, 3:59 a.m.

Question

HOTSPOT -
You are developing the shopping on-the-go project.
You need to build the Adaptive Card for the chatbot.
How should you complete the code? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Hot Area:
image

Suggested Answer

image
Answer Description Click to expand

Box 1: name.en -
Box 2: "$when": "${stockLevel != 'OK'}"
Product displays must include images and warnings when stock levels are low or out of stock.
Box 3:image.altText.en

Comments 18 comments Click to expand

Comment 1

ID: 738481 User: halfway Badges: Highly Voted Relative Date: 3 years, 3 months ago Absolute Date: Thu 08 Dec 2022 01:25 Selected Answer: - Upvotes: 52

1. name[language] 2. "$when": "${stockLevel} != 'OK'" 3. image.altText[language]

Comment 1.1

ID: 1241519 User: MarceloManhaes Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Wed 03 Jul 2024 18:14 Selected Answer: - Upvotes: 1

I ´m sure if is correct if you see the shopping on the go project requirements for chatbot that says the multi-language approach

Comment 1.2

ID: 1319213 User: nastolgia Badges: - Relative Date: 1 year, 3 months ago Absolute Date: Thu 28 Nov 2024 13:12 Selected Answer: - Upvotes: 4

im agree with you.
1) name[language] - in task we app should support 3 languages
2) $when": "${stockLevel} != 'OK' - If item is not in the stock show
3) image.altText[language] - Show alt.text in different languages

Comment 2

ID: 980874 User: james2033 Badges: Highly Voted Relative Date: 2 years, 6 months ago Absolute Date: Mon 14 Aug 2023 16:29 Selected Answer: - Upvotes: 8

File JSON

{
"sku": "b1",
"name": {
"en": "Bicycle",
"es": "Bicicleta",
"pt": "Bicicleta"
},
"stocklevel": "Out of Stock",
"description": {
"en": "Bicycle",
"es": "Bicicleta",
"pt": "Bicicleta"
},
"image": {
"uri": "https://upload.wordwideimporters.org/bicycle.jpg",
"alttext": {
"en": "Bicycle",
"es": "Bicicleta",
"pt": "Bicicleta"
}
},
"createdUtc": "2020-0214T06:08:39Z",
"language": "en"
}

1. name.en

2. "$when": "${stockLevel != 'OK'}"

3. image.altText.en

Comment 2.1

ID: 1205721 User: upliftinghut Badges: - Relative Date: 1 year, 10 months ago Absolute Date: Thu 02 May 2024 18:44 Selected Answer: - Upvotes: 1

This is correct because the structure of original requirement Json is reflected and replaced by $<parameter name>

Comment 3

ID: 1283995 User: mrg998 Badges: Most Recent Relative Date: 1 year, 5 months ago Absolute Date: Sun 15 Sep 2024 10:42 Selected Answer: - Upvotes: 1

Let's evaluate why the other options for answer3 are not valid:
image.altText.en:
This option hardcodes the language to English (en).
image.altText.language:
This option assumes language is a property under altText. However, language is a variable representing the selected language (e.g., en, es, pt), not a property within altText.
image.altText.[“language”]:
This option incorrectly uses quotation marks around language, treating it as a literal string rather than a variable. This would attempt to look for an alt text property with the name "language", which does not exist.
Correct Option: image.altText.[language]
Explanation: This option correctly uses the language variable to dynamically access the appropriate alt text based on the user's selected language. It ensures that the alt text is in the correct language (English, Spanish, or Portuguese) by referencing the variable directly. This fulfills the requirement for a multilingual customer experience.

Comment 4

ID: 1246792 User: krzkrzkra Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Fri 12 Jul 2024 15:43 Selected Answer: - Upvotes: 1

1. name.en

2. "$when": "${stockLevel != 'OK'}"

3. image.altText.en

Comment 5

ID: 1245045 User: Toby86 Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Tue 09 Jul 2024 20:28 Selected Answer: - Upvotes: 2

Requirement: "Support interactions in English, Spanish, and Portuguese"

If you hard set it on english, how can you satisfy this requirement?
must be name[language] on top and AltText.[language] on bottom.

Comment 6

ID: 1231498 User: fba825b Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sun 16 Jun 2024 20:37 Selected Answer: - Upvotes: 2

How I understood it:
The description in the first table says that the app and the user input is in english => app is in english => adaptive card should be in english
In addition to that, the provided JSON has a language property with value 'en'.
So, my guess is:
1. name.en
2. "$when": "${stockLevel != 'OK'}"
3. image.altText.en

Comment 6.1

ID: 1713907 User: OnlineCH Badges: - Relative Date: 3 weeks, 6 days ago Absolute Date: Thu 12 Feb 2026 20:55 Selected Answer: - Upvotes: 1

The question is "You are developing the shopping on-the-go project." and in "Shopping On-the-Go Requirements" it says "Support interactions in English, Spanish, and Portuguese.". The table you're referencing to is about "Product management" not "shopping on-the-go".

So correct answers should be:
1. name[language]
2. "$when": "${stockLevel != 'OK'}"
3. image.altText[language]

Comment 7

ID: 1230535 User: NagaoShingo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 14 Jun 2024 16:39 Selected Answer: - Upvotes: 1

1. name.en
2. "$when": "${stockLevel != 'OK'}"
3. image.altText.en

Comment 8

ID: 1214431 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Mon 20 May 2024 16:34 Selected Answer: - Upvotes: 1

Is this question still available on May 21, 2024?

Comment 9

ID: 1014599 User: sl_mslconsulting Badges: - Relative Date: 2 years, 5 months ago Absolute Date: Sat 23 Sep 2023 04:40 Selected Answer: - Upvotes: 6

You don’t want code it in a way that the code just works with en - it’s a bad coding practice.

Comment 10

ID: 828470 User: AzureJobsTillRetire Badges: - Relative Date: 3 years ago Absolute Date: Sat 04 Mar 2023 00:49 Selected Answer: - Upvotes: 5

The given answers are correct.

The correct answer pairs for box1 and box3 could be-
name.en and image.altText.en
name["language"] and image.altText["language"]

since name["language"] does not exist as an option in box3, we will have to choose the first pair as the correct answer.

https://www.w3schools.com/js/js_json_syntax.asp

Comment 11

ID: 728859 User: Canyu Badges: - Relative Date: 3 years, 3 months ago Absolute Date: Mon 28 Nov 2022 08:54 Selected Answer: - Upvotes: 3

This is a repeat question, but this time the answer does not give an explanation or reference document. Is this answer correct?

Comment 12

ID: 638131 User: ParkXD Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Wed 27 Jul 2022 16:04 Selected Answer: - Upvotes: 4

it seems a repeat question, but with a slight different option. The earlier one with altText.[language], but this altText[language] without .
which one is right?

Comment 13

ID: 620095 User: sdokmak Badges: - Relative Date: 3 years, 8 months ago Absolute Date: Wed 22 Jun 2022 03:59 Selected Answer: - Upvotes: 2

If there was an option for name.[language] instead of name[language] then I'd use that for more than one language option, otherwise we can't. For name[language], language would have to be an integer but it's a string, wouldn't work.
But.. for name.[language], that works.
https://stackoverflow.com/questions/882727/is-there-a-way-to-use-variable-keys-in-a-javascript-object-literal

Comment 13.1

ID: 635823 User: RamonKaus Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Sun 24 Jul 2022 02:44 Selected Answer: - Upvotes: 10

This is a repeat of an earlier question but they changed the answer? Weird.

8. AI-102 Topic 3 Question 50

Sequence
15
Discussion ID
112140
Source URL
https://www.examtopics.com/discussions/microsoft/view/112140-exam-ai-102-topic-3-question-50-discussion/
Posted By
973b658
Posted At
June 14, 2023, 9:12 a.m.

Question

DRAG DROP
-

You have a Docker host named Host1 that contains a container base image.

You have an Azure subscription that contains a custom speech-to-text model named model1.

You need to run model1 on Host1.

Which three actions should you perform in sequence? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.

image

Suggested Answer

image
Answer Description Click to expand


Comments 13 comments Click to expand

Comment 1

ID: 922836 User: 973b658 Badges: Highly Voted Relative Date: 2 years, 9 months ago Absolute Date: Wed 14 Jun 2023 09:12 Selected Answer: - Upvotes: 9

No.
1.Export model1 to Host1.
2.Request approval to run the container.
3.Run the container.

Comment 2

ID: 1712278 User: phvogel Badges: Most Recent Relative Date: 1 month ago Absolute Date: Fri 06 Feb 2026 22:08 Selected Answer: - Upvotes: 1

The issue here, for me, is the order which is indeterminate. You do need to export the model and you do need to run the container (you do not need to retrain the model unless you change the domain). However, you can ask for permission to run the container either before or after exporting the model (though it makes sense to do that just before running the container). So, presumably Request/Export/Run and Export/Run/Request are both correct answers.

Comment 3

ID: 1578852 User: StelSen Badges: - Relative Date: 8 months, 3 weeks ago Absolute Date: Thu 19 Jun 2025 11:24 Selected Answer: - Upvotes: 2

Export model1 to Host1
→ You need to export the custom model from Azure to your Docker host so it’s available locally.

Request approval to run the container
→ Azure Speech container deployment typically requires Microsoft approval (e.g., for speech containers using Custom Speech models).

Run the container
→ Once the model is exported and approval is granted, you can start the Docker container that runs the speech-to-text service

Comment 4

ID: 1356211 User: sukantadey Badges: - Relative Date: 1 year ago Absolute Date: Thu 13 Feb 2025 18:02 Selected Answer: - Upvotes: 2

Request approval to run container
Export model1 to Host1
Run the container

Comment 5

ID: 1281217 User: famco Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Mon 09 Sep 2024 21:25 Selected Answer: - Upvotes: 3

Maybe the question creator was in a hurry (they always are) and saw only the documentation that said there needs to be approval (but it was for disconnected and even though this is not disconnected). Or maybe the question from ET missed disconnected keyword.

Comment 6

ID: 1248508 User: krzkrzkra Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Mon 15 Jul 2024 20:11 Selected Answer: - Upvotes: 2

1. Request approval to run container
2. Export model1 to Host1
3. Run the container

Comment 7

ID: 1230635 User: etellez Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 14 Jun 2024 21:39 Selected Answer: - Upvotes: 4

this is the answer from Copilot:

To run the custom speech-to-text model on your Docker host, you should perform the following actions in sequence:

Export model to Host1: You need to export the model from Azure to your Docker host. This typically involves downloading the model and moving it to the Docker host.

Run the container: Once the model is on the Docker host, you can run the container that uses this model.

Configure disk logging: After the container is running, configure disk logging to keep track of the container's activities and to troubleshoot any issues that might arise.

Comment 8

ID: 1215787 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 22 May 2024 16:02 Selected Answer: - Upvotes: 1

Request
Export
Run

Comment 9

ID: 1043741 User: sl_mslconsulting Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Sun 15 Oct 2023 00:38 Selected Answer: - Upvotes: 3

You only need to request approval if you plan to run the container in a completely disconnected environment. And you might not get approved at all as there are some requirements. The question does not indicate if we need to run the container in such a environment.

Comment 9.1

ID: 1043742 User: sl_mslconsulting Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Sun 15 Oct 2023 00:46 Selected Answer: - Upvotes: 2

Based on what you can choose here I would say it’s a disconnected environment. For a connected environment using docker run allows you to download the model at the same so there is no need to export the model manually and then copy it to the host. Zellck is right in the choices.

Comment 10

ID: 996888 User: M25 Badges: - Relative Date: 2 years, 6 months ago Absolute Date: Sat 02 Sep 2023 18:19 Selected Answer: - Upvotes: 1

https://learn.microsoft.com/en-us/azure/ai-services/speech-service/speech-container-cstt?tabs=custom-model%2Ccontainer&pivots=programming-language-csharp
In this article, you'll learn how to download [from Microsoft Container Registry (MCR)], install, and run a Custom speech to text container.
Get the model ID (to use as the argument to the ModelId parameter of the docker run command): The custom model has to have been trained by using the Speech Studio.
N/a so far, excluding A (Retrain the model).

Comment 10.1

ID: 996889 User: M25 Badges: - Relative Date: 2 years, 6 months ago Absolute Date: Sat 02 Sep 2023 14:41 Selected Answer: - Upvotes: 2

Run the container with docker run:
B (Request approval precedes),
The docker run command will start the container when all three of the following options are provided with valid values: ApiKey, Billing, EULA
https://learn.microsoft.com/en-us/azure/ai-services/speech-service/speech-container-howto#billing-arguments

D (Run container), C (Get model)
Here's an example docker run command with placeholder values. … This command:
• Runs a custom speech to text container from the container image.
• Allocates 4 CPU cores and 8 GB of memory.
• Loads the custom speech to text model from the volume input mount, for example, C:\CustomSpeech.
• Exposes TCP port 5000 and allocates a pseudo-TTY for the container.
• Downloads the model given the ModelId (if not found on the volume mount).

Comment 11

ID: 934027 User: Tin_Tin Badges: - Relative Date: 2 years, 8 months ago Absolute Date: Mon 26 Jun 2023 06:16 Selected Answer: - Upvotes: 1

Not sure.
https://learn.microsoft.com/en-us/azure/cognitive-services/speech-service/speech-container-overview#request-approval-to-run-the-container

9. AI-102 Topic 3 Question 55

Sequence
16
Discussion ID
123688
Source URL
https://www.examtopics.com/discussions/microsoft/view/123688-exam-ai-102-topic-3-question-55-discussion/
Posted By
sl_mslconsulting
Posted At
Oct. 15, 2023, 7:07 p.m.

Question

DRAG DROP
-

You have a collection of Microsoft Word documents and PowerPoint presentations in German.

You need to create a solution to translate the files to French. The solution must meet the following requirements:

• Preserve the original formatting of the files.
• Support the use of a custom glossary.

You create a blob container for German files and a blob container for French files. You upload the original files to the container for German files.

Which three actions should you perform in sequence to complete the solution? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.

image

Suggested Answer

image
Answer Description Click to expand


Comments 21 comments Click to expand

Comment 1

ID: 1712279 User: phvogel Badges: - Relative Date: 1 month ago Absolute Date: Fri 06 Feb 2026 22:27 Selected Answer: - Upvotes: 1

The only reasons I can think of for picking where the glossary file goes
a) The app has read permission for the German resource and write permission for the French resouce, so the glossary has to go in the German resource
b) The resource is being used as a proxy for the language the glossary targets. If so, this would need to be a German-to-French glossary -- see https://learn.microsoft.com/en-us/azure/ai-services/translator/document-translation/how-to-guides/create-use-glossaries -- and would belong with the German files.

Comment 2

ID: 1388645 User: Mattt Badges: - Relative Date: 12 months ago Absolute Date: Thu 13 Mar 2025 14:53 Selected Answer: - Upvotes: 1

The given answer is correct.

Comment 3

ID: 1261900 User: moonlightc Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Wed 07 Aug 2024 03:36 Selected Answer: - Upvotes: 2

1. Upload a Glossary file to the french files container
2. Define a document translation specification that has french target
3. Perform asynchronous translation by using the document specification

This is the answer according to ChatGPT.
Why not Glossary file in German container? Glossary contains the specific terms for the TARGET language. French in this case. So it make sense to keep it in French file
container

Comment 3.1

ID: 1336153 User: Will_Aguilar Badges: - Relative Date: 1 year, 2 months ago Absolute Date: Fri 03 Jan 2025 22:07 Selected Answer: - Upvotes: 2

Maybe I misunderstood. It just seems that we are translating from German, specifically German presentations, so the words to interpret would be in German including any glossary terms. The words read/consumed would be in the target language. I can't imagine a use case for having a French glossary.

Comment 4

ID: 1239648 User: MarceloManhaes Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sun 30 Jun 2024 15:04 Selected Answer: - Upvotes: 4

I think that correct is the original response, Because the glossary file is to be in the source language. On the URL : https://learn.microsoft.com/en-us/azure/ai-services/translator/document-translation/how-to-guides/create-use-glossaries gives an example of english to french and in session
Create, upload, and use a glossary file says : ... "The following English-source glossary contains words that can have different meanings depending upon the context in which they're used". So the glossary needs to be in the source language and not in the target language. So the answer is
1. Upload a glossary file to the container for German files
2. Define a document translation specification with French as the target language
3. Perform an asynchronous translation by using the document translation specification
Ps.: I tested on chatGPT and the response goes to the same.

Comment 4.1

ID: 1281365 User: famco Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Tue 10 Sep 2024 07:39 Selected Answer: - Upvotes: 1

The glossary file is in both languages, please !!!
The question is which container to put it in. It can be in any container and probably in a different container than source or target. If possible contest the question itself if you see this same question

Comment 5

ID: 1235142 User: LM12 Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 07:36 Selected Answer: - Upvotes: 2

was on exam 20.06.24

Comment 6

ID: 1229734 User: PeteColag Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 11:39 Selected Answer: - Upvotes: 1

curl -X POST "https://<your-region>.cognitiveservices.azure.com/translator/text/batch/v1.1/translator/document/batches?api-version=2024-05-01" \
-H "Ocp-Apim-Subscription-Key: <your-subscription-key>" \
-H "Content-Type: application/json" \
-d <insert JSON document here to specific source, target and glossary>

Comment 7

ID: 1229733 User: PeteColag Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 11:37 Selected Answer: - Upvotes: 2

{
"inputs": [
{
"source": {
"sourceUrl": "https://<your-storage-account>.blob.core.windows.net/german-files",
"storageSource": "AzureBlob"
},
"targets": [
{
"targetUrl": "https://<your-storage-account>.blob.core.windows.net/french-files",
"storageSource": "AzureBlob",
"language": "fr",
"glossaries": [
{
"glossaryUrl": "https://<your-storage-account>.blob.core.windows.net/glossaries/de-fr-glossary.tsv",
"format": "TSV"
}
]
}
]
}
]
}

Comment 7.1

ID: 1281366 User: famco Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Tue 10 Sep 2024 07:40 Selected Answer: - Upvotes: 1

Look, their own example put it in a different container

Comment 8

ID: 1216016 User: vovap0vovap Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 22 May 2024 22:46 Selected Answer: - Upvotes: 1

Well, reality is that in MS example glossary in own container. But it is also in "targets" section of json file, so answer "French container" probably safer.

Comment 8.1

ID: 1281368 User: famco Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Tue 10 Sep 2024 07:41 Selected Answer: - Upvotes: 1

wow, people working with are very creative finding some logic in this illogical Microsoft question

Comment 9

ID: 1215774 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 22 May 2024 15:42 Selected Answer: - Upvotes: 2

1. Upload a Glossary file to the french files container
2. Define a document translation specification that has french target
3. Perform asynchronous translation by using the document specification

Comment 10

ID: 1139693 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sun 04 Feb 2024 01:48 Selected Answer: - Upvotes: 3

The given answer is correct : reason is below
Upload the terminology list: Begin by uploading a custom terminology list, essential for accurate translations. Ensure it's placed in a location recognized by the translation service.

Define translation specifications: Specify source (German), target (French), and any special parameters in a translation specification.

Execute asynchronous translation: Initiate the translation process based on the specification, which will preserve original formatting and save translated files in the French document container.

These three steps form the core of the solution, although practical implementation may require additional configuration and permissions. Ensure the terminology list is uploaded to a location compatible with the translation service's requirements.

Comment 11

ID: 1116648 User: dimsok Badges: - Relative Date: 2 years, 2 months ago Absolute Date: Mon 08 Jan 2024 14:47 Selected Answer: - Upvotes: 1

You need a French Glossary in the French container - how else can the translator know the French Jargon

Comment 11.1

ID: 1281370 User: famco Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Tue 10 Sep 2024 07:42 Selected Answer: - Upvotes: 1

It can be in any container. Why do you need it in any specific container?

Comment 12

ID: 1044327 User: sl_mslconsulting Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Sun 15 Oct 2023 19:07 Selected Answer: - Upvotes: 2

The glossary file can be in this own container and that’s tricky as in theory you can put it anywhere. But since we need probably need have different permissions for the source and the target container putting it in the container where the source files would do. Also pay attention to what a glossary file its content looks like and how to specify the location in a request sent to the endpoint. https://learn.microsoft.com/en-us/azure/ai-services/translator/document-translation/how-to-guides/create-use-glossaries

Comment 12.1

ID: 1044334 User: sl_mslconsulting Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Sun 15 Oct 2023 19:11 Selected Answer: - Upvotes: 1

Although I can’t find the term translation specification file in the doc, but you don’t need to create a list of files that’s for sure (pay attention to how the target and source are specified in the link I provided above.

Comment 12.2

ID: 1044414 User: sl_mslconsulting Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Sun 15 Oct 2023 21:49 Selected Answer: - Upvotes: 4

I changed my mind the glossary file should be in the target folder. Reason for that is because you can have multiple target languages each with its own glossary files. Remember the glossary provides the expected translation for each word in the file to help ensure accuracy. In other words, you are providing the translation and should have done so for each target language.

Comment 12.2.1

ID: 1281360 User: famco Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Tue 10 Sep 2024 07:33 Selected Answer: - Upvotes: 1

You can change your mind every minute but there is no right answer to this. These questions exist because the question creator is not a subject matter expert in translation or AI, but his expertise is in making-questions. He did not understand what is a glossary and understood something weird from what he read and asking us to guess what he thought glossary is. If I know his background I can guess, but without that it is impossible.
If questions like this come, raise objections in the exam itself so that you can get points if someone looks at it. And just hope it is someone who knows the subject.

Comment 12.2.1.1

ID: 1281361 User: famco Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Tue 10 Sep 2024 07:37 Selected Answer: - Upvotes: 1

I will not put the glossary either in the source or target but in another container. Why miss the content with the rules? But the people working with Microsoft are hello-world-programmers

10. AI-102 Topic 1 Question 54

Sequence
17
Discussion ID
112133
Source URL
https://www.examtopics.com/discussions/microsoft/view/112133-exam-ai-102-topic-1-question-54-discussion/
Posted By
973b658
Posted At
June 14, 2023, 8:37 a.m.

Question

DRAG DROP -

You are building an app that will scan confidential documents and use the Language service to analyze the contents.

You provision an Azure Cognitive Services resource.

You need to ensure that the app can make requests to the Language service endpoint. The solution must ensure that confidential documents remain on-premises.

Which three actions should you perform in sequence? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.

image

Suggested Answer

image
Answer Description Click to expand


Comments 21 comments Click to expand

Comment 1

ID: 1118448 User: dimsok Badges: Highly Voted Relative Date: 2 years, 2 months ago Absolute Date: Wed 10 Jan 2024 11:21 Selected Answer: - Upvotes: 15

It needs to have internet connectivity to send billing information, so you cannot select an isolated K8s service.
1. K8s with internet
2. Image from MCR
3. AppID and Secret

Comment 2

ID: 1212529 User: JamesKJoker Badges: Highly Voted Relative Date: 1 year, 9 months ago Absolute Date: Thu 16 May 2024 19:38 Selected Answer: - Upvotes: 6

1. Provision on-premise k8 cluster that is connected to Internet
2. Pull image from MCR
3. Run container and specify API key and endpoint URL of Cognitive Services Services

Internet Connectivity is required for billing

Comment 3

ID: 1712080 User: phvogel Badges: Most Recent Relative Date: 1 month ago Absolute Date: Thu 05 Feb 2026 20:20 Selected Answer: - Upvotes: 1

I don't think that internet connectivity is required either to send billing data or to access the service in Azure: You can do both over the VNet connection that your enterprise will have created to Azure. To quote Microsoft: "Docker must be configured to allow the containers to connect with and send billing data to Azure. ailure to allowlist various network channels that the Azure AI containers rely on will prevent the containe
r from working" which doesn't mean internet connectivity.

Comment 4

ID: 1306806 User: Christian_garcia_martin Badges: - Relative Date: 1 year, 4 months ago Absolute Date: Mon 04 Nov 2024 07:33 Selected Answer: - Upvotes: 1

first of all you need to provision cluster , after in a pod or deployment in containers section tag image specify the image to pull from MCR and Run container and specify API key and endpoint URL of Cognitive Services Services , cluster needs internet access to can reach that endpoint.

Comment 5

ID: 1297282 User: Christian_garcia_martin Badges: - Relative Date: 1 year, 4 months ago Absolute Date: Mon 14 Oct 2024 09:21 Selected Answer: - Upvotes: 1

k8s with internet : This is necessary because you still need to reach Azure Cognitive Services, which is hosted on Azure and thus requires internet connectivity.

Comment 6

ID: 1235736 User: rookiee1111 Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sun 23 Jun 2024 10:18 Selected Answer: - Upvotes: 6

1. provision K8 with internet
2. pull from MCR
3. Run container and specify API key and endpoint URL of Cognitive Services Services,

Comment 6.1

ID: 1272068 User: JakeCallham Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Sun 25 Aug 2024 11:03 Selected Answer: - Upvotes: 1

You need to not use cognitive services, that why you deploy on prem container

Comment 6.1.1

ID: 1296945 User: jeffangel28 Badges: - Relative Date: 1 year, 4 months ago Absolute Date: Sun 13 Oct 2024 16:12 Selected Answer: - Upvotes: 1

is just for billing purposes, the documents remain in OnPrem

Comment 7

ID: 1217602 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 16:42 Selected Answer: - Upvotes: 3

1. Provision on-premise k8 cluster that is isolated from Internet
2. Pull image from MCR
3. Run container and specify API key and endpoint URL of Cognitive Services Services

Comment 8

ID: 1187824 User: NullVoider_0 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Tue 02 Apr 2024 05:44 Selected Answer: - Upvotes: 1

1. Provision an on-premises Kubernetes cluster that is isolated from the internet.
2. Pull an image from the Microsoft Container Registry (MCR).
3. Run the container and specify an API key and the Endpoint URL of the Cognitive Services resource.

Comment 9

ID: 1182352 User: varinder82 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Mon 25 Mar 2024 10:24 Selected Answer: - Upvotes: 1

Final Answer:
1. Provision on-premise k8 cluster that is isolated from Internet
2. Pull image from MCR
3. Run container and specify API key and endpoint URL of Cognitive Services Services

Comment 10

ID: 1179544 User: Ody Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Thu 21 Mar 2024 20:26 Selected Answer: - Upvotes: 3

The Azure AI containers are required to submit metering information for billing purposes. Failure to allowlist various network channels that the Azure AI containers rely on will prevent the container from working.

https://learn.microsoft.com/en-us/azure/ai-services/cognitive-services-container-support

Comment 11

ID: 1164261 User: audlindr Badges: - Relative Date: 2 years ago Absolute Date: Sat 02 Mar 2024 18:44 Selected Answer: - Upvotes: 2

This was in 2-Mar-24 exam

Comment 12

ID: 1130955 User: cloudguy1975 Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Wed 24 Jan 2024 18:57 Selected Answer: - Upvotes: 5

The Image needs to be able to send billing info to the Azure AI Service. Also the question does not say anything about no internet traffic. It only states that the docs should remain on-premise.


Build a K8s cluster onprem with internet acecss.
Pull an image
specify the API key and URL

Comment 13

ID: 1027715 User: sl_mslconsulting Badges: - Relative Date: 2 years, 5 months ago Absolute Date: Sun 08 Oct 2023 06:15 Selected Answer: - Upvotes: 4

Request access to use containers in disconnected environments might not get approved what then? A on-premise k8 cluster with internet connectivity is the way to go for billing purpose. The documents still remain on premise.

Comment 13.1

ID: 1027716 User: sl_mslconsulting Badges: - Relative Date: 2 years, 5 months ago Absolute Date: Sun 08 Oct 2023 06:20 Selected Answer: - Upvotes: 2

Also it’s meaningless to specify API key and endpoint URL of Cognitive Services Services while running the container if you don’t have internet connectivity.

Comment 14

ID: 980320 User: abcd9999 Badges: - Relative Date: 2 years, 7 months ago Absolute Date: Mon 14 Aug 2023 00:11 Selected Answer: - Upvotes: 2

if your cluster is isolated from Internet, how can you pull image from MCR?

Comment 14.1

ID: 982982 User: mhauber Badges: - Relative Date: 2 years, 6 months ago Absolute Date: Wed 16 Aug 2023 22:18 Selected Answer: - Upvotes: 1

via the microsoft network

Comment 15

ID: 935866 User: msdfqwerfewf Badges: - Relative Date: 2 years, 8 months ago Absolute Date: Tue 27 Jun 2023 23:55 Selected Answer: - Upvotes: 5

1. Provision an on-premises Kubernetes cluster that is isolated from the internet.
2. Pull an image from MCR (Microsoft Container Registry) or your private container registry.
3. Deploy the app container on the on-premises Kubernetes cluster, specifying the API Key and Endpoint URL of the Cognitive Services resource.

FROM CHATGPT

Comment 16

ID: 925815 User: Hisayuki Badges: - Relative Date: 2 years, 8 months ago Absolute Date: Sat 17 Jun 2023 09:15 Selected Answer: - Upvotes: 6

- Provision an on-prem kubernetes cluster that is isolated from Internet
- Pull an image from MCR
- Run the container and specify an API Key and Endpoint URL of the Cognitive Services resource

Comment 17

ID: 922806 User: 973b658 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Wed 14 Jun 2023 08:37 Selected Answer: - Upvotes: 2

It is true.

11. AI-102 Topic 3 Question 66

Sequence
19
Discussion ID
135016
Source URL
https://www.examtopics.com/discussions/microsoft/view/135016-exam-ai-102-topic-3-question-66-discussion/
Posted By
Harry300
Posted At
March 1, 2024, 7:14 p.m.

Question

HOTSPOT -

You are developing a text processing solution.

You have the function shown below.

image

For the second argument, you call the function and specify the following string.

Our tour of Paris included a visit to the Eiffel Tower

For each of the following statements, select Yes if the statement is true. Otherwise, select No.

image

Suggested Answer

image
Answer Description Click to expand


Comments 9 comments Click to expand

Comment 1

ID: 1163666 User: Harry300 Badges: Highly Voted Relative Date: 2 years ago Absolute Date: Fri 01 Mar 2024 19:14 Selected Answer: - Upvotes: 22

Should by NYN.
Key phrases would be (tested with API): Eiffel Tower, tour, Paris, visit
Entities are: Eiffen Tower, Paris

Comment 2

ID: 1283747 User: AzureGeek79 Badges: Most Recent Relative Date: 1 year, 5 months ago Absolute Date: Sat 14 Sep 2024 20:18 Selected Answer: - Upvotes: 1

Duplicate question. The answer should be NYN.

Comment 3

ID: 1261905 User: moonlightc Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Wed 07 Aug 2024 04:00 Selected Answer: - Upvotes: 3

I ran this in my local, confirmed this is the output,

Key Words:
Paris
Eiffel Tower

So it is N N N

Comment 3.1

ID: 1712128 User: d335823 Badges: - Relative Date: 1 month ago Absolute Date: Fri 06 Feb 2026 01:28 Selected Answer: - Upvotes: 1

That means that the correct answer is N Y N. "and" wasn't meant to be included as one of the entities. It's just how the answer was phrased gramatically.

Comment 4

ID: 1215762 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 22 May 2024 15:31 Selected Answer: - Upvotes: 3

NYN are right answer.

Comment 5

ID: 1191498 User: Murtuza Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Mon 08 Apr 2024 12:08 Selected Answer: - Upvotes: 2

There is a separate function to print key phrase to the console so the answer is NO. They are tricky you into thinking that key phrase is the same thing as key word which its not

Comment 6

ID: 1184118 User: varinder82 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Wed 27 Mar 2024 14:11 Selected Answer: - Upvotes: 2

Final Answer:
N
Y
N

Comment 7

ID: 1178570 User: Murtuza Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Wed 20 Mar 2024 18:56 Selected Answer: - Upvotes: 1

The correct answers are NYY. The last one is yes because it will send info to the console as per the code snippet

Comment 8

ID: 1177235 User: rober13 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Tue 19 Mar 2024 12:03 Selected Answer: - Upvotes: 2

Same : Topic 3, Question 48. NYN.

12. AI-102 Topic 5 Question 50

Sequence
20
Discussion ID
125072
Source URL
https://www.examtopics.com/discussions/microsoft/view/125072-exam-ai-102-topic-5-question-50-discussion/
Posted By
mh_63
Posted At
Nov. 1, 2023, 10:55 a.m.

Question

You build a bot.

You create an Azure Bot resource.

You need to deploy the bot to Azure.

What else should you create?

  • A. only an app registration in Microsoft Azure Active Directory (Azure AD), part of Microsoft Entra, an Azure App Service instance, and an App Service plan
  • B. only an app registration in Microsoft Azure Active Directory (Azure AD), part of Microsoft Entra, an Azure Kubernetes Service (AKS) instance, and a container image
  • C. only an Azure App Service instance, and an App Service plan
  • D. only an Azure Machine Learning workspace and an app registration in Microsoft Azure Active Directory (Azure AD), part of Microsoft Entra

Suggested Answer

A

Answer Description Click to expand


Community Answer Votes

Comments 16 comments Click to expand

Comment 1

ID: 1133325 User: evangelist Badges: Highly Voted Relative Date: 2 years, 1 month ago Absolute Date: Sat 27 Jan 2024 13:08 Selected Answer: A Upvotes: 13

To deploy a bot to Azure, at a minimum you need:

An app registration in Azure AD to represent the bot identity
An App Service instance to host the bot web service
An App Service plan which defines the pricing tier and scale for the App Service

Comment 2

ID: 1712055 User: ysm Badges: Most Recent Relative Date: 1 month ago Absolute Date: Thu 05 Feb 2026 16:16 Selected Answer: C Upvotes: 1

We can not have an azure bot resource created without an App Registration.
Since the statement is: You create an azure bot (means the resource is already created; the app registration already created).

Comment 3

ID: 1331235 User: kennynelcon Badges: - Relative Date: 1 year, 2 months ago Absolute Date: Tue 24 Dec 2024 22:20 Selected Answer: C Upvotes: 4

When you create an Azure Bot resource in the portal, part of what that resource does under the hood is register an application for you in Azure Active Directory. In other words, the Bot resource itself takes care of the AAD app registration automatically. Therefore, you do not need to create a separate app registration yourself—meaning option A includes an extra, unnecessary step.

Comment 3.1

ID: 1617279 User: steven1 Badges: - Relative Date: 4 months, 4 weeks ago Absolute Date: Wed 15 Oct 2025 01:29 Selected Answer: - Upvotes: 1

Correct.
Option A double-counts the app registration that’s already handled by the Azure Bot resource.

Comment 4

ID: 1229911 User: hatanaoki Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 17:12 Selected Answer: A Upvotes: 2

A is the correct answer for this.

Comment 5

ID: 1229855 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 15:43 Selected Answer: A Upvotes: 1

A is answer.

Comment 6

ID: 1218960 User: anto69 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sun 26 May 2024 13:46 Selected Answer: A Upvotes: 1

No doubt is A

Comment 7

ID: 1214240 User: JamesKJoker Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Mon 20 May 2024 11:01 Selected Answer: C Upvotes: 2

App registry is only required for OnPremise App
https://learn.microsoft.com/en-us/azure/bot-service/bot-service-quickstart-registration?view=azure-bot-service-4.0&tabs=userassigned

Comment 7.1

ID: 1257136 User: Shariq Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Mon 29 Jul 2024 02:09 Selected Answer: - Upvotes: 1

but we need to deploy the bot to Azure

Comment 7.2

ID: 1275375 User: JakeCallham Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Sat 31 Aug 2024 07:02 Selected Answer: - Upvotes: 1

yes for on prem, we deploy to azure though

Comment 8

ID: 1206568 User: michaelmorar Badges: - Relative Date: 1 year, 10 months ago Absolute Date: Sat 04 May 2024 18:43 Selected Answer: C Upvotes: 2

Entra ID is surplus to requirement surely.

Comment 9

ID: 1181061 User: Murtuza Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Sat 23 Mar 2024 18:34 Selected Answer: A Upvotes: 2

Therefore, the correct answer is A. only an app registration in Microsoft Azure Active Directory (Azure AD), part of Microsoft Entra, an Azure App Service instance, and an App Service plan. These components work together to ensure your bot is up and running in Azure.

Comment 10

ID: 1164643 User: Harry300 Badges: - Relative Date: 2 years ago Absolute Date: Sun 03 Mar 2024 10:23 Selected Answer: A Upvotes: 2

A is correct

Comment 11

ID: 1104474 User: MelMac Badges: - Relative Date: 2 years, 2 months ago Absolute Date: Sun 24 Dec 2023 07:31 Selected Answer: A Upvotes: 3

https://learn.microsoft.com/en-us/azure/bot-service/provision-and-publish-a-bot?view=azure-bot-service-4.0&tabs=userassigned%2Ccsharp

Your bot will use these types of resources.

The Azure subscription that you'll use to provision, publish, and manage the bot
One or more resource groups
A user-assigned managed identity or an Microsoft Entra ID app registration
An App Service Plan resource
An App Service resource
An Azure Bot resource

Comment 12

ID: 1060110 User: _LAW_ Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Thu 02 Nov 2023 00:02 Selected Answer: C Upvotes: 3

AD it's not necessary, aks doesn't have sense here

Comment 13

ID: 1059536 User: mh_63 Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Wed 01 Nov 2023 10:55 Selected Answer: - Upvotes: 3

I think the correct answer should be C. An Azure Service Plan and an Azure Service instance should are necessary to deploy the bot to Azure.

13. AI-102 Topic 1 Question 30

Sequence
21
Discussion ID
80844
Source URL
https://www.examtopics.com/discussions/microsoft/view/80844-exam-ai-102-topic-1-question-30-discussion/
Posted By
mk1967
Posted At
Sept. 7, 2022, 10:45 a.m.

Question

HOTSPOT -
You are building an app that will process incoming email and direct messages to either French or English language support teams.
Which Azure Cognitive Services API should you use? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Hot Area:
image

Suggested Answer

image
Answer Description Click to expand


Comments 20 comments Click to expand

Comment 1

ID: 758884 User: Adedoyin_Simeon Badges: Highly Voted Relative Date: 3 years, 2 months ago Absolute Date: Tue 27 Dec 2022 18:58 Selected Answer: - Upvotes: 59

The answer is wrong.
Correct answer should be:
Box1:
https://eastus.api.cognitive.microsoft.com

Box2:
/text/analytics/v3.1/languages

REF:
https://learn.microsoft.com/en-us/rest/api/cognitiveservices-textanalytics/3.0/languages/languages?tabs=HTTP.

NOTE:
Pay special attention to the Sample Request provided. Request to the API should be of the form:

POST {Endpoint}/text/analytics/v3.0/languages

Where the {Endpoint} as stated under the sub-heading "URI Parameters" was described as quoted here (see "Description" column of the table):

"Supported Cognitive Services endpoints (protocol and hostname, for example: https://westus.api.cognitive.microsoft.com)."

So the sample given shows the correct format of the *endpoint* as https://{location}.api.cognitive.microsoft.com

Comment 1.1

ID: 917307 User: ziggy1117 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Wed 07 Jun 2023 16:44 Selected Answer: - Upvotes: 1

this is correct: https://learn.microsoft.com/en-us/rest/api/cognitiveservices-textanalytics/3.1preview4/languages/languages?tabs=HTTP

Comment 1.2

ID: 1213268 User: p2006 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sat 18 May 2024 13:02 Selected Answer: - Upvotes: 1

https://westus.dev.cognitive.microsoft.com/docs/services/TextAnalytics-v3-1/operations/Languages

Comment 2

ID: 662256 User: mk1967 Badges: Highly Voted Relative Date: 3 years, 6 months ago Absolute Date: Wed 07 Sep 2022 10:45 Selected Answer: - Upvotes: 11

We only need to detect the language, so the Language Cognitive Service could be used, i.e.:
POST {Endpoint}/text/analytics/v3.0/languages
where Endpoint could be for example "westus"
https://docs.microsoft.com/en-us/rest/api/cognitiveservices-textanalytics/3.0/languages/languages?tabs=HTTP

Comment 3

ID: 1709743 User: LovelyGroovey Badges: Most Recent Relative Date: 1 month, 2 weeks ago Absolute Date: Tue 27 Jan 2026 04:33 Selected Answer: - Upvotes: 1

Microsoft Co-Pilot said,

✅ Correct Solution
You are building an app that will process incoming email/direct messages and route them to English or French support teams.
That means the app must:

Detect language → to know if the message is EN or FR
Translate to English or French → depending on routing needs

So you must use:
Base URL
✔ api.cognitive.microsofttranslator.com
This is the correct endpoint for Translator Service.
*****(Not EastUS Cognitive, not portal.azure.com)*****

Comment 4

ID: 1217624 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 16:55 Selected Answer: - Upvotes: 1

1. eastus.api.cognitive.microsoft.com
2. /text/analytics/v3.1/languages

Comment 5

ID: 1217623 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 16:55 Selected Answer: - Upvotes: 2

1. eastus.api.cognitive.microsoft.com
2. text/analytics/v3.1/languages

Comment 6

ID: 1217340 User: funny_penguin Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 11:21 Selected Answer: - Upvotes: 1

on exam today 24/05/2024, Adedoyin_Simeon's response is the one I selected as well.

Comment 7

ID: 1213797 User: reiwanotora Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sun 19 May 2024 15:22 Selected Answer: - Upvotes: 1

eastus.api.cognitive.microsoft.com and /text/analytics/v3.1/languages

Comment 8

ID: 1182287 User: varinder82 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Mon 25 Mar 2024 08:44 Selected Answer: - Upvotes: 1

Final Answer:
Box1:
https://eastus.api.cognitive.microsoft.com
Box2:
/text/analytics/v3.1/languages

Comment 9

ID: 1164251 User: audlindr Badges: - Relative Date: 2 years ago Absolute Date: Sat 02 Mar 2024 18:40 Selected Answer: - Upvotes: 5

This was in 2-Mar-24 exam

Comment 10

ID: 1077576 User: sca88 Badges: - Relative Date: 2 years, 3 months ago Absolute Date: Wed 22 Nov 2023 17:57 Selected Answer: - Upvotes: 1

https://eastus.dev.cognitive.microsoft.com/docs/services/TextAnalytics.V2.0/operations/56f30ceeeda5650db055a3c7

Comment 11

ID: 936114 User: Pixelmate Badges: - Relative Date: 2 years, 8 months ago Absolute Date: Wed 28 Jun 2023 07:11 Selected Answer: - Upvotes: 3

This came in exam 28/06/2023

Comment 12

ID: 917311 User: ziggy1117 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Wed 07 Jun 2023 16:46 Selected Answer: - Upvotes: 3

The scenario needs a Language service. Specifically just a way to know if its in english or french. No translation is needed. Hence:
Box1:
https://eastus.api.cognitive.microsoft.com
Box2:
/text/analytics/v3.1/languages

Source: https://learn.microsoft.com/en-us/rest/api/cognitiveservices-textanalytics/3.1preview4/languages/languages?tabs=HTTP

Comment 13

ID: 839701 User: RAN_L Badges: - Relative Date: 2 years, 12 months ago Absolute Date: Wed 15 Mar 2023 09:51 Selected Answer: - Upvotes: 3

api.cogitive.microsofttranslator.com
eastus.api.cognitive.microsoft.com
portal.azure.com
Out of these three options, only option 2 is a valid Azure Cognitive Services endpoint. The eastus.api.cognitive.microsoft.com endpoint is used for various Azure Cognitive Services, including the Text Analytics and Translator services.

Option 1, api.cogitive.microsofttranslator.com, is not a valid endpoint for Azure Cognitive Services. It appears to be a typo, as "cogitive" is misspelled.

Option 3, portal.azure.com, is the URL for the Azure portal, which is a web-based management interface for Azure services. It is not an endpoint for Azure Cognitive Services APIs.

-./text/analytics/v3.1/languages
One of the APIs that you can use to process incoming email and direct messages to either French or English language support teams is the Text Analytics API. The specific API endpoint you can use is /text/analytics/v3.1/languages. This API can detect the language of a given text and return the ISO 639-1 language code. You can then use this information to route the message to the appropriate language support team.

Comment 13.1

ID: 1177930 User: Ody Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Wed 20 Mar 2024 06:06 Selected Answer: - Upvotes: 1

api.cognitive.microsofttranslator.com is a valid endpoint for text translation

https://learn.microsoft.com/en-us/azure/ai-services/translator/quickstart-text-rest-api?tabs=csharp#prerequisites

Comment 14

ID: 833515 User: Nicoseal Badges: - Relative Date: 3 years ago Absolute Date: Thu 09 Mar 2023 02:30 Selected Answer: - Upvotes: 4

mk1967 and Adedoyin_Simeon are correct
The correct answer should be:
Box1:
https://eastus.api.cognitive.microsoft.com

Box2:
/text/analytics/v3.1/languages

Comment 15

ID: 758882 User: Adedoyin_Simeon Badges: - Relative Date: 3 years, 2 months ago Absolute Date: Tue 27 Dec 2022 18:57 Selected Answer: - Upvotes: 2

The answer is wrong. Correct answer should be:
https://eastus.api.cognitive.microsoft.com
&
/text/analytics/v3.1/languages

REF:
https://learn.microsoft.com/en-us/rest/api/cognitiveservices-textanalytics/3.0/languages/languages?tabs=HTTP.

NOTE:
Pay special attention to the Sample Request provided. Request to the API should be of the form:

POST {Endpoint}/text/analytics/v3.0/languages

Where the {Endpoint} as stated under the sub-heading "URI Parameters" was described as quoted here (see "Description" column of the table):

"Supported Cognitive Services endpoints (protocol and hostname, for example: https://westus.api.cognitive.microsoft.com)."

So the sample given shows the correct format of the *endpoint* as https://{location}.api.cognitive.microsoft.com

Comment 16

ID: 758880 User: Adedoyin_Simeon Badges: - Relative Date: 3 years, 2 months ago Absolute Date: Tue 27 Dec 2022 18:55 Selected Answer: - Upvotes: 1

The answer is wrong. Correct answer should be:
https://eastus.api.cognitive.microsoft.com
&
/text/analytics/v3.1/languages

REF:
https://learn.microsoft.com/en-us/rest/api/cognitiveservices-textanalytics/3.0/languages/languages?tabs=HTTP.

NOTE:
Pay special attention to the Sample Request provided. Request to the API should be of the form:

POST {Endpoint}/text/analytics/v3.0/languages

Where the {Endpoint} as stated under the sub-heading "URI Parameters" was described as quoted here

"Supported Cognitive Services endpoints (protocol and hostname, for example: https://westus.api.cognitive.microsoft.com)."

So the sample given shows the correct format of the *endpoint* as https://{location}.api.cognitive.microsoft.com

Comment 17

ID: 741850 User: HotDurian Badges: - Relative Date: 3 years, 3 months ago Absolute Date: Sun 11 Dec 2022 16:44 Selected Answer: - Upvotes: 1

Answer is incorrect. Language Cognitive Service is sufficient.
https://learn.microsoft.com/en-us/rest/api/cognitiveservices-textanalytics/3.0/languages/languages?tabs=HTTP

14. AI-102 Topic 1 Question 23

Sequence
35
Discussion ID
76149
Source URL
https://www.examtopics.com/discussions/microsoft/view/76149-exam-ai-102-topic-1-question-23-discussion/
Posted By
Khiem
Posted At
May 23, 2022, 9:08 a.m.

Question

DRAG DROP -
You need to develop an automated call handling system that can respond to callers in their own language. The system will support only French and English.
Which Azure Cognitive Services service should you use to meet each requirement? To answer, drag the appropriate services to the correct requirements. Each service may be used once, more than once, or not at all. You may need to drag the split bat between panes or scroll to view content.
NOTE: Each correct selection is worth one point.
Select and Place:
image

Suggested Answer

image
Answer Description Click to expand


Comments 23 comments Click to expand

Comment 1

ID: 605883 User: Khiem Badges: Highly Voted Relative Date: 3 years, 9 months ago Absolute Date: Mon 23 May 2022 09:08 Selected Answer: - Upvotes: 96

It should be:
- Speech to Text with AutoDetectSourceLanguageConfig. It can't be Text Analytics because the input is callers' voice.
- Text to Speech: the output is voice.

Comment 1.1

ID: 1339064 User: hannah380099 Badges: - Relative Date: 1 year, 2 months ago Absolute Date: Sat 11 Jan 2025 06:01 Selected Answer: - Upvotes: 3

the Azure Speech to Text service does support language detection, which can be used to identify the language being spoken in an audio source. This feature is part of the Speech service and can be used in conjunction with the Translator service for tasks like speech translation1.

So, you can indeed use Speech to Text for detecting the language and then use Translator for translating the detected language if needed.

Comment 1.2

ID: 1319423 User: Alan_CA Badges: - Relative Date: 1 year, 3 months ago Absolute Date: Thu 28 Nov 2024 20:27 Selected Answer: - Upvotes: 1

I agree
https://learn.microsoft.com/en-us/azure/ai-services/speech-service/language-identification

Comment 1.3

ID: 1218227 User: PeteColag Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sat 25 May 2024 13:07 Selected Answer: - Upvotes: 1

I agree with your answer. What is missing in this question is an explanation of how the response text (to feed the text to speech) is being generated. I am assuming this would be based on LUIS or CLU?

Comment 1.4

ID: 1245584 User: Toby86 Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Wed 10 Jul 2024 18:24 Selected Answer: - Upvotes: 1

private static async Task Respond(SpeechRecognitionResult result, SpeechConfig speechConfig, Random random)
{
if (result.Reason.Equals(ResultReason.RecognizedSpeech))
{
var client = new TextAnalyticsClient(languageDetectorEndpoint, credentials);
DetectedLanguage detectedLanguage = client.DetectLanguage(result.Text);

Comment 2

ID: 631394 User: Eltooth Badges: Highly Voted Relative Date: 3 years, 8 months ago Absolute Date: Thu 14 Jul 2022 15:27 Selected Answer: - Upvotes: 14

Speech-to-Text : https://docs.microsoft.com/en-us/azure/cognitive-services/speech-service/speech-to-text

Text-to-Speech : https://docs.microsoft.com/en-us/azure/cognitive-services/speech-service/text-to-speech

Both support common languages, including French.

https://docs.microsoft.com/en-us/azure/cognitive-services/speech-service/language-support?tabs=speechtotext

Comment 3

ID: 1700857 User: ravikumc333 Badges: Most Recent Relative Date: 2 months, 2 weeks ago Absolute Date: Sun 21 Dec 2025 16:52 Selected Answer: - Upvotes: 1

✔ Detect the incoming language:
Translator

Translator includes built‑in language detection for text.
Even though the input is voice, the typical flow is: Speech‑to‑Text → Translator (detect language + translate).
Among the listed services, Translator is the only one capable of detecting language.

Comment 4

ID: 1306852 User: vivekgoel15 Badges: - Relative Date: 1 year, 4 months ago Absolute Date: Mon 04 Nov 2024 10:10 Selected Answer: - Upvotes: 2

1. Speech to Text
2. Text to Speech

Comment 5

ID: 1235093 User: SAMBIT Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 04:32 Selected Answer: - Upvotes: 1

https://learn.microsoft.com/en-us/azure/synapse-analytics/machine-learning/tutorial-text-analytics-use-mmlspark

Comment 6

ID: 1234514 User: HaraTadahisa Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Fri 21 Jun 2024 16:56 Selected Answer: - Upvotes: 1

1. Speech to Text
2. Text to Speech

Comment 7

ID: 1227038 User: reiwanotora Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sun 09 Jun 2024 05:48 Selected Answer: - Upvotes: 1

1. Speech to Text
2. Text to Speech

Comment 8

ID: 1223135 User: anto69 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sun 02 Jun 2024 13:11 Selected Answer: - Upvotes: 1

ChatGPT: speech to text + text to speech

Comment 9

ID: 1221419 User: Tempeck Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 30 May 2024 08:33 Selected Answer: - Upvotes: 1

The answer should be Text Analytics and TTS.

Comment 10

ID: 1217626 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 16:57 Selected Answer: - Upvotes: 1

1. Speech to Text
2. Test to Speech

From Takedajuku perspective, if you study for 4 days and spend 2 days reviewing, you will have a better chance of passing the exam.

Comment 11

ID: 1182283 User: varinder82 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Mon 25 Mar 2024 08:38 Selected Answer: - Upvotes: 3

Final Answer:
1. Speech to Text
2. Text to Speech

Comment 12

ID: 1170903 User: dacchione Badges: - Relative Date: 2 years ago Absolute Date: Mon 11 Mar 2024 09:50 Selected Answer: - Upvotes: 4

Language identification (LID) use cases include:
Speech to text recognition when you need to identify the language in an audio source and then transcribe it to text.
Speech translation when you need to identify the language in an audio source and then translate it to another language.
https://learn.microsoft.com/en-us/azure/ai-services/speech-service/language-identification?tabs=once&pivots=programming-language-python

Comment 13

ID: 1163139 User: Mehe323 Badges: - Relative Date: 2 years ago Absolute Date: Fri 01 Mar 2024 01:59 Selected Answer: - Upvotes: 4

People, read the descriptions, these are describing only parts of the process, not the whole process.
1) DETECT incoming language has got not to do with speech to text, but with text analytics. Yes, speech to text is a part of the whole process but this services TRANSCRIBES and does not do language detection.
2) Respond to caller's language should be Text to speech

Comment 13.1

ID: 1185987 User: Mehe323 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Sat 30 Mar 2024 10:03 Selected Answer: - Upvotes: 5

IGNORE my comment! Speech to text also does language detection:
https://learn.microsoft.com/en-us/azure/ai-services/speech-service/language-identification?tabs=once&pivots=programming-language-csharp

Comment 14

ID: 1154794 User: audlindr Badges: - Relative Date: 2 years ago Absolute Date: Tue 20 Feb 2024 15:59 Selected Answer: - Upvotes: 1

It can't be Text Analytics. It is an incoming call. It should be Speech to Text
https://learn.microsoft.com/en-us/azure/ai-services/speech-service/language-identification?tabs=once&pivots=programming-language-csharp

And then Text to Speech for the response

Comment 15

ID: 1131527 User: suzanne_exam Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Thu 25 Jan 2024 10:03 Selected Answer: - Upvotes: 1

Speech to text - as it's based on a voice call
Text to speech - it's not translator here, as the key thing is that the program needs to respond, not just translate

Comment 16

ID: 980070 User: james2033 Badges: - Relative Date: 2 years, 7 months ago Absolute Date: Sun 13 Aug 2023 15:58 Selected Answer: - Upvotes: 4

At this time, this question is out of date. See https://learn.microsoft.com/en-us/azure/ai-services/language-service/language-detection/overview . In old time, Text Analytics is correct.

Comment 17

ID: 922469 User: endeesa Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Tue 13 Jun 2023 21:06 Selected Answer: - Upvotes: 1

to detect the incoming language you need Text analytics, Speech to text does not have an option to detect language

Comment 17.1

ID: 933232 User: CauchyLee Badges: - Relative Date: 2 years, 8 months ago Absolute Date: Sun 25 Jun 2023 07:17 Selected Answer: - Upvotes: 1

POST /cognitiveservices/v1/speechtotext/recognition HTTP/1.1
Ocp-Apim-Subscription-Key: <subscription_key>
Content-Type: application/json

{
"config": {
"language": "auto",
"enableSeparation": true
},
"format": "audio-16khz-128kbitrate-mono-mp3",
"audio": <binary_audio_file>
}

15. AI-102 Topic 1 Question 3

Sequence
37
Discussion ID
54105
Source URL
https://www.examtopics.com/discussions/microsoft/view/54105-exam-ai-102-topic-1-question-3-discussion/
Posted By
VT100
Posted At
June 1, 2021, 6:27 p.m.

Question

You need to build a chatbot that meets the following requirements:
✑ Supports chit-chat, knowledge base, and multilingual models
✑ Performs sentiment analysis on user messages
✑ Selects the best language model automatically
What should you integrate into the chatbot?

  • A. QnA Maker, Language Understanding, and Dispatch
  • B. Translator, Speech, and Dispatch
  • C. Language Understanding, Text Analytics, and QnA Maker
  • D. Text Analytics, Translator, and Dispatch

Suggested Answer

C

Answer Description Click to expand


Community Answer Votes

Comments 17 comments Click to expand

Comment 1

ID: 1291126 User: AnnaR Badges: Highly Voted Relative Date: 1 year, 5 months ago Absolute Date: Sun 29 Sep 2024 12:58 Selected Answer: - Upvotes: 13

according to https://learn.microsoft.com/en-us/azure/ai-services/language-service/overview
Azure AI Language now unifies the Text Analyrics, QnA Maker and LUIS. So in my opinion, the question is obsolete.

Comment 2

ID: 1699772 User: Raj1982 Badges: Most Recent Relative Date: 2 months, 3 weeks ago Absolute Date: Tue 16 Dec 2025 08:27 Selected Answer: A Upvotes: 3

without dispatch you select the manually decide when call to Qna and LUIS .. thats violates for automation. QnA Maker, Language Understanding, and Dispatch

Comment 3

ID: 1699153 User: Madhusmita Badges: - Relative Date: 2 months, 4 weeks ago Absolute Date: Sat 13 Dec 2025 01:41 Selected Answer: D Upvotes: 1

D. Text Analytics, Translator, and Dispatch
Text Analytics analyzes user messages for sentiment and other insights
Translator handles language translation for multilingual support
Dispatch intelligently routes requests to the appropriate underlying services (knowledge base, chit-chat models, etc.) and selects the best model automatically

Comment 4

ID: 1595049 User: Pankajloms Badges: - Relative Date: 7 months, 1 week ago Absolute Date: Tue 05 Aug 2025 13:27 Selected Answer: D Upvotes: 1

Dispatch is used to route the user's query to the appropriate model (LUIS, QnA Maker, or others).

🔸 Translator is needed for multilingual support.

Comment 5

ID: 1582068 User: Canarys_Automations Badges: - Relative Date: 8 months, 2 weeks ago Absolute Date: Tue 01 Jul 2025 07:16 Selected Answer: C Upvotes: 2

LUIS → Understands user intent and supports multilingual.

Text Analytics → Does sentiment analysis.

QnA Maker → For chit-chat and FAQ knowledge base.

So Other options don’t meet all features. Only C checks all boxes

Comment 6

ID: 1577868 User: StelSen Badges: - Relative Date: 8 months, 4 weeks ago Absolute Date: Mon 16 Jun 2025 08:20 Selected Answer: C Upvotes: 2

Option C: Language Understanding, Text Analytics, and QnA Maker
• Language Understanding = LUIS, handles intent detection.
• Text Analytics = handles sentiment analysis.
• QnA Maker = handles chit-chat and knowledge base.
This meets all three requirements.

Why not A?
Option A: QnA Maker, Language Understanding, and Dispatch
• QnA Maker handles knowledge base and chit-chat.
• Language Understanding (LUIS) enables intent and entity recognition.
• Dispatch routes to different services, including QnA or LUIS, based on user input.

BUT: No sentiment analysis capability included

Comment 7

ID: 1572635 User: man5484 Badges: - Relative Date: 9 months, 2 weeks ago Absolute Date: Tue 27 May 2025 11:47 Selected Answer: C Upvotes: 3

C. Language Understanding, Text Analytics, and QnA Maker
It meets all the functional requirements:

Chitchat + KB → QnA Maker

Multilingual → LUIS and QnA both support it natively

Sentiment → Text Analytics

Model selection → handled by bot framework routing logic or Dispatch (not strictly required if you're only using two models)

Comment 8

ID: 1563899 User: miro1 Badges: - Relative Date: 10 months, 2 weeks ago Absolute Date: Sat 26 Apr 2025 15:42 Selected Answer: D Upvotes: 1

Why the others don’t fully match:
A. QnA Maker, Language Understanding, and Dispatch
No sentiment analysis or multilingual support.
B. Translator, Speech, and Dispatch
No sentiment analysis or knowledge base.
C. Language Understanding, Text Analytics, and QnA Maker
Lacks multilingual capability and automatic model selection.

Comment 9

ID: 1562654 User: kakarooky Badges: - Relative Date: 10 months, 3 weeks ago Absolute Date: Tue 22 Apr 2025 08:17 Selected Answer: D Upvotes: 1

A: no sentiment analysis
B: don't need speech and no sentiment analysis
C: can't select the best language model automatically. -> no dispatch

Comment 10

ID: 1357201 User: gyaansastra Badges: - Relative Date: 1 year ago Absolute Date: Sun 16 Feb 2025 11:36 Selected Answer: A Upvotes: 2

The best choice here would be Option A: QnA Maker, Language Understanding, and Dispatch.

Here's why:
- QnA Maker supports the knowledge base requirement.
- Language Understanding (LUIS) supports chit-chat by interpreting user intents and entities.
- Dispatch helps in selecting the best language model automatically and routing the requests to the appropriate service.

The combination of these three components covers all the necessary requirements for your chatbot. Option B and D do not include the knowledge base component, and Option C lacks the automatic selection of the best language model.

Comment 11

ID: 1339039 User: hannah380099 Badges: - Relative Date: 1 year, 2 months ago Absolute Date: Sat 11 Jan 2025 03:21 Selected Answer: A Upvotes: 3

Option A (QnA Maker, Language Understanding, and Dispatch): This combination supports chit-chat, knowledge base, multilingual models, and automatic language model selection through Dispatch.

Option C (Language Understanding, Text Analytics, and QnA Maker): This option includes Text Analytics for sentiment analysis but lacks Dispatch for automatic model selection.

Option D (Text Analytics, Translator, and Dispatch): This option includes Text Analytics and Dispatch but lacks QnA Maker for knowledge bases.

Given these considerations, Option A would be the most suitable choice as it covers all requirements through the combined capabilities of QnA Maker, Language Understanding, and Dispatch.

Comment 12

ID: 1332751 User: jiyoung22 Badges: - Relative Date: 1 year, 2 months ago Absolute Date: Sat 28 Dec 2024 04:55 Selected Answer: A Upvotes: 1

Language Understanding include Sentiment analysis.
And use the Bot framework Dispatch CLI tool to help build a process to work with both services(Language Understanding, QnA Maker).
So Answer is A
https://learn.microsoft.com/en-us/azure/ai-services/luis/luis-reference-prebuilt-sentiment?tabs=english

Comment 13

ID: 1263552 User: anto69 Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Sat 10 Aug 2024 16:43 Selected Answer: C Upvotes: 1

C is the correct answer

Comment 14

ID: 1242628 User: Aouatef Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Fri 05 Jul 2024 10:41 Selected Answer: - Upvotes: 1

I think the answer option is A (see https://learn.microsoft.com/en-us/azure/ai-services/qnamaker/choose-natural-language-processing-service)

Comment 15

ID: 1219583 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Mon 27 May 2024 15:25 Selected Answer: C Upvotes: 2

C is right answer.

Comment 16

ID: 1213814 User: reiwanotora Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sun 19 May 2024 15:37 Selected Answer: C Upvotes: 2

C is right.

Comment 17

ID: 1211295 User: JamesKJoker Badges: - Relative Date: 1 year, 10 months ago Absolute Date: Tue 14 May 2024 09:55 Selected Answer: C Upvotes: 2

As Luis is outdated (with its Dispatch option), I must be C

16. AI-102 Topic 3 Question 7

Sequence
39
Discussion ID
80186
Source URL
https://www.examtopics.com/discussions/microsoft/view/80186-exam-ai-102-topic-3-question-7-discussion/
Posted By
Internal_Koala
Posted At
Sept. 5, 2022, 5 a.m.

Question

You are building a Conversational Language Understanding model for an e-commerce chatbot. Users can speak or type their billing address when prompted by the chatbot.
You need to construct an entity to capture billing addresses.
Which entity type should you use?

  • A. machine learned
  • B. Regex
  • C. list
  • D. Pattern.any

Suggested Answer

A

Answer Description Click to expand


Community Answer Votes

Comments 13 comments Click to expand

Comment 1

ID: 659689 User: Internal_Koala Badges: Highly Voted Relative Date: 3 years, 6 months ago Absolute Date: Mon 05 Sep 2022 05:00 Selected Answer: A Upvotes: 19

The link provided mentions addresses under 'ML Entities with Structure'. Will be hard to identify all possible international addresses with RegEx.

Comment 2

ID: 672170 User: praticewizards Badges: Highly Voted Relative Date: 3 years, 5 months ago Absolute Date: Sun 18 Sep 2022 11:34 Selected Answer: A Upvotes: 10

A - MAchine Learned

in documentation
https://learn.microsoft.com/en-us/azure/cognitive-services/LUIS/concepts/entities#ml-entity-with-structure

- ML Entity with Structure
An ML entity can be composed of smaller sub-entities, each of which can have its own properties. For example, an Address entity could have the following structure:
Address: 4567 Main Street, NY, 98052, USA
Building Number: 4567
Street Name: Main Street
State: NY
Zip Code: 98052
Country: USA

Comment 3

ID: 1699437 User: rbh1090 Badges: Most Recent Relative Date: 2 months, 3 weeks ago Absolute Date: Mon 15 Dec 2025 04:26 Selected Answer: D Upvotes: 1

For the multiple-choice exam question you showed earlier, the correct answer is D. Pattern.any, because the scenario only asks to capture billing addresses — not to break them down into sub-entities.

Comment 4

ID: 1599777 User: JayatiRC Badges: - Relative Date: 6 months, 3 weeks ago Absolute Date: Tue 19 Aug 2025 12:52 Selected Answer: D Upvotes: 4

Should be pattern Any

Comment 5

ID: 1285107 User: mrg998 Badges: - Relative Date: 1 year, 5 months ago Absolute Date: Tue 17 Sep 2024 10:54 Selected Answer: - Upvotes: 1

Azure bot service no longer part of the exam, no need to revise this topic - https://trainingsupport.microsoft.com/en-us/mcp/forum/all/azure-ai102-chatbots-service/798c0cfa-3475-474e-b4ec-8ab7fc790e81#:~:text=Yes%2C%20Julian%20is%20correct%2C%20based,not%20appear%20in%20the%20exam.

Comment 6

ID: 1237171 User: anto69 Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Wed 26 Jun 2024 05:25 Selected Answer: - Upvotes: 2

For me e GPT 3.5: machine learned

Comment 7

ID: 1229245 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 12 Jun 2024 16:26 Selected Answer: A Upvotes: 1

I know you don't know what I'm talking about, but if you think as Crossroads leads you, the answer is A.

Comment 8

ID: 1217673 User: Arun20212021 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 18:57 Selected Answer: - Upvotes: 1

As per Gemini chatbot, the answer is B. Regex. The following is the explanation it gives:
Regex offers a powerful and efficient way to define patterns for capturing specific text formats, making it ideal for structured data like billing addresses. You can design a regex pattern to capture elements like street address, city, state, zip code, and potentially country code.

; and why the answer is not 'machine learned':
While machine learning can be powerful for complex entity recognition, it might be overkill for a well-defined structure like billing addresses. Training a machine learning model could be time-consuming and resource-intensive compared to a simpler solution.

Comment 9

ID: 1217537 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 15:56 Selected Answer: A Upvotes: 1

It must be A.

Comment 10

ID: 1217352 User: funny_penguin Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 11:36 Selected Answer: A Upvotes: 2

on exam, I selected machine learned

Comment 11

ID: 1183895 User: Mehe323 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Wed 27 Mar 2024 06:43 Selected Answer: A Upvotes: 4

Regex would be a good solution if the addresses were in a specific area (e.g, USA). But since that information is not given, you have to assume any type of address. So the correct answer in this case is A.

Comment 12

ID: 672278 User: Richi0907 Badges: - Relative Date: 3 years, 5 months ago Absolute Date: Sun 18 Sep 2022 13:23 Selected Answer: A Upvotes: 2

it should be A

Comment 13

ID: 664889 User: Sharks82 Badges: - Relative Date: 3 years, 6 months ago Absolute Date: Fri 09 Sep 2022 20:54 Selected Answer: - Upvotes: 2

A. machine learned

17. AI-102 Topic 1 Question 20

Sequence
40
Discussion ID
61023
Source URL
https://www.examtopics.com/discussions/microsoft/view/61023-exam-ai-102-topic-1-question-20-discussion/
Posted By
weasel97
Posted At
Aug. 30, 2021, 12:53 p.m.

Question

You plan to provision a QnA Maker service in a new resource group named RG1.
In RG1, you create an App Service plan named AP1.
Which two Azure resources are automatically created in RG1 when you provision the QnA Maker service? Each correct answer presents part of the solution.
NOTE: Each correct selection is worth one point.

  • A. Language Understanding
  • B. Azure SQL Database
  • C. Azure Storage
  • D. Azure Cognitive Search
  • E. Azure App Service

Suggested Answer

DE

Answer Description Click to expand


Community Answer Votes

Comments 18 comments Click to expand

Comment 1

ID: 575280 User: ghoppa Badges: Highly Voted Relative Date: 3 years, 11 months ago Absolute Date: Fri 25 Mar 2022 23:13 Selected Answer: - Upvotes: 22

Another one to memorise. A quick trick:
Q sounds like a C, for Cognitive
n
A for App Service

Comment 1.1

ID: 781123 User: dev2dev Badges: - Relative Date: 3 years, 1 month ago Absolute Date: Thu 19 Jan 2023 13:32 Selected Answer: - Upvotes: 8

lol then there is a risk of selecting A and C as answers :)

Comment 2

ID: 440098 User: methodidacte Badges: Highly Voted Relative Date: 4 years, 6 months ago Absolute Date: Mon 06 Sep 2021 06:39 Selected Answer: - Upvotes: 17

Right answer : at the creation, we have to precise Azure Search an Azure Web App details.
"When you create a QnAMaker resource, you host the data in your own Azure subscription. Azure Search is used to index your data." & "When you create a QnAMaker resource, you host the runtime in your own Azure subscription. App Service is the compute engine that runs the QnA Maker queries for you."

Comment 3

ID: 1699152 User: Madhusmita Badges: Most Recent Relative Date: 2 months, 4 weeks ago Absolute Date: Sat 13 Dec 2025 01:40 Selected Answer: CD Upvotes: 2

The correct answers are C and D:

C. Azure Storage - Required for storing knowledge base data, configurations, and logs
D. Azure Cognitive Search - Required for indexing and searching through the knowledge base content
These two services are fundamental dependencies that QnA Maker automatically provisions to support its core functionality: storing knowledge bases (Storage) and enabling fast question matching (Cognitive Search).

Comment 4

ID: 1624748 User: CMal Badges: - Relative Date: 4 months ago Absolute Date: Mon 10 Nov 2025 18:44 Selected Answer: CD Upvotes: 1

When you provision the Azure AI Language – Question Answering feature (the successor to QnA Maker), the following resources are automatically created in the specified resource group:
Azure Cognitive Search - Used to index and query your knowledge base documents.
Essential for enabling semantic search and retrieval-based question answering.

Azure Storage - Stores the source documents (e.g., PDFs, Word files, web pages) that are ingested into your knowledge base.

Comment 5

ID: 1605143 User: Volcano2cool Badges: - Relative Date: 6 months, 1 week ago Absolute Date: Mon 01 Sep 2025 15:01 Selected Answer: BD Upvotes: 1

The Azure SQL Database is used by QnA Maker to store the knowledge base data.
Azure Cognitive Search is used to index and search the QnA pairs efficiently.

Comment 6

ID: 1585101 User: sureyyy Badges: - Relative Date: 8 months ago Absolute Date: Thu 10 Jul 2025 01:58 Selected Answer: DE Upvotes: 2

DE are correct answer

Comment 7

ID: 1411599 User: yukkki Badges: - Relative Date: 11 months, 2 weeks ago Absolute Date: Sat 29 Mar 2025 04:01 Selected Answer: DE Upvotes: 5

QnAMaker was terminated

Comment 8

ID: 1353939 User: miai74uu Badges: - Relative Date: 1 year, 1 month ago Absolute Date: Sun 09 Feb 2025 15:08 Selected Answer: DE Upvotes: 1

You can check directly in azure.portal by creating a Q&A Maker. Now besides D and E it creates an App Insights service but this is not in the quiestion options.

Comment 9

ID: 1306454 User: noahsark Badges: - Relative Date: 1 year, 4 months ago Absolute Date: Sun 03 Nov 2024 09:11 Selected Answer: DE Upvotes: 2

Apart from the above pricing, you also pay for resources consumed as described below. When you create a QnA Maker resource, you host the data and the runtime in your own Azure subscription. These are powered by Azure AI Search and App Service.

https://azure.microsoft.com/en-us/pricing/details/cognitive-services/qna-maker/

Comment 10

ID: 1291129 User: AnnaR Badges: - Relative Date: 1 year, 5 months ago Absolute Date: Sun 29 Sep 2024 13:20 Selected Answer: - Upvotes: 2

I see only Azure AI Search service being provided and the only correct answer.

Comment 11

ID: 1282633 User: csdodo Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Thu 12 Sep 2024 14:45 Selected Answer: DE Upvotes: 4

REF: https://azure.microsoft.com/en-us/pricing/details/cognitive-services/qna-maker/

"When you create a QnA Maker resource, you host the data and the runtime in your own Azure subscription. These are powered by Azure AI Search and App Service."

Comment 12

ID: 1265510 User: anto69 Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Wed 14 Aug 2024 07:32 Selected Answer: - Upvotes: 6

QnA Maker is now retired

Comment 13

ID: 1222998 User: anto69 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sun 02 Jun 2024 05:20 Selected Answer: DE Upvotes: 1

D and E for Perplexity AI and Claude. Sadly from portal this resource cannot be provisioned anymore

Comment 14

ID: 1222995 User: anto69 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sun 02 Jun 2024 05:16 Selected Answer: - Upvotes: 1

C and D for ChatGPT

Comment 15

ID: 1219577 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Mon 27 May 2024 15:21 Selected Answer: DE Upvotes: 1

DE are right answer.

Comment 16

ID: 1217638 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 17:08 Selected Answer: DE Upvotes: 1

DE is power.

Comment 17

ID: 1213806 User: reiwanotora Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sun 19 May 2024 15:32 Selected Answer: DE Upvotes: 1

D and E are right.

18. AI-102 Topic 1 Question 21

Sequence
41
Discussion ID
86971
Source URL
https://www.examtopics.com/discussions/microsoft/view/86971-exam-ai-102-topic-1-question-21-discussion/
Posted By
ArchMelody
Posted At
Nov. 4, 2022, 10:11 a.m.

Question

You are building a language model by using a Language Understanding (classic) service.
You create a new Language Understanding (classic) resource.
You need to add more contributors.
What should you use?

  • A. a conditional access policy in Azure Active Directory (Azure AD)
  • B. the Access control (IAM) page for the authoring resources in the Azure portal
  • C. the Access control (IAM) page for the prediction resources in the Azure portal

Suggested Answer

B

Answer Description Click to expand


Community Answer Votes

Comments 10 comments Click to expand

Comment 1

ID: 1699289 User: PrasadMP Badges: - Relative Date: 2 months, 4 weeks ago Absolute Date: Sun 14 Dec 2025 07:33 Selected Answer: B Upvotes: 1

B is the answer
In Language Understanding (LUIS) classic, there are two separate resources:

Authoring resource – used to create, edit, train, and publish the language model

Prediction resource – used only at runtime to query the published model

To add more contributors who can build or modify the language model, you must grant permissions on the authoring resource.

Comment 2

ID: 1222999 User: anto69 Badges: - Relative Date: 1 year, 3 months ago Absolute Date: Mon 02 Dec 2024 06:24 Selected Answer: B Upvotes: 1

It's B. Check with AI

Comment 3

ID: 1217619 User: nanaw770 Badges: - Relative Date: 1 year, 3 months ago Absolute Date: Sun 24 Nov 2024 17:52 Selected Answer: B Upvotes: 3

I selected B, but it does not make sense. It is like there is nothing sadder than having a breakup with a lover.

Comment 4

ID: 1213794 User: reiwanotora Badges: - Relative Date: 1 year, 3 months ago Absolute Date: Tue 19 Nov 2024 16:17 Selected Answer: B Upvotes: 1

B is right.

Comment 5

ID: 1193507 User: CDL_Learner Badges: - Relative Date: 1 year, 5 months ago Absolute Date: Fri 11 Oct 2024 09:00 Selected Answer: B Upvotes: 1

The best answer is B. the Access control (IAM) page for the authoring resources in the Azure portal., to add more contributors to a Language Understanding (classic) resource, we should use the Access control (IAM) page for the authoring resources in the Azure portal. This allows we to manage access and permissions for the resource at a granular level, including adding new contributors.

Comment 5.1

ID: 1193508 User: CDL_Learner Badges: - Relative Date: 1 year, 5 months ago Absolute Date: Fri 11 Oct 2024 09:01 Selected Answer: - Upvotes: 2

Here’s why the other options are not suitable:
• Option A: A conditional access policy in Azure Active Directory (Azure AD) is used to enforce certain conditions before granting access to resources. It doesn’t directly allow us to add contributors to a specific resource.
• Option C: The Access control (IAM) page for the prediction resources in the Azure portal is used to manage access to prediction resources, not authoring resources. Since we’re looking to add contributors to the Language Understanding resource we’ve created (which is an authoring resource), this option wouldn’t be suitable.

Comment 6

ID: 1147831 User: evangelist Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Mon 12 Aug 2024 05:29 Selected Answer: B Upvotes: 2

In Azure's Language Understanding (LUIS) service, adding more contributors or managing permissions is usually done through the Access Control (IAM) page in the Azure portal. This is because, for the management and development of LUIS services, especially in the process of designing and building language models, the required permission management mainly focuses on authoring resources. Through the "Access Control (IAM)" page, you can assign different roles to different users or user groups, such as "Contributor" or "Reader", to control their access and operation permissions to resources.

Comment 7

ID: 839672 User: RAN_L Badges: - Relative Date: 2 years, 5 months ago Absolute Date: Fri 15 Sep 2023 08:25 Selected Answer: B Upvotes: 2

B. the Access control (IAM) page for the authoring resources in the Azure portal is the correct option to add more contributors to a Language Understanding (classic) resource. IAM (Identity and Access Management) in Azure allows you to manage access to your Azure resources by assigning roles to users, groups, and applications. By using the Access control (IAM) page for the authoring resources in the Azure portal, you can add new users or groups to your Language Understanding (classic) resource and grant them the necessary permissions to edit and publish your language model.

Comment 8

ID: 778433 User: KingChuang Badges: - Relative Date: 2 years, 7 months ago Absolute Date: Mon 17 Jul 2023 01:10 Selected Answer: - Upvotes: 4

on my exam (2023-01-16 Passed)

My Answer:B

Comment 9

ID: 711011 User: ArchMelody Badges: - Relative Date: 2 years, 10 months ago Absolute Date: Thu 04 May 2023 09:11 Selected Answer: B Upvotes: 3

Correct Answer, indeed!

19. AI-102 Topic 3 Question 49

Sequence
44
Discussion ID
112137
Source URL
https://www.examtopics.com/discussions/microsoft/view/112137-exam-ai-102-topic-3-question-49-discussion/
Posted By
973b658
Posted At
June 14, 2023, 9:09 a.m.

Question

HOTSPOT
-

You are building an Azure web app named App1 that will translate text from English to Spanish.

You need to use the Text Translation REST API to perform the translation. The solution must ensure that you have data sovereignty in the United States.

How should you complete the URI? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

image

Suggested Answer

image
Answer Description Click to expand


Comments 8 comments Click to expand

Comment 1

ID: 1698886 User: falconai102 Badges: - Relative Date: 3 months ago Absolute Date: Thu 11 Dec 2025 16:55 Selected Answer: - Upvotes: 1

https://eastus.api.cognitive.microsoft.com/translate?api-version=3.0&to=es

api.cognitive.microsofttranslator.com <-- this is a Global endpoints / multi-geo front doors
api-nam.cognitive.microsofttranslator.com <-- this is a Global endpoints / multi-geo front doors
api-nam.cognitiveservices.azure.com <-- this is a Global endpoints / multi-geo front doors

These are global front door endpoints, which:
- Do not guarantee single-region processing
- Route internally based on load/performance
- Are not compliant with strict data sovereignty requirements

While eastus.api.cognitive.microsoft.com <-- Region-specific endpoint which
✔ Guarantee regional data processing
✔ Required for data residency / sovereignty

Comment 2

ID: 1388231 User: Mattt Badges: - Relative Date: 12 months ago Absolute Date: Thu 13 Mar 2025 09:47 Selected Answer: - Upvotes: 1

Repeated more than 5 times in the question bank.

Comment 3

ID: 1225503 User: NagaoShingo Badges: - Relative Date: 1 year, 3 months ago Absolute Date: Fri 06 Dec 2024 15:44 Selected Answer: - Upvotes: 2

1. api-nam.cognitive.microsofttranslator.com
2. translate

Comment 4

ID: 1215789 User: takaimomoGcup Badges: - Relative Date: 1 year, 3 months ago Absolute Date: Fri 22 Nov 2024 17:08 Selected Answer: - Upvotes: 3

api-nam.cognitive.microsofttranslator.com
translate

Comment 5

ID: 1211305 User: abelarda Badges: - Relative Date: 1 year, 3 months ago Absolute Date: Thu 14 Nov 2024 11:43 Selected Answer: - Upvotes: 1

Why not api-nam.cognitiveservice.azure.com/translate?

Comment 6

ID: 1139200 User: evangelist Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Sat 03 Aug 2024 12:01 Selected Answer: - Upvotes: 3

nam=North America

Comment 7

ID: 1081641 User: sca88 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Mon 27 May 2024 14:24 Selected Answer: - Upvotes: 1

The response is correct
1. api-nam.cognitive.microsofttranslator.com
2. translate

Comment 8

ID: 922830 User: 973b658 Badges: - Relative Date: 2 years, 2 months ago Absolute Date: Thu 14 Dec 2023 10:09 Selected Answer: - Upvotes: 2

It is true.

20. AI-102 Topic 3 Question 47

Sequence
47
Discussion ID
108116
Source URL
https://www.examtopics.com/discussions/microsoft/view/108116-exam-ai-102-topic-3-question-47-discussion/
Posted By
EliteAllen
Posted At
May 1, 2023, 2:33 p.m.

Question

You have a text-based chatbot.

You need to enable content moderation by using the Text Moderation API of Content Moderator.

Which two service responses should you use? Each correct answer presents part of the solution.

NOTE: Each correct selection is worth one point.

  • A. personal data
  • B. the adult classification score
  • C. text classification
  • D. optical character recognition (OCR)
  • E. the racy classification score

Suggested Answer

BE

Answer Description Click to expand


Community Answer Votes

Comments 12 comments Click to expand

Comment 1

ID: 886225 User: EliteAllen Badges: Highly Voted Relative Date: 2 years, 10 months ago Absolute Date: Mon 01 May 2023 14:33 Selected Answer: BE Upvotes: 9

B. the adult classification score
E. the racy classification score

To enable content moderation in a text-based chatbot using the Text Moderation API of Content Moderator, you should use the adult classification score (B) and the racy classification score (E). These scores will help you determine if the content is adult or racy in nature, enabling you to take appropriate action for moderation purposes.

Comment 2

ID: 903722 User: DDD6 Badges: Highly Voted Relative Date: 2 years, 9 months ago Absolute Date: Mon 22 May 2023 07:12 Selected Answer: - Upvotes: 5

Answer is correct.
The reference URL:
https://learn.microsoft.com/en-us/azure/cognitive-services/content-moderator/api-reference

Comment 3

ID: 1698509 User: Madhusmita Badges: Most Recent Relative Date: 3 months ago Absolute Date: Wed 10 Dec 2025 10:19 Selected Answer: BE Upvotes: 2

The two correct answers are: B. The adult classification score - Core content moderation for explicit content E. The racy classification score - Core content moderation for suggestive content

These two classification scores are the primary responses from the Content Moderator Text API for identifying inappropriate content in text-based conversations, providing the essential content filtering capabilities needed for chatbot moderation.

Comment 4

ID: 1362558 User: gyaansastra Badges: - Relative Date: 1 year ago Absolute Date: Thu 27 Feb 2025 15:31 Selected Answer: BE Upvotes: 2

* B. the adult classification score
* E. the racy classification score

These two responses from the Content Moderator API will help you detect and filter potentially inappropriate content in your chatbot:

The adult classification score provides a probability rating for sexually explicit or adult content, while the racy classification score indicates potentially suggestive or provocative content that might not be explicitly adult but could still be inappropriate depending on your audience.

Together, these scores allow you to implement appropriate content filtering rules based on your chatbot's usage context and audience requirements.

Comment 5

ID: 1247886 User: krzkrzkra Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sun 14 Jul 2024 19:44 Selected Answer: AC Upvotes: 1

AC is the answer.

Comment 6

ID: 1229219 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 12 Jun 2024 16:15 Selected Answer: AC Upvotes: 2

AC is answer.

Comment 7

ID: 1202917 User: Jimmy1017 Badges: - Relative Date: 1 year, 10 months ago Absolute Date: Sat 27 Apr 2024 04:02 Selected Answer: - Upvotes: 4

C. Text classification
A. Personal data

Here's why these two are important:

Text classification: This functionality allows the API to analyze the text for potentially harmful content like hate speech, bullying, threats, etc. This is crucial for ensuring a safe and positive environment in your chatbot.
Personal data: This helps identify and potentially mask sensitive information like names, addresses, phone numbers, etc., which users might accidentally or intentionally reveal during conversations. This protects user privacy.
Let's break down the other options:

B. The adult classification score: This functionality is not available in the Text Moderation API. It's likely part of the Content Moderator's Image Moderation API for identifying inappropriate visuals.
D. Optical character recognition (OCR): This is not relevant for text-based chatbots as OCR deals with converting images containing text into machine-readable format.
E. The racy classification score: Similar to adult classification score, this functionality is likely intended for image moderation and not directly applicable to text analysis.

Comment 8

ID: 1139199 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 03 Feb 2024 12:58 Selected Answer: AC Upvotes: 1

The appropriate service responses for content moderation using the Text Moderation API of Content Moderator are:

A. Personal Data and C. Text Classification.

These features help identify sensitive information and categorize text content based on its potential appropriateness, including detecting profanity, personal data, and classifying text into categories related to potentially undesired content​

PLEASE DO NOT SELECT A,C!!!

Comment 9

ID: 920314 User: ziggy1117 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Sun 11 Jun 2023 01:30 Selected Answer: AC Upvotes: 3

answer is correct:
https://westus.dev.cognitive.microsoft.com/docs/services/57cf753a3f9b070c105bd2c1/operations/57cf753a3f9b070868a1f66f

request parameters: PII and classify

Comment 10

ID: 916052 User: ziggy1117 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Tue 06 Jun 2023 10:31 Selected Answer: - Upvotes: 2

answer is correct:
https://westus.dev.cognitive.microsoft.com/docs/services/57cf753a3f9b070c105bd2c1/operations/57cf753a3f9b070868a1f66f

request parameters: PII and classify

Comment 11

ID: 910961 User: EliteAllen Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Wed 31 May 2023 09:22 Selected Answer: AC Upvotes: 4

Based on the information from the official Azure Cognitive Services documentation, it seems that the Text Moderation API indeed returns profanity terms and personal data (A), which can be used for content moderation. It also performs text classification (C), which can be used to categorize and filter content.

So, the correct answers according to the official Azure documentation are A. personal data and C. text classification.

Comment 12

ID: 909811 User: examworld Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Tue 30 May 2023 03:45 Selected Answer: - Upvotes: 1

Image Moderation API

Scan images and detect potential adult and racy content by using tags, confidence scores, and other extracted information.

Text Moderation API

Scan text content. Profanity terms and personal data are returned.

21. AI-102 Topic 1 Question 57

Sequence
49
Discussion ID
122901
Source URL
https://www.examtopics.com/discussions/microsoft/view/122901-exam-ai-102-topic-1-question-57-discussion/
Posted By
sl_mslconsulting
Posted At
Oct. 9, 2023, 12:09 a.m.

Question

You are building an app that will use the Speech service.

You need to ensure that the app can authenticate to the service by using a Microsoft Azure Active Directory (Azure AD), part of Microsoft Entra, token.

Which two actions should you perform? Each correct answer presents part of the solution.

NOTE: Each correct selection is worth one point.

  • A. Enable a virtual network service endpoint.
  • B. Configure a custom subdomain.
  • C. Request an X.509 certificate.
  • D. Create a private endpoint.
  • E. Create a Conditional Access policy.

Suggested Answer

BD

Answer Description Click to expand


Community Answer Votes

Comments 18 comments Click to expand

Comment 1

ID: 1130081 User: radzio123 Badges: Highly Voted Relative Date: 2 years, 1 month ago Absolute Date: Tue 23 Jan 2024 23:49 Selected Answer: BD Upvotes: 19

https://learn.microsoft.com/en-us/azure/ai-services/speech-service/role-based-access-control#authentication-with-keys-and-tokens

Comment 2

ID: 1261943 User: marras Badges: Highly Voted Relative Date: 1 year, 7 months ago Absolute Date: Wed 07 Aug 2024 07:12 Selected Answer: BD Upvotes: 6

under MS Documentation link: https://learn.microsoft.com/en-us/azure/ai-services/speech-service/role-based-access-control#authentication-with-keys-and-token
you will find: "To authenticate with a Microsoft Entra token, the Speech resource must have a custom subdomain and use a private endpoint. The Speech service uses custom subdomains with private endpoints only."

Comment 3

ID: 1629872 User: Madhusmita Badges: Most Recent Relative Date: 3 months, 1 week ago Absolute Date: Tue 02 Dec 2025 13:04 Selected Answer: BE Upvotes: 1

The correct answers are B. Configure a custom subdomain and E. Create a Conditional Access policy.

These two actions work together to enable Azure AD token authentication: the custom subdomain provides the technical capability for the Speech service to accept Azure AD tokens, while the Conditional Access policy provides the Azure AD-side configuration to control how and when those tokens are issued and validated.

Comment 4

ID: 1628812 User: harshad883 Badges: - Relative Date: 3 months, 2 weeks ago Absolute Date: Thu 27 Nov 2025 19:02 Selected Answer: BE Upvotes: 1

B and E

Comment 5

ID: 1628380 User: harshad883 Badges: - Relative Date: 3 months, 2 weeks ago Absolute Date: Tue 25 Nov 2025 19:46 Selected Answer: BE Upvotes: 1

B & E
B. Configure a custom subdomain

Required because Azure AD authentication for Cognitive Services works only with a custom subdomain endpoint (e.g., https://myspeech.cognitiveservices.azure.com).
This ensures token-based authentication instead of subscription key.

E. Create a Conditional Access policy

Optional but recommended for enforcing security and compliance when using Azure AD authentication.
Allows you to control access conditions (e.g., MFA, location-based restrictions).

Comment 6

ID: 1283986 User: 9c652a0 Badges: - Relative Date: 1 year, 5 months ago Absolute Date: Sun 15 Sep 2024 10:30 Selected Answer: BD Upvotes: 2

To authenticate with a Microsoft Entra token, the Speech resource must have a custom subdomain and use a private endpoint. The Speech service uses custom subdomains with private endpoints only.

Comment 7

ID: 1246400 User: SAMBIT Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Fri 12 Jul 2024 03:05 Selected Answer: - Upvotes: 1

Options A, B, and C are not directly related to Azure AD token authentication for the Speech service. Option A (enabling a virtual network service endpoint) is about network connectivity, option B (configuring a custom subdomain) is unrelated to authentication, and option C (requesting an X.509 certificate) is not necessary for Azure AD token-based authentication. (DE)

Comment 8

ID: 1233527 User: cxw23 Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Thu 20 Jun 2024 10:15 Selected Answer: - Upvotes: 3

BD.
To authenticate with a Microsoft Entra token, the Speech resource must have a custom subdomain and use a private endpoint. The Speech service uses custom subdomains with private endpoints only.

Comment 8.1

ID: 1265566 User: anto69 Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Wed 14 Aug 2024 10:15 Selected Answer: - Upvotes: 1

Writing down the question in a different way ChatGPT comes to this conclusion too

Comment 9

ID: 1230905 User: gary_cooper Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sat 15 Jun 2024 13:10 Selected Answer: BD Upvotes: 2

To authenticate with a Microsoft Entra token, the Speech resource must have a custom subdomain and use a private endpoint. The Speech service uses custom subdomains with private endpoints only.

Comment 10

ID: 1223018 User: ManikNeelan Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sun 02 Jun 2024 06:25 Selected Answer: - Upvotes: 3

To authenticate with Speech resource keys, all you need is the key and region. To authenticate with a Microsoft Entra token, the Speech resource must have a custom subdomain and use a private endpoint. The Speech service uses custom subdomains with private endpoints only.

Comment 11

ID: 1221490 User: Tempeck Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 30 May 2024 11:58 Selected Answer: - Upvotes: 1

B. Configure a custom subdomain: This is necessary because Azure AD authentication requires a custom domain for the Speech service.
D. Create a private endpoint: This ensures secure communication between the app and the Speech service.
These actions align with the best practices for securing Azure services and enable the app to use Azure AD tokens for authentication.

Comment 12

ID: 1219909 User: PeteColag Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Tue 28 May 2024 02:31 Selected Answer: - Upvotes: 1

When you authenticate using Microsoft Entra ID (formerly Azure Active Directory) tokens, having a private endpoint along with a custom subdomain ensures full feature access. Without a private endpoint, some features may be limited. Specifically, if your Speech resource uses a Microsoft Entra token without a custom subdomain and private endpoint, you may encounter limitations, such as not being able to use custom speech models to transcribe audio files​

Comment 13

ID: 1219908 User: PeteColag Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Tue 28 May 2024 02:28 Selected Answer: BD Upvotes: 1

B. Configure a custom subdomain: This is necessary when using Azure AD authentication with a private endpoint. However, it's also a requirement for using Azure AD authentication in general with the Speech service.
C. Request an X.509 certificate: This is not needed for Azure AD token-based authentication. It's typically used for client certificate authentication.

Hence, B is absolutely required. The others are optional.

Comment 14

ID: 1217601 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 16:40 Selected Answer: BD Upvotes: 2

Wrong again. It's frustrating.

Comment 15

ID: 1213781 User: reiwanotora Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sun 19 May 2024 14:55 Selected Answer: BD Upvotes: 1

B and D.

Comment 16

ID: 1195169 User: GirishGururani Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Sun 14 Apr 2024 02:14 Selected Answer: - Upvotes: 2

correct answer : B and D
https://learn.microsoft.com/en-us/azure/ai-services/speech-service/role-based-access-control#authentication-with-keys-and-tokens

To authenticate with a Microsoft Entra token, the Speech resource must have a custom subdomain and use a private endpoint. The Speech service uses custom subdomains with private endpoints only.

Comment 17

ID: 1190658 User: Murtuza Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Sun 07 Apr 2024 01:15 Selected Answer: BD Upvotes: 2

B and D is the correct answer

22. AI-102 Topic 1 Question 29

Sequence
50
Discussion ID
75222
Source URL
https://www.examtopics.com/discussions/microsoft/view/75222-exam-ai-102-topic-1-question-29-discussion/
Posted By
JTWang
Posted At
May 6, 2022, 3:38 a.m.

Question

You have a Language Understanding resource named lu1.
You build and deploy an Azure bot named bot1 that uses lu1.
You need to ensure that bot1 adheres to the Microsoft responsible AI principle of inclusiveness.
How should you extend bot1?

  • A. Implement authentication for bot1.
  • B. Enable active learning for lu1.
  • C. Host lu1 in a container.
  • D. Add Direct Line Speech to bot1.

Suggested Answer

D

Answer Description Click to expand


Community Answer Votes

Comments 15 comments Click to expand

Comment 1

ID: 636036 User: RamonKaus Badges: Highly Voted Relative Date: 3 years, 1 month ago Absolute Date: Tue 24 Jan 2023 15:11 Selected Answer: D Upvotes: 33

I disagree with JTWang because Direct Line Speech lets the bot speak outloud. This supports Microsoft's goal of AI Inclusiveness because you can include people with disabilities.

Comment 1.1

ID: 650361 User: ninjia Badges: - Relative Date: 3 years ago Absolute Date: Wed 22 Feb 2023 19:50 Selected Answer: - Upvotes: 2

Agreed the answer should be D, which would people who don't text.
Active learning would improve the overall quality, but not necessary make it more inclusive.

Comment 2

ID: 597506 User: JTWang Badges: Highly Voted Relative Date: 3 years, 4 months ago Absolute Date: Sun 06 Nov 2022 04:38 Selected Answer: - Upvotes: 9

Answer is B: user can participate the content of QnA knowledge base
https://docs.microsoft.com/en-us/azure/cognitive-services/qnamaker/how-to/use-active-learning
The Active learning suggestions feature allows you to improve the quality of your knowledge base by suggesting alternative questions, based on user-submissions, to your question and answer pair. You review those suggestions, either adding them to existing questions or rejecting them.

Comment 3

ID: 1629541 User: hasyatan Badges: Most Recent Relative Date: 3 months, 1 week ago Absolute Date: Mon 01 Dec 2025 12:41 Selected Answer: D Upvotes: 1

D is correct

Comment 4

ID: 1600040 User: tanu03 Badges: - Relative Date: 6 months, 3 weeks ago Absolute Date: Wed 20 Aug 2025 03:32 Selected Answer: D Upvotes: 1

https://azure.microsoft.com/en-us/blog/direct-line-with-speech-now-available-in-preview/ - This will enable endusers to converse with bots without having to type.

Comment 5

ID: 1217628 User: nanaw770 Badges: - Relative Date: 1 year, 3 months ago Absolute Date: Sun 24 Nov 2024 17:59 Selected Answer: D Upvotes: 1

Add Direct Line Speech

Comment 6

ID: 1200407 User: brajesh8684 Badges: - Relative Date: 1 year, 4 months ago Absolute Date: Wed 23 Oct 2024 01:37 Selected Answer: - Upvotes: 1

B. Enable active learning for lu1.

Explanation:

Active learning is a mechanism that allows the Language Understanding (LU) service to improve over time by actively seeking feedback from users.
By enabling active learning for lu1, the bot can continuously learn from interactions with users, improving its understanding and inclusiveness over time.
While options like implementing authentication, hosting lu1 in a container, or adding Direct Line Speech may be relevant for other aspects of the bot's functionality, enabling active learning directly addresses the goal of improving inclusiveness through continuous learning and adaptation.

Comment 7

ID: 1197199 User: trato Badges: - Relative Date: 1 year, 4 months ago Absolute Date: Thu 17 Oct 2024 13:44 Selected Answer: - Upvotes: 1

D. Add Direct Line Speech to bot1.
A, B, and C might be useful for other purposes, they do not directly contribute to the principle of inclusiveness. Authentication (A) is more about security, active learning (B) is about improving the bot’s understanding over time, and hosting in a container. (C) is about deployment flexibility. None of these directly enhance the inclusiveness of the bot.

Comment 8

ID: 1147852 User: evangelist Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Mon 12 Aug 2024 06:23 Selected Answer: D Upvotes: 2

Other options apparently have nothing to do with "inclusiveness"

Comment 9

ID: 1140066 User: Nihilist11 Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Sun 04 Aug 2024 12:12 Selected Answer: - Upvotes: 2

It was in today's exam - Feb-04-2023

Comment 10

ID: 839688 User: RAN_L Badges: - Relative Date: 2 years, 5 months ago Absolute Date: Fri 15 Sep 2023 08:33 Selected Answer: D Upvotes: 3

Enabling Direct Line Speech for bot1 would improve accessibility by providing users with the ability to interact with the bot using natural language and speech. Therefore, D is a valid option to ensure that bot1 adheres to the Microsoft responsible AI principle of inclusiveness.

Comment 11

ID: 718983 User: Shaka711 Badges: - Relative Date: 2 years, 10 months ago Absolute Date: Mon 15 May 2023 17:24 Selected Answer: D Upvotes: 1

Direct line speech enables the bots to speak

Comment 12

ID: 635080 User: Eltooth Badges: - Relative Date: 3 years, 1 month ago Absolute Date: Sun 22 Jan 2023 10:55 Selected Answer: B Upvotes: 1

B is correct answer.

Comment 12.1

ID: 671435 User: STH Badges: - Relative Date: 2 years, 12 months ago Absolute Date: Fri 17 Mar 2023 13:48 Selected Answer: - Upvotes: 4

No, because even if enabling active learning could improve the bot answers over time with more accurate understanding of questions... it does not help people that are not able to use a textual bot.

Question is about how to include larger audience, like blind users or others, not how to improve bot quality

so D is the right one

Comment 13

ID: 618459 User: sdokmak Badges: - Relative Date: 3 years, 2 months ago Absolute Date: Mon 19 Dec 2022 03:04 Selected Answer: B Upvotes: 1

from JTWang's comment

23. AI-102 Topic 1 Question 2

Sequence
55
Discussion ID
74339
Source URL
https://www.examtopics.com/discussions/microsoft/view/74339-exam-ai-102-topic-1-question-2-discussion/
Posted By
kiassi1998
Posted At
April 24, 2022, 5:32 p.m.

Question

DRAG DROP -
You plan to use a Language Understanding application named app1 that is deployed to a container.
App1 was developed by using a Language Understanding authoring resource named lu1.
App1 has the versions shown in the following table.
image
You need to create a container that uses the latest deployable version of app1.
Which three actions should you perform in sequence? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.
Select and Place:
image

Suggested Answer

image
Answer Description Click to expand


Comments 18 comments Click to expand

Comment 1

ID: 632001 User: Eltooth Badges: Highly Voted Relative Date: 3 years, 7 months ago Absolute Date: Sat 16 Jul 2022 06:27 Selected Answer: - Upvotes: 27

Select v1.1
Export
Run and mount

Comment 2

ID: 602919 User: DingDongSingSong Badges: Highly Voted Relative Date: 3 years, 9 months ago Absolute Date: Tue 17 May 2022 14:36 Selected Answer: - Upvotes: 5

Why select v1.1 and not v1.0. Deployable model is a published model, not trained model.

Comment 2.1

ID: 617540 User: sdokmak Badges: - Relative Date: 3 years, 8 months ago Absolute Date: Fri 17 Jun 2022 07:10 Selected Answer: - Upvotes: 12

from solution link: "The Language Understanding (LUIS) container loads your trained or published Language Understanding model." Trained alone is deployable, v1.1

Comment 3

ID: 1627791 User: nkcodescribbler Badges: Most Recent Relative Date: 3 months, 2 weeks ago Absolute Date: Sun 23 Nov 2025 02:23 Selected Answer: - Upvotes: 3

Select v1.0 of app1, since it is the latest trained AND published version, and deployable one.

I believe
1) Select v1.0
2) Export
3) Run and mount

Comment 4

ID: 1572634 User: man5484 Badges: - Relative Date: 9 months, 2 weeks ago Absolute Date: Tue 27 May 2025 11:42 Selected Answer: - Upvotes: 2

Select V1.1 of app1
→ It’s trained, and it’s the most recent usable version.

Export the model by using the Export for container (GZIP) option
→ Only this format works for containers (not JSON).

Run a container and mount the model file
→ You mount the .gz model file into the container at runtime.

Comment 5

ID: 1289379 User: shanakrs Badges: - Relative Date: 1 year, 5 months ago Absolute Date: Thu 26 Sep 2024 12:10 Selected Answer: - Upvotes: 2

Answer

1. Select v1.1
2. Export as GZIP
3. Run and mount

Export versioned app's package from LUIS portal
The versioned app's package is available from the Versions list page.

Sign on to the LUIS Azure portal.
Select the app in the list.
Select Manage in the app's navigation bar.
Select Versions in the left navigation bar.
Select the checkbox to the left of the version name in the list.
Select the Export item from the contextual toolbar above the list.
Select Export for container (GZIP).
The package is downloaded from the browser.

Comment 6

ID: 1227037 User: reiwanotora Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sun 09 Jun 2024 05:45 Selected Answer: - Upvotes: 3

1. Select v1.1
2. Export as GZIP
3. Run and mount

Comment 7

ID: 1221551 User: aelhalawany Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 30 May 2024 13:45 Selected Answer: - Upvotes: 2

Select v1.0 of app1.
Export the model by using the Export for containers (GZIP) option.
Run a container and mount the model file.

Comment 8

ID: 1209921 User: demonite Badges: - Relative Date: 1 year, 10 months ago Absolute Date: Sat 11 May 2024 21:07 Selected Answer: - Upvotes: 2

Sign on to the LUIS portal.
Select the app in the list.
Select Manage in the app's navigation bar.
Select Versions in the left navigation bar.
Select the checkbox to the left of the version name in the list.
Select the Export item from the contextual toolbar above the list.
Select Export for container (GZIP).
The package is downloaded from the browser.

Comment 9

ID: 1194704 User: sivapolam90 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Sat 13 Apr 2024 08:46 Selected Answer: - Upvotes: 1

1. Export as GZIP
2. Select v1.1
3. Run and mount

Comment 10

ID: 1182261 User: varinder82 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Mon 25 Mar 2024 07:42 Selected Answer: - Upvotes: 2

Final Anser:
1. Select v1.1
2. Export as GZIP
3.. Run and mount

Comment 11

ID: 1182236 User: varinder82 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Mon 25 Mar 2024 07:01 Selected Answer: - Upvotes: 2

Select v1.1
Export
Run and mount

Comment 12

ID: 1146998 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sun 11 Feb 2024 06:57 Selected Answer: - Upvotes: 4

you have to select first and then export and run:
Select v1.1 of app1: This step is necessary to ensure you are working with the correct version of the application that you intend to deploy.
Export the model by using the Export for containers (GZIP) option: This action is crucial because deploying an app to a container requires the model to be in a format that is optimized for container deployment, which is typically a GZIP file that includes all necessary components of the Language Understanding model.
Run a container and mount the model file: Once you have the model exported in the correct format, you run a container instance and mount the exported model file to it. This allows the container to use the Language Understanding model version 1.1 for processing natural language inputs.

Comment 13

ID: 1107800 User: costas99 Badges: - Relative Date: 2 years, 2 months ago Absolute Date: Thu 28 Dec 2023 14:28 Selected Answer: - Upvotes: 3

hi, hope it helps someone looking for the comments section. From the documentation, using a LUIS container it's possible to mount a gzip VERSIONED (just trained model) or PUBLISHED (trained and published) model. In the question it says "latest deployable version", so the app version to export as GZIP is v1.1. So just trained model is deployable, but also deployed model is deployable in a LUIS contrainer

Comment 14

ID: 985232 User: sepenthi Badges: - Relative Date: 2 years, 6 months ago Absolute Date: Sat 19 Aug 2023 15:17 Selected Answer: - Upvotes: 1

The Language Understanding (LUIS) container loads your trained or published

Comment 15

ID: 976870 User: sl_mslconsulting Badges: - Relative Date: 2 years, 7 months ago Absolute Date: Wed 09 Aug 2023 19:05 Selected Answer: - Upvotes: 1

I checked the link https://learn.microsoft.com/en-us/azure/ai-services/luis/luis-container-howto?tabs=v3 and it says you can no longer create new LUIS resources. Wonder why it still appeared in the exam per acsoma's comment.

Comment 16

ID: 611182 User: jampack Badges: - Relative Date: 3 years, 9 months ago Absolute Date: Fri 03 Jun 2022 19:17 Selected Answer: - Upvotes: 2

Select v1.1 is not published a version. It should be Select v1.0

Comment 17

ID: 604000 User: ThomasKong Badges: - Relative Date: 3 years, 9 months ago Absolute Date: Thu 19 May 2022 17:05 Selected Answer: - Upvotes: 1

how come can export first before select version ? imagination ?

24. AI-102 Topic 7 Question 15

Sequence
60
Discussion ID
143643
Source URL
https://www.examtopics.com/discussions/microsoft/view/143643-exam-ai-102-topic-7-question-15-discussion/
Posted By
Toby86
Posted At
July 9, 2024, 7:32 p.m.

Question

HOTSPOT
-

In Azure OpenAI Studio, you are prototyping a chatbot by using Chat playground.

You need to configure the chatbot to meet the following requirements:

• Reduce the repetition of words in conversations.
• Reduce the randomness of each response.

Which two parameters should you modify? To answer, select the appropriate parameters in the answer area.

NOTE: Each correct answer is worth one point.

image

Suggested Answer

image
Answer Description Click to expand


Comments 4 comments Click to expand

Comment 1

ID: 1291619 User: fawzi008 Badges: Highly Voted Relative Date: 1 year, 5 months ago Absolute Date: Mon 30 Sep 2024 18:24 Selected Answer: - Upvotes: 10

-temperature
-frequency penalty

Comment 2

ID: 1245013 User: Toby86 Badges: Highly Voted Relative Date: 1 year, 8 months ago Absolute Date: Tue 09 Jul 2024 19:32 Selected Answer: - Upvotes: 7

Reduce repetition:
Increase Frequency penalty to suppress repetition of words:
Frequency Penalty modifies the probability of words seen frequently during training, making them less likely.

NOT Presence Penalty:
Presence Penalty modifies the probability of words in the input text, making them less likely to repeat in the output.

Reduce Randomness of each response:
Decrease the Temperature to make the answers more deterministic

Comment 3

ID: 1608926 User: cesaraldo Badges: Most Recent Relative Date: 5 months, 4 weeks ago Absolute Date: Sun 14 Sep 2025 08:12 Selected Answer: - Upvotes: 1

Frequency penalty -> prevents repetition of words.
Both Temperature and Top P will redudce the randomness of the response: Temperature will "smooth" the probability distribution, with higher values making rare tokens more likely; Top-P wil filter out low-probability tokens entirely. Both answeres should be accepted, though I will stick with Top-p when doing the exam.

I hate finding so many of these questions, where there is really no RIGHT answer :(

Comment 4

ID: 1275559 User: testmaillo020 Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Sat 31 Aug 2024 13:12 Selected Answer: - Upvotes: 2

1. Frequency penalty
- This parameter reduces the model's likelihood of repeating the same word or phrase. Increasing the frequency penalty will help in reducing the repetition of words in the chatbot's responses.

2. Top P (or possibly Temperature)
- To reduce the randomness of each response, you can adjust the Top P parameter, which controls the diversity of the tokens generated by narrowing down the token pool to the most likely ones. Lowering the Top P value (or Temperature) will result in more deterministic responses, thus reducing randomness.

25. AI-102 Topic 3 Question 35

Sequence
61
Discussion ID
111618
Source URL
https://www.examtopics.com/discussions/microsoft/view/111618-exam-ai-102-topic-3-question-35-discussion/
Posted By
ziggy1117
Posted At
June 9, 2023, 4:31 a.m.

Question

HOTSPOT
-

You are building a call handling system that will receive calls from French-speaking and German-speaking callers. The system must perform the following tasks:

• Capture inbound voice messages as text.
• Replay messages in English on demand.

Which Azure Cognitive Services services should you use? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

image

Suggested Answer

image
Answer Description Click to expand


Comments 5 comments Click to expand

Comment 1

ID: 1606021 User: dgowrishankar Badges: - Relative Date: 6 months, 1 week ago Absolute Date: Thu 04 Sep 2025 05:35 Selected Answer: - Upvotes: 2

1. Speech-to-text
2. Text-to-speech and Translator

This question was asked in Sep 2025 exam

Comment 2

ID: 1247878 User: krzkrzkra Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sun 14 Jul 2024 19:39 Selected Answer: - Upvotes: 1

1. Speech-to-text
2. Text-to-speech and Translator

Comment 3

ID: 1215793 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 22 May 2024 16:11 Selected Answer: - Upvotes: 2

Speech-to-text
Text-to-speech and Translator

Comment 4

ID: 1102036 User: Gvalli Badges: - Relative Date: 2 years, 2 months ago Absolute Date: Thu 21 Dec 2023 00:31 Selected Answer: - Upvotes: 2

A modified version of this was in the exam today.

Comment 5

ID: 918834 User: ziggy1117 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Fri 09 Jun 2023 04:31 Selected Answer: - Upvotes: 3

answer is correct

26. AI-102 Topic 5 Question 14

Sequence
64
Discussion ID
75647
Source URL
https://www.examtopics.com/discussions/microsoft/view/75647-exam-ai-102-topic-5-question-14-discussion/
Posted By
Isidro
Posted At
May 15, 2022, 8:36 a.m.

Question

You are building a chatbot by using the Microsoft Bot Framework Composer as shown in the exhibit. (Click the Exhibit tab.)
image
The chatbot contains a dialog named GetUserDetails. GetUserDetails contains a TextInput control that prompts users for their name.
The user input will be stored in a property named name.
You need to ensure that you can dispose of the property when the last active dialog ends.
Which scope should you assign to name?

  • A. dialog
  • B. user
  • C. turn
  • D. conversation

Suggested Answer

A

Answer Description Click to expand


Community Answer Votes

Comments 8 comments Click to expand

Comment 1

ID: 1576574 User: qqaws Badges: - Relative Date: 9 months ago Absolute Date: Wed 11 Jun 2025 16:08 Selected Answer: D Upvotes: 2

I think that it is "Conversation" because it must be retained until the "last active dialog ends". So the last dialog of the active conversation.

Comment 2

ID: 1229900 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 17:02 Selected Answer: A Upvotes: 1

A is answer.

Comment 3

ID: 1222689 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sat 01 Jun 2024 16:14 Selected Answer: A Upvotes: 1

A is right answer.

Comment 4

ID: 601934 User: Isidro Badges: - Relative Date: 3 years, 10 months ago Absolute Date: Sun 15 May 2022 08:36 Selected Answer: - Upvotes: 1

Shouldn't it be USER? If Dialog is selected, one the last dialog is concluded, the information will be lost.

Comment 4.1

ID: 610503 User: Moody_L Badges: - Relative Date: 3 years, 9 months ago Absolute Date: Thu 02 Jun 2022 10:35 Selected Answer: - Upvotes: 4

Dialog scope are retained until the "last active dialog ends". That's what the question asked for.
Ref: Under "Composer v1.x" tab, https://docs.microsoft.com/en-us/composer/concept-memory?tabs=v2x

Comment 4.1.1

ID: 648165 User: ninjia Badges: - Relative Date: 3 years, 6 months ago Absolute Date: Wed 17 Aug 2022 19:06 Selected Answer: - Upvotes: 2

Agreed. A. dialog

Comment 4.1.2

ID: 1605684 User: cesaraldo Badges: - Relative Date: 6 months, 1 week ago Absolute Date: Tue 02 Sep 2025 21:20 Selected Answer: - Upvotes: 1

So it was true for v1.x, but not for v2.x? (The wording changed from "dialog is associated with the dialog stack. Properties in the dialog scope are retained until the last active dialog ends." in v1 to "The dialog scope associates properties with the active dialog. Properties in the dialog scope are retained until the dialog ends." in v2. My guess is that was a correction on the documentation because otherwise, the "dialog" scope would not make any sense. I'll go for "conversation".

Comment 4.2

ID: 1211554 User: abelarda Badges: - Relative Date: 1 year, 10 months ago Absolute Date: Tue 14 May 2024 18:19 Selected Answer: - Upvotes: 1

True, when it says "when the dialog ends", one could say that it means "after the dialog ends", or "until". Why are you so sure?

27. AI-102 Topic 7 Question 1

Sequence
65
Discussion ID
75514
Source URL
https://www.examtopics.com/discussions/microsoft/view/75514-exam-ai-102-topic-7-question-1-discussion/
Posted By
ManAtWorkAtNight
Posted At
May 12, 2022, 10:25 a.m.

Question

You are developing the chatbot.
You create the following components:
✑ A QnA Maker resource
✑ A chatbot by using the Azure Bot Framework SDK
You need to integrate the components to meet the chatbot requirements.
Which property should you use?

  • A. QnAMakerOptions.StrictFilters
  • B. QnADialogResponseOptions.CardNoMatchText
  • C. QnAMakerOptions.RankerType
  • D. QnAMakerOptions.ScoreThreshold

Suggested Answer

D

Answer Description Click to expand


Community Answer Votes

Comments 16 comments Click to expand

Comment 1

ID: 600530 User: ManAtWorkAtNight Badges: Highly Voted Relative Date: 3 years, 10 months ago Absolute Date: Thu 12 May 2022 10:25 Selected Answer: - Upvotes: 25

Answer should be D. QnAMakerOptions.ScoreThreshold
Technical Requirements says "AI solution responses must have a confidence score that is equal to or greater than 70 percent" and "When the response confidence score is low, ensure that the chatbot can provide other response options to the customers"
https://docs.microsoft.com/en-us/azure/cognitive-services/qnamaker/concepts/confidence-score#set-threshold

Comment 2

ID: 1603127 User: 395d135 Badges: Most Recent Relative Date: 6 months, 2 weeks ago Absolute Date: Wed 27 Aug 2025 09:30 Selected Answer: C Upvotes: 1

Answer should be C. QnAMakerOptions.RankerType

Comment 3

ID: 1284585 User: mustafaalhnuty Badges: - Relative Date: 1 year, 5 months ago Absolute Date: Mon 16 Sep 2024 12:08 Selected Answer: D Upvotes: 2

D 100%

Comment 4

ID: 1252197 User: anto69 Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Sun 21 Jul 2024 05:28 Selected Answer: D Upvotes: 1

QnAMakerOptions.ScoreThreshold, others make no sense

Comment 5

ID: 1246794 User: krzkrzkra Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Fri 12 Jul 2024 15:44 Selected Answer: - Upvotes: 1

Selected Answer: D

Comment 6

ID: 1235227 User: HaraTadahisa Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 08:59 Selected Answer: D Upvotes: 1

I say this answer is D.
I assume this question is part of the lab type question.

Comment 7

ID: 1230481 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 14 Jun 2024 14:21 Selected Answer: D Upvotes: 1

D is the correct answer.

Comment 8

ID: 1181934 User: Murtuza Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Sun 24 Mar 2024 20:23 Selected Answer: D Upvotes: 1

D is correct

Comment 9

ID: 1145964 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 10 Feb 2024 08:09 Selected Answer: D Upvotes: 2

D. QnAMakerOptions.ScoreThreshold

This property allows you to set a minimum confidence score for answers. If an answer's confidence score is below this threshold, you can trigger alternative actions, such as providing different response options or escalating to a human agent, directly addressing the specified chatbot requirements.

Comment 10

ID: 1133353 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 27 Jan 2024 14:05 Selected Answer: D Upvotes: 2

QnAMakerOptions.ScoreThreshold is the property that controls the minimum confidence score required for a QnA Maker response to be considered valid. This directly addresses the requirement that AI responses must have a confidence score of at least 70%.
By setting QnAMakerOptions.ScoreThreshold to 0.7 (or 70%), you ensure that the chatbot only provides answers from QnA Maker that have a high level of confidence, meeting the specified requirement.

Comment 11

ID: 1057032 User: propanther Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Sun 29 Oct 2023 18:30 Selected Answer: - Upvotes: 1

Complete chatbot constraints from Case Study:

Contoso identifies the following requirements for the chatbot:
* Provide customers with answers to the FAQs.
* Ensure that the customers can chat to a customer service agent.
* Ensure that the members of a group named Management-Accountants can approve the FAQs.
* Ensure that the members of a group named Consultant-Accountants can create and amend the FAQs.
* Ensure that the members of a group named the Agent-CustomerServices can browse the FAQs.
* Ensure that access to the customer service agents is managed by using Omnichannel for Customer Service.
* When the response confidence score is low. ensure that the chatbot can provide other response options to the customers.

Comment 12

ID: 984782 User: james2033 Badges: - Relative Date: 2 years, 6 months ago Absolute Date: Fri 18 Aug 2023 23:11 Selected Answer: C Upvotes: 2

"AI solution responses must have a confidence score that is equal to or greater than 70 percent." --> So choose "D. QnAMakerOptions.ScoreThreshold" . See https://learn.microsoft.com/en-us/dotnet/api/microsoft.bot.builder.ai.qna.qnamakeroptions.scorethreshold?view=botbuilder-dotnet-stable

Comment 13

ID: 937924 User: Tin_Tin Badges: - Relative Date: 2 years, 8 months ago Absolute Date: Thu 29 Jun 2023 12:55 Selected Answer: D Upvotes: 1

should be D

Comment 14

ID: 749191 User: SSJA Badges: - Relative Date: 3 years, 2 months ago Absolute Date: Sun 18 Dec 2022 22:08 Selected Answer: D Upvotes: 2

Correct answer is D

Comment 15

ID: 635939 User: Eltooth Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Sun 24 Jul 2022 10:46 Selected Answer: D Upvotes: 2

I'm going for answer D based on
https://docs.microsoft.com/en-us/azure/cognitive-services/qnamaker/concepts/best-practices

Comment 16

ID: 635805 User: RamonKaus Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Sun 24 Jul 2022 01:59 Selected Answer: D Upvotes: 1

https://www.google.com/url?sa=t&rct=j&q=&esrc=s&source=web&cd=&cad=rja&uact=8&ved=2ahUKEwilta3-tZD5AhXpGFkFHWGzAMAQFnoECAoQAw&url=https%3A%2F
Choosing Ranker type
" By default, QnA Maker searches through questions and answers. If you want to search through questions only, to generate an answer, use the RankerType=QuestionOnly in the POST body of the GenerateAnswer request. "
from
%2Fdocs.microsoft.com%2Fen-us%2Fazure%2Fcognitive-services%2Fqnamaker%2Fconcepts%2Fbest-practices&usg=AOvVaw3agfuSigl7RLnYFlhA-vb3

I think it is D because "Gets or sets the minimum score threshold, used to filter returned results."



https://docs.microsoft.com/en-us/dotnet/api/microsoft.bot.builder.ai.qna.qnamakeroptions.scorethreshold?view=botbuilder-dotnet-stable

28. AI-102 Topic 5 Question 45

Sequence
69
Discussion ID
112149
Source URL
https://www.examtopics.com/discussions/microsoft/view/112149-exam-ai-102-topic-5-question-45-discussion/
Posted By
973b658
Posted At
June 14, 2023, 10:05 a.m.

Question

DRAG DROP
-

You are building a bot.

You need to test the bot in the Bot Framework Emulator. The solution must ensure that you can debug the bot interactively.

Which three actions should you perform in sequence? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.

image

Suggested Answer

image
Answer Description Click to expand


Comments 7 comments Click to expand

Comment 1

ID: 1145889 User: evangelist Badges: Highly Voted Relative Date: 2 years, 1 month ago Absolute Date: Sat 10 Feb 2024 02:51 Selected Answer: - Upvotes: 5

The order should run bot app on local host first and then create trace activity and send trace activity

Comment 1.1

ID: 1564562 User: tech_rum Badges: - Relative Date: 10 months, 2 weeks ago Absolute Date: Tue 29 Apr 2025 03:29 Selected Answer: - Upvotes: 2

1 Run the bot app on local host
2 In the code for the bot, create a new trace activity
3 In the code for the bot, send a trace activity

For example: Would you open a website after you wrote HTML but before starting the web server?
No — you need the server running first.
Same idea here: Bot must be running first to interact with the Emulator.

Comment 2

ID: 1598403 User: supro Badges: Most Recent Relative Date: 6 months, 4 weeks ago Absolute Date: Sat 16 Aug 2025 05:03 Selected Answer: - Upvotes: 1

In the Bot Framework Emulator context, the exam question is specifically about the sequence of actions when testing/debugging:

You always start the bot locally first (otherwise Emulator cannot connect).

Then you must have trace activity code in place.

Finally, you send the trace activity so you can see debug info.

Comment 3

ID: 1266172 User: anto69 Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Thu 15 Aug 2024 05:39 Selected Answer: - Upvotes: 2

1. In code, create new trace activity
2. In code, send a trace activity
3. Run bot app on local host

Comment 4

ID: 1248526 User: krzkrzkra Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Mon 15 Jul 2024 20:39 Selected Answer: - Upvotes: 1

1. In code, create new trace activity
2. In code, send a trace activity
3. Run bot app on local host

Comment 5

ID: 1133306 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 27 Jan 2024 12:47 Selected Answer: - Upvotes: 1

Connect with the Bot Framework Emulator:

Use the Bot Framework Emulator to connect to your bot running on localhost. The default URL is typically http://localhost:3978/api/messages.
Test Your Bot: Interact with your bot via the Emulator to see the responses and the trace activities.

Comment 6

ID: 922873 User: 973b658 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Wed 14 Jun 2023 10:05 Selected Answer: - Upvotes: 1

It is true.

29. AI-102 Topic 5 Question 4

Sequence
73
Discussion ID
76495
Source URL
https://www.examtopics.com/discussions/microsoft/view/76495-exam-ai-102-topic-5-question-4-discussion/
Posted By
Moody_L
Posted At
June 2, 2022, 9:49 a.m.

Question

DRAG DROP -
You plan to build a chatbot to support task tracking.
You create a Language Understanding service named lu1.
You need to build a Language Understanding model to integrate into the chatbot. The solution must minimize development time to build the model.
Which four actions should you perform in sequence? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.
Select and Place:
image

Suggested Answer

image
Answer Description Click to expand

Step 1: Add a new application -

Create a new app -
1. Sign in to the LUIS portal with the URL of https://www.luis.ai.
2. Select Create new app.
3. Etc.
Step 2: Add example utterances.
In order to classify an utterance, the intent needs examples of user utterances that should be classified with this intent.

Step 3: Train the application -

Step 4: Publish the application -
In order to receive a LUIS prediction in a chat bot or other client application, you need to publish the app to the prediction endpoint.
Reference:
https://docs.microsoft.com/en-us/azure/cognitive-services/luis/tutorial-intents-only

Comments 14 comments Click to expand

Comment 1

ID: 652647 User: Nebary Badges: Highly Voted Relative Date: 3 years, 6 months ago Absolute Date: Sat 27 Aug 2022 19:34 Selected Answer: - Upvotes: 37

As per Udemy it is:
- Add a new application.
- Add the prebuilt domain ToDo.
- Train the application.
- Publish the application

Comment 1.1

ID: 1273585 User: JakeCallham Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Tue 27 Aug 2024 18:35 Selected Answer: - Upvotes: 2

Why would ToDO have anything todo with this? doesnt ake sense at all

Comment 2

ID: 633591 User: Eltooth Badges: Highly Voted Relative Date: 3 years, 7 months ago Absolute Date: Tue 19 Jul 2022 15:21 Selected Answer: - Upvotes: 12

Given answer is correct.

(Taken from Udemy practice exam question)
The order should be :
1. Add a new app,
2. Add example utterances,
3. Train the app,
4. Publish the app.

Comment 3

ID: 1579480 User: StelSen Badges: Most Recent Relative Date: 8 months, 3 weeks ago Absolute Date: Sun 22 Jun 2025 05:06 Selected Answer: - Upvotes: 1

Answer Area

Add a new application. (You need to start by creating a new LUIS application.)
Add the prebuilt domain ToDo. (This is crucial for minimizing development time as it provides pre-built intents and entities related to task management, which aligns with the "task tracking" requirement.)
Add example utterances. (Even with a prebuilt domain, you'll want to add your own specific examples to tailor the model to your chatbot's needs and user language.)
Train the application. (After adding data, you need to train the model so it can learn and make predictions.)
(Publish the application would be the next step after training and testing, but only four actions were requested.)

Comment 4

ID: 1363936 User: gyaansastra Badges: - Relative Date: 1 year ago Absolute Date: Sun 02 Mar 2025 13:19 Selected Answer: - Upvotes: 2

To build a Language Understanding model for a task tracking chatbot while minimizing development time, the correct sequence of four actions is:

1.Add a new application.
2. Add the prebuilt domain ToDo.
3. Add example utterances.
4. Train the application.

This sequence provides the most efficient path because:

- First, you create the application in the Language Understanding service
- Next, you leverage the prebuilt ToDo domain which comes with pre-configured intents and
entities for task tracking (saving significant development time)
- Then, you add custom example utterances to tailor the model to your specific needs
- Finally, you train the model so it can recognize the patterns in your utterances

Publishing would be the next step after these four, but since the question asks for exactly four actions in sequence, these represent the core model building process that minimizes development time

Comment 5

ID: 1296661 User: aa18a1a Badges: - Relative Date: 1 year, 5 months ago Absolute Date: Sun 13 Oct 2024 01:14 Selected Answer: - Upvotes: 2

In the case that this question actually requires 3 steps as opposed to 4 (per pyguy's comment at the bottom), I believe the correct sequence would be:
1. Add a new app
2. add the prebuilt domain ToDo
3. Publish the application
Prebuilt domains don't require training (it's ready to go out of the box, why would it need trained?)
It's also worth noting that LUIS is retired next year, and you actually haven't been able to make any new LUIS resources since April 2023. I really doubt this would show up on the current exams (at least I hope) since I see no mention of prebuilt domains in the CLU documentation. If I missed it somewhere, please correct me.

Comment 6

ID: 1229640 User: mon2002 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 07:43 Selected Answer: - Upvotes: 2

Add a new app,
Add example utterances,
Train the app,
Publish the app.

Comment 7

ID: 1222685 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sat 01 Jun 2024 16:11 Selected Answer: - Upvotes: 3

1. Add a new app
2. Add example utterances
3. Train the app
4. Publish the app

Comment 8

ID: 1184881 User: varinder82 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Thu 28 Mar 2024 17:20 Selected Answer: - Upvotes: 2

Final Answer
1. Add a new application
2. Add the prebuilt domain ToDo
3. Publish the application

Comment 9

ID: 1139854 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sun 04 Feb 2024 08:03 Selected Answer: - Upvotes: 5

Corrrection to answer:
1.add a new application
2. add the prebuilt domain ToDo
3. Train the application
4. Publish the application
Select a pre-built domain (e.g., "Todo"), which provides a predefined set of intents and entities designed specifically for scenarios like task tracking.
Train the application, even when using a pre-built domain. Training the application is necessary because it optimizes the model to better understand user queries specific to your data and use case.
Even if you choose a pre-built domain, training the application remains a necessary step. Pre-built domains provide a good starting point, including a set of related intents and entities, but through training, you can optimize the model's performance for your specific data and scenario.

Comment 10

ID: 610482 User: Moody_L Badges: - Relative Date: 3 years, 9 months ago Absolute Date: Thu 02 Jun 2022 09:49 Selected Answer: - Upvotes: 7

The question requires minimal time to build a model. Should we use prebuilt model with a custom domain ("ToDo") here?
Ref: https://docs.microsoft.com/en-us/azure/cognitive-services/luis/luis-reference-prebuilt-domains#prebuilt-domains-per-language

Comment 10.1

ID: 620245 User: sdokmak Badges: - Relative Date: 3 years, 8 months ago Absolute Date: Wed 22 Jun 2022 10:28 Selected Answer: - Upvotes: 3

True, but then we only have either 3 steps (don't need to need train prebuilt domain models), or can end up with 5 steps (if we add example utterances and train model)... so I'm stumped. Maybe the question will be different in exam.

Comment 10.1.1

ID: 620246 User: sdokmak Badges: - Relative Date: 3 years, 8 months ago Absolute Date: Wed 22 Jun 2022 10:29 Selected Answer: - Upvotes: 6

Ideally we'd do:
1. Add a new application
2. Add the prebuilt domain ToDo
3. Publish the application
https://docs.microsoft.com/en-us/azure/cognitive-services/luis/howto-add-prebuilt-models

Comment 10.1.1.1

ID: 794064 User: Pyguy Badges: - Relative Date: 3 years, 1 month ago Absolute Date: Tue 31 Jan 2023 12:26 Selected Answer: - Upvotes: 17

It was in Real Exam , and yes actually were 3 options there (Question said : Chose 3 action, NOT 4 , I also made like yours. :

1. Add a new application
2. Add the prebuilt domain ToDo
3. Publish the application

30. AI-102 Topic 5 Question 26

Sequence
74
Discussion ID
102920
Source URL
https://www.examtopics.com/discussions/microsoft/view/102920-exam-ai-102-topic-5-question-26-discussion/
Posted By
marti_tremblay000
Posted At
March 17, 2023, 12:38 p.m.

Question

Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.

After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.

You have a chatbot that uses question answering in Azure Cognitive Service for Language.

Users report that the responses of the chatbot lack formality when answering spurious questions.

You need to ensure that the chatbot provides formal responses to spurious questions.

Solution: From Language Studio, you change the chitchat source to qna_chitchat_friendly.tsv, and then retrain and republish the model.

Does this meet the goal?

  • A. Yes
  • B. No

Suggested Answer

B

Answer Description Click to expand


Community Answer Votes

Comments 6 comments Click to expand

Comment 1

ID: 1242908 User: Gorkon2022 Badges: Highly Voted Relative Date: 1 year, 8 months ago Absolute Date: Fri 05 Jul 2024 18:53 Selected Answer: - Upvotes: 10

Had to Look up the Word "spurious" .. MSFT Should Not have used this Word .. Could have used a more common term .. "Something false or inauthentic" , For Some People around the World English is not their Primary Language

Comment 1.1

ID: 1579501 User: StelSen Badges: - Relative Date: 8 months, 3 weeks ago Absolute Date: Sun 22 Jun 2025 07:12 Selected Answer: - Upvotes: 2

Agree. MS should test technical. Not English...

Comment 2

ID: 1206257 User: 9H3zmT6 Badges: Most Recent Relative Date: 1 year, 10 months ago Absolute Date: Fri 03 May 2024 21:56 Selected Answer: B Upvotes: 3

This file is used to provide responses suitable for a friendly or casual tone. You need to use qna_chitchat_professional.tsv.

Comment 3

ID: 1178422 User: rober13 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Wed 20 Mar 2024 16:03 Selected Answer: A Upvotes: 3

Based on reference Professional = formal and Question 28.

https://learn.microsoft.com/en-us/azure/cognitive-services/language-service/question-answering/how-to/chit-chat

Comment 4

ID: 1140084 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sun 04 Feb 2024 13:38 Selected Answer: B Upvotes: 3

The qna_chitchat_professional.tsv is specifically designed to provide responses with a formal tone compared to other chitchat sources. It contains a curated set of question-and-answer pairs that are structured to reflect a professional and formal manner of communication. This makes it more suitable for environments or user interactions where formal responses are preferred, ensuring that the chatbot's replies to spurious or off-topic questions maintain the desired level of professionalism.

Comment 5

ID: 841936 User: marti_tremblay000 Badges: - Relative Date: 2 years, 12 months ago Absolute Date: Fri 17 Mar 2023 12:38 Selected Answer: B Upvotes: 2

ChatGPT answer :
No, changing the chitchat source to "qna_chitchat_friendly.tsv" and retraining and republishing the model will not necessarily meet the goal of ensuring that the chatbot provides formal responses to spurious questions.

The "qna_chitchat_friendly.tsv" file is a source file for casual chitchat, which includes conversational responses for informal topics, such as hobbies and movies. This file is not designed to provide formal responses to spurious questions. Therefore, changing the source file to this file and retraining the model will not necessarily improve the formality of responses to spurious questions.

31. AI-102 Topic 3 Question 48

Sequence
75
Discussion ID
112136
Source URL
https://www.examtopics.com/discussions/microsoft/view/112136-exam-ai-102-topic-3-question-48-discussion/
Posted By
973b658
Posted At
June 14, 2023, 9:05 a.m.

Question

HOTSPOT -

You are developing a text processing solution.

You have the function shown below.

image

For the second argument, you call the function and specify the following string.

Our tour of Paris included a visit to the Eiffel Tower

For each of the following statements, select Yes if the statement is true. Otherwise, select No.

image

Suggested Answer

image
Answer Description Click to expand


Comments 18 comments Click to expand

Comment 1

ID: 953922 User: StrateqEBS Badges: Highly Voted Relative Date: 2 years, 7 months ago Absolute Date: Mon 17 Jul 2023 09:38 Selected Answer: - Upvotes: 33

Should be NYN:

https://learn.microsoft.com/en-us/dotnet/api/azure.ai.textanalytics.textanalyticsclient.recognizeentities?view=azure-dotnet

Definition:
Runs a predictive model to identify a collection of named entities in the passed-in document, and categorize those entities into types such as person, location, or organization.

This method does not extract phrases.

Comment 1.1

ID: 996848 User: M25 Badges: - Relative Date: 2 years, 6 months ago Absolute Date: Sat 02 Sep 2023 13:34 Selected Answer: - Upvotes: 3

Correct! Examples:
Output NER: trip, Seattle, last week
https://learn.microsoft.com/en-us/azure/ai-services/language-service/named-entity-recognition/quickstart?tabs=ga-api&pivots=programming-language-csharp#output

Output Key phrase extraction: modern medical office, Dr. Smith, great staff
https://learn.microsoft.com/en-us/azure/ai-services/language-service/key-phrase-extraction/quickstart?pivots=programming-language-csharp#output

Comment 2

ID: 1578849 User: StelSen Badges: Most Recent Relative Date: 8 months, 3 weeks ago Absolute Date: Thu 19 Jun 2025 11:18 Selected Answer: - Upvotes: 1

1. N
2. Y
3. N
The output will include the following words: our and included." No – These are not named entities; they are common words and won't be recognized by RecognizeEntities.

"The output will include the following words: Paris, Eiffel, and Tower." Yes – RecognizeEntities will detect Paris (a city) and Eiffel Tower (a landmark). Most likely, "Eiffel Tower" will be returned as a single entity, but even if split, "Eiffel" and "Tower" would still be considered part of named entities.

"The function will output all the key phrases from the input string to the console." No – This function uses RecognizeEntities, not ExtractKeyPhrases, so it only outputs recognized entities, not all key phrases.

Comment 3

ID: 1281206 User: famco Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Mon 09 Sep 2024 21:08 Selected Answer: - Upvotes: 1

I'm not sure if it is going to be Eiffel and Tower or just "Eiffel Tower"

Comment 3.1

ID: 1281210 User: famco Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Mon 09 Sep 2024 21:15 Selected Answer: - Upvotes: 1

On closer look at this trick (e.v.i.l) question, I will answer No for Eiffel and Tower as separate.
Under location there is :
Structural - Manmade structures - en (supported language)

So, I assume the question creator read this and thought it can be used to trick people

Comment 4

ID: 1275145 User: JakeCallham Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Fri 30 Aug 2024 17:27 Selected Answer: - Upvotes: 2

I beleive NNN. Second is N because it will output two keys, not three.

Comment 4.1

ID: 1275147 User: JakeCallham Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Fri 30 Aug 2024 17:32 Selected Answer: - Upvotes: 1

i get back to it: You're correct. Based on the given C# code using TextAnalyticsClient and the provided input string, the output would indeed include the words "Paris", "Eiffel", and "Tower". Let's break this down:

The function getkeywords takes a TextAnalyticsClient and a string as input.
It uses the RecognizeEntities method (note: there's a typo in the original code; it should be RecognizeEntities, not RecognizeEntitites) to analyze the text.
The input string is: "Our tour of Paris included a visit to the Eiffel Tower"
The TextAnalyticsClient's entity recognition would identify named entities in this text.
"Paris" is recognized as a location entity.
"Eiffel Tower" is recognized as a landmark entity, which would be split into two separate entities: "Eiffel" and "Tower".

Comment 4.1.1

ID: 1281213 User: famco Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Mon 09 Sep 2024 21:19 Selected Answer: - Upvotes: 1

Why will they be split. Eiffel tower is identified as a man-made structure under location-structural.

Comment 5

ID: 1264950 User: anto69 Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Tue 13 Aug 2024 05:20 Selected Answer: - Upvotes: 1

Must be N-Y-N, these are named entities

Comment 6

ID: 1261887 User: moonlightc Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Wed 07 Aug 2024 03:07 Selected Answer: - Upvotes: 4

It should be NNN

According to ChatGPT

Key words:
Paris
Eiffel Tower

Comment 7

ID: 1217361 User: funny_penguin Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 11:44 Selected Answer: - Upvotes: 4

on exam, NYN. For the last one I selected N because the method looks like it's extracting entities, not key phrases.

Comment 8

ID: 1215790 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 22 May 2024 16:09 Selected Answer: - Upvotes: 3

NYN is right answer.

Comment 9

ID: 1186324 User: Murtuza Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Sat 30 Mar 2024 20:15 Selected Answer: - Upvotes: 2

The function will output all the key phrases from the input string to the console. No, the function will output the recognized entities, not all key phrases. Key phrases could include other important words or phrases in the text that are not necessarily entities. For key phrase extraction, a different method would be used.

Comment 10

ID: 1177342 User: Murtuza Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Tue 19 Mar 2024 14:14 Selected Answer: - Upvotes: 2

For the last choice that seems to be a topic of dicussion
Console Output:
The code prints the header “Key words:” to the console.
It then iterates through the response.Value (presumably a collection of categorized entites

Comment 11

ID: 1116600 User: lastget Badges: - Relative Date: 2 years, 2 months ago Absolute Date: Mon 08 Jan 2024 13:27 Selected Answer: - Upvotes: 3

Agree NYN

Comment 12

ID: 1047623 User: katrang Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Thu 19 Oct 2023 08:41 Selected Answer: - Upvotes: 1

I think NYY. I could not see where all keywords would be output, but after checking the documentation they would also be identified as entities along with the Eiffel Tower (tour, paris and visit)

Comment 13

ID: 922825 User: 973b658 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Wed 14 Jun 2023 09:05 Selected Answer: - Upvotes: 1

It is true.

Comment 13.1

ID: 1235110 User: anjanc Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 07:01 Selected Answer: - Upvotes: 2

not sure why this guy writes 'it is true' for all questions!!!

32. AI-102 Topic 3 Question 1

Sequence
77
Discussion ID
55511
Source URL
https://www.examtopics.com/discussions/microsoft/view/55511-exam-ai-102-topic-3-question-1-discussion/
Posted By
Jenny1
Posted At
June 17, 2021, 3:23 p.m.

Question

Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.
You build a language model by using a Language Understanding service. The language model is used to search for information on a contact list by using an intent named FindContact.
A conversational expert provides you with the following list of phrases to use for training.
✑ Find contacts in London.
✑ Who do I know in Seattle?
✑ Search for contacts in Ukraine.
You need to implement the phrase list in Language Understanding.
Solution: You create a new pattern in the FindContact intent.
Does this meet the goal?

  • A. Yes
  • B. No

Suggested Answer

B

Answer Description Click to expand


Community Answer Votes

Comments 21 comments Click to expand

Comment 1

ID: 394978 User: azurelearner666 Badges: Highly Voted Relative Date: 4 years, 8 months ago Absolute Date: Wed 30 Jun 2021 19:00 Selected Answer: - Upvotes: 37

Using a pattern could be a good solution IMHO...
✑ Find contacts in London.
✑ Who do I know in Seattle?
✑ Search for contacts in Ukraine.

Like
Where is {FormName}[?]
Who authored {FormName}[?]
{FormName} is published in French[?]
(taken from https://docs.microsoft.com/en-us/azure/cognitive-services/luis/luis-concept-patterns)

we could do:
✑ Find contacts in {CityOrCountry}.
✑ Who do I know in {CityOrCountry}[?]
✑ Search for contacts in {CityOrCountry}[?].

So, to me a pattern is a Solution (A)

Comment 1.1

ID: 415285 User: YipingRuan Badges: - Relative Date: 4 years, 7 months ago Absolute Date: Tue 27 Jul 2021 10:10 Selected Answer: - Upvotes: 2

Agree, but Entity is also good https://docs.microsoft.com/bs-cyrl-ba/azure/cognitive-services/luis/luis-concept-intent#intent-compared-to-entity

Comment 1.2

ID: 672158 User: praticewizards Badges: - Relative Date: 3 years, 5 months ago Absolute Date: Sun 18 Sep 2022 11:16 Selected Answer: - Upvotes: 1

I agree. The intent here is "search for contact", for example. The location is an entity type "location", as you can see in the example What's the weather like in Seattle tomorrow? on the link given by @YipingRuan

Comment 1.3

ID: 1329539 User: jiantao_john_pan Badges: - Relative Date: 1 year, 2 months ago Absolute Date: Fri 20 Dec 2024 16:32 Selected Answer: - Upvotes: 2

https://learn.microsoft.com/en-us/training/modules/build-language-understanding-model/4-use-patterns-differentiate-similar-utterances

Comment 2

ID: 673895 User: STH Badges: Highly Voted Relative Date: 3 years, 5 months ago Absolute Date: Tue 20 Sep 2022 09:21 Selected Answer: A Upvotes: 13

According to MS learn, answer should be yes (A)

https://learn.microsoft.com/en-us/training/modules/create-language-understanding-app/5-use-patterns-to-differentiate-similar-utterances

This is a FindContact intent with a location entity pattern

Comment 3

ID: 1578717 User: 098ca78 Badges: Most Recent Relative Date: 8 months, 3 weeks ago Absolute Date: Wed 18 Jun 2025 20:29 Selected Answer: B Upvotes: 2

Explanation:
Creating a pattern in the FindContact intent is not the correct way to incorporate multiple diverse user phrases for training.

Patterns are used to define a specific sentence structure with parameterized parts (e.g., "Find contacts in {Location}"). They are more rigid and less flexible for general training of varied user utterances.

To properly train the Language Understanding (LUIS) model, you should add the provided phrases as example utterances to the FindContact intent. This helps the model learn how users might express that intent in natural language.

Correct Action:
You should add the phrases like:

"Find contacts in London"

"Who do I know in Seattle?"

"Search for contacts in Ukraine"

as example utterances, not as a single pattern.

Comment 4

ID: 1576993 User: qqaws Badges: - Relative Date: 9 months ago Absolute Date: Thu 12 Jun 2025 20:19 Selected Answer: B Upvotes: 1

it is "No" 100%

Comment 5

ID: 1576691 User: azuretrainer1 Badges: - Relative Date: 9 months ago Absolute Date: Thu 12 Jun 2025 01:32 Selected Answer: B Upvotes: 1

❌ No, this does not meet the goal.

Creating a pattern in the FindContact intent is not the most appropriate way to implement the provided phrases for training a Language Understanding (LUIS) model.

Here's why:
The provided sentences are training utterances, not patterns.

Training utterances help the model learn the variety of ways users may express an intent (in this case, FindContact).

Patterns in LUIS are used when you need to define a specific structure with pattern.any entities or placeholders (e.g., "Find contacts in {location}").

✅ Correct Approach:
You should add the provided phrases as training utterances under the FindContact intent. This will help the model recognize the intent when users ask similar questions in different ways.

Comment 6

ID: 1399801 User: SunilB Badges: - Relative Date: 11 months, 4 weeks ago Absolute Date: Mon 17 Mar 2025 20:29 Selected Answer: B Upvotes: 1

This should be utterances and not pattern

Comment 7

ID: 1361858 User: gyaansastra Badges: - Relative Date: 1 year ago Absolute Date: Wed 26 Feb 2025 11:11 Selected Answer: B Upvotes: 1

Creating a new pattern in the FindContact intent alone does not fully meet the goal of implementing the phrase list in Language Understanding (LUIS). Instead, you should add these phrases as utterances under the FindContact intent. By adding them as utterances, the model will learn to recognize these specific phrases and similar variations when users are searching for contacts.

Comment 8

ID: 1340119 User: BenALGuhl Badges: - Relative Date: 1 year, 1 month ago Absolute Date: Tue 14 Jan 2025 02:57 Selected Answer: B Upvotes: 1

In my opinion, example utterances are missing. The stated steps should provide a complete solution, which A does not. So it is B.

Comment 9

ID: 1333354 User: pabsinaz Badges: - Relative Date: 1 year, 2 months ago Absolute Date: Sun 29 Dec 2024 06:55 Selected Answer: B Upvotes: 1

No, this solution does not fully meet the goal. Creating a new pattern in the FindContact intent is helpful, but it's not sufficient on its own for implementing the given phrase list effectively.

To properly implement these phrases, you should add them as example utterances within the FindContact intent. This allows the Language Understanding (LUIS) service to learn from these specific examples and better understand the variations in the queries.

In summary, you should:

Add the given phrases as example utterances in the FindContact intent.

Optionally, create patterns to handle similar phrases or variations more effectively.

Comment 10

ID: 1319910 User: Alan_CA Badges: - Relative Date: 1 year, 3 months ago Absolute Date: Fri 29 Nov 2024 21:37 Selected Answer: B Upvotes: 1

Copilot says :
The solution of creating a new pattern in the FindContact intent is not the most appropriate way to implement these training phrases. Instead, you should use these phrases as examples directly within the FindContact intent
So I would say NO

Comment 11

ID: 1308153 User: jolimon Badges: - Relative Date: 1 year, 4 months ago Absolute Date: Wed 06 Nov 2024 21:45 Selected Answer: A Upvotes: 2

A is the right answer

Comment 12

ID: 1291544 User: RajuTS Badges: - Relative Date: 1 year, 5 months ago Absolute Date: Mon 30 Sep 2024 14:06 Selected Answer: B Upvotes: 2

Th answer would be to add example utterances that trains the model to recognize similar phrases

Comment 13

ID: 1279023 User: famco Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Thu 05 Sep 2024 17:34 Selected Answer: - Upvotes: 1

Phrases are like synonyms. Here it is about entities (location to be specific) that is required along with the intent to do the action. Unless of course there are some hidden tricks to make people who know how it is used fail. This (certification) department in the dysfunctional microsoft is about making more money in that department (and missing the bigger picture). Intra company fights is microsoft

Comment 13.1

ID: 1279029 User: famco Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Thu 05 Sep 2024 17:43 Selected Answer: - Upvotes: 1

Wait a minute, of course there is a catch. It is a trap. They do not talk about entities, which will come to mind. The question is formulated in a complex way I can't make out what breakfast the creator of this marvellous question ate.

I have to say no because it is just creating the pattern, and not creating the phrase and in the phrase the pattern. But god knows what is the intent of this question. Microsoft !!

Comment 14

ID: 1274907 User: testmaillo020 Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Fri 30 Aug 2024 06:39 Selected Answer: B Upvotes: 2

**Answer: B. No**

Creating a new pattern in the FindContact intent is not the correct approach. Patterns in Language Understanding services are used to match specific sentence structures rather than training the model with example phrases. To properly implement the phrase list for training, you should add these phrases as example utterances within the FindContact intent. This way, the model learns to recognize different ways users might ask to find contacts.

So, the correct approach would be to add these phrases as example utterances, not just creating a new pattern.

Comment 15

ID: 1235716 User: HVardhini Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sun 23 Jun 2024 09:26 Selected Answer: B Upvotes: 1

No. Find contacts in {Location} where location is an entity

Comment 16

ID: 1234478 User: HaraTadahisa Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Fri 21 Jun 2024 16:37 Selected Answer: A Upvotes: 2

I say A is correct! HipHop!

Comment 17

ID: 1230991 User: fuck_india Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 15 Jun 2024 16:03 Selected Answer: - Upvotes: 2

This issue is not working because everyone's opinion is 50-50. I don't know the correct answer, so I will be staring at MS Learn when this question comes up on the real exam.Examtopic should scrutinize and post the correct answer.

33. AI-102 Topic 5 Question 18

Sequence
84
Discussion ID
200012
Source URL
https://www.examtopics.com/discussions/microsoft/view/200012-exam-ai-102-topic-5-question-18-discussion/
Posted By
-
Posted At
April 4, 2025, 7:18 a.m.

Question

SIMULATION -
You need to create and publish a bot that will use Language Understanding and QnA Maker. The bot must be named bot12345678. You must publish the bot by using the [email protected] account.
NOTE: Complete this task first. It may take several minutes to complete the required deployment steps. While this is taking place, you can complete tasks 2-6 in this lab during the deployment.
To complete this task, use the Microsoft Bot Framework Composer.

Suggested Answer

image
Answer Description Click to expand


Comments 1 comment Click to expand

Comment 1

ID: 1576654 User: osx Badges: - Relative Date: 9 months ago Absolute Date: Wed 11 Jun 2025 21:39 Selected Answer: - Upvotes: 1

QnA Maker service is being retired on 31st March, 2025. A newer version of this capability is now available as a part of Azure AI service for Language called question answering. To use this service, you need to provision a Language resource. For question answering capability within the Language service, see question answering and its pricing page. You can't create new QnA Maker resources anymore. For information on migrating your existing QnA Maker knowledge bases to question answering, consult the migration guide.

34. AI-102 Topic 5 Question 22

Sequence
85
Discussion ID
102918
Source URL
https://www.examtopics.com/discussions/microsoft/view/102918-exam-ai-102-topic-5-question-22-discussion/
Posted By
marti_tremblay000
Posted At
March 17, 2023, 12:34 p.m.

Question

You are designing a conversational interface for an app that will be used to make vacation requests. The interface must gather the following data:

• The start date of a vacation
• The end date of a vacation
• The amount of required paid time off

The solution must minimize dialog complexity.

Which type of dialog should you use?

  • A. adaptive
  • B. skill
  • C. waterfall
  • D. component

Suggested Answer

C

Answer Description Click to expand


Community Answer Votes

Comments 18 comments Click to expand

Comment 1

ID: 1139929 User: evangelist Badges: Highly Voted Relative Date: 2 years, 1 month ago Absolute Date: Sun 04 Feb 2024 10:18 Selected Answer: C Upvotes: 9

The answer is C:
For designing a conversational interface that gathers data in a structured manner with minimal dialog complexity, you should use C. waterfall dialog. Waterfall dialogs are designed to guide the user through a series of steps or questions, one after another, which is ideal for collecting specific pieces of information sequentially, like vacation start and end dates, and the amount of required paid time off.

Comment 2

ID: 1283189 User: mrg998 Badges: Most Recent Relative Date: 1 year, 6 months ago Absolute Date: Fri 13 Sep 2024 15:24 Selected Answer: C Upvotes: 2

its C, the least complex way to do it

Comment 3

ID: 1235221 User: HaraTadahisa Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 08:54 Selected Answer: C Upvotes: 1

The solution must minimize dialog complexity. so I say this answer is C.

Comment 4

ID: 1229895 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 17:00 Selected Answer: C Upvotes: 1

C is answer.

Comment 5

ID: 1222670 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sat 01 Jun 2024 15:49 Selected Answer: C Upvotes: 1

C is right answer.

Comment 6

ID: 1150172 User: elizebeth2 Badges: - Relative Date: 2 years ago Absolute Date: Wed 14 Feb 2024 12:57 Selected Answer: C Upvotes: 3

https://learn.microsoft.com/en-us/azure/bot-service/bot-builder-concept-waterfall-dialogs?view=azure-bot-service-4.0#waterfall-dialogs

Comment 7

ID: 1135732 User: trysec Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Tue 30 Jan 2024 12:44 Selected Answer: C Upvotes: 3

C. waterfall

A waterfall dialog is best suited for scenarios where a conversational flow requires a sequence of steps to collect information or guide the user through a set process. For gathering specific pieces of data like start date, end date, and the amount of required paid time off, a waterfall dialog allows you to prompt the user for each piece of information one after the other, which can help minimize the complexity of the dialog interactions.

Comment 8

ID: 1115876 User: dimsok Badges: - Relative Date: 2 years, 2 months ago Absolute Date: Sun 07 Jan 2024 14:48 Selected Answer: C Upvotes: 2

Opinionated, but waterfall seems like a better UX than component

Comment 9

ID: 1008624 User: sl_mslconsulting Badges: - Relative Date: 2 years, 5 months ago Absolute Date: Fri 15 Sep 2023 19:42 Selected Answer: D Upvotes: 1

Should be D. You need to take a step further to reduce the complexity - https://learn.microsoft.com/en-us/azure/bot-service/bot-builder-compositcontrol?view=azure-bot-service-4.0&tabs=csharp

Comment 9.1

ID: 1576581 User: qqaws Badges: - Relative Date: 9 months ago Absolute Date: Wed 11 Jun 2025 16:28 Selected Answer: - Upvotes: 1

Component helps managing complexity. But in this case we only need 4 waterfall questions. So, the correct answer is C

Comment 10

ID: 998214 User: M25 Badges: - Relative Date: 2 years, 6 months ago Absolute Date: Mon 04 Sep 2023 06:51 Selected Answer: C Upvotes: 3

https://learn.microsoft.com/en-us/azure/bot-service/bot-builder-concept-waterfall-dialogs?view=azure-bot-service-4.0
The interface must gather … date: Prompt dialogs (prompts) are dialogs designed to ask the user for specific types of information, such as a number, a date, or a name, and so on. Prompts are designed to work with waterfall dialogs in a component dialog.
A waterfall dialog (or waterfall) … are designed to work within the context of a component dialog.
The component dialog provides a strategy for creating independent dialogs to handle specific scenarios, breaking a large dialog set into more manageable pieces. Each of these pieces has its own dialog set, and avoids any name collisions with the dialog set that contains it.

Comment 10.1

ID: 998215 User: M25 Badges: - Relative Date: 2 years, 6 months ago Absolute Date: Mon 04 Sep 2023 06:51 Selected Answer: - Upvotes: 1

https://learn.microsoft.com/en-us/azure/bot-service/bot-builder-concept-dialog?view=azure-bot-service-4.0#dialog-types
adaptive dialog: A type of container dialog used by Composer to provide more natural conversational flows. Not intended to be used directly in an SDK-first bot.
skill dialog: Automates the management of one or more skill bots from a skill consumer. Composer directly supports skills as actions.
Important
Adaptive dialogs were first added in version 4.9 of the C# SDK. Adaptive dialogs support the Bot Framework Composer and are not intended to be used directly in an SDK-first bot.
Bot Framework Composer is a visual authoring tool for building Conversational AI applications.

Comment 11

ID: 990752 User: tranatrana Badges: - Relative Date: 2 years, 6 months ago Absolute Date: Sat 26 Aug 2023 14:58 Selected Answer: - Upvotes: 1

What is it? C or D?

Comment 12

ID: 913955 User: kail85 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Sun 04 Jun 2023 01:23 Selected Answer: - Upvotes: 1

C. waterfall

Waterfall dialogs are a good choice for this scenario because they are designed to guide users through a linear, step-by-step process. In this case, you can use a waterfall dialog to gather the vacation start date, end date, and required paid time off sequentially. This approach can help minimize dialog complexity while ensuring that all necessary data is collected in an orderly manner.

Comment 13

ID: 912100 User: EliteAllen Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Thu 01 Jun 2023 14:35 Selected Answer: C Upvotes: 1

C. waterfall

A waterfall dialog is a type of dialog that guides the user through a series of steps or prompts in a specific order. This makes it a good choice for gathering a set of related data points, like the start date, end date, and amount of required paid time off for a vacation request. By using a waterfall dialog, you can ensure that all necessary information is collected in a structured and predictable manner, which can help minimize dialog complexity.

Option A, adaptive dialog, is a more flexible type of dialog that can handle more complex and dynamic conversation flows, but it might be overkill for this relatively straightforward data-gathering task. Option B, skill dialog, is used to manage the invocation of a bot skill, which is not relevant to this scenario. Option D, component dialog, is a reusable dialog that encapsulates its own state, but it doesn't inherently simplify the dialog structure.

Comment 14

ID: 861652 User: odisor Badges: - Relative Date: 2 years, 11 months ago Absolute Date: Wed 05 Apr 2023 03:00 Selected Answer: - Upvotes: 3

I think the answer could be C or D, but in order to minimize complexity, I think D is correct.
ChatGPT answer:
In Azure CLU, there are three main types of dialogs:

Root Dialog: The Root Dialog is the main entry point for the bot. It handles incoming messages and dispatches them to other dialogs as appropriate. The Root Dialog can be thought of as the top-level container for all other dialogs in the bot.

Component Dialog: A Component Dialog is a reusable dialog that can be embedded within other dialogs. It encapsulates a specific set of related functionality, such as a login dialog or a payment dialog. Component Dialogs can be used to create more modular and maintainable bot code.

Waterfall Dialog: A Waterfall Dialog is a type of dialog that guides the user through a series of steps or prompts to complete a task. Each step in the Waterfall Dialog can be implemented as a separate method, making it easy to create complex conversational flows. Waterfall Dialogs can be used for tasks like booking appointments, ordering food, or making reservations.

Comment 15

ID: 856453 User: ptmk Badges: - Relative Date: 2 years, 11 months ago Absolute Date: Fri 31 Mar 2023 02:19 Selected Answer: - Upvotes: 1

I think the answer is correct. The key requirement is "The solution must minimize dialog complexity".

My answer is C. Component

Waterfall dialog is used to manage linear and more complex conversation flows.
https://learn.microsoft.com/en-us/azure/bot-service/bot-builder-dialog-manage-conversation-flow?view=azure-bot-service-4.0&tabs=csharp

Component dialog is used to create independent dialogs to handle specific scenarios, breaking a large dialog set into more manageable pieces.
https://learn.microsoft.com/en-us/azure/bot-service/bot-builder-compositcontrol?view=azure-bot-service-4.0&tabs=csharp

Comment 16

ID: 841930 User: marti_tremblay000 Badges: - Relative Date: 2 years, 12 months ago Absolute Date: Fri 17 Mar 2023 12:34 Selected Answer: C Upvotes: 3

ChatGPT answer :
To minimize dialog complexity and gather the required data in a conversational interface for vacation requests, the best type of dialog to use would be a "waterfall" dialog.

A waterfall dialog is a simple, linear conversation flow where the bot prompts the user for information in a structured sequence. Each prompt is presented to the user one at a time, in a predefined order, until all the required information is collected. The waterfall dialog makes it easy to collect data in a conversational interface, while keeping the dialog simple and easy to follow.

In this case, the waterfall dialog can be designed to prompt the user for the start date, end date, and amount of required paid time off, in that order. The dialog can then confirm the details with the user and submit the request.

Skill and component dialogs are not relevant to this scenario, as they are used for integrating with external services and breaking down large bots into smaller, reusable parts, respectively.

35. AI-102 Topic 3 Question 64

Sequence
88
Discussion ID
134939
Source URL
https://www.examtopics.com/discussions/microsoft/view/134939-exam-ai-102-topic-3-question-64-discussion/
Posted By
audlindr
Posted At
Feb. 29, 2024, 8:10 p.m.

Question

You have an Azure subscription that contains a multi-service Azure Cognitive Services Translator resource named Translator1.

You are building an app that will translate text and documents by using Translator1.

You need to create the REST API request for the app.

Which headers should you include in the request?

  • A. the access control request, the content type, and the content length
  • B. the subscription key and the client trace ID
  • C. the resource ID and the content language
  • D. the subscription key, the subscription region, and the content type

Suggested Answer

D

Answer Description Click to expand


Community Answer Votes

Comments 8 comments Click to expand

Comment 1

ID: 1162945 User: audlindr Badges: Highly Voted Relative Date: 2 years ago Absolute Date: Thu 29 Feb 2024 20:10 Selected Answer: D Upvotes: 10

Answer is correct

https://learn.microsoft.com/en-us/azure/ai-services/translator/reference/v3-0-translate
content-type is required header

https://learn.microsoft.com/en-us/azure/ai-services/translator/reference/v3-0-reference
When you use a multi-service secret key, you must include two authentication headers with your request. There are two headers that you need to call the Translator.
Ocp-Apim-Subscription-Key The value is the Azure secret key for your multi-service resource.
Ocp-Apim-Subscription-Region The value is the region of the multi-service resource.
Region is required for the multi-service Text API subscription. The region you select is the only region that you can use for text translation when using the multi-service key. It must be the same region you selected when you signed up for your multi-service subscription through the Azure portal.
D is the only option with both the headers

Comment 2

ID: 1177775 User: Murtuza Badges: Highly Voted Relative Date: 1 year, 11 months ago Absolute Date: Tue 19 Mar 2024 22:57 Selected Answer: - Upvotes: 5

To create the REST API request for your app that will translate text and documents using the Azure Cognitive Services Translator resource (Translator1), you should include the following headers:

Subscription Key: This key serves as the authentication token to access the service.
Subscription Region: The region where your Translator resource is deployed.
Content Type: Specify the type of content you are sending (e.g., JSON, plain text).
Therefore, the correct answer is D. the subscription key, the subscription region, and the content type. When making the API request, ensure that you include these headers to connect your application to the Translator service.

Comment 3

ID: 1575462 User: prasioso Badges: Most Recent Relative Date: 9 months, 1 week ago Absolute Date: Sat 07 Jun 2025 11:35 Selected Answer: D Upvotes: 1

D. Example for swiss:

curl -X POST "https://my-swiss-n.cognitiveservices.azure.com/translator/text/v3.0/translate?to=fr" \
-H "Ocp-Apim-Subscription-Key: xxx" \
-H "Ocp-Apim-Subscription-Region: switzerlandnorth" \
-H "Content-Type: application/json" \
-d "[{'Text':'Hello'}]" -v

Comment 4

ID: 1265830 User: anto69 Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Wed 14 Aug 2024 17:01 Selected Answer: D Upvotes: 1

D for Copilot too

Comment 5

ID: 1235183 User: HaraTadahisa Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 08:20 Selected Answer: D Upvotes: 1

I say this answer is D. Please hurry up and transport the meat.

Comment 6

ID: 1229211 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 12 Jun 2024 16:11 Selected Answer: D Upvotes: 1

D is answer.

Comment 7

ID: 1215766 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 22 May 2024 15:36 Selected Answer: D Upvotes: 1

the subscription key, the subscription region, and the content type

Comment 8

ID: 1208107 User: kay1101 Badges: - Relative Date: 1 year, 10 months ago Absolute Date: Wed 08 May 2024 02:44 Selected Answer: - Upvotes: 2

D.
https://learn.microsoft.com/en-us/azure/ai-services/translator/quickstart-text-rest-api?tabs=csharp#headers

36. AI-102 Topic 3 Question 19

Sequence
93
Discussion ID
74742
Source URL
https://www.examtopics.com/discussions/microsoft/view/74742-exam-ai-102-topic-3-question-19-discussion/
Posted By
SamedKia
Posted At
April 28, 2022, 10:19 a.m.

Question

Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.
You build a language model by using a Language Understanding service. The language model is used to search for information on a contact list by using an intent named FindContact.
A conversational expert provides you with the following list of phrases to use for training.
✑ Find contacts in London.
✑ Who do I know in Seattle?
Search for contacts in Ukraine.
image
You need to implement the phrase list in Language Understanding.
Solution: You create a new entity for the domain.
Does this meet the goal?

  • A. Yes
  • B. No

Suggested Answer

B

Answer Description Click to expand


Community Answer Votes

Comments 19 comments Click to expand

Comment 1

ID: 593671 User: SamedKia Badges: Highly Voted Relative Date: 3 years, 10 months ago Absolute Date: Thu 28 Apr 2022 10:19 Selected Answer: - Upvotes: 19

Wrong! We create a new location entity for domain to keep the location of FindContact intent.

Comment 2

ID: 1092274 User: abbey127 Badges: Highly Voted Relative Date: 2 years, 3 months ago Absolute Date: Sun 10 Dec 2023 05:01 Selected Answer: - Upvotes: 10

B. No

Creating a new entity for the domain does not directly address the goal of implementing the provided phrase list for the "FindContact" intent. Entities are typically used to extract specific pieces of information from user input, like names, locations, etc. In this case, the goal is to improve the understanding of the intent "FindContact" by providing training phrases related to searching for contacts in specific locations. Instead of creating a new entity, you should focus on training the language model within the existing "FindContact" intent and include the provided phrases to enhance its ability to recognize and understand user queries related to finding contacts in different locations.

Comment 3

ID: 1361913 User: gyaansastra Badges: Most Recent Relative Date: 1 year ago Absolute Date: Wed 26 Feb 2025 13:40 Selected Answer: B Upvotes: 1

Creating a new entity for the domain does not meet the goal. Instead, you should add these phrases as utterances under the existing FindContact intent. This allows the model to learn and recognize these specific phrases and similar variations when users are searching for contacts.

Comment 3.1

ID: 1570963 User: 8d82b67 Badges: - Relative Date: 9 months, 3 weeks ago Absolute Date: Wed 21 May 2025 15:25 Selected Answer: - Upvotes: 1

Utterances with a pattern for the location entity which can be a pre-built entity Geography.Location should be the full solution for this. This answer must be 'No'.

Comment 4

ID: 1333356 User: pabsinaz Badges: - Relative Date: 1 year, 2 months ago Absolute Date: Sun 29 Dec 2024 06:55 Selected Answer: B Upvotes: 1

No, this solution does not meet the goal. Creating a new entity for the domain alone is not sufficient for implementing the given phrase list effectively. Entities are used to identify specific data points within a query, such as locations or names.

To properly train the language model with these phrases, you should add them as example utterances within the FindContact intent. This helps the Language Understanding (LUIS) service learn from the specific examples and improve its ability to recognize similar queries.

So, you should:

Add the given phrases as example utterances in the FindContact intent.

Optionally, create entities to capture specific details like locations (e.g., London, Seattle, Ukraine).

Comment 5

ID: 1329836 User: pabsinaz Badges: - Relative Date: 1 year, 2 months ago Absolute Date: Sat 21 Dec 2024 06:50 Selected Answer: B Upvotes: 1

Right answer is no

Comment 6

ID: 1295063 User: 4371883 Badges: - Relative Date: 1 year, 5 months ago Absolute Date: Wed 09 Oct 2024 11:22 Selected Answer: B Upvotes: 2

B. it should be phrase list. https://learn.microsoft.com/en-us/azure/ai-services/luis/concepts/patterns-features

Comment 7

ID: 1288155 User: mrg998 Badges: - Relative Date: 1 year, 5 months ago Absolute Date: Mon 23 Sep 2024 16:04 Selected Answer: B Upvotes: 5

NO - you need a pattern

Comment 8

ID: 1283430 User: AzureGeek79 Badges: - Relative Date: 1 year, 5 months ago Absolute Date: Sat 14 Sep 2024 02:56 Selected Answer: - Upvotes: 1

ChatGPT answer is B. No, here is the explanation:
Creating a new entity for the domain does not meet the goal. To implement the phrase list in Language Understanding, you should add these phrases as example utterances for the FindContact intent. This helps the model recognize the intent based on similar phrases.

Comment 9

ID: 1281504 User: mrg998 Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Tue 10 Sep 2024 13:46 Selected Answer: B Upvotes: 2

the answer is a pattern

Comment 10

ID: 1246821 User: krzkrzkra Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Fri 12 Jul 2024 16:30 Selected Answer: A Upvotes: 1

Selected Answer: A

Comment 11

ID: 1234476 User: HaraTadahisa Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Fri 21 Jun 2024 16:36 Selected Answer: B Upvotes: 2

"create a new entity for the domain." is not "to implement the phrase list".
I say B is correct! HipHop!

Comment 12

ID: 1230989 User: fuck_india Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 15 Jun 2024 16:02 Selected Answer: - Upvotes: 1

This issue is not working because everyone's opinion is 50-50. I don't know the correct answer, so I will be staring at MS Learn when this question comes up on the real exam.Examtopic should scrutinize and post the correct answer.

Comment 13

ID: 1230984 User: fuck_india Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 15 Jun 2024 16:02 Selected Answer: - Upvotes: 1

This issue is not working because everyone's opinion is 50-50. I don't know the correct answer, so I will be staring at MS Learn when this question comes up on the real exam.Examtopic should scrutinize and post the correct answer.

Comment 14

ID: 1185097 User: f2c587e Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Fri 29 Mar 2024 00:37 Selected Answer: - Upvotes: 3

This same question (topic 3, question 1) seems to be indicated by the community that this question has two affirmative answers. I'm not sure if this is correct, normally in this type of question there is only one affirmative answer and the rest are negative. Is anyone clear on the real answer to this question?

Comment 14.1

ID: 1203081 User: lesiris Badges: - Relative Date: 1 year, 10 months ago Absolute Date: Sat 27 Apr 2024 13:25 Selected Answer: - Upvotes: 1

Indeed !!

Comment 15

ID: 1173041 User: chandiochan Badges: - Relative Date: 1 year, 12 months ago Absolute Date: Thu 14 Mar 2024 02:59 Selected Answer: B Upvotes: 1

must be B

Comment 16

ID: 1168609 User: Mehe323 Badges: - Relative Date: 2 years ago Absolute Date: Fri 08 Mar 2024 08:06 Selected Answer: B Upvotes: 2

The question says
"You need to implement the phrase list in Language Understanding.
Solution: You create a new entity for the domain.
Does this meet the goal?"

Creating an entity is not implementing a phrase list, so I vote NO.

Comment 17

ID: 1133037 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 27 Jan 2024 04:48 Selected Answer: - Upvotes: 4

The ANSWER is NO !
Instead of creating a new intent for location, you should:
Add these phrases to the "FindContact" intent: This helps the model understand that these are different ways a user might express the intent to find contacts.

Use Entities for Locations: In addition to training the intent with these phrases, you should use entities to capture the location information within the phrases. In LUIS, you can define an entity (like "Location") and then annotate locations in your training phrases. This way, the model learns to recognize and extract location details from user inputs.

By combining a well-trained "FindContact" intent with a location entity, your language model will be better equipped to understand and process user queries about finding contacts in specific locations.

37. AI-102 Topic 5 Question 8

Sequence
104
Discussion ID
57093
Source URL
https://www.examtopics.com/discussions/microsoft/view/57093-exam-ai-102-topic-5-question-8-discussion/
Posted By
TheB
Posted At
July 4, 2021, 4 p.m.

Question

HOTSPOT -
You are building a chatbot for a Microsoft Teams channel by using the Microsoft Bot Framework SDK. The chatbot will use the following code.
image
For each of the following statements, select Yes if the statement is true. Otherwise, select No.
NOTE: Each correct selection is worth one point.
Hot Area:
image

Suggested Answer

image
Answer Description Click to expand


Comments 24 comments Click to expand

Comment 1

ID: 430625 User: guruaws2021 Badges: Highly Voted Relative Date: 4 years, 6 months ago Absolute Date: Tue 24 Aug 2021 10:50 Selected Answer: - Upvotes: 27

2nd one should ne no, when a new member will join the chatbot will only great the new member and not all existing member in the chat

Comment 1.1

ID: 1334443 User: Happiman Badges: - Relative Date: 1 year, 2 months ago Absolute Date: Mon 30 Dec 2024 23:05 Selected Answer: - Upvotes: 2

OnMembersAddedAsync will be triggered when a user joins the conversation.
Answer: Yes

The OnMembersAddedAsync method is specifically triggered when a new member is added to the conversation in a Microsoft Teams channel. This is indicated by the membersAdded parameter, which contains the list of newly added members.
When a new user joins the conversation, the existing users in the conversation will see the chatbot greeting.
Answer: No

The code is written to send a welcome message only to the newly added user (member.Id != turnContext.Activity.Recipient.Id ensures this). Existing users will not see this greeting.
OnMembersAddedAsync will be initialized when a user sends a message.
Answer: No

The OnMembersAddedAsync method is only triggered when a new member is added to the conversation. It is not triggered by regular messages sent by users.

Comment 1.2

ID: 437299 User: Viktur Badges: - Relative Date: 4 years, 6 months ago Absolute Date: Wed 01 Sep 2021 18:02 Selected Answer: - Upvotes: 27

I feel the answer is correct. The members are all in a "group". So when a new member is added, the old members will see the chatbot greet the new member ("Hello ${new.member}), and not every other existing members.

Comment 1.2.1

ID: 1256563 User: Labani1987 Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Sun 28 Jul 2024 04:59 Selected Answer: - Upvotes: 2

It will greet the only new user who is added, but other users will be able to see that greeting ( someone new is added, this feature is sometimes used to welcome a new member in a group chat or Teams channel )

Comment 2

ID: 559231 User: reachmymind Badges: Highly Voted Relative Date: 4 years ago Absolute Date: Wed 02 Mar 2022 08:24 Selected Answer: - Upvotes: 22

A1. Yes
https://docs.microsoft.com/en-us/dotnet/api/microsoft.bot.builder.activityhandler.onmembersaddedasync?view=botbuilder-dotnet-stable
ActivityHandler.OnMembersAddedAsync(IList<ChannelAccount>, ITurnContext<IConversationUpdateActivity>, CancellationToken)
Method invoked when members other than the bot join the conversation

A2. No
https://docs.microsoft.com/en-us/dotnet/api/microsoft.bot.builder.turncontext.sendactivityasync?view=botbuilder-dotnet-stable
Sends a message activity to the sender of the incoming activity in turncontext

A3: No
https://docs.microsoft.com/en-us/dotnet/api/microsoft.bot.builder.activityhandler.onconversationupdateactivityasync?view=botbuilder-dotnet-stable
ActivityHandler.OnConversationUpdateActivityAsync(ITurnContext<IConversationUpdateActivity>, CancellationToken)
Method invoked when a conversation update activity that indicates one or more users other than the bot are joining the conversation

Comment 2.1

ID: 756061 User: ninjapunk Badges: - Relative Date: 3 years, 2 months ago Absolute Date: Sun 25 Dec 2022 23:34 Selected Answer: - Upvotes: 3

This is correct, as per the docs SendActivityAsync "sends an activity to the sender of the incoming activity", thus not a public message to be seen by everyone in the group.

Comment 3

ID: 1265105 User: anto69 Badges: Most Recent Relative Date: 1 year, 7 months ago Absolute Date: Tue 13 Aug 2024 12:23 Selected Answer: - Upvotes: 2

I read many comments. No reason why the second should be N.
Y-Y-N

Comment 3.1

ID: 1275356 User: JakeCallham Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Sat 31 Aug 2024 06:06 Selected Answer: - Upvotes: 3

Are you a .net developer? I guess not. its no

Comment 4

ID: 1246802 User: krzkrzkra Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Fri 12 Jul 2024 15:53 Selected Answer: - Upvotes: 1

YNN is the answer.

Comment 5

ID: 1234505 User: HaraTadahisa Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Fri 21 Jun 2024 16:52 Selected Answer: - Upvotes: 1

My answer is that
Yes
No
No

Comment 6

ID: 1230855 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sat 15 Jun 2024 10:54 Selected Answer: - Upvotes: 2

Yes
No
No

Comment 7

ID: 1222698 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sat 01 Jun 2024 16:46 Selected Answer: - Upvotes: 2

Will this question still be asked on June 2, 2024?

Comment 7.1

ID: 1399652 User: Mattt Badges: - Relative Date: 12 months ago Absolute Date: Mon 17 Mar 2025 14:23 Selected Answer: - Upvotes: 1

No, that's outside the scope of the exam.

Comment 8

ID: 1186912 User: Murtuza Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Sun 31 Mar 2024 19:14 Selected Answer: - Upvotes: 1

Yes, the OnMembersAddedAsync method is triggered when a new user joins the conversation.
Yes, when a new user joins the conversation, the existing users in the conversation will see the chatbot greeting. This is because the OnMembersAddedAsync method sends a welcome message to each new member.
No, the OnMembersAddedAsync method is not initialized when a user sends a message. It is only triggered when a new user joins the conversation. The method that handles incoming messages is typically OnMessageActivityAsyn

Comment 9

ID: 1144485 User: idcanymore Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Thu 08 Feb 2024 14:36 Selected Answer: - Upvotes: 2

Discord sends this exact message to people joining groups and everyone in the chat sees it. don't let chatgpt do your thinking for you.

Comment 10

ID: 1143890 User: PCRamirez Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Wed 07 Feb 2024 23:59 Selected Answer: - Upvotes: 1

YNN: Asking Windows Copilot about second answer:
No, the existing users in the conversation will not see the chatbot greeting. The code snippet provided shows a method `OnMembersAddedAsync` that is triggered when new members are added to the conversation. Inside this method, there is a conditional statement `if (member.Id != turnContext.Activity.Recipient.Id)` which ensures that the welcome message is sent only to the new member who has joined, and not to the existing users or the bot itself. The welcome message is personalized with the new member's name and is sent directly to them.

Comment 11

ID: 1139859 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sun 04 Feb 2024 08:18 Selected Answer: - Upvotes: 1

given answer is NOT Correct!
why:
Enable Active Learning: LUIS's Active Learning feature automatically identifies vague or low-confidence user inputs that may require further training.
Review and Add Utterances Logged for Review: Utterances collected through Active Learning are marked for review. Regularly reviewing and categorizing these utterances into the relevant intents is a crucial step in enhancing model performance.
Train and Republish the Language Understanding Model: After adding new utterances and potentially adjusting intents and entities, you need to retrain the LUIS model to incorporate these updates.

Comment 12

ID: 1115839 User: dimsok Badges: - Relative Date: 2 years, 2 months ago Absolute Date: Sun 07 Jan 2024 14:00 Selected Answer: - Upvotes: 5

YYN
- for the second one, everyone will see that the bot said hi to a new user , the message sent by the bot is not private, its posted in the channel

Comment 13

ID: 1047483 User: sl_mslconsulting Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Thu 19 Oct 2023 05:23 Selected Answer: - Upvotes: 3

The answer to the second question is no. The membersAdded will have two guys : you and the bot. The OnMembersAddedAsync will only be triggered when non bot member joined the conversation which is you. The logic coded here will not send the welcome message to the bot which is also a existing user. You and the bot are both members. I think we are reading to much to this - it’s not a chat room where a new guy joined everyone will be notified.

Comment 13.1

ID: 1047484 User: sl_mslconsulting Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Thu 19 Oct 2023 05:25 Selected Answer: - Upvotes: 1

You can actually write the code to verify this yourself and comment out the statement that check if a member is a bot to see what would happen.

Comment 14

ID: 814624 User: AzureJobsTillRetire Badges: - Relative Date: 3 years ago Absolute Date: Sun 19 Feb 2023 23:26 Selected Answer: - Upvotes: 3

The given answers are correct.

On the second question:-
The sender is the Microsoft Teams channel.
"from" Identifies the sender of the message. example: from: ChannelAccount
The recipients are the members

According to the code,
If new member <> existing member, then send out welcome message
if (member.Id != turnContext.Activity.Recipient.Id)

Comment 14.1

ID: 862716 User: saadashraf Badges: - Relative Date: 2 years, 11 months ago Absolute Date: Thu 06 Apr 2023 07:57 Selected Answer: - Upvotes: 2

Thanks, spent quite a time on this question. Your answer is absolutely correct

Comment 15

ID: 468973 User: Adedoyin_Simeon Badges: - Relative Date: 4 years, 4 months ago Absolute Date: Thu 28 Oct 2021 06:26 Selected Answer: - Upvotes: 3

The answer is correct.

Comment 16

ID: 398398 User: TheB Badges: - Relative Date: 4 years, 8 months ago Absolute Date: Sun 04 Jul 2021 16:00 Selected Answer: - Upvotes: 7

The answer looks correct.

38. AI-102 Topic 3 Question 9

Sequence
105
Discussion ID
60311
Source URL
https://www.examtopics.com/discussions/microsoft/view/60311-exam-ai-102-topic-3-question-9-discussion/
Posted By
alan007
Posted At
Aug. 23, 2021, 4:51 a.m.

Question

HOTSPOT -
You are developing an application that includes language translation.
The application will translate text retrieved by using a function named getTextToBeTranslated. The text can be in one of many languages. The content of the text must remain within the Americas Azure geography.
You need to develop code to translate the text to a single language.
How should you complete the code? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Hot Area:
image

Suggested Answer

image
Answer Description Click to expand


Comments 15 comments Click to expand

Comment 1

ID: 527675 User: Shourya007 Badges: Highly Voted Relative Date: 4 years, 1 month ago Absolute Date: Wed 19 Jan 2022 16:42 Selected Answer: - Upvotes: 30

Wrong , correct answer is both last option.
api-nam.cognitive.microsofttranslator.com
'?to=en'

Comment 2

ID: 1146291 User: anto69 Badges: Highly Voted Relative Date: 2 years, 1 month ago Absolute Date: Sat 10 Feb 2024 12:40 Selected Answer: - Upvotes: 8

insane way to ask a question

Comment 2.1

ID: 1241104 User: eskimolight Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Wed 03 Jul 2024 03:18 Selected Answer: - Upvotes: 1

Hhahhahha

Comment 3

ID: 1399535 User: 35e5f6e Badges: Most Recent Relative Date: 12 months ago Absolute Date: Mon 17 Mar 2025 06:24 Selected Answer: - Upvotes: 3

A and C
https://api.cognitive.microsofttranslator.com/translate?api-version=3.0&to=en

https://learn.microsoft.com/en-us/training/modules/translate-text-with-translator-service/3-understand-language-detection-translation-transliteration

Comment 4

ID: 1286642 User: master9 Badges: - Relative Date: 1 year, 5 months ago Absolute Date: Fri 20 Sep 2024 04:17 Selected Answer: - Upvotes: 2

var endpoint = "https://api-nam.cognitive.microsofttranslator.com/translate";
var apiKey = "FF956C68B83821838691ABD200A4C606";
var text = getTextToBeTranslated();
var body = '[{"Text":"' + text + '"}]';
var client = new HttpClient();
client.DefaultRequestHeaders.Add("Ocp-Apim-Subscription-Key", apiKey);
var uri = endpoint + "?to=en";
HttpResponseMessage response;
var content = new StringContent(body, Encoding.UTF8, "application/json");
response = await client.PutAsync(uri, content);

Comment 5

ID: 1217566 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 16:14 Selected Answer: - Upvotes: 3

api-nam/translate
to=en

Comment 6

ID: 1217353 User: funny_penguin Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 11:37 Selected Answer: - Upvotes: 2

on exam, /translate and to=en

Comment 7

ID: 1183272 User: varinder82 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Tue 26 Mar 2024 12:55 Selected Answer: - Upvotes: 1

Final Answer
1. /translate
2. to=en

Comment 8

ID: 1046873 User: katrang Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Wed 18 Oct 2023 13:52 Selected Answer: - Upvotes: 2

api-nam....../translate
?to=en

Comment 9

ID: 839892 User: RAN_L Badges: - Relative Date: 2 years, 12 months ago Absolute Date: Wed 15 Mar 2023 14:00 Selected Answer: - Upvotes: 1

var endpoint = "https://api-nam.cognitive.microsofttranslator.com/translate";
var apiKey = "FF956C68883821838691A8D200A4C606";
var text = getTextToBeTranslated();
var body = "[{\"Text\":\"" + text + "\"}]";
var client = new HttpClient();
client.DefaultRequestHeaders.Add("Ocp-Apim-Subscription-Key", apiKey);
var uri = endpoint + "?to=en";
var content = new StringContent(body, Encoding.UTF8, "application/json");
HttpResponseMessage response = await client.PutAsync(uri, content);
string translatedText = await response.Content.ReadAsStringAsync();

Comment 10

ID: 633090 User: Eltooth Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Mon 18 Jul 2022 18:33 Selected Answer: - Upvotes: 1

Box 1: api-nam/translate : https://docs.microsoft.com/en-us/azure/cognitive-services/translator/reference/v3-0-reference#base-urls

Box 2: "?to=en";

Comment 11

ID: 612716 User: PHD_CHENG Badges: - Relative Date: 3 years, 9 months ago Absolute Date: Tue 07 Jun 2022 13:30 Selected Answer: - Upvotes: 2

Was on exam 7 Jun 2022

Comment 12

ID: 468491 User: Adedoyin_Simeon Badges: - Relative Date: 4 years, 4 months ago Absolute Date: Wed 27 Oct 2021 12:14 Selected Answer: - Upvotes: 3

SuperPetey is correct.

ref:
https://docs.microsoft.com/en-us/azure/cognitive-services/translator/reference/v3-0-reference#base-urls

Comment 13

ID: 429620 User: alan007 Badges: - Relative Date: 4 years, 6 months ago Absolute Date: Mon 23 Aug 2021 04:51 Selected Answer: - Upvotes: 7

wrong for first choice,should be API.cognitive……./tanslate

Comment 13.1

ID: 430587 User: SuperPetey Badges: - Relative Date: 4 years, 6 months ago Absolute Date: Tue 24 Aug 2021 10:07 Selected Answer: - Upvotes: 43

incorrect - the question specifies it should be routed to the Americas region. The correct answer for box 1 is api-nam.cognitive.microsofttranslator.com/translate according to doc @

https://docs.microsoft.com/en-us/azure/cognitive-services/translator/reference/v3-0-reference#base-urls

I agree second drop-down is '?to=en'

39. AI-102 Topic 4 Question 1

Sequence
109
Discussion ID
55451
Source URL
https://www.examtopics.com/discussions/microsoft/view/55451-exam-ai-102-topic-4-question-1-discussion/
Posted By
WillyMac
Posted At
June 16, 2021, 4:19 p.m.

Question

HOTSPOT -
You are developing a text processing solution.
You develop the following method.
image
You call the method by using the following code.
GetKeyPhrases(textAnalyticsClient, "the cat sat on the mat");
For each of the following statements, select Yes if the statement is true. Otherwise, select No.
NOTE: Each correct selection is worth one point.
Hot Area:
image

Suggested Answer

image
Answer Description Click to expand

Box 1: Yes -
The Key Phrase Extraction API evaluates unstructured text, and for each JSON document, returns a list of key phrases.

Box 2: No -
'the' is not a key phrase.
This capability is useful if you need to quickly identify the main points in a collection of documents. For example, given input text "The food was delicious and there were wonderful staff", the service returns the main talking points: "food" and "wonderful staff".

Box 3: No -
Key phrase extraction does not have confidence levels.
Reference:
https://docs.microsoft.com/en-us/azure/cognitive-services/text-analytics/how-tos/text-analytics-how-to-keyword-extraction

Comments 10 comments Click to expand

Comment 1

ID: 392678 User: joecruz Badges: Highly Voted Relative Date: 4 years, 8 months ago Absolute Date: Mon 28 Jun 2021 09:24 Selected Answer: - Upvotes: 35

I think the answer is correct

Comment 1.1

ID: 1368534 User: ABI2091 Badges: - Relative Date: 1 year ago Absolute Date: Sun 09 Mar 2025 08:24 Selected Answer: - Upvotes: 1

Can someone tell me why it wont display confidence level?

Comment 2

ID: 1282653 User: mrg998 Badges: Most Recent Relative Date: 1 year, 6 months ago Absolute Date: Thu 12 Sep 2024 15:28 Selected Answer: - Upvotes: 1

YNN the way forward

Comment 3

ID: 1218715 User: reiwanotora Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sun 26 May 2024 05:37 Selected Answer: - Upvotes: 2

Yes
No
No

Comment 4

ID: 1177464 User: rober13 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Tue 19 Mar 2024 15:55 Selected Answer: - Upvotes: 2

Same Question 67, Topic 3

Comment 5

ID: 633406 User: Eltooth Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Tue 19 Jul 2022 08:46 Selected Answer: - Upvotes: 2

Answer is correct.
Yes
No
No

Comment 6

ID: 398500 User: Adedoyin_Simeon Badges: - Relative Date: 4 years, 8 months ago Absolute Date: Sun 04 Jul 2021 18:17 Selected Answer: - Upvotes: 4

The answer is correct

Comment 7

ID: 395051 User: azurelearner666 Badges: - Relative Date: 4 years, 8 months ago Absolute Date: Wed 30 Jun 2021 19:54 Selected Answer: - Upvotes: 4

The response is correct.
Check the sample code at https://github.com/Azure/azure-sdk-for-net/blob/master/sdk/textanalytics/Azure.AI.TextAnalytics/samples/Sample3_ExtractKeyPhrases.md

Comment 8

ID: 383470 User: WillyMac Badges: - Relative Date: 4 years, 8 months ago Absolute Date: Wed 16 Jun 2021 16:19 Selected Answer: - Upvotes: 1

The First option is also NO.
The code doesn't function, because "response" is of Type "KeyPhraseCollection", and this Class is a ReadOnlyCollection<string> and does not have a definition for: "Value"

Comment 8.1

ID: 387120 User: cavefish Badges: - Relative Date: 4 years, 8 months ago Absolute Date: Mon 21 Jun 2021 15:00 Selected Answer: - Upvotes: 2

response is of type Response<KeyPhraseCollection>, so value is the field with the result collection https://docs.microsoft.com/en-us/azure/cognitive-services/text-analytics/quickstarts/client-libraries-rest-api?tabs=version-3-1&pivots=programming-language-csharp#key-phrase-extraction

40. AI-102 Topic 1 Question 63

Sequence
113
Discussion ID
135106
Source URL
https://www.examtopics.com/discussions/microsoft/view/135106-exam-ai-102-topic-1-question-63-discussion/
Posted By
Harry300
Posted At
March 3, 2024, 10:41 a.m.

Question

You are building a solution in Azure that will use Azure Cognitive Service for Language to process sensitive customer data.

You need to ensure that only specific Azure processes can access the Language service. The solution must minimize administrative effort.

What should you include in the solution?

  • A. IPsec rules
  • B. Azure Application Gateway
  • C. a virtual network gateway
  • D. virtual network rules

Suggested Answer

D

Answer Description Click to expand


Community Answer Votes

Comments 6 comments Click to expand

Comment 1

ID: 1164663 User: Harry300 Badges: Highly Voted Relative Date: 2 years ago Absolute Date: Sun 03 Mar 2024 10:41 Selected Answer: D Upvotes: 6

Correct.
Source: https://learn.microsoft.com/en-us/azure/ai-services/cognitive-services-virtual-networks?tabs=portal

Comment 2

ID: 1365425 User: Mattt Badges: Most Recent Relative Date: 1 year ago Absolute Date: Wed 05 Mar 2025 15:18 Selected Answer: D Upvotes: 1

D is correct

Comment 3

ID: 1235662 User: rookiee1111 Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sun 23 Jun 2024 03:10 Selected Answer: - Upvotes: 4

answer is virtual network rules. why?
because it helps configure IP range address to restrict access to an end point only to services deployed in those IP ranges. this way the access can be restricted. Gateway can work but too much of an effort, and IPsec is for securing an already established channel of communication between 2 points.

Comment 4

ID: 1214371 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Mon 20 May 2024 15:55 Selected Answer: D Upvotes: 3

I believe D is right answer.

Comment 5

ID: 1214010 User: dragongoseki Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Mon 20 May 2024 00:33 Selected Answer: D Upvotes: 1

D is OK.

Comment 6

ID: 1194122 User: michaelmorar Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Fri 12 Apr 2024 05:43 Selected Answer: D Upvotes: 2

This question is more relevant to AZ104 (Azure Administration) rather than AI, but anyway, I reckon the answer is D.

41. AI-102 Topic 5 Question 6

Sequence
114
Discussion ID
75984
Source URL
https://www.examtopics.com/discussions/microsoft/view/75984-exam-ai-102-topic-5-question-6-discussion/
Posted By
g2000
Posted At
May 20, 2022, 10:22 p.m.

Question

DRAG DROP -
You are using a Language Understanding service to handle natural language input from the users of a web-based customer agent.
The users report that the agent frequently responds with the following generic response: "Sorry, I don't understand that."
You need to improve the ability of the agent to respond to requests.
Which three actions should you perform in sequence? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.
Select and Place:
image

Suggested Answer

image
Answer Description Click to expand


Comments 12 comments Click to expand

Comment 1

ID: 604626 User: g2000 Badges: Highly Voted Relative Date: 3 years, 9 months ago Absolute Date: Fri 20 May 2022 22:22 Selected Answer: - Upvotes: 69

1. Enable active learning
2. Validate the utterances logged for review and modify the model
3. Train and republish the language understanding model
not sure how the pre built model can improve accuracy....

Comment 1.1

ID: 633597 User: Eltooth Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Tue 19 Jul 2022 15:30 Selected Answer: - Upvotes: 3

Agreed

Comment 1.2

ID: 708489 User: GigaCaster Badges: - Relative Date: 3 years, 4 months ago Absolute Date: Mon 31 Oct 2022 15:28 Selected Answer: - Upvotes: 4

But the question isn't asking for accuracy but rather making it able to give a response instead of the current "Sorry, I don't understand that." response the customers are receiving. Thus the answer wants a short-term quick fix so that the team can work on the accuracy later, leading to a pre-built model being used.

Comment 1.3

ID: 646961 User: ninjia Badges: - Relative Date: 3 years, 6 months ago Absolute Date: Mon 15 Aug 2022 02:25 Selected Answer: - Upvotes: 2

Agreed

Comment 2

ID: 1363939 User: gyaansastra Badges: Most Recent Relative Date: 1 year ago Absolute Date: Sun 02 Mar 2025 13:26 Selected Answer: - Upvotes: 2

Absolutely wrong. To improve the ability of a Language Understanding service that's frequently responding with "Sorry, I don't understand that," I need to select three actions in the correct sequence.

The appropriate sequence is:

1. Enable log collection by using Log Analytics.
2. Validate the utterances logged for review and modify the model.
3. Train and republish the Language Understanding model.

This approach follows a logical improvement cycle:

First, you need data about what users are actually saying that the model doesn't understand, which Log Analytics will provide by collecting user interactions.

Second, you examine these logged utterances to identify patterns in what's being missed, then modify your model to address these gaps (adding intents, entities, or example utterances as needed).

Finally, after making these improvements, you train the updated model to incorporate the changes and republish it so users can benefit from the improvements.

This sequence creates a data-driven approach to iteratively improving the model based on real user interactions.

Comment 3

ID: 1248574 User: SAMBIT Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Mon 15 Jul 2024 22:48 Selected Answer: - Upvotes: 2

a) Add prebuilt domain models as required. Validate the utterances logged for review and modify the model. (2)
b) Migrate authoring to an Azure resource authoring key.
c) Enable active learning.
d) Enable log collection by using Log Analytics. (1)
e) Train and republish the Language (3)
f) Understanding model.

Comment 4

ID: 1231450 User: etellez Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sun 16 Jun 2024 18:34 Selected Answer: - Upvotes: 4

Copilot says

To improve the ability of the agent to respond to requests, you should perform the following actions in sequence:

Enable log collection using Log Analytics: This will allow you to collect and analyze the utterances that the agent is not understanding.

Enable active learning: Active learning will help the model to learn from the utterances it didn't understand and improve over time.

Train and republish the language understanding model: After enabling active learning and collecting more data, you should retrain the model and publish it again. This will incorporate the new learnings into the model and should improve its ability to understand user requests.

Comment 4.1

ID: 1233669 User: etellez Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Thu 20 Jun 2024 16:02 Selected Answer: - Upvotes: 1

Today the answer has been different
Enable active learning.
Validate the utterances logged for review and modify the model.
Train and republish the Language Understanding model.

Comment 5

ID: 1222688 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sat 01 Jun 2024 16:13 Selected Answer: - Upvotes: 1

1. Enable active learning
2. Validate the utterances logged for review and modify the model
3. Train and republish the language understanding model

Comment 6

ID: 1184883 User: varinder82 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Thu 28 Mar 2024 17:22 Selected Answer: - Upvotes: 3

Final Answer:

1. Enable active learning
2. Validate the utterances logged for review and modify the model
3. Train and republish the language understanding model

Comment 7

ID: 1139860 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sun 04 Feb 2024 08:19 Selected Answer: - Upvotes: 4

given answer is NOT Correct!
why:
Enable Active Learning: LUIS's Active Learning feature automatically identifies vague or low-confidence user inputs that may require further training.
Review and Add Utterances Logged for Review: Utterances collected through Active Learning are marked for review. Regularly reviewing and categorizing these utterances into the relevant intents is a crucial step in enhancing model performance.
Train and Republish the Language Understanding Model: After adding new utterances and potentially adjusting intents and entities, you need to retrain the LUIS model to incorporate these updates.

Comment 8

ID: 633598 User: Eltooth Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Tue 19 Jul 2022 15:30 Selected Answer: - Upvotes: 4

1. Enable active learning
2. Validate the utterances logged for review and modify the model
3. Train and republish the language understanding model

42. AI-102 Topic 5 Question 9

Sequence
115
Discussion ID
55006
Source URL
https://www.examtopics.com/discussions/microsoft/view/55006-exam-ai-102-topic-5-question-9-discussion/
Posted By
kubikjakubik
Posted At
June 9, 2021, 9:36 p.m.

Question

HOTSPOT -
You are reviewing the design of a chatbot. The chatbot includes a language generation file that contains the following fragment.
# Greet(user)
- ${Greeting()}, ${user.name}
For each of the following statements, select Yes if the statement is true. Otherwise, select No.
NOTE: Each correct selection is worth one point.
Hot Area:
image

Suggested Answer

image
Answer Description Click to expand


Comments 13 comments Click to expand

Comment 1

ID: 1139875 User: evangelist Badges: Highly Voted Relative Date: 2 years, 1 month ago Absolute Date: Sun 04 Feb 2024 08:45 Selected Answer: - Upvotes: 13

COrrect answer is : NYY
why:
${user.name} retrieves the user's name using a prompt.-->No.
In this context, ${user.name} directly retrieves the user's name from the incoming user object. It doesn't obtain the user's name through a prompt but assumes that the user object already contains the name property.

Greet() is the name of the language generation template.-->Yes.
# Greet(user) defines a language generation template named Greet. This template takes a user parameter, indicating that Greet is the template name used to generate personalized greeting messages.

${Greeting()} is a reference to a template in the language generation file.-->Yes.
${Greeting()} is a reference to another template in the language generation file. This syntax indicates that Greeting is a template that will be invoked to generate all or part of a greeting when messages are generated. This allows developers to reuse and combine templates in the language generation file to create more complex messages.

Comment 2

ID: 1363943 User: gyaansastra Badges: Most Recent Relative Date: 1 year ago Absolute Date: Sun 02 Mar 2025 13:36 Selected Answer: - Upvotes: 2

1. "${user.name} retrieves the user name by using a prompt."
No - The ${user.name} syntax is accessing a property of the user object passed to the template. It's not retrieving the name through a prompt but using data that's already available in the context.

2. "Greet() is the name of the language generation template."
Yes - The "#" symbol indicates a template definition in LG files, and "Greet(user)" is the name of this template with a parameter called "user".

3. "${Greeting()} is a reference to a template in the language generation file."
Yes - The ${} syntax indicates a template reference in LG files, so ${Greeting()} is calling another template named "Greeting" that must be defined elsewhere in the same language generation file.

Comment 3

ID: 1234504 User: HaraTadahisa Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Fri 21 Jun 2024 16:51 Selected Answer: - Upvotes: 1

My answer is that
No
Yes
Yes

Comment 4

ID: 1230853 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sat 15 Jun 2024 10:51 Selected Answer: - Upvotes: 1

No
Yes
Yes

Comment 5

ID: 1057132 User: Lion007 Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Sun 29 Oct 2023 22:08 Selected Answer: - Upvotes: 1

${user.name} retrieves the “name” parameter from the “user” by using PARAMETER (NOT Prompt). Check out the Parametrization of Templates section https://learn.microsoft.com/en-us/azure/bot-service/file-format/bot-builder-lg-file-format?view=azure-bot-service-4.0#parametrization-of-templates

Greet() is indeed the NAME of the language generation template. Why not Greet(user) is the name, I hear you asking? Not including the parameter “user” in the question doesn’t change the fact that what comes after the hashtag # is the NAME of the template, only without referring to its parameter. Again, check out Parametrization of Templates section https://learn.microsoft.com/en-us/azure/bot-service/file-format/bot-builder-lg-file-format?view=azure-bot-service-4.0#parametrization-of-templates

${Greeting()} is a reference to another template’s name called Greeting() template.

Comment 5.1

ID: 1057135 User: Lion007 Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Sun 29 Oct 2023 22:09 Selected Answer: - Upvotes: 1

N.Y.Y.

Comment 6

ID: 1048351 User: sl_mslconsulting Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Fri 20 Oct 2023 05:49 Selected Answer: - Upvotes: 1

Just think of $ as the eval in JavaScript. It evals an expression. A template is just like a function you can invoke. The answers are then quite evident.

Comment 7

ID: 912590 User: kail85 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Fri 02 Jun 2023 08:53 Selected Answer: - Upvotes: 1

No, Yes, Yes

Comment 8

ID: 820612 User: NNU Badges: - Relative Date: 3 years ago Absolute Date: Fri 24 Feb 2023 15:40 Selected Answer: - Upvotes: 3

the user.name retrieves from .log file, the anwser is No
second is yes Greet() is the name of LG template
# TemplateName
- You can say cheese and tomato \[toppings are optional\]
third yes is a reference to a template
Variation text can include references to another named template to aid with composition and resolution of sophisticated responses. References to other named templates are denoted using braces, such as ${<TemplateName>()}.
https://learn.microsoft.com/en-us/azure/bot-service/file-format/bot-builder-lg-file-format?view=azure-bot-service-4.0

Comment 9

ID: 635199 User: Eltooth Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Fri 22 Jul 2022 14:30 Selected Answer: - Upvotes: 2

No
No
Yes

Comment 10

ID: 378500 User: kubikjakubik Badges: - Relative Date: 4 years, 9 months ago Absolute Date: Wed 09 Jun 2021 21:36 Selected Answer: - Upvotes: 4

First is true. You even have it in your reference.
https://docs.microsoft.com/en-us/composer/how-to-ask-for-user-input
"Hello ${user.name}, nice to talk to you!" ... so the part ${user.name} in fact does retrieve the name of the user.

Comment 10.1

ID: 814627 User: AzureJobsTillRetire Badges: - Relative Date: 3 years ago Absolute Date: Sun 19 Feb 2023 23:32 Selected Answer: - Upvotes: 1

One way of getting the user name is by using a prompt, but there are other ways as well

Comment 10.2

ID: 395201 User: azurelearner666 Badges: - Relative Date: 4 years, 8 months ago Absolute Date: Wed 30 Jun 2021 22:46 Selected Answer: - Upvotes: 28

it does retrieve the user name, but NOT USING A PROMPT. retrieves the already stored property.
But without a PROMPT.

so, the response is correct, it's a NO.
1. NO
2. NO
3. YES

43. AI-102 Topic 5 Question 11

Sequence
116
Discussion ID
54965
Source URL
https://www.examtopics.com/discussions/microsoft/view/54965-exam-ai-102-topic-5-question-11-discussion/
Posted By
motu
Posted At
June 9, 2021, 11:41 a.m.

Question

HOTSPOT -
You are building a chatbot that will provide information to users as shown in the following exhibit.
image
Use the drop-down menus to select the answer choice that completes each statement based on the information presented in the graphic.
NOTE: Each correct selection is worth one point.
Hot Area:
image

Suggested Answer

image
Answer Description Click to expand


Comments 22 comments Click to expand

Comment 1

ID: 378144 User: motu Badges: Highly Voted Relative Date: 4 years, 9 months ago Absolute Date: Wed 09 Jun 2021 11:41 Selected Answer: - Upvotes: 46

ET answer is wrong. It's an adaptive card using an image.

Comment 1.1

ID: 414720 User: PaPaTee Badges: - Relative Date: 4 years, 7 months ago Absolute Date: Mon 26 Jul 2021 15:31 Selected Answer: - Upvotes: 7

Yes, the answer is Adaptive Card. The image is a combination of many texts, images, with date and currency fields.

Comment 1.2

ID: 506433 User: plalwa Badges: - Relative Date: 4 years, 2 months ago Absolute Date: Tue 21 Dec 2021 21:57 Selected Answer: - Upvotes: 1

its Thumbnail, Adaptive cards require buttons. Like in MS teams approve PR :)

Comment 2

ID: 506560 User: snilu Badges: Highly Voted Relative Date: 4 years, 2 months ago Absolute Date: Wed 22 Dec 2021 02:06 Selected Answer: - Upvotes: 31

Answer is Adaptive Card
https://docs.microsoft.com/en-us/composer/how-to-send-cards?tabs=v1x

Comment 2.1

ID: 522387 User: Trumpenstein Badges: - Relative Date: 4 years, 1 month ago Absolute Date: Wed 12 Jan 2022 19:57 Selected Answer: - Upvotes: 7

Good Find - the image at the and is nearly identical with the screenshot!

Comment 3

ID: 1363949 User: gyaansastra Badges: Most Recent Relative Date: 1 year ago Absolute Date: Sun 02 Mar 2025 13:59 Selected Answer: - Upvotes: 1

Based on the flight information display shown in the chatbot interface:

For the first dropdown:
The chatbot is showing - an Adaptive Card

Adaptive Cards are the most appropriate choice for this complex, structured layout that displays detailed flight information with formatting, sections, and visual elements.

For the second dropdown:
The card includes - an action set

The interface appears to allow users to interact with different flight options (2 Stops vs. Non-Stop), which would be implemented using an action set to enable selection and interaction with the presented flight choices.

These components together create an interactive flight booking experience within the chatbot.

Comment 4

ID: 1288839 User: master9 Badges: - Relative Date: 1 year, 5 months ago Absolute Date: Wed 25 Sep 2024 03:38 Selected Answer: - Upvotes: 2

it should be:
1. An Adaptive card
2. an action set

because both are link with each other and more suitable for given scenario.

ction sets are collections of buttons or other interactive elements that allow users to perform actions directly within the card, without having to navigate to a separate screen or page. This can improve the user experience by making it easier for users to complete tasks.

For example, an action set in a flight booking card might include buttons for selecting a different flight, changing the number of passengers, or proceeding to payment. By including action sets in their cards, chatbots can provide a more seamless and efficient user experience.

Comment 5

ID: 1247923 User: krzkrzkra Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sun 14 Jul 2024 20:19 Selected Answer: - Upvotes: 1

1. Adaptive Card
2. image

Comment 6

ID: 1222712 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sat 01 Jun 2024 17:02 Selected Answer: - Upvotes: 1

1. an Adaptive Card
2. an image

Comment 7

ID: 1184938 User: varinder82 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Thu 28 Mar 2024 19:11 Selected Answer: - Upvotes: 2

Final Answer:
1- Adaptive Card
2- Image

Comment 8

ID: 1139885 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sun 04 Feb 2024 08:56 Selected Answer: - Upvotes: 3

The answer is adaptive card and it is the same on Microsoft Documentation:
https://learn.microsoft.com/en-us/composer/how-to-send-cards?tabs=v1x#adaptive-card

Comment 9

ID: 1049148 User: sl_mslconsulting Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Fri 20 Oct 2023 23:54 Selected Answer: - Upvotes: 1

Here is how you compose this card: https://adaptivecards.io/samples/FlightItinerary.html.
There are many text blocks and two images. Now you know what to choose.

Comment 9.1

ID: 1049151 User: sl_mslconsulting Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Sat 21 Oct 2023 00:00 Selected Answer: - Upvotes: 1

You really need to dig a bit deeper to find the right answer and get your hands dirty. It’s very easy to show during your interviews that you are just guessing and don’t really understand the topic.

Comment 9.2

ID: 1049154 User: sl_mslconsulting Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Sat 21 Oct 2023 00:05 Selected Answer: - Upvotes: 1

*Same image source but shows up twice https://adaptivecards.io/content/airplane.png

Comment 10

ID: 984771 User: james2033 Badges: - Relative Date: 2 years, 6 months ago Absolute Date: Fri 18 Aug 2023 22:37 Selected Answer: - Upvotes: 1

1. is showing Adaptive card https://learn.microsoft.com/en-us/composer/how-to-send-cards?tabs=v1x#adaptive-card

2. image (just 1 image) of icon file.

Comment 11

ID: 814654 User: AzureJobsTillRetire Badges: - Relative Date: 3 years ago Absolute Date: Mon 20 Feb 2023 00:16 Selected Answer: - Upvotes: 2

The given answers are correct.

First of all, everyone agrees that the answer to the second question is an imagine and not an action set, or an image group, or media. Hence the card includes an image and not many images. If there is only one image, the answer to the first question has to be a Thumbnail card, as an adaptive card is highly customizable and can contain any combination of text, speech, images, buttons, and input fields, which is not required.

Comment 11.1

ID: 827312 User: AzureJobsTillRetire Badges: - Relative Date: 3 years ago Absolute Date: Thu 02 Mar 2023 20:19 Selected Answer: - Upvotes: 5

I was wrong. I think it is Adaptive Card. Although an adaptive card contains any combination of text, speech, images, buttons, and input fields, it can also be rendered into a single PNG image.

https://learn.microsoft.com/en-us/adaptive-cards/sdk/rendering-cards/net-image/getting-started

Comment 12

ID: 633603 User: Eltooth Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Tue 19 Jul 2022 15:47 Selected Answer: - Upvotes: 2

Answer is correct.

Adaptive Card
Image

https://docs.microsoft.com/en-us/composer/how-to-send-cards?tabs=v1x

Comment 13

ID: 559356 User: reachmymind Badges: - Relative Date: 4 years ago Absolute Date: Wed 02 Mar 2022 12:24 Selected Answer: - Upvotes: 2

an Adaptive Card
an image

an Adaptive Card that is a JSON-serialized card object model rendered into a PNG image

https://docs.microsoft.com/en-us/azure/bot-service/bot-service-design-user-experience?view=azure-bot-service-4.0
https://docs.microsoft.com/en-us/adaptive-cards/sdk/rendering-cards/net-image/render-a-card

Comment 14

ID: 513560 User: RamonRW Badges: - Relative Date: 4 years, 2 months ago Absolute Date: Thu 30 Dec 2021 18:41 Selected Answer: - Upvotes: 2

The question is a bit confusing, as it is not entirely clear, if this is a sent screenshot. I think, if it is a screenshot Thumbnail and image makes absolutely sense, but if it is a "constructed" message, adaptive card would make sense. But due to the complexity of the adaptive card (images of the plane, information about passenger) several answers need to be right in the second element. Therefore, I believe the answer is correct. But the question is not really a good one.

Comment 14.1

ID: 637902 User: ParkXD Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Wed 27 Jul 2022 09:45 Selected Answer: - Upvotes: 1

I agree with what you say.
And for the difference between cards: https://docs.microsoft.com/en-us/microsoftteams/platform/task-modules-and-cards/cards/cards-reference.

Comment 15

ID: 396862 User: SAANCODE Badges: - Relative Date: 4 years, 8 months ago Absolute Date: Fri 02 Jul 2021 14:25 Selected Answer: - Upvotes: 3

The answer is correct

44. AI-102 Topic 3 Question 18

Sequence
118
Discussion ID
54520
Source URL
https://www.examtopics.com/discussions/microsoft/view/54520-exam-ai-102-topic-3-question-18-discussion/
Posted By
hiabcd
Posted At
June 4, 2021, 12:54 p.m.

Question

Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.
You build a language model by using a Language Understanding service. The language model is used to search for information on a contact list by using an intent named FindContact.
A conversational expert provides you with the following list of phrases to use for training.
✑ Find contacts in London.
✑ Who do I know in Seattle?
✑ Search for contacts in Ukraine.
You need to implement the phrase list in Language Understanding.
Solution: You create a new intent for location.
Does this meet the goal?

  • A. Yes
  • B. No

Suggested Answer

B

Answer Description Click to expand


Community Answer Votes

Comments 18 comments Click to expand

Comment 1

ID: 374271 User: hiabcd Badges: Highly Voted Relative Date: 4 years, 9 months ago Absolute Date: Fri 04 Jun 2021 12:54 Selected Answer: - Upvotes: 34

This should be "No"

Comment 1.1

ID: 398460 User: iyiola_daniel Badges: - Relative Date: 4 years, 8 months ago Absolute Date: Sun 04 Jul 2021 17:07 Selected Answer: - Upvotes: 8

I think the answer should be 'No' too. The intent is for FindContact, not location really.

Comment 2

ID: 445805 User: zubfaruq Badges: Highly Voted Relative Date: 4 years, 5 months ago Absolute Date: Thu 16 Sep 2021 11:40 Selected Answer: - Upvotes: 12

The answer should be NO.
An utterance having wo intents? This is illogical.
The model should have an Entity "Location" that will help in finding the contacts.

Reference: https://docs.microsoft.com/en-us/azure/cognitive-services/luis/luis-concept-intent

Comment 3

ID: 1361901 User: gyaansastra Badges: Most Recent Relative Date: 1 year ago Absolute Date: Wed 26 Feb 2025 13:28 Selected Answer: B Upvotes: 2

Creating a new intent for location does not meet the goal. Instead, you should add these phrases as utterances under the existing FindContact intent. This allows the model to learn and recognize these specific phrases and similar variations when users are searching for contacts.

Comment 4

ID: 1279646 User: famco Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Fri 06 Sep 2024 16:54 Selected Answer: - Upvotes: 1

And for the other question about whether patterns are enough:
Do not expect to see improved entity prediction if you collapse multiple utterances into a single pattern. For simple entities to be utilized by your app, you need to add utterances or use list entities.

Comment 5

ID: 1261844 User: moonlightc Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Wed 07 Aug 2024 00:06 Selected Answer: B Upvotes: 1

Simply 'location' is an entity not an intent

Comment 6

ID: 1234477 User: HaraTadahisa Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Fri 21 Jun 2024 16:37 Selected Answer: B Upvotes: 1

I say B is correct! HipHop!

Comment 7

ID: 1230990 User: fuck_india Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 15 Jun 2024 16:02 Selected Answer: - Upvotes: 1

This issue is not working because everyone's opinion is 50-50. I don't know the correct answer, so I will be staring at MS Learn when this question comes up on the real exam.Examtopic should scrutinize and post the correct answer.

Comment 8

ID: 1230985 User: fuck_india Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 15 Jun 2024 16:02 Selected Answer: - Upvotes: 1

This issue is not working because everyone's opinion is 50-50. I don't know the correct answer, so I will be staring at MS Learn when this question comes up on the real exam.Examtopic should scrutinize and post the correct answer.

Comment 9

ID: 1133036 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 27 Jan 2024 04:47 Selected Answer: - Upvotes: 2

The ANSWER is NO !
Instead of creating a new intent for location, you should:
Add these phrases to the "FindContact" intent: This helps the model understand that these are different ways a user might express the intent to find contacts.

Use Entities for Locations: In addition to training the intent with these phrases, you should use entities to capture the location information within the phrases. In LUIS, you can define an entity (like "Location") and then annotate locations in your training phrases. This way, the model learns to recognize and extract location details from user inputs.

By combining a well-trained "FindContact" intent with a location entity, your language model will be better equipped to understand and process user queries about finding contacts in specific locations.

Comment 10

ID: 932382 User: Hisayuki Badges: - Relative Date: 2 years, 8 months ago Absolute Date: Sat 24 Jun 2023 11:00 Selected Answer: B Upvotes: 1

Location should be for entity not for Intent

Comment 11

ID: 915749 User: ziggy1117 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Tue 06 Jun 2023 00:15 Selected Answer: - Upvotes: 1

Should be NO. An intent represents a task or action the user wants to perform. It is a purpose or goal expressed in a user's utterance. In this case, the Intent is to find the Contact. And it is already created.

The utterances specified have the same Intent which is to find/search for contacts. Hence they should be added to the list and an Entity should be created to capture location.

Comment 12

ID: 817659 User: Marilena96 Badges: - Relative Date: 3 years ago Absolute Date: Wed 22 Feb 2023 10:25 Selected Answer: - Upvotes: 3

Think its Yes, creating a new intent for location would meet the goal of implementing the phrase list in Language Understanding. You can add the phrases to the new intent and label the location entity in each phrase. This will help the language model understand the user's intent to search for contacts in a specific location.

Comment 13

ID: 779287 User: ap1234pa Badges: - Relative Date: 3 years, 1 month ago Absolute Date: Tue 17 Jan 2023 20:51 Selected Answer: B Upvotes: 1

This is "NO"

Comment 14

ID: 635178 User: Eltooth Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Fri 22 Jul 2022 13:07 Selected Answer: B Upvotes: 2

B is correct answer.

Needs Entity/type for location instead.

Comment 15

ID: 597427 User: zb99 Badges: - Relative Date: 3 years, 10 months ago Absolute Date: Thu 05 May 2022 20:15 Selected Answer: B Upvotes: 1

Definitely no.

Comment 16

ID: 517834 User: sumanshu Badges: - Relative Date: 4 years, 2 months ago Absolute Date: Wed 05 Jan 2022 23:10 Selected Answer: - Upvotes: 2

intent (means what task we need to perform) - So here task is same i.e. Find Contacts. So No need to add an new intent.

But need to add type/entity : location

Comment 17

ID: 485824 User: TanujitRoy Badges: - Relative Date: 4 years, 3 months ago Absolute Date: Wed 24 Nov 2021 10:31 Selected Answer: B Upvotes: 2

This should be NO

45. AI-102 Topic 1 Question 1

Sequence
123
Discussion ID
54103
Source URL
https://www.examtopics.com/discussions/microsoft/view/54103-exam-ai-102-topic-1-question-1-discussion/
Posted By
VT100
Posted At
June 1, 2021, 6:23 p.m.

Question

DRAG DROP -
You have 100 chatbots that each has its own Language Understanding model.
Frequently, you must add the same phrases to each model.
You need to programmatically update the Language Understanding models to include the new phrases.
How should you complete the code? To answer, drag the appropriate values to the correct targets. Each value may be used once, more than once, or not at all.
You may need to drag the split bar between panes or scroll to view content.
NOTE: Each correct selection is worth one point.
Select and Place:
image

Suggested Answer

image
Answer Description Click to expand

Box 1: AddPhraseListAsync -

Example: Add phraselist feature -
var phraselistId = await client.Features.AddPhraseListAsync(appId, versionId, new PhraselistCreateObject
{
EnabledForAllModels = false,
IsExchangeable = true,
Name = "QuantityPhraselist",
Phrases = "few,more,extra"
});

Box 2: PhraselistCreateObject -
Reference:
https://docs.microsoft.com/en-us/azure/cognitive-services/luis/client-libraries-rest-api

Comments 7 comments Click to expand

Comment 1

ID: 561183 User: reachmymind Badges: Highly Voted Relative Date: 4 years ago Absolute Date: Sat 05 Mar 2022 05:43 Selected Answer: - Upvotes: 16

Box 1: AddPhraseListAsync
Box 2: PhraselistCreateObject

var phraselistId = await client.Features.AddPhraseListAsync(appId, versionId, new PhraselistCreateObject
{
EnabledForAllModels = false,
IsExchangeable = true,
Name = "QuantityPhraselist",
Phrases = "few,more,extra"
});

Mapping :
Model - Entity - Async
Feature - PhraseList - Async

Comment 2

ID: 515964 User: sumanshu Badges: Highly Voted Relative Date: 4 years, 2 months ago Absolute Date: Mon 03 Jan 2022 19:37 Selected Answer: - Upvotes: 6

Vote for 'C' as per the provided Link (in Solution itself). But Better go with elimination method.

We want to add the Phrases, So most suitable function name is : AddPhraseListAsync (out of all other options)

Now, for an argument, If we look options (Phrases (it's already passed in body, SavePhrases and UploadPhrases will also get eliminated, because we are manually passing the phrases with in function body). So 2 options left i.e. PhraseList and PhraseListCreateObject, Here in C# code, we are using "new" keyword which is used to instance the object, So "PhraseListCreateObject" more suitable.

Comment 3

ID: 1361424 User: gyaansastra Badges: Most Recent Relative Date: 1 year ago Absolute Date: Tue 25 Feb 2025 12:56 Selected Answer: - Upvotes: 1

Correct Answer. Ref https://learn.microsoft.com/en-us/dotnet/api/microsoft.azure.cognitiveservices.language.luis.authoring.featuresextensions.addphraselistasync?view=azure-dotnet

Comment 4

ID: 1248537 User: krzkrzkra Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Mon 15 Jul 2024 21:09 Selected Answer: - Upvotes: 3

Box 1: AddPhraseListAsync
Box 2: PhraselistCreateObject

Comment 5

ID: 1135998 User: Florin83 Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Tue 30 Jan 2024 18:48 Selected Answer: - Upvotes: 1

// Add phraselist feature
var phraselistId = await client.Features.AddPhraseListAsync(appId, versionId, new PhraselistCreateObject
{
EnabledForAllModels = false,
IsExchangeable = true,
Name = "QuantityPhraselist",
Phrases = "few,more,extra"
});

Comment 6

ID: 384038 User: akinscube Badges: - Relative Date: 4 years, 8 months ago Absolute Date: Thu 17 Jun 2021 10:34 Selected Answer: - Upvotes: 2

Please where can I find an explanation for this. I have gone through the modules but I couldn't get it, and the reference is not that straight forward.

Comment 7

ID: 380549 User: WillyMac Badges: - Relative Date: 4 years, 9 months ago Absolute Date: Sat 12 Jun 2021 17:19 Selected Answer: - Upvotes: 3

Answer is correct.

46. AI-102 Topic 3 Question 40

Sequence
125
Discussion ID
108982
Source URL
https://www.examtopics.com/discussions/microsoft/view/108982-exam-ai-102-topic-3-question-40-discussion/
Posted By
WinzigWeich
Posted At
May 11, 2023, 5:58 p.m.

Question

DRAG DROP -

You develop an app in C# named App1 that performs speech-to-speech translation.

You need to configure App1 to translate English to German.

How should you complete the SpeechTranslationConfig object? To answer, drag the appropriate values to the correct targets. Each value may be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view content.

NOTE: Each correct selection is worth one point.

image

Suggested Answer

image
Answer Description Click to expand


Comments 11 comments Click to expand

Comment 1

ID: 898933 User: MaliSanFuu Badges: Highly Voted Relative Date: 2 years, 10 months ago Absolute Date: Tue 16 May 2023 08:44 Selected Answer: - Upvotes: 34

Agreeing with @WinzigWeich

Answer should be:

1) SpeechRecognitionLanguage
2) AddTargetLanguage

Comment 2

ID: 916032 User: ziggy1117 Badges: Highly Voted Relative Date: 2 years, 9 months ago Absolute Date: Tue 06 Jun 2023 09:58 Selected Answer: - Upvotes: 13

1) SpeechRecognitionLanguage
2) AddTargetLanguage

in the exercise of AI-102 online learning

Comment 3

ID: 1356206 User: sukantadey Badges: Most Recent Relative Date: 1 year ago Absolute Date: Thu 13 Feb 2025 17:58 Selected Answer: - Upvotes: 2

1) SpeechRecognitionLanguage
2) AddTargetLanguage

Comment 4

ID: 1317368 User: jkkkkkaaaaa Badges: - Relative Date: 1 year, 3 months ago Absolute Date: Mon 25 Nov 2024 08:41 Selected Answer: - Upvotes: 1

https://learn.microsoft.com/en-us/azure/ai-services/speech-service/how-to-translate-speech?tabs=terminal&pivots=programming-language-csharp

Comment 5

ID: 1295729 User: Skyhawks Badges: - Relative Date: 1 year, 5 months ago Absolute Date: Thu 10 Oct 2024 22:08 Selected Answer: - Upvotes: 1

In this scenario, addTargetLanguage is not the best choice because you are configuring speech-to-speech translation with Azure Speech Translation API, and you need to set both the speech recognition language (input) and the speech synthesis language (output). If we were working with text translation, addTargetLanguage would be relevant. Because of this the correct answers are: speechRecognitionLanguage to "en-US" and speechSynthesisLanguage to "de"

Comment 6

ID: 1279785 User: famco Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Fri 06 Sep 2024 22:06 Selected Answer: - Upvotes: 2

Wait a minute, Microsoft did it again. They made the first one an assignment and the second one a method call and gave the SpeechSynthesizer as an option. That's e v i l. I would just not notice that trick. How else can you explain that. Just to make the person who knows what SpeechSynthesis and with a time pressure fail.

Comment 7

ID: 1247881 User: krzkrzkra Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sun 14 Jul 2024 19:41 Selected Answer: - Upvotes: 1

1. SpeechRecognitionLanguage
2. AddTargetLanguage

Comment 8

ID: 1234519 User: HaraTadahisa Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Fri 21 Jun 2024 16:59 Selected Answer: - Upvotes: 1

1. speech_recognition_language
2. add_target_language

Comment 9

ID: 1215798 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 22 May 2024 16:17 Selected Answer: - Upvotes: 2

SpeechRecognitionLanguage
AddTargetLanguage

Comment 10

ID: 1081438 User: sca88 Badges: - Relative Date: 2 years, 3 months ago Absolute Date: Mon 27 Nov 2023 11:57 Selected Answer: - Upvotes: 1

1) SpeechRecognitionLanguage
2) AddTargetLanguage

https://learn.microsoft.com/en-us/dotnet/api/microsoft.cognitiveservices.speech.speechtranslationconfig?view=azure-dotnet

Comment 11

ID: 895201 User: WinzigWeich Badges: - Relative Date: 2 years, 10 months ago Absolute Date: Thu 11 May 2023 17:58 Selected Answer: - Upvotes: 6

https://microsoftlearning.github.io/AI-102-AIEngineer.de-de/Instructions/08-translate-speech.html
C# Part
translationConfig.SpeechRecognitionLanguage
translationConfig.AddTargetLanguage

47. AI-102 Topic 3 Question 71

Sequence
126
Discussion ID
135018
Source URL
https://www.examtopics.com/discussions/microsoft/view/135018-exam-ai-102-topic-3-question-71-discussion/
Posted By
Harry300
Posted At
March 1, 2024, 7:21 p.m.

Question

HOTSPOT -

You are building an app that will answer customer calls about the status of an order. The app will query a database for the order details and provide the customers with a spoken response.

You need to identify which Azure AI service APIs to use. The solution must minimize development effort.

Which object should you use for each requirement? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

image

Suggested Answer

image
Answer Description Click to expand


Comments 9 comments Click to expand

Comment 1

ID: 1163670 User: Harry300 Badges: Highly Voted Relative Date: 2 years ago Absolute Date: Fri 01 Mar 2024 19:21 Selected Answer: - Upvotes: 19

SpeechRecognizer / SpeechSynthesizer
doesn't say anything about translation recognizer, which needs more code

Comment 1.1

ID: 1163871 User: audlindr Badges: - Relative Date: 2 years ago Absolute Date: Sat 02 Mar 2024 03:32 Selected Answer: - Upvotes: 6

Agreed
https://learn.microsoft.com/en-us/azure/ai-services/speech-service/get-started-speech-to-text?tabs=windows%2Cterminal&pivots=programming-language-csharp

https://learn.microsoft.com/en-us/azure/ai-services/speech-service/how-to-speech-synthesis?tabs=browserjs%2Cterminal&pivots=programming-language-csharp

Comment 2

ID: 1186792 User: Murtuza Badges: Highly Voted Relative Date: 1 year, 11 months ago Absolute Date: Sun 31 Mar 2024 14:49 Selected Answer: - Upvotes: 7

To convert customer calls into text queries, you should use the SpeechRecognizer object. This is part of the Speech service in Azure Cognitive Services and it converts spoken language into written text.

To provide customers with the order details in a spoken response, you should use the SpeechSynthesizer object. This service converts text into lifelike speech, which can be used to deliver information to customers in a more interactive way.

So, the answer is:

Convert customer calls into text queries: SpeechRecognizer
Provide customers with the order details: SpeechSynthesizer

Comment 3

ID: 1356217 User: sukantadey Badges: Most Recent Relative Date: 1 year ago Absolute Date: Thu 13 Feb 2025 18:10 Selected Answer: - Upvotes: 1

SpeechRecognizer
2. SpeechSynthesizer

Comment 4

ID: 1246815 User: krzkrzkra Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Fri 12 Jul 2024 16:12 Selected Answer: - Upvotes: 4

1. SpeechRecognizer
2. SpeechSynthesizer

Comment 5

ID: 1225493 User: NagaoShingo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 06 Jun 2024 14:40 Selected Answer: - Upvotes: 4

1. SpeechRecognizer
2. SpeechSynthesizer

Comment 6

ID: 1215754 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 22 May 2024 15:25 Selected Answer: - Upvotes: 3

It must be SpeechRecognizer and SpeechSynthesizer.

Comment 7

ID: 1188297 User: anto69 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Tue 02 Apr 2024 22:37 Selected Answer: - Upvotes: 3

SpeechRecognizer / SpeechSynthesizer

Comment 8

ID: 1184123 User: varinder82 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Wed 27 Mar 2024 14:14 Selected Answer: - Upvotes: 2

Final Answer;
1. SpeechRecognizer
2. SpeechSynthesizer

48. AI-102 Topic 3 Question 37

Sequence
128
Discussion ID
102802
Source URL
https://www.examtopics.com/discussions/microsoft/view/102802-exam-ai-102-topic-3-question-37-discussion/
Posted By
marti_tremblay000
Posted At
March 16, 2023, 12:16 p.m.

Question

You develop a custom question answering project in Azure Cognitive Service for Language. The project will be used by a chatbot.

You need to configure the project to engage in multi-turn conversations.

What should you do?

  • A. Add follow-up prompts.
  • B. Enable active learning.
  • C. Add alternate questions.
  • D. Enable chit-chat.

Suggested Answer

A

Answer Description Click to expand


Community Answer Votes

Comments 13 comments Click to expand

Comment 1

ID: 1350719 User: abain89 Badges: - Relative Date: 1 year, 1 month ago Absolute Date: Mon 03 Feb 2025 02:52 Selected Answer: A Upvotes: 1

Use follow-up prompts and context to manage the multiple turns, known as multi-turn, for your bot from one question to another.
https://learn.microsoft.com/en-us/azure/ai-services/qnamaker/how-to/multi-turn

Comment 2

ID: 1279776 User: famco Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Fri 06 Sep 2024 21:25 Selected Answer: - Upvotes: 1

The expected answer of Microsoft is active-learning. Not follow-up prompt. Chatgpt is not correct and it is going for the literal meaning of follow-up prompt and that is not right

Comment 2.1

ID: 1281154 User: famco Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Mon 09 Sep 2024 19:41 Selected Answer: - Upvotes: 2

I'm wrong as per Microsoft documentation:
"Use follow-up prompts to create multiple turns of a conversation"
Answer is follow-up prompts

Comment 3

ID: 1247880 User: krzkrzkra Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sun 14 Jul 2024 19:40 Selected Answer: - Upvotes: 1

Selected Answer: A

Comment 4

ID: 1236634 User: JacobZ Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Tue 25 Jun 2024 01:02 Selected Answer: - Upvotes: 2

Got this in the exam, Jun 2024.

Comment 5

ID: 1235189 User: HaraTadahisa Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 08:23 Selected Answer: A Upvotes: 1

I say this answer is A. The rising of the shield hero.

Comment 6

ID: 1229224 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 12 Jun 2024 16:16 Selected Answer: A Upvotes: 1

A is answer.

Comment 7

ID: 1217508 User: reiwanotora Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 15:32 Selected Answer: A Upvotes: 1

A is right answer.

Comment 8

ID: 1139045 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 03 Feb 2024 08:08 Selected Answer: A Upvotes: 1

no doubt answer is A

Comment 9

ID: 1134634 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Mon 29 Jan 2024 04:34 Selected Answer: A Upvotes: 1

no doubt follow up prompt

Comment 10

ID: 1133059 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 27 Jan 2024 05:35 Selected Answer: A Upvotes: 1

Multi-turn conversations in the context of chatbots or conversational AI refer to interactions where the dialogue between the user and the bot extends beyond a single question and response. In these conversations, the bot and the user exchange multiple messages, and the bot's responses are contextually dependent on the previous parts of the conversation.

Comment 11

ID: 856708 User: tzuyichao Badges: - Relative Date: 2 years, 11 months ago Absolute Date: Fri 31 Mar 2023 09:24 Selected Answer: - Upvotes: 2

Ref: https://learn.microsoft.com/en-us/azure/cognitive-services/qnamaker/how-to/multi-turn

Comment 12

ID: 840873 User: marti_tremblay000 Badges: - Relative Date: 2 years, 12 months ago Absolute Date: Thu 16 Mar 2023 12:16 Selected Answer: A Upvotes: 1

ChatGPT confirms :
To configure the project to engage in multi-turn conversations, you should add follow-up prompts. Follow-up prompts are a way to ask additional questions or provide more information to help the user clarify their intent. By adding follow-up prompts, the chatbot can engage in a back-and-forth conversation with the user to gather additional information and ultimately provide a better answer.

Therefore, the correct answer is A. Add follow-up prompts.

49. AI-102 Topic 3 Question 32

Sequence
133
Discussion ID
102570
Source URL
https://www.examtopics.com/discussions/microsoft/view/102570-exam-ai-102-topic-3-question-32-discussion/
Posted By
marti_tremblay000
Posted At
March 14, 2023, 12:31 p.m.

Question

You have a Language service resource that performs the following:

• Sentiment analysis
• Named Entity Recognition (NER)
• Personally Identifiable Information (PII) identification

You need to prevent the resource from persisting input data once the data is analyzed.

Which query parameter in the Language service API should you configure?

  • A. model-version
  • B. piiCategories
  • C. showStats
  • D. loggingOptOut

Suggested Answer

D

Answer Description Click to expand


Community Answer Votes

Comments 9 comments Click to expand

Comment 1

ID: 838782 User: marti_tremblay000 Badges: Highly Voted Relative Date: 2 years, 12 months ago Absolute Date: Tue 14 Mar 2023 12:31 Selected Answer: D Upvotes: 9

The LoggingOptOut parameter is true by default for the PII and health feature endpoints.
Reference https://learn.microsoft.com/en-us/legal/cognitive-services/language-service/data-privacy

Comment 1.1

ID: 840866 User: marti_tremblay000 Badges: - Relative Date: 2 years, 12 months ago Absolute Date: Thu 16 Mar 2023 12:12 Selected Answer: - Upvotes: 3

ChatGPT confirms :
To prevent the resource from persisting input data once the data is analyzed, you should configure the loggingOptOut query parameter in the Language service API. Setting the value of loggingOptOut to true will prevent the service from logging or storing the input data after analysis.

Therefore, the correct answer is D. loggingOptOut.

Comment 2

ID: 1281148 User: famco Badges: Most Recent Relative Date: 1 year, 6 months ago Absolute Date: Mon 09 Sep 2024 19:36 Selected Answer: - Upvotes: 2

Another of the Microsoft greatness in naming conventions. loggingOutput=false means it will keep the data, if it is true it will not keep the data. Now make sense of that.
https://learn.microsoft.com/en-us/legal/cognitive-services/language-service/data-privacy#how-is-data-retained-and-what-customer-controls-are-available

"By default, this parameter is set to false for Language Detection, Key Phrase Extraction, Sentiment Analysis and Named Entity Recognition endpoints"

Comment 2.1

ID: 1342479 User: Architect_CTO Badges: - Relative Date: 1 year, 1 month ago Absolute Date: Sat 18 Jan 2025 12:11 Selected Answer: - Upvotes: 2

its not output in options rather OPTout

Comment 3

ID: 1229228 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 12 Jun 2024 16:18 Selected Answer: D Upvotes: 1

D is answer.

Comment 4

ID: 1217516 User: reiwanotora Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 15:41 Selected Answer: D Upvotes: 1

loggingOptOut Yes

Comment 5

ID: 1139032 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 03 Feb 2024 07:59 Selected Answer: D Upvotes: 2

no doubt answer is D

Comment 6

ID: 1102035 User: Gvalli Badges: - Relative Date: 2 years, 2 months ago Absolute Date: Thu 21 Dec 2023 00:30 Selected Answer: - Upvotes: 1

A modified version of this was in the exam today.

Comment 6.1

ID: 1103816 User: shuklabond007 Badges: - Relative Date: 2 years, 2 months ago Absolute Date: Sat 23 Dec 2023 02:21 Selected Answer: - Upvotes: 2

Hi @Gavlli, May know how many simulation questions comes in AI 102 exam? also read through all the ET questions should be enough to clear the exam?

50. AI-102 Topic 1 Question 48

Sequence
138
Discussion ID
111815
Source URL
https://www.examtopics.com/discussions/microsoft/view/111815-exam-ai-102-topic-1-question-48-discussion/
Posted By
973b658
Posted At
June 10, 2023, 10:14 a.m.

Question

You are building an AI solution that will use Sentiment Analysis results from surveys to calculate bonuses for customer service staff.

You need to ensure that the solution meets the Microsoft responsible AI principles.

What should you do?

  • A. Add a human review and approval step before making decisions that affect the staff's financial situation.
  • B. Include the Sentiment Analysis results when surveys return a low confidence score.
  • C. Use all the surveys, including surveys by customers who requested that their account be deleted and their data be removed.
  • D. Publish the raw survey data to a central location and provide the staff with access to the location.

Suggested Answer

A

Answer Description Click to expand


Community Answer Votes

Comments 8 comments Click to expand

Comment 1

ID: 1335308 User: amax702 Badges: - Relative Date: 1 year, 2 months ago Absolute Date: Wed 01 Jan 2025 23:41 Selected Answer: D Upvotes: 1

Answer A is biased lol. The person reviewing and approving the results could be biased. Then making a decision based on the financial situation and whether they can afford it or not?? That makes it even worse. D should be the more likely answer as it includes transparency

Comment 2

ID: 1297846 User: Sujeeth Badges: - Relative Date: 1 year, 4 months ago Absolute Date: Tue 15 Oct 2024 00:36 Selected Answer: - Upvotes: 1

The correct choice is A. Add a human review and approval step before making decisions that affect the staff's financial situation.

This aligns with Microsoft's responsible AI principles, particularly fairness and accountability. AI systems should have human oversight, especially when decisions significantly impact people's lives, such as determining bonuses. By adding a human review, you ensure that the AI's output is verified for fairness and accuracy, reducing the risk of unintended bias or errors affecting staff members' financial rewards.

B (using low-confidence results) could introduce inaccurate or unfair results.
C (using data from deleted accounts) would violate privacy principles.
D (publishing raw survey data) could lead to privacy violations and does not ensure fairness or accountability.

Comment 3

ID: 1235505 User: Dazigster Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 18:32 Selected Answer: A Upvotes: 2

I went with A and got 100%, so its definitely A

Comment 4

ID: 1217604 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 16:43 Selected Answer: A Upvotes: 2

A is good answer. From Takedajuku perspective, if you study for 4 days and spend 2 days reviewing, you will have a better chance of passing the exam.

Comment 5

ID: 1213787 User: reiwanotora Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sun 19 May 2024 15:05 Selected Answer: A Upvotes: 1

A is right. This question is redundant.

Comment 6

ID: 1213785 User: reiwanotora Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sun 19 May 2024 15:04 Selected Answer: A Upvotes: 1

A is right.

Comment 7

ID: 995886 User: kiro_kocha Badges: - Relative Date: 2 years, 6 months ago Absolute Date: Fri 01 Sep 2023 11:21 Selected Answer: A Upvotes: 2

I also think is A

Comment 8

ID: 919907 User: 973b658 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Sat 10 Jun 2023 10:14 Selected Answer: A Upvotes: 1

It is A.

51. AI-102 Topic 3 Question 51

Sequence
141
Discussion ID
112141
Source URL
https://www.examtopics.com/discussions/microsoft/view/112141-exam-ai-102-topic-3-question-51-discussion/
Posted By
973b658
Posted At
June 14, 2023, 9:13 a.m.

Question

Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.

After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.

You build a language model by using a Conversational Language Understanding. The language model is used to search for information on a contact list by using an intent named FindContact.

A conversational expert provides you with the following list of phrases to use for training.

• Find contacts in London.
• Who do I know in Seattle?
• Search for contacts in Ukraine.

You need to implement the phrase list in Conversational Language Understanding.

Solution: You create a new utterance for each phrase in the FindContact intent.

Does this meet the goal?

  • A. Yes
  • B. No

Suggested Answer

A

Answer Description Click to expand


Community Answer Votes

Comments 17 comments Click to expand

Comment 1

ID: 1139220 User: evangelist Badges: Highly Voted Relative Date: 2 years, 1 month ago Absolute Date: Sat 03 Feb 2024 13:23 Selected Answer: A Upvotes: 14

A. Yes

Creating a new utterance for each phrase in the FindContact intent is a correct approach to implement the phrase list in Conversational Language Understanding. This method trains the language model to recognize variations of how users might express the intent to find contacts in different locations, thereby improving the model's accuracy in identifying the FindContact intent.

Comment 1.1

ID: 1139221 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 03 Feb 2024 13:25 Selected Answer: - Upvotes: 1

Yes, creating a new utterance for each phrase in the FindContact intent aligns with the recommended practice for designing applications in Conversational Language Understanding, as detailed in the documentation. This approach helps in accurately capturing the intent by providing diverse examples of how users might express their request, thus enhancing the model's ability to understand and classify user queries correctly. For more detailed guidelines, refer to the section on creating example utterances for each intent in the documentation

Comment 2

ID: 1333351 User: pabsinaz Badges: Most Recent Relative Date: 1 year, 2 months ago Absolute Date: Sun 29 Dec 2024 06:53 Selected Answer: A Upvotes: 1

Absolutely yes

Comment 3

ID: 1329738 User: kennynelcon Badges: - Relative Date: 1 year, 2 months ago Absolute Date: Sat 21 Dec 2024 00:06 Selected Answer: B Upvotes: 1

Phrase List

Comment 4

ID: 1246812 User: krzkrzkra Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Fri 12 Jul 2024 16:06 Selected Answer: B Upvotes: 2

Selected Answer: B

Comment 5

ID: 1235749 User: HVardhini Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sun 23 Jun 2024 10:45 Selected Answer: A Upvotes: 2

A is the correct answer

Comment 6

ID: 1235114 User: anjanc Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 07:16 Selected Answer: B Upvotes: 1

it is B

Comment 7

ID: 1230987 User: fuck_india Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 15 Jun 2024 16:02 Selected Answer: - Upvotes: 2

This issue is not working because everyone's opinion is 50-50. I don't know the correct answer, so I will be staring at MS Learn when this question comes up on the real exam.Examtopic should scrutinize and post the correct answer.

Comment 8

ID: 1230983 User: fuck_india Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 15 Jun 2024 16:02 Selected Answer: - Upvotes: 1

This issue is not working because everyone's opinion is 50-50. I don't know the correct answer, so I will be staring at MS Learn when this question comes up on the real exam. Examtopic should scrutinize and post the correct answer.

Comment 9

ID: 1217579 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 16:21 Selected Answer: A Upvotes: 2

It MUST be A.

Comment 10

ID: 1143587 User: PCRamirez Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Wed 07 Feb 2024 18:36 Selected Answer: - Upvotes: 3

According to Windows Copilot: B. No

Creating a new utterance for each phrase in the FindContact intent is not the most efficient approach. Instead, you can use phrase lists in Conversational Language Understanding (LUIS) to group similar phrases together. By defining a phrase list, you can handle variations of the same intent more effectively. In this case, you can create a phrase list containing the cities (London, Seattle, Ukraine) and use it within the FindContact intent. This way, LUIS will recognize any variation of these cities as part of the same intent without creating individual utterances for each location. 😊

Comment 10.1

ID: 1224128 User: demonite Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Tue 04 Jun 2024 14:25 Selected Answer: - Upvotes: 2

LUIS is deprecated, CLU will handle creating a new utterance for each phrase

Comment 11

ID: 1141436 User: Tactable Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Mon 05 Feb 2024 21:19 Selected Answer: B Upvotes: 2

According to ChatGPT:
B. No

Explanation: While creating a new utterance for each phrase in the FindContact intent is a step in the right direction, it may not be sufficient to fully meet the goal. To effectively implement the phrase list in Conversational Language Understanding, it's essential to consider variations in how users might express the same intent. The provided phrases cover different scenarios (finding contacts in different locations), but there may be additional variations and nuances to consider. Therefore, merely creating a new utterance for each provided phrase might not capture all possible ways users could express the intent to find contacts. A more comprehensive approach to training the language model might involve incorporating synonyms, alternative phrasings, and potential variations that users might use when searching for contacts.

Comment 12

ID: 1115378 User: dimsok Badges: - Relative Date: 2 years, 2 months ago Absolute Date: Sat 06 Jan 2024 19:35 Selected Answer: A Upvotes: 2

It's "Yes" actually, not the more efficient way to do it, but it will work

Comment 13

ID: 1043747 User: sl_mslconsulting Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Sun 15 Oct 2023 01:00 Selected Answer: B Upvotes: 2

I picked B because you need the entity to retrieve the location from the utterances for you app to be able to know which contacts to retrieve from the store.

Comment 13.1

ID: 1043751 User: sl_mslconsulting Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Sun 15 Oct 2023 01:03 Selected Answer: - Upvotes: 2

The utterances provided are different enough. But if your picky and insist that you should have at least 15 utterances that you would pick B anyway.

Comment 14

ID: 922837 User: 973b658 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Wed 14 Jun 2023 09:13 Selected Answer: B Upvotes: 2

B. Same question.

52. AI-102 Topic 1 Question 61

Sequence
146
Discussion ID
135070
Source URL
https://www.examtopics.com/discussions/microsoft/view/135070-exam-ai-102-topic-1-question-61-discussion/
Posted By
Razvan_C
Posted At
March 2, 2024, 10:35 p.m.

Question

You build a bot by using the Microsoft Bot Framework SDK.

You start the bot on a local computer.

You need to validate the functionality of the bot.

What should you do before you connect to the bot?

  • A. Run the Bot Framework Emulator.
  • B. Run the Bot Framework Composer.
  • C. Register the bot with Azure Bot Service.
  • D. Run Windows Terminal.

Suggested Answer

A

Answer Description Click to expand


Community Answer Votes

Comments 9 comments Click to expand

Comment 1

ID: 1164511 User: Mehe323 Badges: Highly Voted Relative Date: 2 years ago Absolute Date: Sun 03 Mar 2024 06:30 Selected Answer: - Upvotes: 12

The answer looks correct:

https://learn.microsoft.com/en-us/azure/bot-service/bot-service-debug-emulator?view=azure-bot-service-4.0&tabs=csharp

Comment 2

ID: 1253318 User: nithin_reddy Badges: Highly Voted Relative Date: 1 year, 7 months ago Absolute Date: Tue 23 Jul 2024 01:00 Selected Answer: - Upvotes: 9

BOT Framework is now out of scope of exam, please check before you waste any time

Comment 3

ID: 1330814 User: Sridevijain Badges: Most Recent Relative Date: 1 year, 2 months ago Absolute Date: Mon 23 Dec 2024 14:04 Selected Answer: A Upvotes: 1

A is right.

Comment 4

ID: 1301353 User: Sujeeth Badges: - Relative Date: 1 year, 4 months ago Absolute Date: Tue 22 Oct 2024 03:27 Selected Answer: - Upvotes: 1

A. Run the Bot Framework Emulator.

The Bot Framework Emulator allows you to connect to and test your bot locally before deploying it. You don't need to register the bot with Azure Bot Service for local testing, and the Emulator is designed specifically for this purpose.

Comment 5

ID: 1217588 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 16:32 Selected Answer: A Upvotes: 2

A is good answer.

Comment 6

ID: 1213777 User: reiwanotora Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sun 19 May 2024 14:45 Selected Answer: A Upvotes: 1

A is right.

Comment 7

ID: 1187865 User: NullVoider_0 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Tue 02 Apr 2024 07:33 Selected Answer: - Upvotes: 2

Correct answer.

Comment 8

ID: 1184908 User: Murtuza Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Thu 28 Mar 2024 18:14 Selected Answer: A Upvotes: 2

Before connecting to the bot and validating its functionality, you should run the Bot Framework Emulator. The Bot Framework Emulator is a powerful tool that allows you to test and debug your bot locally

Comment 9

ID: 1164375 User: Razvan_C Badges: - Relative Date: 2 years ago Absolute Date: Sat 02 Mar 2024 22:35 Selected Answer: - Upvotes: 2

Answer seems to be correct

53. AI-102 Topic 3 Question 11

Sequence
149
Discussion ID
80189
Source URL
https://www.examtopics.com/discussions/microsoft/view/80189-exam-ai-102-topic-3-question-11-discussion/
Posted By
Internal_Koala
Posted At
Sept. 5, 2022, 5:15 a.m.

Question

HOTSPOT -
You run the following command.
image
For each of the following statements, select Yes if the statement is true. Otherwise, select No.
NOTE: Each correct selection is worth one point.
Hot Area:
image

Suggested Answer

image
Answer Description Click to expand


Comments 13 comments Click to expand

Comment 1

ID: 659694 User: Internal_Koala Badges: Highly Voted Relative Date: 3 years, 6 months ago Absolute Date: Mon 05 Sep 2022 05:15 Selected Answer: - Upvotes: 36

Yes
No
Yes

Log location is not mounted. The ET answer relates to an example provided on the given website which DOES mount a log location.

Comment 1.1

ID: 1326235 User: Alan_CA Badges: - Relative Date: 1 year, 2 months ago Absolute Date: Fri 13 Dec 2024 20:42 Selected Answer: - Upvotes: 1

Correct
Logging is not enabled. See :
https://learn.microsoft.com/en-us/azure/ai-services/language-service/text-analytics-for-health/how-to/configure-containers#logging-settings

Comment 2

ID: 863989 User: MDawson Badges: Highly Voted Relative Date: 2 years, 11 months ago Absolute Date: Fri 07 Apr 2023 16:36 Selected Answer: - Upvotes: 7

Documentation says it will NOT cause an endpoint query, so I think the first one should be NO

Comment 3

ID: 1248501 User: krzkrzkra Badges: Most Recent Relative Date: 1 year, 7 months ago Absolute Date: Mon 15 Jul 2024 20:02 Selected Answer: - Upvotes: 2

YNY is the answer.

Comment 4

ID: 1217535 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 15:55 Selected Answer: - Upvotes: 1

Yes No Yes

Comment 5

ID: 1217354 User: funny_penguin Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 11:37 Selected Answer: - Upvotes: 2

on exam, YNY

Comment 6

ID: 1183278 User: varinder82 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Tue 26 Mar 2024 13:01 Selected Answer: - Upvotes: 2

Fintal Answer:
Yes
No
Yes

Comment 7

ID: 1132972 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 27 Jan 2024 02:28 Selected Answer: - Upvotes: 4

Going to http://localhost:5000/status will query the Azure endpoint to verify whether the API key used to start the container is valid.

Yes. Typically, Azure Cognitive Services containers provide a /status endpoint that can be used to check the status of the service, including the validity of the API key. Since the service is mapped to localhost:5000, accessing this URL should provide the status of the containerized service, including the API key's validity.

The container logging provider will write log data.

No (Assuming). This statement is somewhat ambiguous and depends on the configuration of the Docker container and the Azure Cognitive Services container.

Going to http://localhost:5000/swagger will provide the details to access the documentation for the available endpoints.

Yes. It is a common practice for web services and APIs, including those provided by Azure Cognitive Services, to offer a Swagger UI at a /swagger endpoint.

Comment 8

ID: 1102030 User: Gvalli Badges: - Relative Date: 2 years, 2 months ago Absolute Date: Thu 21 Dec 2023 00:27 Selected Answer: - Upvotes: 5

A modified version of this was in the exam today.

Comment 9

ID: 996005 User: M25 Badges: - Relative Date: 2 years, 6 months ago Absolute Date: Fri 01 Sep 2023 13:36 Selected Answer: - Upvotes: 4

No ("without causing an endpoint query"), No, Yes

Comment 9.1

ID: 1322260 User: chrillelundmark Badges: - Relative Date: 1 year, 3 months ago Absolute Date: Thu 05 Dec 2024 08:29 Selected Answer: - Upvotes: 1

Can't see the logger setting being specified in the command, which is mandatory according to this:
https://learn.microsoft.com/es-es/azure/ai-services/language-service/concepts/configure-containers#logging-settings

Comment 10

ID: 911215 User: mmaguero Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Wed 31 May 2023 14:08 Selected Answer: - Upvotes: 4

1 and 3, is true: https://learn.microsoft.com/es-es/azure/cognitive-services/language-service/sentiment-opinion-mining/how-to/use-containers#validate-that-a-container-is-running

2, i think, is yes... (by docker settings)

Comment 11

ID: 875175 User: odisor Badges: - Relative Date: 2 years, 10 months ago Absolute Date: Thu 20 Apr 2023 02:16 Selected Answer: - Upvotes: 1

I think the first one is YES
https://learn.microsoft.com/en-us/azure/cognitive-services/language-service/text-analytics-for-health/how-to/use-containers?tabs=language#validate-that-a-container-is-running

54. AI-102 Topic 3 Question 80

Sequence
164
Discussion ID
152249
Source URL
https://www.examtopics.com/discussions/microsoft/view/152249-exam-ai-102-topic-3-question-80-discussion/
Posted By
Terk0
Posted At
Nov. 28, 2024, 11:56 a.m.

Question

You have the following Python function.

image

You call the function by using the following code.

my_function(text_analytics_client, "the quick brown fox jumps over the lazy dog")

Following 'Key phrases', what output will you receive?

  • A. The quick -
    The lazy
  • B. jumps over the
  • C. quick brown fox
    lazy dog
  • D. the quick brown fox jumps over the lazy dog

Suggested Answer

C

Answer Description Click to expand


Community Answer Votes

Comments 1 comment Click to expand

Comment 1

ID: 1319179 User: Terk0 Badges: - Relative Date: 1 year, 3 months ago Absolute Date: Thu 28 Nov 2024 11:56 Selected Answer: C Upvotes: 3

Correct answer: C

55. AI-102 Topic 5 Question 19

Sequence
166
Discussion ID
125598
Source URL
https://www.examtopics.com/discussions/microsoft/view/125598-exam-ai-102-topic-5-question-19-discussion/
Posted By
rdemontis
Posted At
Nov. 7, 2023, 7:41 p.m.

Question

SIMULATION -
You need to create a QnA Maker service named QNA12345678 in the East US Azure region. QNA12345678 must contain a knowledge base that uses the questions and answers available at https://support.microsoft.com/en-us/help/12435/windows-10-upgrade-faq.
To complete this task, sign in to the Azure portal and the QnA Maker portal.

Suggested Answer

image
Answer Description Click to expand


Comments 3 comments Click to expand

Comment 1

ID: 1218942 User: reiwanotora Badges: - Relative Date: 1 year, 3 months ago Absolute Date: Tue 26 Nov 2024 14:04 Selected Answer: - Upvotes: 3

Simulation question will not appear on the actual exam as of May 26, 2024.

Comment 2

ID: 1215801 User: takaimomoGcup Badges: - Relative Date: 1 year, 3 months ago Absolute Date: Fri 22 Nov 2024 17:19 Selected Answer: - Upvotes: 2

If the simulation question is not on the real exam, delete this question and the related simulation question.

Comment 3

ID: 1173415 User: GHill1982 Badges: - Relative Date: 1 year, 5 months ago Absolute Date: Sat 14 Sep 2024 12:53 Selected Answer: - Upvotes: 1

The QnA Maker service is being retired and you can no longer creare a new QnA Marker resource. Instead as of March 2024:
1. Create a Language service reosource in Azure AI services with Custom question answering selected
2. Launch Language Studio, select Custom question answering
3. Create a new Project and add the provided URL as a data source

56. AI-102 Topic 5 Question 20

Sequence
167
Discussion ID
125599
Source URL
https://www.examtopics.com/discussions/microsoft/view/125599-exam-ai-102-topic-5-question-20-discussion/
Posted By
rdemontis
Posted At
Nov. 7, 2023, 7:43 p.m.

Question

SIMULATION -
You need to add a question pair to the published knowledge base used by a QnA Maker service named QNA12345678. The question must be: `What will be the next version of Windows?`
The answer must be: `Windows 11`.
To complete this task, sign in to the QnA Maker portal.

Suggested Answer

image
Answer Description Click to expand


Comments 3 comments Click to expand

Comment 1

ID: 1218941 User: reiwanotora Badges: - Relative Date: 1 year, 3 months ago Absolute Date: Tue 26 Nov 2024 14:04 Selected Answer: - Upvotes: 2

Simulation question will not appear on the actual exam as of May 26 2024.

Comment 2

ID: 1177122 User: AlviraTony Badges: - Relative Date: 1 year, 5 months ago Absolute Date: Thu 19 Sep 2024 08:34 Selected Answer: - Upvotes: 2

QnAMaker is no longer in the syllabus of Exam AI-102

Comment 3

ID: 1132447 User: AnonymousJhb Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Fri 26 Jul 2024 10:35 Selected Answer: - Upvotes: 2

The given answer is deprecated.
This question follows on from pre-built question 5.19,
go back to your project QNA12345678 in language studio > edit knowledge base > + Add >
select the url as source > add your given q and a from the question > done > save >
edit knowledge base > click on test icon > select use deployed kb > paste your new question > What will be the next version of Windows? > scroll down to find a bold font Windows 11 answer
All done :)

57. AI-102 Topic 3 Question 21

Sequence
169
Discussion ID
77006
Source URL
https://www.examtopics.com/discussions/microsoft/view/77006-exam-ai-102-topic-3-question-21-discussion/
Posted By
sdokmak
Posted At
June 22, 2022, 9:18 a.m.

Question

DRAG DROP -
You are building a Language Understanding model for purchasing tickets.
You have the following utterance for an intent named PurchaseAndSendTickets.
Purchase [2 audit business] tickets to [Paris] [next Monday] and send tickets to [[email protected]]
You need to select the entity types. The solution must use built-in entity types to minimize training data whenever possible.
Which entity type should you use for each label? To answer, drag the appropriate entity types to the correct labels. Each entity type may be used once, more than once, or not at all.
You may need to drag the split bar between panes or scroll to view content.
Select and Place:
image

Suggested Answer

image
Answer Description Click to expand

Box 1: GeographyV2 -
The prebuilt geographyV2 entity detects places. Because this entity is already trained, you do not need to add example utterances containing GeographyV2 to the application intents.

Box 2: Email -
Email prebuilt entity for a LUIS app: Email extraction includes the entire email address from an utterance. Because this entity is already trained, you do not need to add example utterances containing email to the application intents.

Box 3: Machine learned -
The machine-learning entity is the preferred entity for building LUIS applications.
Reference:
https://docs.microsoft.com/en-us/azure/cognitive-services/luis/luis-reference-prebuilt-geographyv2 https://docs.microsoft.com/en-us/azure/cognitive-services/luis/luis-reference-prebuilt-email https://docs.microsoft.com/en-us/azure/cognitive-services/luis/reference-entity-machine-learned-entity

Comments 5 comments Click to expand

Comment 1

ID: 936184 User: Pixelmate Badges: Highly Voted Relative Date: 2 years, 2 months ago Absolute Date: Thu 28 Dec 2023 08:40 Selected Answer: - Upvotes: 12

THIS WAS ON EXAM 28/06

Comment 2

ID: 1217549 User: nanaw770 Badges: Most Recent Relative Date: 1 year, 3 months ago Absolute Date: Sun 24 Nov 2024 17:02 Selected Answer: - Upvotes: 2

Paris: GeographyV2 [email protected]: Email 2 audit business: Machine learned

Comment 3

ID: 1130207 User: evangelist Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Wed 24 Jul 2024 04:16 Selected Answer: - Upvotes: 3

The answer is correct

Comment 4

ID: 633395 User: Eltooth Badges: - Relative Date: 3 years, 1 month ago Absolute Date: Thu 19 Jan 2023 09:25 Selected Answer: - Upvotes: 4

Answer is correct :
Geography v2 :
https://docs.microsoft.com/en-us/azure/cognitive-services/luis/luis-reference-prebuilt-geographyv2?tabs=V3

Email :
https://docs.microsoft.com/en-us/azure/cognitive-services/luis/luis-reference-prebuilt-email?tabs=V3-verbose

Machine Learned :
https://docs.microsoft.com/en-us/azure/cognitive-services/luis/reference-entity-machine-learned-entity?tabs=V3

https://docs.microsoft.com/en-us/azure/cognitive-services/luis/concepts/entities

Comment 5

ID: 620224 User: sdokmak Badges: - Relative Date: 3 years, 2 months ago Absolute Date: Thu 22 Dec 2022 10:18 Selected Answer: - Upvotes: 2

Correct.
https://docs.microsoft.com/en-us/azure/cognitive-services/luis/luis-reference-prebuilt-email?tabs=V3

58. AI-102 Topic 3 Question 22

Sequence
170
Discussion ID
75322
Source URL
https://www.examtopics.com/discussions/microsoft/view/75322-exam-ai-102-topic-3-question-22-discussion/
Posted By
PHD_CHENG
Posted At
May 9, 2022, 6:45 a.m.

Question

You have the following C# method.
image
You need to deploy an Azure resource to the East US Azure region. The resource will be used to perform sentiment analysis.
How should you call the method?

  • A. create_resource("res1", "ContentModerator", "S0", "eastus")
  • B. create_resource("res1", "TextAnalytics", "S0", "eastus")
  • C. create_resource("res1", "ContentModerator", "Standard", "East US")
  • D. create_resource("res1", "TextAnalytics", "Standard", "East US")

Suggested Answer

B

Answer Description Click to expand


Community Answer Votes

Comments 9 comments Click to expand

Comment 1

ID: 1217555 User: nanaw770 Badges: - Relative Date: 1 year, 3 months ago Absolute Date: Sun 24 Nov 2024 17:06 Selected Answer: B Upvotes: 2

SKU is S0.
Region is eastus.
Sentiment analysis uses TextAnalytics

Comment 2

ID: 1133043 User: evangelist Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Sat 27 Jul 2024 03:51 Selected Answer: - Upvotes: 3

free tier==>S0, region eastus not East US, sentiment analysis=>TextAnalytics

Comment 3

ID: 1130209 User: evangelist Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Wed 24 Jul 2024 04:17 Selected Answer: B Upvotes: 1

region has to be in format of: eastus
ContentModerator cannot do sentimental Analsysis

Comment 4

ID: 1045256 User: SaviB Badges: - Relative Date: 1 year, 10 months ago Absolute Date: Tue 16 Apr 2024 22:50 Selected Answer: - Upvotes: 2

Is there any way to get this entire Q & A for free? I'm unable to go past page #23, even though I've signed up for an account.

Comment 4.1

ID: 1094840 User: Mobidic Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 12 Jun 2024 18:51 Selected Answer: - Upvotes: 1

@SaviB, did you found a solution? Accessed more than page #23?

Comment 5

ID: 778394 User: ap1234pa Badges: - Relative Date: 2 years, 7 months ago Absolute Date: Sun 16 Jul 2023 23:53 Selected Answer: B Upvotes: 1

B is correct

Comment 6

ID: 633398 User: Eltooth Badges: - Relative Date: 3 years, 1 month ago Absolute Date: Thu 19 Jan 2023 09:34 Selected Answer: B Upvotes: 3

Correct answer is B.
create_resource("res1", "TextAnalytics", "S0", "eastus")

Note TextAnalysis will be rebranded into Cognitive Services for Language Service

Comment 7

ID: 612719 User: PHD_CHENG Badges: - Relative Date: 3 years, 3 months ago Absolute Date: Wed 07 Dec 2022 14:30 Selected Answer: - Upvotes: 2

Was on exam 7 Jun 2022

Comment 8

ID: 598839 User: PHD_CHENG Badges: - Relative Date: 3 years, 4 months ago Absolute Date: Wed 09 Nov 2022 07:45 Selected Answer: B Upvotes: 3

Answer is correct

59. AI-102 Topic 3 Question 44

Sequence
172
Discussion ID
111250
Source URL
https://www.examtopics.com/discussions/microsoft/view/111250-exam-ai-102-topic-3-question-44-discussion/
Posted By
ziggy1117
Posted At
June 6, 2023, 10:12 a.m.

Question

DRAG DROP
-

You have a Language Understanding solution that runs in a Docker container.

You download the Language Understanding container image from the Microsoft Container Registry (MCR).

You need to deploy the container image to a host computer.

Which three actions should you perform in sequence? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.

image

Suggested Answer

image
Answer Description Click to expand


Comments 5 comments Click to expand

Comment 1

ID: 1215795 User: takaimomoGcup Badges: - Relative Date: 1 year, 3 months ago Absolute Date: Fri 22 Nov 2024 17:15 Selected Answer: - Upvotes: 4

export
move
run

Comment 2

ID: 1139198 User: evangelist Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Sat 03 Aug 2024 11:52 Selected Answer: - Upvotes: 1

Export the LUIS application as a package file from the Azure portal: This involves downloading the LUIS model you've developed and want to run locally in a container.

Move the exported package file to the Docker input directory on the host computer: This step involves transferring the downloaded LUIS application package to a specific directory that the Docker container will use as its input source.

Run the Docker container and specify the input directory: This involves using Docker commands to start the container with the necessary parameters, including the location of the LUIS application package file in the input directory.

Retraining the model is not mentioned as a step for deploying the container image to a host computer because the model should already be trained and exported from the LUIS portal before deployment.

Comment 3

ID: 934021 User: Tin_Tin Badges: - Relative Date: 2 years, 2 months ago Absolute Date: Tue 26 Dec 2023 07:02 Selected Answer: - Upvotes: 2

The answer seems correct.
https://learn.microsoft.com/en-us/azure/cognitive-services/luis/luis-container-howto?tabs=v3

Comment 4

ID: 920313 User: 973b658 Badges: - Relative Date: 2 years, 3 months ago Absolute Date: Mon 11 Dec 2023 02:27 Selected Answer: - Upvotes: 1

No.
Export,Docker,Output directory.

Comment 5

ID: 916038 User: ziggy1117 Badges: - Relative Date: 2 years, 3 months ago Absolute Date: Wed 06 Dec 2023 11:12 Selected Answer: - Upvotes: 3

answer is correct

60. AI-102 Topic 5 Question 10

Sequence
174
Discussion ID
66453
Source URL
https://www.examtopics.com/discussions/microsoft/view/66453-exam-ai-102-topic-5-question-10-discussion/
Posted By
Voxo
Posted At
Nov. 21, 2021, 3:07 p.m.

Question

HOTSPOT -
You are building a chatbot by using the Microsoft Bot Framework SDK.
You use an object named UserProfile to store user profile information and an object named ConversationData to store information related to a conversation.
You create the following state accessors to store both objects in state. var userStateAccessors = _userState.CreateProperty<UserProfile>(nameof(UserProfile)); var conversationStateAccessors = _conversationState.CreateProperty<ConversationData>(nameof(ConversationData));
The state storage mechanism is set to Memory Storage.
For each of the following statements, select Yes if the statement is true. Otherwise, select No.
NOTE: Each correct selection is worth one point.
Hot Area:
image

Suggested Answer

image
Answer Description Click to expand

Box 1: Yes -
You create property accessors using the CreateProperty method that provides a handle to the BotState object. Each state property accessor allows you to get or set the value of the associated state property.

Box 2: Yes -

Box 3: No -
Before you exit the turn handler, you use the state management objects' SaveChangesAsync() method to write all state changes back to storage.
Reference:
https://docs.microsoft.com/en-us/azure/bot-service/bot-builder-howto-v4-state

Comments 12 comments Click to expand

Comment 1

ID: 483317 User: Voxo Badges: Highly Voted Relative Date: 4 years, 3 months ago Absolute Date: Sun 21 Nov 2021 15:07 Selected Answer: - Upvotes: 11

Looks correct

Comment 2

ID: 1313683 User: 3fbc31b Badges: Most Recent Relative Date: 1 year, 3 months ago Absolute Date: Sun 17 Nov 2024 19:03 Selected Answer: - Upvotes: 3

Save some time and ignore Bot Framework. It's no longer on the AI-102 exam.

Comment 3

ID: 1234503 User: HaraTadahisa Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Fri 21 Jun 2024 16:50 Selected Answer: - Upvotes: 1

My answer is that
Yes
Yes
No

Comment 4

ID: 1230851 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sat 15 Jun 2024 10:49 Selected Answer: - Upvotes: 1

This is the same question as topic 5 No.52.

Comment 5

ID: 1184889 User: varinder82 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Thu 28 Mar 2024 17:35 Selected Answer: - Upvotes: 2

Final Answer
YYN

Comment 6

ID: 1184633 User: Mehe323 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Thu 28 Mar 2024 09:20 Selected Answer: - Upvotes: 1

The first two are YES, the last is NO.

"Memory storage
The Bot Framework SDK allows you to store user inputs using in-memory storage. Since in-memory storage is cleared each time the bot is restarted, it's best suited for testing purposes and isn't intended for production use. Persistent storage types, such as database storage, are best for production bots."

https://learn.microsoft.com/en-us/azure/bot-service/bot-builder-howto-v4-storage?view=azure-bot-service-4.0&tabs=csharp

Comment 7

ID: 1048340 User: sl_mslconsulting Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Fri 20 Oct 2023 05:32 Selected Answer: - Upvotes: 2

Ok seems everyone get this wrong. The answer are N, N, N. CreateProperty method provides a handle to the BotState object. BotState is part of the cache. Nothing will save to the storage layer unless you explicitly do so in the code. The storage layer is in memory in this case as a result nothing will be persisted once the bot is destroyed.

Comment 7.1

ID: 1115846 User: dimsok Badges: - Relative Date: 2 years, 2 months ago Absolute Date: Sun 07 Jan 2024 14:07 Selected Answer: - Upvotes: 1

you are right, but i think since you dont the full code to check, save is implied

Comment 8

ID: 914022 User: kail85 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Sun 04 Jun 2023 03:16 Selected Answer: - Upvotes: 1

Yes, No, No

Comment 9

ID: 913952 User: kail85 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Sun 04 Jun 2023 01:07 Selected Answer: - Upvotes: 1

First one is Yes.

The code provided creates state accessors for both UserProfile and ConversationData objects. These state accessors are used to store and retrieve the respective objects in the underlying state storage mechanism, which in this case is Memory Storage. When you use the state accessors to set or get the UserProfile object, the Microsoft Bot Framework SDK will create and maintain the UserProfile object in the Memory Storage, allowing you to persist user profile information across multiple turns in the conversation.

Comment 10

ID: 710120 User: flutterb Badges: - Relative Date: 3 years, 4 months ago Absolute Date: Thu 03 Nov 2022 00:04 Selected Answer: - Upvotes: 4

Answer is :
No
No
No
If the user profile and conversation states are committed to memory, then they are in fact stored in cache memory and not on the underlying storage layer (disk). Hence, the user profile and conversation data will get deleted once the session runtime is over.

Comment 10.1

ID: 756077 User: ninjapunk Badges: - Relative Date: 3 years, 2 months ago Absolute Date: Sun 25 Dec 2022 23:45 Selected Answer: - Upvotes: 2

Wrong. The first two answers are true.

As per the documentation, the BotState Class "defines a state management object and automates the reading and writing of associated state properties to a storage layer."

CreateProperty is a method of the BotState Class, which "creates a named state property within the scope of a BotState and returns an accessor for the property."

Reference: https://learn.microsoft.com/en-us/dotnet/api/microsoft.bot.builder.botstate?view=botbuilder-dotnet-stable

61. AI-102 Topic 3 Question 24

Sequence
177
Discussion ID
147102
Source URL
https://www.examtopics.com/discussions/microsoft/view/147102-exam-ai-102-topic-3-question-24-discussion/
Posted By
famco
Posted At
Sept. 6, 2024, 5:14 p.m.

Question

SIMULATION -
You need to configure and publish bot12345678 to support task management. The intent must be named TaskReminder. The LUDown for the intent is in the C:
\Resources\LU folder.
To complete this task, use the Microsoft Bot Framework Composer.

Suggested Answer

image
Answer Description Click to expand


Comments 2 comments Click to expand

Comment 1

ID: 1308582 User: a8da4af Badges: - Relative Date: 1 year, 4 months ago Absolute Date: Thu 07 Nov 2024 23:46 Selected Answer: - Upvotes: 2

Anyone seen simulator / labs questions recently (Nov 2024)?

Comment 2

ID: 1279664 User: famco Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Fri 06 Sep 2024 17:14 Selected Answer: - Upvotes: 3

bot is out of syllabus

62. AI-102 Topic 3 Question 30

Sequence
178
Discussion ID
111617
Source URL
https://www.examtopics.com/discussions/microsoft/view/111617-exam-ai-102-topic-3-question-30-discussion/
Posted By
ziggy1117
Posted At
June 9, 2023, 3:57 a.m.

Question

SIMULATION -
Use the following login credentials as needed:
To enter your username, place your cursor in the Sign in box and click on the username below.
To enter your password, place your cursor in the Enter password box and click on the password below.

Azure Username: [email protected] -

Azure Password: XXXXXXXXXXXX -
The following information is for technical support purposes only:

Lab Instance: 12345678 -

Task -
You need to create and publish a Language Understanding (classic) model named 1u12345678. The model will contain an intent of Travel that has an utterance of
Boat.
To complete this task, sign in to the Language Understanding portal at httptc//www.luis-ai/.

Suggested Answer

image
Answer Description Click to expand


Comments 5 comments Click to expand

Comment 1

ID: 1308587 User: a8da4af Badges: - Relative Date: 1 year, 4 months ago Absolute Date: Thu 07 Nov 2024 23:58 Selected Answer: - Upvotes: 1

Anyone know if simulation questions are in the exam as of NOV 2024?

Comment 2

ID: 1229246 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 12 Jun 2024 16:27 Selected Answer: - Upvotes: 2

I saw in the Microsoft question and answer that this question will not be asked, Mr. Big Tits.

Comment 3

ID: 1215794 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 22 May 2024 16:14 Selected Answer: - Upvotes: 2

If the simulation question is not on the actual exam, delete this question.

Comment 4

ID: 1079228 User: AnonymousJhb Badges: - Relative Date: 2 years, 3 months ago Absolute Date: Fri 24 Nov 2023 12:44 Selected Answer: - Upvotes: 2

will / is this querstion updated? Since LUIS is deprecating and already encouraging users to begin using CLU rather when accessing https://www.luis.ai/.

Comment 5

ID: 918826 User: ziggy1117 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Fri 09 Jun 2023 03:57 Selected Answer: - Upvotes: 3

1. go to https://language.cognitive.azure.com/
2. create an intent Travel
3. create sample utterances with boat
4. train, test, deploy

63. AI-102 Topic 15 Question 1

Sequence
182
Discussion ID
89941
Source URL
https://www.examtopics.com/discussions/microsoft/view/89941-exam-ai-102-topic-15-question-1-discussion/
Posted By
AdarshKumarKhare
Posted At
Dec. 4, 2022, 9:38 a.m.

Question

HOTSPOT -
You are developing the shopping on-the-go project.
You need to build the Adaptive Card for the chatbot.
How should you complete the code? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Hot Area:
image

Suggested Answer

image
Answer Description Click to expand

Box 1: name [language]
Chatbot must support interactions in English, Spanish, and Portuguese.
Box 2: "$when:${stockLevel != 'OK'}"
Product displays must include images and warnings when stock levels are low or out of stock.
Box 3: image.altText[language]

Comments 11 comments Click to expand

Comment 1

ID: 778577 User: KingChuang Badges: Highly Voted Relative Date: 3 years, 1 month ago Absolute Date: Tue 17 Jan 2023 06:05 Selected Answer: - Upvotes: 31

on my exam. (2023-01-16 passed)
My Answer:
1. name[language]
2. != OK
3. image.altText.[language]

Comment 1.1

ID: 1306783 User: Christian_garcia_martin Badges: - Relative Date: 1 year, 4 months ago Absolute Date: Mon 04 Nov 2024 06:28 Selected Answer: - Upvotes: 1

correct but do not forget double quotas ["language"]

Comment 2

ID: 1273855 User: JakeCallham Badges: Most Recent Relative Date: 1 year, 6 months ago Absolute Date: Wed 28 Aug 2024 08:04 Selected Answer: - Upvotes: 2

Let's evaluate why the other options for answer3 are not valid:
image.altText.en:
This option hardcodes the language to English (en).
image.altText.language:
This option assumes language is a property under altText. However, language is a variable representing the selected language (e.g., en, es, pt), not a property within altText.
image.altText.[“language”]:
This option incorrectly uses quotation marks around language, treating it as a literal string rather than a variable. This would attempt to look for an alt text property with the name "language", which does not exist.
Correct Option: image.altText.[language]
Explanation: This option correctly uses the language variable to dynamically access the appropriate alt text based on the user's selected language. It ensures that the alt text is in the correct language (English, Spanish, or Portuguese) by referencing the variable directly. This fulfills the requirement for a multilingual customer experience.

Comment 3

ID: 1252447 User: CellCS Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Sun 21 Jul 2024 14:57 Selected Answer: - Upvotes: 1

Answer is correct. language is a identifier, could be 'en', ...,

Comment 4

ID: 1214430 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Mon 20 May 2024 16:34 Selected Answer: - Upvotes: 1

Is this question still available on May 21, 2024?

Comment 5

ID: 828471 User: AzureJobsTillRetire Badges: - Relative Date: 3 years ago Absolute Date: Sat 04 Mar 2023 00:52 Selected Answer: - Upvotes: 1

The correct answers are those that are given in answers to question 3.

The correct answer pairs for box1 and box3 could be-
1) name.en and image.altText.en
2) name["language"] and image.altText["language"]

Since name["language"] does not exist as an option in box1, we will have to choose the first pair as the correct answer.
https://www.w3schools.com/js/js_json_syntax.asp

Comment 5.1

ID: 828476 User: AzureJobsTillRetire Badges: - Relative Date: 3 years ago Absolute Date: Sat 04 Mar 2023 01:01 Selected Answer: - Upvotes: 2

Looking at the given JSON sample for a product again, I actually cannot find either the data for name["language"]/name.language or image.altText["language"]/image.altText.language.

Comment 5.2

ID: 828478 User: AzureJobsTillRetire Badges: - Relative Date: 3 years ago Absolute Date: Sat 04 Mar 2023 01:04 Selected Answer: - Upvotes: 1

Since name["language"]/name.language and image.altText["language"]/image.altText.language do not exist, we will have to choose the first pair as the answers

Comment 6

ID: 738480 User: halfway Badges: - Relative Date: 3 years, 3 months ago Absolute Date: Thu 08 Dec 2022 01:25 Selected Answer: - Upvotes: 2

My last comment was wrong. There is no correct answer for the last question. The correct answers should be:
1. name[language] 2. "$when": "${stockLevel} != 'OK'" 3. image.altText[language]

Comment 7

ID: 738467 User: halfway Badges: - Relative Date: 3 years, 3 months ago Absolute Date: Thu 08 Dec 2022 00:37 Selected Answer: - Upvotes: 1

1. name[language] 2. "$when": "${stockLevel} != 'OK'" 3. image.altText.language

Comment 8

ID: 734947 User: AdarshKumarKhare Badges: - Relative Date: 3 years, 3 months ago Absolute Date: Sun 04 Dec 2022 09:38 Selected Answer: - Upvotes: 4

Repeat Question. Answers are different. I don't know which one is correct

64. AI-102 Topic 3 Question 52

Sequence
188
Discussion ID
112143
Source URL
https://www.examtopics.com/discussions/microsoft/view/112143-exam-ai-102-topic-3-question-52-discussion/
Posted By
973b658
Posted At
June 14, 2023, 9:18 a.m.

Question

DRAG DROP
-

You have a question answering project in Azure Cognitive Service for Language.

You need to move the project to a Language service instance in a different Azure region.

Which three actions should you perform in sequence? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.

image

Suggested Answer

image
Answer Description Click to expand


Comments 6 comments Click to expand

Comment 1

ID: 1303512 User: Christian_garcia_martin Badges: - Relative Date: 1 year, 4 months ago Absolute Date: Sun 27 Oct 2024 08:53 Selected Answer: - Upvotes: 2

options are wrong last one should be train and publish model instead train and publish the project . Fix that.

Comment 2

ID: 1248512 User: krzkrzkra Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Mon 15 Jul 2024 20:12 Selected Answer: - Upvotes: 1

1. From original instance, export existing project.
2. From new instance, import the project file.
3. From new instance, train and publish model.

Comment 3

ID: 1236636 User: JacobZ Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Tue 25 Jun 2024 01:07 Selected Answer: - Upvotes: 3

Got this in the exam, Jun 2024.

Comment 4

ID: 1215784 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 22 May 2024 15:57 Selected Answer: - Upvotes: 2

original, export existing project.
new, import the project file.
new, train and publish model.

Comment 5

ID: 1139222 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 03 Feb 2024 13:27 Selected Answer: - Upvotes: 1

Answer is correct
first from the source instance, export the source project;
at the new instance, import the exported source project;
to deploy, one has to train and publish the model from imported project, the training is needed before the model can work in new project from different Azure region

Comment 6

ID: 922842 User: 973b658 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Wed 14 Jun 2023 09:18 Selected Answer: - Upvotes: 2

It is true.

65. AI-102 Topic 1 Question 49

Sequence
192
Discussion ID
108848
Source URL
https://www.examtopics.com/discussions/microsoft/view/108848-exam-ai-102-topic-1-question-49-discussion/
Posted By
Rob77
Posted At
May 10, 2023, 9:20 a.m.

Question

You have an Azure subscription that contains a Language service resource named ta1 and a virtual network named vnet1.

You need to ensure that only resources in vnet1 can access ta1.

What should you configure?

  • A. a network security group (NSG) for vnet1
  • B. Azure Firewall for vnet1
  • C. the virtual network settings for ta1
  • D. a Language service container for ta1

Suggested Answer

C

Answer Description Click to expand


Community Answer Votes

Comments 8 comments Click to expand

Comment 1

ID: 893706 User: Rob77 Badges: Highly Voted Relative Date: 2 years, 10 months ago Absolute Date: Wed 10 May 2023 09:20 Selected Answer: - Upvotes: 15

Correct
https://learn.microsoft.com/en-us/azure/cognitive-services/cognitive-services-virtual-networks?tabs=portal

Comment 2

ID: 1056568 User: trashbox Badges: Highly Voted Relative Date: 2 years, 4 months ago Absolute Date: Sun 29 Oct 2023 05:10 Selected Answer: - Upvotes: 6

Appeared on Oct/29/2023.

Comment 3

ID: 1297849 User: Sujeeth Badges: Most Recent Relative Date: 1 year, 4 months ago Absolute Date: Tue 15 Oct 2024 00:41 Selected Answer: - Upvotes: 1

The correct choice is C. the virtual network settings for ta1.
To ensure that only resources in vnet1 can access the Language service resource (ta1), you need to configure private endpoint access by setting up the virtual network settings for ta1. This will allow ta1 to be accessible only from resources within the specified virtual network, ensuring network isolation and security.
A. Network Security Group (NSG) manages traffic rules within a virtual network but doesn't limit access to specific Azure resources like Language services.
B. Azure Firewall can filter traffic but is not needed for restricting access to a single resource in this scenario.
D. A Language service container is not relevant to restricting access via virtual networks.

Comment 4

ID: 1235507 User: Dazigster Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 18:34 Selected Answer: C Upvotes: 2

I went with C and got 100% so that's definitely correct.

Comment 5

ID: 1217606 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 16:45 Selected Answer: C Upvotes: 1

We can use virtual network.

Comment 6

ID: 1213789 User: reiwanotora Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sun 19 May 2024 15:07 Selected Answer: C Upvotes: 1

C is right.

Comment 7

ID: 1194711 User: sivapolam90 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Sat 13 Apr 2024 08:57 Selected Answer: C Upvotes: 2

C. the virtual network settings for ta1

Comment 8

ID: 1147896 User: evangelist Badges: - Relative Date: 2 years ago Absolute Date: Mon 12 Feb 2024 09:22 Selected Answer: C Upvotes: 3

configure vent1 as "trustworthy network" for Language service Ta1, so no resource except from trustworthy network Vnet1 can access ta1.

66. AI-102 Topic 5 Question 47

Sequence
196
Discussion ID
112151
Source URL
https://www.examtopics.com/discussions/microsoft/view/112151-exam-ai-102-topic-5-question-47-discussion/
Posted By
973b658
Posted At
June 14, 2023, 10:13 a.m.

Question

You are building a flight booking bot by using the Microsoft Bot Framework SDK.

The bot will ask users for the departure date. The bot must repeat the question until a valid date is given, or the users cancel the transaction.

Which type of dialog should you use?

  • A. prompt
  • B. adaptive
  • C. waterfall
  • D. action

Suggested Answer

A

Answer Description Click to expand


Community Answer Votes

Comments 9 comments Click to expand

Comment 1

ID: 1229914 User: hatanaoki Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 17:13 Selected Answer: A Upvotes: 1

A is the correct answer for this.
Le A est la bonne réponse à cette question.

Comment 2

ID: 1229856 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 15:44 Selected Answer: A Upvotes: 1

A is answer.

Comment 3

ID: 1145892 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 10 Feb 2024 02:54 Selected Answer: A Upvotes: 1

The same question has appeared more than once, please choose "prompt"

Comment 4

ID: 1145079 User: PCRamirez Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Thu 08 Feb 2024 23:34 Selected Answer: - Upvotes: 1

According to Windows Copilot:
To create a flight booking bot that repeatedly prompts users for the departure date until a valid date is provided or the transaction is canceled, you should use a waterfall dialog.

Here’s why:

Waterfall Dialogs:
A waterfall dialog is a type of container dialog that guides the conversation through a series of steps or prompts.
It allows you to define a sequence of actions or questions that occur in a linear order.
You can easily manage the flow of conversation by chaining multiple steps together.
In your case, the bot can start by asking for the departure date, validate the input, and repeat the question if an invalid date is provided.
Once a valid date is obtained, the waterfall dialog can proceed to the next step or end the conversation.

Comment 4.1

ID: 1145080 User: PCRamirez Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Thu 08 Feb 2024 23:34 Selected Answer: - Upvotes: 1

Other Options:
Prompt Dialogs: These are useful for asking a single question and collecting a specific type of input (e.g., asking for a date). However, they don’t handle multi-turn conversations as effectively as waterfall dialogs.
Adaptive Dialogs: While adaptive dialogs provide more flexibility, they are better suited for complex scenarios involving dynamic branching and adaptive behavior. For a straightforward task like collecting a departure date, a waterfall dialog is simpler and more appropriate.
Action Dialogs: Action dialogs are not specifically designed for managing conversation flow. They are more focused on executing specific actions based on user input.

Comment 4.2

ID: 1296713 User: aa18a1a Badges: - Relative Date: 1 year, 5 months ago Absolute Date: Sun 13 Oct 2024 04:45 Selected Answer: - Upvotes: 1

EXTREMELY important to understand here:
A waterfall dialog will, like you said, allow for you to PROMPT users for the departure date until a valid date is provided.
HOWEVER, you should refer to this documentation from Microsoft:
https://learn.microsoft.com/en-us/azure/bot-service/bot-builder-concept-waterfall-dialogs?view=azure-bot-service-4.0
Take a look at this part:
--
"At each step, the bot prompts the user for input (or can begin a child dialog, but that it's often a prompt), waits for a response, and then passes the result to the next step. The result of the first function is passed as an argument into the next function, and so on."
--
IMO, this key part referring to a child dialog is exactly what this question is asking. Therefore, the answer should be A. This is an example of that child dialog, which Microsoft deliberately states is "often a prompt". Like evangelist's comment below, anything that needs repeated until a valid response is given is always a Prompt

Comment 5

ID: 1133310 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 27 Jan 2024 12:56 Selected Answer: A Upvotes: 1

for anything that needs repeated until a valid response is given, it leads to " Prompt"

Comment 6

ID: 937885 User: Tin_Tin Badges: - Relative Date: 2 years, 8 months ago Absolute Date: Thu 29 Jun 2023 12:11 Selected Answer: C Upvotes: 1

C. Waterfall
same question 22 of topic5

Comment 7

ID: 922879 User: 973b658 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Wed 14 Jun 2023 10:13 Selected Answer: A Upvotes: 1

A.
Same Question.

67. AI-102 Topic 1 Question 18

Sequence
197
Discussion ID
62494
Source URL
https://www.examtopics.com/discussions/microsoft/view/62494-exam-ai-102-topic-1-question-18-discussion/
Posted By
htolajide
Posted At
Sept. 21, 2021, 4:31 p.m.

Question

HOTSPOT -
You are developing a streaming Speech to Text solution that will use the Speech SDK and MP3 encoding.
You need to develop a method to convert speech to text for streaming MP3 data.
How should you complete the code? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Hot Area:
image

Suggested Answer

image
Answer Description Click to expand

Reference:
https://docs.microsoft.com/en-us/azure/cognitive-services/speech-service/how-to-use-codec-compressed-audio-input-streams?tabs=debian&pivots=programming- language-csharp

Comments 9 comments Click to expand

Comment 1

ID: 448960 User: htolajide Badges: Highly Voted Relative Date: 3 years, 11 months ago Absolute Date: Mon 21 Mar 2022 17:31 Selected Answer: - Upvotes: 34

The answer is correct

Comment 1.1

ID: 473330 User: Mll1975 Badges: - Relative Date: 3 years, 10 months ago Absolute Date: Fri 06 May 2022 08:13 Selected Answer: - Upvotes: 9

I agree

GetCompressedFormat
https://docs.microsoft.com/en-us/dotnet/api/microsoft.cognitiveservices.speech.audio.audiostreamformat.getcompressedformat?view=azure-dotnet

SpeechRecognizer
https://docs.microsoft.com/en-us/dotnet/api/microsoft.cognitiveservices.speech.speechrecognizer.-ctor?view=azure-dotnet#Microsoft_CognitiveServices_Speech_SpeechRecognizer__ctor_Microsoft_CognitiveServices_Speech_SpeechConfig_Microsoft_CognitiveServices_Speech_Audio_AudioConfig_

Comment 2

ID: 517004 User: sumanshu Badges: Highly Voted Relative Date: 3 years, 8 months ago Absolute Date: Mon 04 Jul 2022 21:30 Selected Answer: - Upvotes: 15

We need to convert Streaming MP3 Data.
So, option 1 is eliminated, as it's not a Audio Streaming function. rest 3 contains the keyword Stream.

But, default audio stream format is WAV, and here we are passing MP3, So other then WAV, we need to pass compressed audio format, So correct answer is : AudioStreamFormat.GetCompressedFormat.

and we need to recognize the speech to convert into the text - so, Speech Recognizer.

Comment 3

ID: 1193438 User: CDL_Learner Badges: Most Recent Relative Date: 1 year, 5 months ago Absolute Date: Fri 11 Oct 2024 06:52 Selected Answer: - Upvotes: 1

AudioStreamFormat.GetCompressedFormat(AudioStreamContainerFormat.MP3): This option is selected because the audio data is in MP3 format, which is a compressed audio format. The GetCompressedFormat method is used to get the format of the compressed audio data.

AudioConfig.SetProperty: This is used to set a property of the AudioConfig object, not to get the format of the audio data.
AudioStreamFormat.GetWaveFormatPCM: This is used to get the format of PCM audio data, not MP3.
PullAudioInputStream: This is used to create a pull audio input stream, not to get the format of the audio data.

Comment 4

ID: 1067005 User: Prodyna Badges: - Relative Date: 1 year, 10 months ago Absolute Date: Fri 10 May 2024 07:17 Selected Answer: - Upvotes: 4

appeared in november exam

Comment 5

ID: 969805 User: acsoma Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Fri 02 Feb 2024 09:51 Selected Answer: - Upvotes: 4

appeared in august exam

Comment 6

ID: 631383 User: Eltooth Badges: - Relative Date: 3 years, 1 month ago Absolute Date: Sat 14 Jan 2023 16:08 Selected Answer: - Upvotes: 2

AudioStream.GetCompressedFormat (MP3 is compressed audio file)
SpeechRecognizer

Comment 7

ID: 494431 User: ashu789 Badges: - Relative Date: 3 years, 9 months ago Absolute Date: Sun 05 Jun 2022 14:32 Selected Answer: - Upvotes: 1

was on exam 05/12/2021

Comment 8

ID: 491275 User: Phong0411 Badges: - Relative Date: 3 years, 9 months ago Absolute Date: Wed 01 Jun 2022 04:49 Selected Answer: - Upvotes: 1

WAs on exam 30/11/2021

68. AI-102 Topic 1 Question 33

Sequence
198
Discussion ID
84645
Source URL
https://www.examtopics.com/discussions/microsoft/view/84645-exam-ai-102-topic-1-question-33-discussion/
Posted By
praticewizards
Posted At
Oct. 7, 2022, 6:44 p.m.

Question

SIMULATION -
You need to create a Text Analytics service named Text12345678, and then enable logging for Text12345678. The solution must ensure that any changes to
Text12345678 will be stored in a Log Analytics workspace.
To complete this task, sign in to the Azure portal.

Suggested Answer

image
Answer Description Click to expand


Comments 8 comments Click to expand

Comment 1

ID: 1115480 User: puchuveryfatty Badges: Highly Voted Relative Date: 2 years, 2 months ago Absolute Date: Sat 06 Jan 2024 23:38 Selected Answer: - Upvotes: 11

Will this "SIMULATION" type of questions come in exam? If yes, how to answer? It says to login to azure environment and then follow next steps.. How to perform those tasks in the exam?

Comment 2

ID: 917314 User: ziggy1117 Badges: Highly Voted Relative Date: 2 years, 9 months ago Absolute Date: Wed 07 Jun 2023 16:51 Selected Answer: - Upvotes: 7

For any resource that needs this requirement:
1. First create a log analytics workspace.
2. Go to Diagnostic settings -> add diagnostic setting -> click audit, alllogs, allmetrics and choose the workspace created in 1

Comment 2.1

ID: 1010626 User: AnonymousJhb Badges: - Relative Date: 2 years, 5 months ago Absolute Date: Mon 18 Sep 2023 15:15 Selected Answer: - Upvotes: 2

#Step 1 - create your language service:
azure portal > search > azure ai services,
+ create language services,
continue to create your resource,
create your resource and accept the responsible AI notice,
Review & Create,
#Step 2 - add your LAW:
open your new language service,
search for diagnostic settings,
you may need to search across the top for your specific resource,
now + add your diagnostic setting,
name, desination log analytics workspace,
select your logs and metrics as required to minimize storage costs.

Comment 2.1.1

ID: 1123082 User: niru_1993 Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Mon 15 Jan 2024 06:37 Selected Answer: - Upvotes: 1

Is this type of answers will come in objective

Comment 2.1.2

ID: 1245992 User: AZ1468 Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Thu 11 Jul 2024 09:48 Selected Answer: - Upvotes: 1

how to choose Language service under azure ai service

Comment 3

ID: 1295889 User: Christian_garcia_martin Badges: Most Recent Relative Date: 1 year, 5 months ago Absolute Date: Fri 11 Oct 2024 09:25 Selected Answer: - Upvotes: 2

the explanation is a bit poor , step one don't sign in in any qna portal , you need to do everything in portal.azure

Comment 4

ID: 1217620 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 16:53 Selected Answer: - Upvotes: 3

Simulation questions will not appear on the actual exam as of May 25, 2024; ET should remove this type of question. Ochinchin.

Comment 5

ID: 688818 User: praticewizards Badges: - Relative Date: 3 years, 5 months ago Absolute Date: Fri 07 Oct 2022 18:44 Selected Answer: - Upvotes: 4

This cannot be done anymore. Is not supported in azure since QnA is being retired on 31st March, 2025.
"QnA Maker service is being retired on 31st March, 2025. A newer version of this capability is now available as a part of Azure Cognitive Service for Language called question answering. To use this service, you need to provision a Language resource. For question answering capability within the Language service, see question answering and its pricing page. You can't create new QnA Maker resources anymore. For information on migrating your existing QnA Maker knowledge bases to question answering, consult the migration guide."

69. AI-102 Topic 3 Question 58

Sequence
199
Discussion ID
122483
Source URL
https://www.examtopics.com/discussions/microsoft/view/122483-exam-ai-102-topic-3-question-58-discussion/
Posted By
jangotango
Posted At
Oct. 5, 2023, 6:39 a.m.

Question

DRAG DROP -

You develop a Python app named App1 that performs speech-to-speech translation.

You need to configure App1 to translate English to German.

How should you complete the SpeechTranslationConfig object? To answer, drag the appropriate values to the correct targets. Each value may be used once, more than once or not at all. You may need to drag the split bar between panes or scroll to view content.

NOTE: Each correct selection is worth one point.

image

Suggested Answer

image
Answer Description Click to expand


Comments 13 comments Click to expand

Comment 1

ID: 1025278 User: jangotango Badges: Highly Voted Relative Date: 2 years, 5 months ago Absolute Date: Thu 05 Oct 2023 06:39 Selected Answer: - Upvotes: 29

Pretty sure the second one is add_target_language

Comment 1.1

ID: 1105245 User: nastolgia Badges: - Relative Date: 2 years, 2 months ago Absolute Date: Mon 25 Dec 2023 13:01 Selected Answer: - Upvotes: 3

you do speech-to-speech. so your output should be voice, but proposed answer is incomplete. You also need to specify the type of voice for the synthesizer

Comment 1.1.1

ID: 1204164 User: upliftinghut Badges: - Relative Date: 1 year, 10 months ago Absolute Date: Mon 29 Apr 2024 20:17 Selected Answer: - Upvotes: 2

Agree, the requirement should changed to translate from speech to text, look at the function name

Comment 2

ID: 1139681 User: evangelist Badges: Highly Voted Relative Date: 2 years, 1 month ago Absolute Date: Sun 04 Feb 2024 01:20 Selected Answer: - Upvotes: 18

The answer is WRONG!
Correct Answer is below:
def translate_speech_to_text():
translation_config = speechsdk.translation.SpeechTranslationConfig(subscription=speech_key, region=service_region)
translation_config.speech_recognition_language = "en-US"
translation_config.add_target_language("de")

Comment 3

ID: 1295748 User: Skyhawks Badges: Most Recent Relative Date: 1 year, 5 months ago Absolute Date: Thu 10 Oct 2024 23:22 Selected Answer: - Upvotes: 4

There is a distinction between speech-to-speech translation and speech-to-text-to-speech translation. Some questions refer natural language processing (NLP) and this simple mention changes the approach completely. Based on my experience, speech-to-speech translation requires both SpeechRecognitionLanguage and SpeechSynthesisLanguage. However, when performing speech-to-text-to-speech translation—where NLP is implicitly involved, since the speech is first converted to text before translation—you also need to specify the AddTargetLanguage. In this particular case, the answer would be: SpeechRecognitionLanguage and SpeechSynthesisLanguage because there is no mention to NLP.

Comment 4

ID: 1264951 User: anto69 Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Tue 13 Aug 2024 05:30 Selected Answer: - Upvotes: 1

Repeated question

Comment 5

ID: 1248515 User: krzkrzkra Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Mon 15 Jul 2024 20:15 Selected Answer: - Upvotes: 3

1. speech_recognition_language
2. add_target_language

Comment 6

ID: 1234516 User: HaraTadahisa Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Fri 21 Jun 2024 16:59 Selected Answer: - Upvotes: 6

1. speech_recognition_language
2. add_target_language

Comment 7

ID: 1215782 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 22 May 2024 15:54 Selected Answer: - Upvotes: 2

It MUST be translation_config.speech_recognition_language = "en-US" and translation_config.add_target_language("de").
What exactly have you studied? Memorization is not enough.

Comment 8

ID: 1135984 User: Florin83 Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Tue 30 Jan 2024 18:23 Selected Answer: - Upvotes: 2

Similar with Question #40 (language is C#, not Python)
I believe it's
1.translation_config.speech_recognition_value = "en-us"
2.translation_config.add_target_language("de")
the second one is a method, not a property, thus the "add"

Comment 9

ID: 1085121 User: tdctdc Badges: - Relative Date: 2 years, 3 months ago Absolute Date: Fri 01 Dec 2023 11:00 Selected Answer: - Upvotes: 3

How is it even possible that ET provides us witth SO MANY wrong answers? I agree with rdemontis and jangotango, the second one is add_target_language

Comment 9.1

ID: 1144441 User: idcanymore Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Thu 08 Feb 2024 13:46 Selected Answer: - Upvotes: 2

rdemontis should be getting paid

Comment 10

ID: 1025279 User: jangotango Badges: - Relative Date: 2 years, 5 months ago Absolute Date: Thu 05 Oct 2023 06:42 Selected Answer: - Upvotes: 3

Proof - https://learn.microsoft.com/en-us/azure/ai-services/speech-service/get-started-speech-translation?tabs=windows%2Cterminal&pivots=programming-language-python

70. AI-102 Topic 5 Question 32

Sequence
201
Discussion ID
111024
Source URL
https://www.examtopics.com/discussions/microsoft/view/111024-exam-ai-102-topic-5-question-32-discussion/
Posted By
kail85
Posted At
June 4, 2023, 1:35 a.m.

Question

Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.

After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.

You are building a chatbot that will use question answering in Azure Cognitive Service for Language.

You have a PDF named Doc1.pdf that contains a product catalogue and a price list.

You upload Doc1.pdf and train the model.

During testing, users report that the chatbot responds correctly to the following question: What is the price of ?

The chatbot fails to respond to the following question: How much does cost?

You need to ensure that the chatbot responds correctly to both questions.

Solution: From Language Studio, you create an entity for price, and then retrain and republish the model.

Does this meet the goal?

  • A. Yes
  • B. No

Suggested Answer

B

Answer Description Click to expand


Community Answer Votes

Comments 2 comments Click to expand

Comment 1

ID: 913959 User: kail85 Badges: Highly Voted Relative Date: 2 years, 3 months ago Absolute Date: Mon 04 Dec 2023 02:35 Selected Answer: - Upvotes: 5

B. No

Creating an entity for price and retraining the model in Language Studio is not the correct approach to solve the issue with Azure Cognitive Service for Language's question-answering capabilities.

Instead, you should use Language Studio to create and train a synonym for the term "price" or build a more comprehensive list of question variations that the chatbot should be able to handle. For example, you can include phrases like "How much does it cost?" or "What is the cost of?" to ensure the model can properly recognize and respond to different ways users might ask about the price. Retrain and republish the model after making these changes to improve the chatbot's ability to answer both questions correctly.

Comment 2

ID: 1191862 User: Murtuza Badges: Most Recent Relative Date: 1 year, 5 months ago Absolute Date: Wed 09 Oct 2024 01:22 Selected Answer: B Upvotes: 2

B is the answer.

71. AI-102 Topic 3 Question 36

Sequence
202
Discussion ID
102574
Source URL
https://www.examtopics.com/discussions/microsoft/view/102574-exam-ai-102-topic-3-question-36-discussion/
Posted By
marti_tremblay000
Posted At
March 14, 2023, 12:48 p.m.

Question

You are building a social media extension that will convert text to speech. The solution must meet the following requirements:

• Support messages of up to 400 characters.
• Provide users with multiple voice options.
• Minimize costs.

You create an Azure Cognitive Services resource.

Which Speech API endpoint provides users with the available voice options?

  • A. https://uksouth.api.cognitive.microsoft.com/speechtotext/v3.0/models/base
  • B. https://uksouth.customvoice.api.speech.microsoft.com/api/texttospeech/v3.0/longaudiosynthesis/voices
  • C. https://uksouth.tts.speech.microsoft.com/cognitiveservices/voices/list
  • D. https://uksouth.voice.speech.microsoft.com/cognitiveservices/v1?deploymentId={deploymentId}

Suggested Answer

C

Answer Description Click to expand


Community Answer Votes

Comments 11 comments Click to expand

Comment 1

ID: 838795 User: marti_tremblay000 Badges: Highly Voted Relative Date: 2 years, 12 months ago Absolute Date: Tue 14 Mar 2023 12:48 Selected Answer: C Upvotes: 15

The correct answer is C
The question is about providing users with all the available voice options.
Get a list of voices
You can use the tts.speech.microsoft.com/cognitiveservices/voices/list endpoint to get a full list of voices for a specific region or endpoint.
Reference : https://learn.microsoft.com/en-us/azure/cognitive-services/speech-service/rest-text-to-speech?tabs=streaming

Comment 1.1

ID: 898928 User: MaliSanFuu Badges: - Relative Date: 2 years, 10 months ago Absolute Date: Tue 16 May 2023 08:35 Selected Answer: - Upvotes: 2

Agreeing, as the only important question is: Which Speech API endpoint provides users with the available voice options?

Therefor answer C should be correct for this one

Comment 2

ID: 1291891 User: AnnaR Badges: Most Recent Relative Date: 1 year, 5 months ago Absolute Date: Tue 01 Oct 2024 11:44 Selected Answer: - Upvotes: 1

Not A: endpoint has speechtotext, but we need text sto speech
Not B: Customvoice.api, but we need to minimize costs
Not D: does not provide users with multiple voice options
--> C is the only reasonable answer

Comment 3

ID: 1247879 User: krzkrzkra Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sun 14 Jul 2024 19:40 Selected Answer: C Upvotes: 1

Selected Answer: C

Comment 4

ID: 1235197 User: HaraTadahisa Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 08:25 Selected Answer: C Upvotes: 1

I say this answer is C.

Comment 5

ID: 1229226 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 12 Jun 2024 16:17 Selected Answer: C Upvotes: 1

C is answer.

Comment 6

ID: 1217511 User: reiwanotora Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 15:36 Selected Answer: C Upvotes: 1

C is right answer.

Comment 7

ID: 1139044 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 03 Feb 2024 08:08 Selected Answer: C Upvotes: 1

only C meets demand of minimum cost and support 400 characters with voice options

Comment 8

ID: 1102038 User: Gvalli Badges: - Relative Date: 2 years, 2 months ago Absolute Date: Thu 21 Dec 2023 00:31 Selected Answer: - Upvotes: 3

A modified version of this was in the exam today.

Comment 9

ID: 984293 User: james2033 Badges: - Relative Date: 2 years, 6 months ago Absolute Date: Fri 18 Aug 2023 10:25 Selected Answer: C Upvotes: 1

The answer is very clear at here https://learn.microsoft.com/en-us/azure/ai-services/speech-service/rest-text-to-speech?tabs=streaming#get-a-list-of-voices

You can use the tts.speech.microsoft.com/cognitiveservices/voices/list endpoint

For example, to get a list of voices for the westus region, use the https://westus.tts.speech.microsoft.com/cognitiveservices/voices/list endpoint.

https://learn.microsoft.com/en-us/azure/ai-services/speech-service/regions

Europe - UK South - uksouth

Comment 10

ID: 879189 User: Pffffff Badges: - Relative Date: 2 years, 10 months ago Absolute Date: Mon 24 Apr 2023 11:35 Selected Answer: B Upvotes: 1

ChatGPT: The Speech API endpoint that provides users with the available voice options is B.

72. AI-102 Topic 15 Question 2

Sequence
204
Discussion ID
57607
Source URL
https://www.examtopics.com/discussions/microsoft/view/57607-exam-ai-102-topic-15-question-2-discussion/
Posted By
aakash_0086
Posted At
July 11, 2021, 10:26 a.m.

Question

HOTSPOT -
You are developing the shopping on-the-go project.
You are configuring access to the QnA Maker (classic) resources.
Which role should you assign to AllUsers and LeadershipTeam? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Hot Area:
image

Suggested Answer

image
Answer Description Click to expand

Box 1: QnA Maker Editor -
Scenario: Provide all employees with the ability to edit Q&As.
The QnA Maker Editor (read/write) has the following permissions:
✑ Create KB API
✑ Update KB API
✑ Replace KB API
✑ Replace Alterations
✑ "Train API" [in new service model v5]

Box 2: Contributor -
Scenario: Only senior managers must be able to publish updates.
Contributor permission: All except ability to add new members to roles
Reference:
https://docs.microsoft.com/en-us/azure/cognitive-services/qnamaker/reference-role-based-access-control

Comments 21 comments Click to expand

Comment 1

ID: 604604 User: g2000 Badges: Highly Voted Relative Date: 3 years, 9 months ago Absolute Date: Fri 20 May 2022 20:57 Selected Answer: - Upvotes: 24

How about Cognitive Service User? the link explicitly says publish is given to this role.
https://docs.microsoft.com/en-us/azure/cognitive-services/qnamaker/concepts/role-based-access-control#access-is-provided-by-a-defined-role

Comment 1.1

ID: 631591 User: AusAv Badges: - Relative Date: 3 years, 8 months ago Absolute Date: Fri 15 Jul 2022 04:28 Selected Answer: - Upvotes: 1

I agree it is: Cognitive Service Userm Contributor can also create new resources

Comment 1.2

ID: 677867 User: firewind Badges: - Relative Date: 3 years, 5 months ago Absolute Date: Sat 24 Sep 2022 15:50 Selected Answer: - Upvotes: 5

Agreed, from a least privilege principle perspective, it should be Cognitive Service User.

Comment 2

ID: 435130 User: SuperPetey Badges: Highly Voted Relative Date: 4 years, 6 months ago Absolute Date: Mon 30 Aug 2021 07:55 Selected Answer: - Upvotes: 16

The Owner role would give the leadership team the ability to publish changes, but also the additional capability to give others access to QnA maker, which they do not require. security best practices (in case of a compromised account etc) dictate the principles of least privilege; leadership team should only be given access to the capabilities they require to do their job and nothing more.

Since the problem only mentions publishing, the leadership team requires the role contributor. In other RBAC exam questions for other certs MSFT remembers to preface such a question with "bearing in mind the principle of least privilege..." but they neglected that hint here.

Comment 2.1

ID: 920415 User: ziggy1117 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Sun 11 Jun 2023 06:54 Selected Answer: - Upvotes: 7

Should be Cognitive Service User -> All access to Cognitive Services resource except for ability to:
1. Add new members to roles.
2. Create new resources.

Contributor role provides Leadership team access to create new resources. And they shouldnt be allowed to do that given that they arent technical people

Comment 3

ID: 1286896 User: mrg998 Badges: Most Recent Relative Date: 1 year, 5 months ago Absolute Date: Fri 20 Sep 2024 16:11 Selected Answer: - Upvotes: 8

AllUsers: QnA Maker Editor
Leadership team: Cognitive Service User

Comment 4

ID: 1257788 User: anto69 Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Tue 30 Jul 2024 04:37 Selected Answer: - Upvotes: 1

I think given answers are correct. ChatGPT 3.5 confirms it

Comment 5

ID: 1246789 User: krzkrzkra Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Fri 12 Jul 2024 15:41 Selected Answer: - Upvotes: 3

1. QnA Maker Editor
2. Cognitive Service User

Comment 6

ID: 1235338 User: reiwanotora Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 12:54 Selected Answer: - Upvotes: 2

1. QnA Maker Editor
2. Cognitive Service User

Comment 7

ID: 1230538 User: NagaoShingo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 14 Jun 2024 16:41 Selected Answer: - Upvotes: 3

1. QnA Maker Editor
2. Cognitive Service User

Comment 8

ID: 1214432 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Mon 20 May 2024 16:34 Selected Answer: - Upvotes: 4

Is this question still available on May 21, 2024?

Comment 9

ID: 1145900 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 10 Feb 2024 03:44 Selected Answer: - Upvotes: 3

The leadership should have permission to publish!!! here is why
AllUsers: QnA Maker Editor
Leadership team: Cognitive Service User

Key Differences in Privilege Scope
Specificity: The QnA Maker Editor role is specific to the QnA Maker service with a focus on content editing, while the Cognitive Service User role is more general and can encompass a broader range of tasks across Azure Cognitive Services.
Publishing the Knowledge Base: The ability to publish the knowledge base to make it live is a critical task. The QnA Maker Editor might have restricted access in this regard, requiring administrative approval, whereas a Cognitive Service User with the right permissions could directly publish changes.

Comment 10

ID: 980883 User: james2033 Badges: - Relative Date: 2 years, 6 months ago Absolute Date: Mon 14 Aug 2023 16:42 Selected Answer: - Upvotes: 6

1. Cognitive Services QnA Maker Editor

2. Cognitive Services User

See https://learn.microsoft.com/en-us/azure/ai-services/qnamaker/concepts/role-based-access-control#access-is-provided-by-a-defined-role

Comment 11

ID: 919068 User: ziggy1117 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Fri 09 Jun 2023 10:07 Selected Answer: - Upvotes: 2

AllUsers: QnA Maker Editor
(read/write)
Leadership Team: Cognitive Service User
(read/write/publish)

Leadership team does not need to add resources

Comment 12

ID: 778578 User: KingChuang Badges: - Relative Date: 3 years, 1 month ago Absolute Date: Tue 17 Jan 2023 06:05 Selected Answer: - Upvotes: 8

on my exam. (2023-01-16 passed)
My Answer:
1. QnA Maker Editor
2. Cognitive Service User

Comment 13

ID: 647317 User: ninjia Badges: - Relative Date: 3 years, 6 months ago Absolute Date: Mon 15 Aug 2022 20:00 Selected Answer: - Upvotes: 8

To address the requirement: Provide all employees with the ability to edit Q&As. Only senior managers must be able to publish updates.

AllUsers: QnA Maker Editor
(read/write)
Leadership Team: Cognitive Service User
(read/write/publish)

Cognitive Service User has less permissions than Contributor.

Comment 13.1

ID: 647318 User: ninjia Badges: - Relative Date: 3 years, 6 months ago Absolute Date: Mon 15 Aug 2022 20:01 Selected Answer: - Upvotes: 1

Reference: https://docs.microsoft.com/en-us/azure/cognitive-services/qnamaker/concepts/role-based-access-control#access-is-provided-by-a-defined-role

Comment 14

ID: 634012 User: Eltooth Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Wed 20 Jul 2022 13:26 Selected Answer: - Upvotes: 1

AllUsers: QnA Maker Editor
LeadershipTeam: Contributor

Comment 15

ID: 464498 User: rikku33 Badges: - Relative Date: 4 years, 4 months ago Absolute Date: Tue 19 Oct 2021 09:09 Selected Answer: - Upvotes: 2

from case study: Provide all employees with the ability to edit Q&As. Only senior managers must be able to publish updates, so the given answer is correct

Comment 16

ID: 403781 User: aakash_0086 Badges: - Relative Date: 4 years, 8 months ago Absolute Date: Sun 11 Jul 2021 10:26 Selected Answer: - Upvotes: 3

Shouldn't Leadership team have owner rights which has right to publish the updates ?

Comment 17

ID: 403779 User: aakash_0086 Badges: - Relative Date: 4 years, 8 months ago Absolute Date: Sun 11 Jul 2021 10:26 Selected Answer: - Upvotes: 2

Shouldn't Leadership eam have owner rights which has right to publish the updates ?

73. AI-102 Topic 1 Question 59

Sequence
207
Discussion ID
135092
Source URL
https://www.examtopics.com/discussions/microsoft/view/135092-exam-ai-102-topic-1-question-59-discussion/
Posted By
Mehe323
Posted At
March 3, 2024, 6:21 a.m.

Question

DRAG DROP -

You have an app that manages feedback.

You need to ensure that the app can detect negative comments by using the Sentiment Analysis API in Azure AI Language. The solution must ensure that the managed feedback remains on your company’s internal network.

Which three actions should you perform in sequence? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.

NOTE: More than one order of answer choices is correct. You will receive credit for any of the correct orders you select.

image

Suggested Answer

image
Answer Description Click to expand


Comments 11 comments Click to expand

Comment 1

ID: 1187862 User: NullVoider_0 Badges: Highly Voted Relative Date: 1 year, 11 months ago Absolute Date: Tue 02 Apr 2024 07:25 Selected Answer: - Upvotes: 29

1. Provision the Language service resource in Azure. This step involves creating the Azure Language service resource, which will provide you with the necessary credentials and endpoint URL to use the Sentiment Analysis API.
2. Deploy a Docker container to an on-premises server. By deploying a Docker container on-premises, you can run the Sentiment Analysis API locally, ensuring that the feedback data does not leave your internal network.
3. Run the container and query the prediction endpoint. Once the container is running on your on-premises server, you can start sending feedback data to the Sentiment Analysis API by querying the prediction endpoint provided by the Language service.

Comment 2

ID: 1194446 User: Jimmy1017 Badges: Highly Voted Relative Date: 1 year, 11 months ago Absolute Date: Fri 12 Apr 2024 19:21 Selected Answer: - Upvotes: 8

Provision the Language service resource in Azure.
Deploy a Docker container to an on-premises server.
Identify the Language service endpoint URL and query the prediction endpoint.

Comment 3

ID: 1284720 User: mrg998 Badges: Most Recent Relative Date: 1 year, 5 months ago Absolute Date: Mon 16 Sep 2024 15:13 Selected Answer: - Upvotes: 1

1) Provision the language service in Azure
2) Deploy a docker container to container instance - this is because you can inject a container instance into a VNET (remember there is no requirement for anything to be on-prem, it just says internal which could be a internal VNET)
3) Run the container and query the prediction endpoint

Comment 4

ID: 1248466 User: krzkrzkra Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Mon 15 Jul 2024 19:13 Selected Answer: - Upvotes: 1

Provision the Language service resource in Azure
Deploy a Docker container to an on-premises server
Run the container and query the prediction endpoint

Comment 5

ID: 1217585 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 16:31 Selected Answer: - Upvotes: 2

Deploy on-premises
Provision
Run

Comment 6

ID: 1190661 User: Murtuza Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Sun 07 Apr 2024 01:28 Selected Answer: - Upvotes: 5

Provision the Language service resource in Azure: This is the first step where you set up the Language service resource in Azure. This service will provide you with the Sentiment Analysis API.
Deploy a Docker container to an on-premises server: After provisioning the Language service, you should deploy a Docker container on an on-premises server. This container will host the Azure AI Language service and ensure that the managed feedback remains on your company’s internal network.
Identify the Language service endpoint URL and query the prediction endpoint: Once the Docker container is running on your on-premises server, you can identify the Language service endpoint URL. You can then query the prediction endpoint to analyze the sentiment of the comments.

Comment 7

ID: 1184894 User: Murtuza Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Thu 28 Mar 2024 17:53 Selected Answer: - Upvotes: 7

Remember that the order matters: provision the language service first, then Identify the Language Service Endpoint URL and Query the Prediction Endpoint
and finally deploy the container based on your chosen deployment target which is on-premises 🚀

Comment 7.1

ID: 1219840 User: vovap0vovap Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Mon 27 May 2024 22:37 Selected Answer: - Upvotes: 1

Question stated that more ten 1 correct order possible. You can provision recourses first or deploy container first.
Now Identify the Language Service Endpoint URL and Query the Prediction Endpoint should not be correct as indirectly assumed Azure Endpoint rather then local from Docker

Comment 8

ID: 1164505 User: Mehe323 Badges: - Relative Date: 2 years ago Absolute Date: Sun 03 Mar 2024 06:21 Selected Answer: - Upvotes: 5

That is correct, see prerequisites where you need to have 1) Docker installed and 2) provisioned a Language resource:
https://learn.microsoft.com/en-us/azure/ai-services/language-service/sentiment-opinion-mining/how-to/use-containers

Comment 8.1

ID: 1179634 User: Ody Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Thu 21 Mar 2024 21:47 Selected Answer: - Upvotes: 7

Right, but the question says Deploy a Docker Container. It is not saying setup a Docker host.

You can't run the container without having the API Key and Endpoint URI.

Provision the Language service (get the API Key and Endpoint URI)
Deploy a Docker container to an on-premise host
Run the container and query the prediction endpoint.

Comment 8.1.1

ID: 1183686 User: Mehe323 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Wed 27 Mar 2024 00:04 Selected Answer: - Upvotes: 3

It says 'deploy a docker container to an on-premises server', you forgot the last part. A synonym for deploy is install.

74. AI-102 Topic 3 Question 77

Sequence
210
Discussion ID
144085
Source URL
https://www.examtopics.com/discussions/microsoft/view/144085-exam-ai-102-topic-3-question-77-discussion/
Posted By
5503824
Posted At
July 18, 2024, 11:15 a.m.

Question

You are developing an app that will use the Speech and Language APIs.

You need to provision resources for the app. The solution must ensure that each service is accessed by using a single endpoint and credential.

Which type of resource should you create?

  • A. Azure AI Language
  • B. Azure AI Speech
  • C. Azure AI Services
  • D. Azure AI Content Safety

Suggested Answer

C

Answer Description Click to expand


Community Answer Votes

Comments 3 comments Click to expand

Comment 1

ID: 1250268 User: 5503824 Badges: Highly Voted Relative Date: 1 year, 7 months ago Absolute Date: Thu 18 Jul 2024 11:15 Selected Answer: - Upvotes: 8

Azure AI Services since more than one services are used

Comment 2

ID: 1284220 User: 9c652a0 Badges: Most Recent Relative Date: 1 year, 5 months ago Absolute Date: Sun 15 Sep 2024 18:50 Selected Answer: C Upvotes: 1

Selected answer C

Comment 3

ID: 1282650 User: mrg998 Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Thu 12 Sep 2024 15:26 Selected Answer: C Upvotes: 1

C 1000%

75. AI-102 Topic 5 Question 52

Sequence
215
Discussion ID
135602
Source URL
https://www.examtopics.com/discussions/microsoft/view/135602-exam-ai-102-topic-5-question-52-discussion/
Posted By
Murtuza
Posted At
March 9, 2024, 8:02 p.m.

Question

Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.

After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.

You are building a chatbot that will use question answering in Azure Cognitive Service for Language.

You have a PDF named Doc1.pdf that contains a product catalogue and a price list.

You upload Doc1.pdf and train the model.

During testing, users report that the chatbot responds correctly to the following question: What is the price of ?

The chatbot fails to respond to the following question: How much does cost?

You need to ensure that the chatbot responds correctly to both questions.

Solution: From Language Studio, you create an entity for cost, and then retrain and republish the model.

Does this meet the goal?

  • A. Yes
  • B. No

Suggested Answer

B

Answer Description Click to expand


Community Answer Votes

Comments 2 comments Click to expand

Comment 1

ID: 1169734 User: Murtuza Badges: Highly Voted Relative Date: 2 years ago Absolute Date: Sat 09 Mar 2024 20:02 Selected Answer: - Upvotes: 6

Solution: From Language Studio, you add alternative phrasing to the question and answer pair, and then retrain and republish the model.
Choice is B

Comment 2

ID: 1284126 User: mustafaalhnuty Badges: Most Recent Relative Date: 1 year, 5 months ago Absolute Date: Sun 15 Sep 2024 15:04 Selected Answer: B Upvotes: 1

B 100%

76. AI-102 Topic 3 Question 68

Sequence
216
Discussion ID
136760
Source URL
https://www.examtopics.com/discussions/microsoft/view/136760-exam-ai-102-topic-3-question-68-discussion/
Posted By
Murtuza
Posted At
March 20, 2024, 8:26 p.m.

Question

HOTSPOT
-

You are developing a service that records lectures given in English (United Kingdom).

You have a method named append_to_transcript_file that takes translated text and a language identifier.

You need to develop code that will provide transcripts of the lectures to attendees in their respective language. The supported languages are English, French, Spanish, and German.

How should you complete the code? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

image

Suggested Answer

image
Answer Description Click to expand


Comments 6 comments Click to expand

Comment 1

ID: 1178636 User: Murtuza Badges: Highly Voted Relative Date: 1 year, 11 months ago Absolute Date: Wed 20 Mar 2024 20:26 Selected Answer: - Upvotes: 9

The given answers are correct

Comment 2

ID: 1283943 User: famco Badges: Most Recent Relative Date: 1 year, 5 months ago Absolute Date: Sun 15 Sep 2024 08:58 Selected Answer: - Upvotes: 1

Why is that language assignment in brackets?
What's Microsoft intending with useless code? Don't they get real python programmers and not re-trained .Net piece of xx

Comment 3

ID: 1247894 User: krzkrzkra Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sun 14 Jul 2024 19:55 Selected Answer: - Upvotes: 3

fr,de,es
TranslationRecognizer

Comment 4

ID: 1232642 User: nanaw770 Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Wed 19 Jun 2024 01:22 Selected Answer: - Upvotes: 1

The actual exam will choose between C# or Python programming languages; these types of questions, which seem to be duplicates in ET, differ in the programming language used. Please check.

Comment 5

ID: 1230520 User: NagaoShingo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 14 Jun 2024 15:50 Selected Answer: - Upvotes: 2

Given answer is correct.

Comment 6

ID: 1214970 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Tue 21 May 2024 15:27 Selected Answer: - Upvotes: 3

fr,de,es and TranslationRecognizer

77. AI-102 Topic 3 Question 69

Sequence
217
Discussion ID
135068
Source URL
https://www.examtopics.com/discussions/microsoft/view/135068-exam-ai-102-topic-3-question-69-discussion/
Posted By
Harry300
Posted At
March 2, 2024, 8:47 p.m.

Question

You are developing an app that will use the text-to-speech capability of the Azure AI Speech service. The app will be used in motor vehicles.

You need to optimize the quality of the synthesized voice output.

Which Speech Synthesis Markup Language (SSML) attribute should you configure?

  • A. the style attribute of the mstts:express-as element
  • B. the effect attribute of the voice element
  • C. the pitch attribute of the prosody element
  • D. the level attribute of the emphasis element

Suggested Answer

B

Answer Description Click to expand


Community Answer Votes

Comments 9 comments Click to expand

Comment 1

ID: 1164438 User: chandiochan Badges: Highly Voted Relative Date: 2 years ago Absolute Date: Sun 03 Mar 2024 02:02 Selected Answer: - Upvotes: 11

Answer is correct:

The audio effect processor that's used to optimize the quality of the synthesized speech output for specific scenarios on devices.

For some scenarios in production environments, the auditory experience might be degraded due to the playback distortion on certain devices. For example, the synthesized speech from a car speaker might sound dull and muffled due to environmental factors such as speaker response, room reverberation, and background noise. The passenger might have to turn up the volume to hear more clearly. To avoid manual operations in such a scenario, the audio effect processor can make the sound clearer by compensating the distortion of playback.

The following values are supported:
eq_car – Optimize the auditory experience when providing high-fidelity speech in cars, buses, and other enclosed automobiles.
eq_telecomhp8k – Optimize the auditory experience for narrowband speech in telecom or telephone scenarios. You should use a sampling rate of 8 kHz. If the sample rate isn't 8 kHz, the auditory quality of the output speech isn't optimized.

Comment 2

ID: 1283947 User: famco Badges: Most Recent Relative Date: 1 year, 5 months ago Absolute Date: Sun 15 Sep 2024 09:06 Selected Answer: - Upvotes: 3

Answer is correct
The Microsoft guy read the following text:
"The audio effect processor that's used to optimize the quality of the synthesized speech output for specific scenarios on devices.

For some scenarios in production environments, the auditory experience might be degraded due to the playback distortion on certain devices. For example, the synthesized speech from a car speaker might sound dull "

Comment 3

ID: 1265832 User: anto69 Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Wed 14 Aug 2024 17:05 Selected Answer: B Upvotes: 1

Copilot:
To optimize the quality of the synthesized voice output for an app used in motor vehicles, you should configure B. the effect attribute of the voice element. This attribute allows you to apply audio effects that enhance the auditory experience in specific environments, such as cars, by compensating for playback distortions.

Comment 4

ID: 1247895 User: krzkrzkra Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sun 14 Jul 2024 19:55 Selected Answer: B Upvotes: 1

Selected Answer: B

Comment 5

ID: 1233623 User: etellez Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Thu 20 Jun 2024 15:15 Selected Answer: - Upvotes: 2

Copilot says:

A. the style attribute of the mstts:express-as element

Comment 6

ID: 1225948 User: TaiNguyen097 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 07 Jun 2024 09:11 Selected Answer: - Upvotes: 1

https://learn.microsoft.com/en-us/azure/ai-services/speech-service/speech-synthesis-markup-voice
The audio effect processor that's used to optimize the quality of the synthesized speech output for specific scenarios on devices.

Comment 7

ID: 1215740 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 22 May 2024 15:11 Selected Answer: B Upvotes: 1

voice element

Comment 8

ID: 1164442 User: chandiochan Badges: - Relative Date: 2 years ago Absolute Date: Sun 03 Mar 2024 02:11 Selected Answer: B Upvotes: 3

https://learn.microsoft.com/en-us/azure/ai-services/speech-service/speech-synthesis-markup-voice

Comment 9

ID: 1164346 User: Harry300 Badges: - Relative Date: 2 years ago Absolute Date: Sat 02 Mar 2024 20:47 Selected Answer: B Upvotes: 1

correct
source: https://learn.microsoft.com/en-us/azure/ai-services/speech-service/speech-synthesis-markup-voice:
"Optimize the auditory experience when providing high-fidelity speech in cars, buses, and other enclosed automobiles."

78. AI-102 Topic 3 Question 78

Sequence
218
Discussion ID
145783
Source URL
https://www.examtopics.com/discussions/microsoft/view/145783-exam-ai-102-topic-3-question-78-discussion/
Posted By
moonlightc
Posted At
Aug. 14, 2024, 11:44 p.m.

Question

HOTSPOT
-

You are building an app that will automatically translate speech from English to French, German, and Spanish by using Azure AI service.

You need to define the output languages and configure the Azure AI Speech service.

How should you complete the code? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

image

Suggested Answer

image
Answer Description Click to expand


Comments 2 comments Click to expand

Comment 1

ID: 1283670 User: mustafaalhnuty Badges: - Relative Date: 1 year, 5 months ago Absolute Date: Sat 14 Sep 2024 16:26 Selected Answer: - Upvotes: 2

box1: 3
box2: 4

Comment 2

ID: 1282651 User: mrg998 Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Thu 12 Sep 2024 15:27 Selected Answer: - Upvotes: 3

Correct, rewording of a old question.

79. AI-102 Topic 3 Question 12

Sequence
219
Discussion ID
54402
Source URL
https://www.examtopics.com/discussions/microsoft/view/54402-exam-ai-102-topic-3-question-12-discussion/
Posted By
Omobonike
Posted At
June 3, 2021, 1:45 p.m.

Question

You are building a Language Understanding model for an e-commerce platform.
You need to construct an entity to capture billing addresses.
Which entity type should you use for the billing address?

  • A. machine learned
  • B. Regex
  • C. geographyV2
  • D. Pattern.any
  • E. list

Suggested Answer

A

Answer Description Click to expand


Community Answer Votes

Comments 20 comments Click to expand

Comment 1

ID: 377496 User: ExamPrep2021 Badges: Highly Voted Relative Date: 4 years, 9 months ago Absolute Date: Tue 08 Jun 2021 14:04 Selected Answer: - Upvotes: 39

My guess is A.

An ML entity can be composed of smaller sub-entities, each of which can have its own properties. For example, Address could have the following structure:

Address: 4567 Main Street, NY, 98052, USA
Building Number: 4567
Street Name: Main Street
State: NY
Zip Code: 98052
Country: USA

Comment 1.1

ID: 1242295 User: MarceloManhaes Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Thu 04 Jul 2024 23:17 Selected Answer: - Upvotes: 1

I agree it is clear that is ML entity, the sample above is on the URL
https://learn.microsoft.com/en-us/azure/ai-services/LUIS/concepts/entities

Comment 2

ID: 381209 User: LKLK10 Badges: Highly Voted Relative Date: 4 years, 9 months ago Absolute Date: Sun 13 Jun 2021 18:25 Selected Answer: - Upvotes: 10

ML. Answer is A

Comment 2.1

ID: 394833 User: azurelearner666 Badges: - Relative Date: 4 years, 8 months ago Absolute Date: Wed 30 Jun 2021 16:32 Selected Answer: - Upvotes: 10

Right! (the correct response is A, Machine Learned)
See
https://docs.microsoft.com/en-us/azure/cognitive-services/luis/luis-concept-entity-types
It is a Machine Learned Entity (check ML Entity with Structure in the link, as it is an Address example… )

Comment 3

ID: 1283375 User: AzureGeek79 Badges: Most Recent Relative Date: 1 year, 5 months ago Absolute Date: Fri 13 Sep 2024 22:30 Selected Answer: - Upvotes: 1

Correct answer is D as per ChatGPT. Here is the response, "For capturing billing addresses in a Language Understanding model, the best choice would be Pattern.any (Option D). This is because billing addresses can vary greatly in format and content, and using Pattern.any allows for the flexibility needed to capture this variability effectively."

Comment 4

ID: 1247875 User: krzkrzkra Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sun 14 Jul 2024 19:36 Selected Answer: A Upvotes: 1

Selected Answer: A

Comment 5

ID: 1235202 User: HaraTadahisa Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 08:28 Selected Answer: A Upvotes: 1

I say this answer is A.

Comment 6

ID: 1230350 User: etellez Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 14 Jun 2024 11:16 Selected Answer: - Upvotes: 1

Copilot says Pattern.any

The Pattern.any entity type is designed to capture free-form text, which makes it suitable for capturing billing addresses that can come in various formats. It uses pattern matching to predict and extract data.

Comment 7

ID: 1229242 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 12 Jun 2024 16:24 Selected Answer: A Upvotes: 1

I know you don't know what I'm talking about, but if you think as Crossroads leads you, the answer is A.

Comment 8

ID: 1217581 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 16:23 Selected Answer: A Upvotes: 1

A is right answer.

Comment 9

ID: 1132988 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 27 Jan 2024 03:43 Selected Answer: - Upvotes: 1

Given these options, A. Machine Learned is the most appropriate choice for capturing billing addresses. Billing addresses are complex entities with a lot of variability in their format and structure. A machine-learned entity is capable of understanding and extracting such complex information from natural language inputs, which makes it suitable for this purpose. It can learn from examples and capture the billing address as an entity based on the context in which it appears, which is essential for handling the wide range of ways in which addresses can be presented.

Comment 10

ID: 1025737 User: jakespeed Badges: - Relative Date: 2 years, 5 months ago Absolute Date: Thu 05 Oct 2023 16:03 Selected Answer: A Upvotes: 1

ML Entity with Structure
An ML entity can be composed of smaller sub-entities, each of which can have its own properties. For example, an Address entity could have the following structure:

Address: 4567 Main Street, NY, 98052, USA
Building Number: 4567
Street Name: Main Street
State: NY
Zip Code: 98052
Country: USA

Comment 11

ID: 923265 User: EliteAllen Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Wed 14 Jun 2023 15:53 Selected Answer: C Upvotes: 3

C. geographyV2
The geographyV2 prebuilt entity in Language Understanding (LUIS) is designed to recognize and label entities that are geographical locations, such as city, state, or country. This would be suitable for capturing billing addresses in an e-commerce platform.

Comment 11.1

ID: 996018 User: M25 Badges: - Relative Date: 2 years, 6 months ago Absolute Date: Fri 01 Sep 2023 13:54 Selected Answer: - Upvotes: 2

https://learn.microsoft.com/en-us/azure/ai-services/luis/luis-reference-prebuilt-geographyv2?tabs=V3

The prebuilt geographyV2 entity detects places.
The geographical locations have subtypes:
poi point of interest
city name of city
countryRegion name of country or region
continent name of continent
state name of state or province

I guess you could charge a bill for the Statue of Liberty on Ellis Island as a (fixed) “poi”, but a more generalized rule would rather look for an Address entity with sub-entities (variable) as an ML Entity with Structure type

Comment 12

ID: 779274 User: ap1234pa Badges: - Relative Date: 3 years, 1 month ago Absolute Date: Tue 17 Jan 2023 20:35 Selected Answer: A Upvotes: 4

Wherever it is address it is ML

Comment 13

ID: 698645 User: David_ml Badges: - Relative Date: 3 years, 4 months ago Absolute Date: Wed 19 Oct 2022 05:37 Selected Answer: A Upvotes: 3

A is correct

Comment 14

ID: 652386 User: Nebary Badges: - Relative Date: 3 years, 6 months ago Absolute Date: Sat 27 Aug 2022 02:08 Selected Answer: A Upvotes: 3

It is 100% A

Comment 15

ID: 643832 User: ExamGuruBhai Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Sun 07 Aug 2022 21:46 Selected Answer: A Upvotes: 2

https://docs.microsoft.com/en-us/azure/cognitive-services/luis/concepts/entities

Comment 16

ID: 634867 User: RamonKaus Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Fri 22 Jul 2022 00:50 Selected Answer: B Upvotes: 5

Its regex. Udemy agrees.

Comment 17

ID: 633100 User: Eltooth Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Mon 18 Jul 2022 19:03 Selected Answer: A Upvotes: 2

A is correct answer.

80. AI-102 Topic 5 Question 25

Sequence
220
Discussion ID
105994
Source URL
https://www.examtopics.com/discussions/microsoft/view/105994-exam-ai-102-topic-5-question-25-discussion/
Posted By
Sachz88
Posted At
April 12, 2023, 6:32 a.m.

Question

HOTSPOT
-

You build a bot named app1 by using the Microsoft Bot Framework.

You prepare app1 for deployment.

You need to deploy app1 to Azure.

How should you complete the command? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

image

Suggested Answer

image
Answer Description Click to expand


Comments 10 comments Click to expand

Comment 1

ID: 979399 User: james2033 Badges: Highly Voted Relative Date: 2 years, 7 months ago Absolute Date: Sat 12 Aug 2023 16:11 Selected Answer: - Upvotes: 29

webapp

config-zip

See command

az webapp deployment source config-zip --resource-group "<resource-group-name>" --name "<name-of-app-service>" --src "<project-zip-path>"

at https://learn.microsoft.com/en-us/azure/bot-service/provision-and-publish-a-bot?view=azure-bot-service-4.0&tabs=userassigned%2Ccsharp#publish-your-bot-to-azure

Comment 1.1

ID: 1044689 User: AnonymousJhb Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Mon 16 Oct 2023 07:45 Selected Answer: - Upvotes: 7

answer is correct, if you agree,simply click the like button.

Comment 2

ID: 1283193 User: mrg998 Badges: Most Recent Relative Date: 1 year, 5 months ago Absolute Date: Fri 13 Sep 2024 15:33 Selected Answer: - Upvotes: 2

bot service questions are no longer on the exam

Comment 3

ID: 1243298 User: anto69 Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 06 Jul 2024 11:35 Selected Answer: - Upvotes: 1

no doubt: "webapp" + "config-zip"

Comment 4

ID: 1140079 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sun 04 Feb 2024 13:28 Selected Answer: - Upvotes: 3

To deploy app1 to Azure using the Microsoft Bot Framework, the correct command components are az webapp deployment source config-zip -- this command specifies that you're deploying a web app to Azure, using a zip file as the source of the deployment. The config-zip argument is used to deploy the application from a zipped source file.

app1/
├── .env (or appsettings.json for configuration settings)
├── index.js (entry point of your bot)
├── package.json (lists package dependencies)
├── node_modules/ (dependencies, can be excluded if using Azure for installation)
└── dialogs/ (or any other directories with additional bot logic)

Comment 5

ID: 929098 User: Tin_Tin Badges: - Relative Date: 2 years, 8 months ago Absolute Date: Wed 21 Jun 2023 07:49 Selected Answer: - Upvotes: 3

the answer seems correct
https://learn.microsoft.com/en-us/azure/bot-service/provision-and-publish-a-bot?view=azure-bot-service-4.0&tabs=userassigned%2Cpython#publish-your-bot-to-azure

Comment 6

ID: 912654 User: kail85 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Fri 02 Jun 2023 10:49 Selected Answer: - Upvotes: 3

first one should be bot

Comment 7

ID: 877716 User: odisor Badges: - Relative Date: 2 years, 10 months ago Absolute Date: Sun 23 Apr 2023 02:23 Selected Answer: - Upvotes: 3

the answers is correct, this is asking how to deploy the app for the bot , not the bot resource:
https://learn.microsoft.com/en-us/azure/bot-service/provision-and-publish-a-bot?view=azure-bot-service-4.0&tabs=userassigned%2Cpython#publish-your-bot-to-azure

to deploy a bot resource:
https://learn.microsoft.com/en-us/azure/bot-service/provision-azure-bot?view=azure-bot-service-4.0&tabs=userassigned

Comment 7.1

ID: 898987 User: MaliSanFuu Badges: - Relative Date: 2 years, 10 months ago Absolute Date: Tue 16 May 2023 10:17 Selected Answer: - Upvotes: 3

i think its a bit tricky with the wording because the bot is named "app1", therefore i think bot should be correct

Comment 8

ID: 867897 User: Sachz88 Badges: - Relative Date: 2 years, 11 months ago Absolute Date: Wed 12 Apr 2023 06:32 Selected Answer: - Upvotes: 2

The first box is incorrect. It should be "bot"
ChatGPT:
If you want to build a bot named app1 using the Microsoft Bot Framework, the deployment source you should use depends on your specific requirements and preferences.

If you want to deploy your bot on Microsoft Azure, you can use the "azbot" deployment source. This deployment source provides a pre-configured Azure Bot Service that allows you to easily deploy your bot to the cloud.

On the other hand, if you want to deploy your bot on a web server or hosting platform, you can use the "webapp" deployment source. This deployment source allows you to deploy your bot as a web application on a web server, using a programming language of your choice.

The question clearly states: "You need to deploy app1 to Azure."
Answer to the first box is "bot"

81. AI-102 Topic 5 Question 31

Sequence
221
Discussion ID
117064
Source URL
https://www.examtopics.com/discussions/microsoft/view/117064-exam-ai-102-topic-5-question-31-discussion/
Posted By
acsoma
Posted At
Aug. 2, 2023, 9:11 a.m.

Question

Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.

After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.

You are building a chatbot that will use question answering in Azure Cognitive Service for Language.

You have a PDF named Doc1.pdf that contains a product catalogue and a price list.

You upload Doc1.pdf and train the model.

During testing, users report that the chatbot responds correctly to the following question: What is the price of ?

The chatbot fails to respond to the following question: How much does cost?

You need to ensure that the chatbot responds correctly to both questions.

Solution: From Language Studio, you enable chit-chat, and then retrain and republish the model.

Does this meet the goal?

  • A. Yes
  • B. No

Suggested Answer

B

Answer Description Click to expand


Community Answer Votes

Comments 3 comments Click to expand

Comment 1

ID: 1283202 User: mrg998 Badges: - Relative Date: 1 year, 5 months ago Absolute Date: Fri 13 Sep 2024 15:43 Selected Answer: - Upvotes: 1

no is correct

Comment 2

ID: 1191861 User: Murtuza Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Tue 09 Apr 2024 01:21 Selected Answer: B Upvotes: 1

clearly B, ChitChat is for pleasantries

Comment 3

ID: 984780 User: james2033 Badges: - Relative Date: 2 years, 6 months ago Absolute Date: Fri 18 Aug 2023 23:00 Selected Answer: B Upvotes: 1

PDF file use for training, it is not rationale.

82. AI-102 Topic 5 Question 37

Sequence
222
Discussion ID
107309
Source URL
https://www.examtopics.com/discussions/microsoft/view/107309-exam-ai-102-topic-5-question-37-discussion/
Posted By
Codia
Posted At
April 24, 2023, 1:57 p.m.

Question

You are building a bot by using Microsoft Bot Framework.

You need to configure the bot to respond to spoken requests. The solution must minimize development effort.

What should you do?

  • A. Deploy the bot to Azure and register the bot with a Direct Line Speech channel.
  • B. Integrate the bot with Cortana by using the Bot Framework SDK.
  • C. Create an Azure function that will call the Speech service and connect the bot to the function.
  • D. Deploy the bot to Azure and register the bot with a Microsoft Teams channel.

Suggested Answer

A

Answer Description Click to expand


Community Answer Votes

Comments 12 comments Click to expand

Comment 1

ID: 913032 User: kail85 Badges: Highly Voted Relative Date: 2 years, 9 months ago Absolute Date: Fri 02 Jun 2023 21:55 Selected Answer: - Upvotes: 13

A. Deploy the bot to Azure and register the bot with a Direct Line Speech channel.

To enable the bot to respond to spoken requests with minimal development effort, you should deploy the bot to Azure and register the bot with a Direct Line Speech channel. Direct Line Speech provides an integrated speech and bot experience without requiring additional development. It combines both the speech-to-text and text-to-speech capabilities with the bot's logic to handle spoken requests and generate spoken responses.

Comment 2

ID: 934479 User: EliteAllen Badges: Highly Voted Relative Date: 2 years, 8 months ago Absolute Date: Mon 26 Jun 2023 15:04 Selected Answer: A Upvotes: 6

A. Deploy the bot to Azure and register the bot with a Direct Line Speech channel.

The Direct Line Speech channel in Azure Bot Service provides an integrated speech and bot experience. It combines the Bot Framework's Direct Line and Speech services into a single service that enables your bot to speak and listen to users. This is the simplest way to enable your bot to respond to spoken requests, as it doesn't require additional coding or integration with other services.

Comment 3

ID: 1283206 User: mrg998 Badges: Most Recent Relative Date: 1 year, 5 months ago Absolute Date: Fri 13 Sep 2024 15:47 Selected Answer: A Upvotes: 1

answer is 100% A

Comment 4

ID: 1235216 User: HaraTadahisa Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 08:51 Selected Answer: A Upvotes: 1

I say this answer is A.

Comment 5

ID: 1229865 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 15:53 Selected Answer: A Upvotes: 1

A is answer.

Comment 6

ID: 1145883 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 10 Feb 2024 02:28 Selected Answer: A Upvotes: 1

checked with documentation and the answer is A

Comment 7

ID: 1133171 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 27 Jan 2024 09:40 Selected Answer: A Upvotes: 2

Direct Line Speech Channel: This channel is specifically designed to handle voice interactions. It provides integrated speech recognition and text-to-speech capabilities, which are essential for handling spoken requests. By using this channel, you can leverage Azure's built-in capabilities without the need for extensive custom development.

Comment 8

ID: 1050004 User: sl_mslconsulting Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Sun 22 Oct 2023 02:48 Selected Answer: B Upvotes: 2

Check the link here https://blog.botframework.com/2017/06/26/speech-to-text/ and then decide the answer yourself. The direct line speech channel the development effort is not trivial.

Comment 9

ID: 920779 User: ziggy1117 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Sun 11 Jun 2023 16:38 Selected Answer: B Upvotes: 2

B. Cortana.

Using Direct Line Speech needs more work because we still need to develop the speech recognition. however, cortana has some built in features already but is not going to be as flexible as Direct line speech

Comment 10

ID: 920438 User: 973b658 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Sun 11 Jun 2023 07:34 Selected Answer: A Upvotes: 4

kail85 is correct.

Comment 11

ID: 916768 User: ziggy1117 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Wed 07 Jun 2023 03:35 Selected Answer: - Upvotes: 2

B. Cortana.

Using Direct Line Speech needs more work because we still need to develop the speech recognition. however, cortana has some built in features already but is not going to be as flexible as Direct line speech

Comment 12

ID: 879286 User: Codia Badges: - Relative Date: 2 years, 10 months ago Absolute Date: Mon 24 Apr 2023 13:57 Selected Answer: - Upvotes: 2

minimize development effort -> Cortana (we can connect the Cortana channel and integrate a Bot Application to Cortana)

83. AI-102 Topic 3 Question 75

Sequence
223
Discussion ID
135062
Source URL
https://www.examtopics.com/discussions/microsoft/view/135062-exam-ai-102-topic-3-question-75-discussion/
Posted By
audlindr
Posted At
March 2, 2024, 7 p.m.

Question

HOTSPOT
-

You have a collection of press releases stored as PDF files.

You need to extract text from the files and perform sentiment analysis.

Which service should you use for each task? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

image

Suggested Answer

image
Answer Description Click to expand


Comments 16 comments Click to expand

Comment 1

ID: 1215746 User: takaimomoGcup Badges: Highly Voted Relative Date: 1 year, 9 months ago Absolute Date: Wed 22 May 2024 15:18 Selected Answer: - Upvotes: 15

AI Document Intelligence and AI Language

Comment 2

ID: 1221861 User: Madeira Badges: Highly Voted Relative Date: 1 year, 9 months ago Absolute Date: Fri 31 May 2024 02:19 Selected Answer: - Upvotes: 7

Updated version of topic 3 question 46.
So according to the discussion in that question the correct answer is:
AI Vision
AI Language

Comment 3

ID: 1282649 User: mrg998 Badges: Most Recent Relative Date: 1 year, 6 months ago Absolute Date: Thu 12 Sep 2024 15:23 Selected Answer: - Upvotes: 1

AI Document Intelligence and AI Language

Comment 4

ID: 1247897 User: krzkrzkra Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sun 14 Jul 2024 19:56 Selected Answer: - Upvotes: 2

AI Vision
AI Language

Comment 5

ID: 1244953 User: Toby86 Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Tue 09 Jul 2024 16:25 Selected Answer: - Upvotes: 1

It can't be Document Intelligence. We don't know how the Press Releases are going to look like, if they have the same format and we don't need to extract anything specific, just text. And that you can achieve with AI Vision.

Comment 5.1

ID: 1244956 User: Toby86 Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Tue 09 Jul 2024 16:28 Selected Answer: - Upvotes: 1

nvm you also have to do sentiment analysis and document intelligence has also just OCR

Comment 5.2

ID: 1261916 User: moonlightc Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Wed 07 Aug 2024 04:40 Selected Answer: - Upvotes: 2

Format is not required when using Read API in document intelligence

Comment 6

ID: 1229577 User: KELCHOSPENSE Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 03:17 Selected Answer: - Upvotes: 5

1. AI Document Intelligence
2. AI Language

Comment 7

ID: 1225491 User: NagaoShingo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 06 Jun 2024 14:38 Selected Answer: - Upvotes: 2

1. AI Document Intelligence
2. AI Language

Comment 8

ID: 1201113 User: GHill1982 Badges: - Relative Date: 1 year, 10 months ago Absolute Date: Wed 24 Apr 2024 05:43 Selected Answer: - Upvotes: 4

For extracting text from PDF files, Azure Document Intelligence would be the better choice. It’s specifically optimized for text-heavy documents like PDFs and includes features such as higher-resolution scanning and paragraph detection. In addition, Azure Computer Vision API does not have direct PDF integration, and you would need to convert PDFs to images before text extraction, which adds an extra step to the process.

Comment 9

ID: 1191513 User: Murtuza Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Mon 08 Apr 2024 12:36 Selected Answer: - Upvotes: 1

Sentiment Analysis = AI Language

Comment 10

ID: 1186817 User: Murtuza Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Sun 31 Mar 2024 15:23 Selected Answer: - Upvotes: 1

the key word to answer this question is press release which implies computer vision

Comment 11

ID: 1178688 User: Murtuza Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Wed 20 Mar 2024 21:43 Selected Answer: - Upvotes: 2

Extract Text = AI Vision

Comment 11.1

ID: 1179967 User: fabrea Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Fri 22 Mar 2024 10:17 Selected Answer: - Upvotes: 1

I disagree, the Computer Vision for text extraction is best suited for unstructured documents
IDP is the right choice

Comment 11.2

ID: 1184544 User: Mehe323 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Thu 28 Mar 2024 04:47 Selected Answer: - Upvotes: 1

Document Intelligence also does text extraction:
"Document Intelligence Read Optical Character Recognition (OCR) model runs at a higher resolution than Azure AI Vision Read and extracts print and handwritten text from PDF documents and scanned images."

https://learn.microsoft.com/en-us/azure/ai-services/document-intelligence/concept-read?view=doc-intel-4.0.0

Comment 12

ID: 1173604 User: Razvan_C Badges: - Relative Date: 1 year, 12 months ago Absolute Date: Thu 14 Mar 2024 18:45 Selected Answer: - Upvotes: 1

Answer seems to be correct.

84. AI-102 Topic 3 Question 54

Sequence
228
Discussion ID
123329
Source URL
https://www.examtopics.com/discussions/microsoft/view/123329-exam-ai-102-topic-3-question-54-discussion/
Posted By
jangotango
Posted At
Oct. 11, 2023, 11:17 p.m.

Question

HOTSPOT
-

You are building an app by using the Speech SDK. The app will translate speech from French to German by using natural language processing.

You need to define the source language and the output language.

How should you complete the code? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

image

Suggested Answer

image
Answer Description Click to expand


Comments 6 comments Click to expand

Comment 1

ID: 1220851 User: hatanaoki Badges: Highly Voted Relative Date: 1 year, 9 months ago Absolute Date: Wed 29 May 2024 13:20 Selected Answer: - Upvotes: 14

SpeechRecognitionLanguage = "fr"
AddTargetLanguage.("de")

Comment 2

ID: 1281222 User: famco Badges: Most Recent Relative Date: 1 year, 6 months ago Absolute Date: Mon 09 Sep 2024 21:38 Selected Answer: - Upvotes: 3

e.v.i.l question.

A person who worked with it will go fast and select recognition and synthesis as the answer. But Microsoft made the second one a function call and with their bad naming convention added a new term "targetLanguage". Now you have to remember that mistake of Microsoft

Comment 3

ID: 1225500 User: NagaoShingo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 06 Jun 2024 14:43 Selected Answer: - Upvotes: 2

Sample answers are correct.

Comment 4

ID: 1185824 User: f2c587e Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Sat 30 Mar 2024 04:39 Selected Answer: - Upvotes: 2

Las respuestas son B y C

Comment 5

ID: 1139960 User: Florin83 Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sun 04 Feb 2024 10:48 Selected Answer: - Upvotes: 2

correct

Comment 6

ID: 1041118 User: jangotango Badges: - Relative Date: 2 years, 5 months ago Absolute Date: Wed 11 Oct 2023 23:17 Selected Answer: - Upvotes: 4

Answer is correct

85. AI-102 Topic 5 Question 38

Sequence
229
Discussion ID
117065
Source URL
https://www.examtopics.com/discussions/microsoft/view/117065-exam-ai-102-topic-5-question-38-discussion/
Posted By
acsoma
Posted At
Aug. 2, 2023, 9:11 a.m.

Question

Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.

After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.

You have a chatbot that uses question answering in Azure Cognitive Service for Language.

Users report that the responses of the chatbot lack formality when answering spurious questions.

You need to ensure that the chatbot provides formal responses to spurious questions.

Solution: From Language Studio, you remove all the chit-chat question and answer pairs, and then retrain and republish the model.

Does this meet the goal?

  • A. Yes
  • B. No

Suggested Answer

B

Answer Description Click to expand


Community Answer Votes

Comments 6 comments Click to expand

Comment 1

ID: 1279791 User: famco Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Fri 06 Sep 2024 22:16 Selected Answer: - Upvotes: 1

spurious?? What is spurious?? Do they provide a dictionary?

Comment 1.1

ID: 1279792 User: famco Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Fri 06 Sep 2024 22:20 Selected Answer: - Upvotes: 1

So, these are the questions they will not let you revisit. So, it is a russian roulette thinking the next question might have the option with professional chitchat selected. If it is not there then Microsoft wins and you lose if you selected No. This might be the closest because the requirement can be interpreted in a way that spurious questions does not need to be answered (or default answer) instead of a "professional reply". Oops.

Comment 2

ID: 1218944 User: reiwanotora Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sun 26 May 2024 13:06 Selected Answer: B Upvotes: 1

No is right answer.

Comment 3

ID: 1193692 User: jv1 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Thu 11 Apr 2024 12:39 Selected Answer: - Upvotes: 1

The proposed solution to ensure that the chatbot provides formal responses to spurious questions is to remove all the chit-chat question and answer pairs, and then retrain and republish the model. Answer=B

Comment 4

ID: 1133174 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 27 Jan 2024 09:43 Selected Answer: B Upvotes: 2

"Professional personality" refers to a specific conversational style or "character" chosen when designing a chatbot. This personality is typically more formal and suitable for use in corporate and official environments. In contrast, there may be conversational styles that are friendlier, more casual, or humorous. By selecting different "personalities," you can adjust the way the chatbot engages with users to better fit its application scenarios and target audience. In Azure AI's chatbot service, various preset personalities are often provided, such as professional, friendly, humorous, etc. Developers can choose the most suitable one based on the chatbot's purpose and user expectations. This flexibility allows the chatbot to exhibit different interaction styles in various application scenarios, thereby enhancing the user experience.

Comment 5

ID: 1056371 User: Lion007 Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Sat 28 Oct 2023 20:11 Selected Answer: B Upvotes: 4

The given answer Correct.

Removing all chitchat QA pairs will not solve the problem of lack of formality.

To solve the issue, change the chitchat source to qna_chitchat_professional.tsv

Check out https://learn.microsoft.com/en-us/azure/ai-services/language-service/question-answering/concepts/best-practices#choosing-a-personality

86. AI-102 Topic 3 Question 5

Sequence
230
Discussion ID
56455
Source URL
https://www.examtopics.com/discussions/microsoft/view/56455-exam-ai-102-topic-3-question-5-discussion/
Posted By
azurelearner666
Posted At
June 30, 2021, 7:11 p.m.

Question

HOTSPOT -
You are developing a service that records lectures given in English (United Kingdom).
You have a method named AppendToTranscriptFile that takes translated text and a language identifier.
You need to develop code that will provide transcripts of the lectures to attendees in their respective language. The supported languages are English, French,
Spanish, and German.
How should you complete the code? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Hot Area:
image

Suggested Answer

image
Answer Description Click to expand

Box 1: {"fr", "de", "es"}
A common task of speech translation is to specify target translation languages, at least one is required but multiples are supported. The following code snippet sets both French and German as translation language targets. static async Task TranslateSpeechAsync()
{
var translationConfig =
SpeechTranslationConfig.FromSubscription(SPEECH__SUBSCRIPTION__KEY, SPEECH__SERVICE__REGION); translationConfig.SpeechRecognitionLanguage = "it-IT";

// Translate to languages. See, https://aka.ms/speech/sttt-languages translationConfig.AddTargetLanguage("fr"); translationConfig.AddTargetLanguage("de");
}

Box 2: TranslationRecognizer -
After you've created a SpeechTranslationConfig, the next step is to initialize a TranslationRecognizer.
Example code:
static async Task TranslateSpeechAsync()
{
var translationConfig =
SpeechTranslationConfig.FromSubscription(SPEECH__SUBSCRIPTION__KEY, SPEECH__SERVICE__REGION); var fromLanguage = "en-US"; var toLanguages = new List { "it", "fr", "de" }; translationConfig.SpeechRecognitionLanguage = fromLanguage; toLanguages.ForEach(translationConfig.AddTargetLanguage); using var recognizer = new TranslationRecognizer(translationConfig);
}

Comments 7 comments Click to expand

Comment 1

ID: 394991 User: azurelearner666 Badges: Highly Voted Relative Date: 4 years, 8 months ago Absolute Date: Wed 30 Jun 2021 19:11 Selected Answer: - Upvotes: 20

Correct!

Comment 2

ID: 398578 User: idrisfl Badges: Highly Voted Relative Date: 4 years, 8 months ago Absolute Date: Sun 04 Jul 2021 20:08 Selected Answer: - Upvotes: 8

Seems correct
https://docs.microsoft.com/en-us/dotnet/api/microsoft.cognitiveservices.speech.translation.translationrecognizer?view=azure-dotnet

Comment 3

ID: 1279551 User: famco Badges: Most Recent Relative Date: 1 year, 6 months ago Absolute Date: Fri 06 Sep 2024 14:30 Selected Answer: - Upvotes: 3

So, they want to make it a trick by asking for english language as well, but it does not need to be passed. So, one point for the person who remembers it does not have to be passed? That knowledge proves they know Azure-AI?

Comment 4

ID: 1235139 User: LM12 Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 07:35 Selected Answer: - Upvotes: 3

was on exam 20.06.24

Comment 5

ID: 1158166 User: audlindr Badges: - Relative Date: 2 years ago Absolute Date: Sat 24 Feb 2024 22:20 Selected Answer: - Upvotes: 1

correct
https://learn.microsoft.com/en-us/azure/ai-services/speech-service/how-to-translate-speech?tabs=terminal&pivots=programming-language-csharp

Comment 6

ID: 1024594 User: ManvaIT Badges: - Relative Date: 2 years, 5 months ago Absolute Date: Wed 04 Oct 2023 10:59 Selected Answer: - Upvotes: 5

Correct,got this in Oct2023 exam

Comment 7

ID: 633034 User: Eltooth Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Mon 18 Jul 2022 15:09 Selected Answer: - Upvotes: 7

Answer is correct.

("fr", "de", "es")
TranslationRecognizer

87. AI-102 Topic 3 Question 8

Sequence
231
Discussion ID
60238
Source URL
https://www.examtopics.com/discussions/microsoft/view/60238-exam-ai-102-topic-3-question-8-discussion/
Posted By
czmiel24
Posted At
Aug. 22, 2021, 1:24 p.m.

Question

You are building an Azure WebJob that will create knowledge bases from an array of URLs.
You instantiate a QnAMakerClient object that has the relevant API keys and assign the object to a variable named client.
You need to develop a method to create the knowledge bases.
Which two actions should you include in the method? Each correct answer presents part of the solution.
NOTE: Each correct selection is worth one point.

  • A. Create a list of FileDTO objects that represents data from the WebJob.
  • B. Call the client.Knowledgebase.CreateAsync method.
  • C. Create a list of QnADTO objects that represents data from the WebJob.
  • D. Create a CreateKbDTO object.

Suggested Answer

BD

Answer Description Click to expand


Community Answer Votes

Comments 22 comments Click to expand

Comment 1

ID: 429229 User: czmiel24 Badges: Highly Voted Relative Date: 4 years, 6 months ago Absolute Date: Sun 22 Aug 2021 13:25 Selected Answer: - Upvotes: 35

It should be BD.

Comment 1.1

ID: 430582 User: SuperPetey Badges: - Relative Date: 4 years, 6 months ago Absolute Date: Tue 24 Aug 2021 09:59 Selected Answer: - Upvotes: 9

Correct - see code example here: https://docs.microsoft.com/en-us/azure/cognitive-services/qnamaker/quickstarts/quickstart-sdk?tabs=v1%2Cversion-1&pivots=programming-language-csharp

Comment 1.2

ID: 464950 User: vominhtri854 Badges: - Relative Date: 4 years, 4 months ago Absolute Date: Wed 20 Oct 2021 08:31 Selected Answer: - Upvotes: 4

A knowledge base stores question and answer pairs for the CreateKbDTO object from three sources:

- For editorial content, use the QnADTO object.
To use metadata and follow-up prompts, use the editorial context, because this data is added at the individual QnA pair level.
- For files, use the FileDTO object. The FileDTO includes the filename as well as the public URL to reach the file.
- For URLs, use a list of strings to represent publicly available URLs.
So I beleive A and C correct

Comment 2

ID: 559051 User: reachmymind Badges: Highly Voted Relative Date: 4 years ago Absolute Date: Tue 01 Mar 2022 23:26 Selected Answer: - Upvotes: 7

Answer :: B & D

A. Create a list of FileDTO objects that represents data from the WebJob.
NO - as it is from URL - so optional
B. Call the client.Knowledgebase.CreateAsync method.
YES - Mandatory to Call the Method
C. Create a list of QnADTO objects that represents data from the WebJob.
NO - as it is from URL - so optional
D. Create a CreateKbDTO object.
YES - Mandatory to Create

Go through the lines starting line 92 at below URL:
https://github.com/Azure-Samples/cognitive-services-qnamaker-csharp/blob/master/documentation-samples/quickstarts/Knowledgebase_Quickstart/Program.cs

Comment 2.1

ID: 646447 User: ninjia Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Sat 13 Aug 2022 22:25 Selected Answer: - Upvotes: 3

I agreed.
You are building an Azure WebJob that will create knowledge bases from an array of ##URLs##.
You could use FileDTO, QnADTO or urls to create the CreateKbDTO. Hence, FileDTO is not mandatory.

Code snippet from the link provided by reachmymind:
var createKbDto = new CreateKbDTO
{
Name = "QnA Maker FAQ from c# quickstart",
QnaList = new List<QnADTO> { qna1 },
//Files = new List<FileDTO> { file1 },
Urls = urls
};

Comment 2.1.1

ID: 825246 User: AzureJobsTillRetire Badges: - Relative Date: 3 years ago Absolute Date: Tue 28 Feb 2023 21:33 Selected Answer: - Upvotes: 1

DTO is data transfer object. You can use CreateKbDTO to create a knowledge base data transfer object that contains URLs and then call the client.Knowledgebase.CreateAsync method to create a knowledge base.

Comment 3

ID: 1235203 User: HaraTadahisa Badges: Most Recent Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 08:29 Selected Answer: BD Upvotes: 2

I say this answer is B and D.

Comment 4

ID: 1229244 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 12 Jun 2024 16:25 Selected Answer: BD Upvotes: 2

BD is answer.

Comment 5

ID: 1217567 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 16:15 Selected Answer: BD Upvotes: 1

client.Knowledgebase.CreateAsync and CreateKbDTO.

Comment 6

ID: 1152277 User: evangelist Badges: - Relative Date: 2 years ago Absolute Date: Sat 17 Feb 2024 00:02 Selected Answer: BD Upvotes: 2

Create a CreateKbDTO object: First, you need to create a CreateKbDTO object, which contains the necessary information for creating a knowledge base, such as an array of URLs and the name of the knowledge base.

Call the client.Knowledgebase.CreateAsync method: Then, use the Knowledgebase.CreateAsync method of the QnAMakerClient object, passing in the previously created CreateKbDTO object, to asynchronously create the knowledge base.

Comment 7

ID: 1049709 User: propanther Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Sat 21 Oct 2023 19:41 Selected Answer: - Upvotes: 1

It should be BD

CreateKbDTO holds list of FileDto, QnADto. In given question, knowledge base accepts URLs which is one of the properties of FileDto. Hence, FileDtos will be created for each returned file URL from WebJob which can be used to create and hold within CreateKbDTO

https://azuresdkdocs.blob.core.windows.net/$web/dotnet/Microsoft.Azure.CognitiveServices.Knowledge.QnAMaker/3.0.0-preview.1/api/Microsoft.Azure.CognitiveServices.Knowledge.QnAMaker.Models/Microsoft.Azure.CognitiveServices.Knowledge.QnAMaker.Models.CreateKbDTO.html

Comment 8

ID: 1046866 User: katrang Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Wed 18 Oct 2023 13:49 Selected Answer: BD Upvotes: 1

The client creates a knowledgebase from a CreateKbDTO object using the CreateAsync method. This object contains a list of FileDTOs and/or a list of QnADTOs. In this scenario we don't know which to use, but we definitely need the CreateAsync and the CreateKbDTO object.

Comment 9

ID: 1040148 User: sl_mslconsulting Badges: - Relative Date: 2 years, 5 months ago Absolute Date: Wed 11 Oct 2023 05:24 Selected Answer: BD Upvotes: 3

You can not even create QnA Maker resource any more. Just read though the code, memorize the answers, and then move on.

Comment 9.1

ID: 1279559 User: famco Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Fri 06 Sep 2024 14:40 Selected Answer: - Upvotes: 1

Isn't it sad that we have to do this with our time? Why do people buy software from Microsoft

Comment 10

ID: 925242 User: nitz14 Badges: - Relative Date: 2 years, 8 months ago Absolute Date: Fri 16 Jun 2023 16:00 Selected Answer: BD Upvotes: 1

ANSWER : B & D

The options A and C are not necessary for creating knowledge bases using the QnAMakerClient object. They mention creating a list of FileDTO and QnADTO objects, which could potentially be used for populating the knowledge base with data, but they are not directly related to the process of creating the knowledge base itself.

Comment 11

ID: 904057 User: EliteAllen Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Mon 22 May 2023 15:02 Selected Answer: BD Upvotes: 1

B. Call the client.Knowledgebase.CreateAsync method.
D. Create a CreateKbDTO object.

To create a knowledge base using the QnAMakerClient, you would need to create a CreateKbDTO object that contains the details of the knowledge base to be created. This object would include information such as the name of the knowledge base and the URLs of the documents to be included in the knowledge base.

After creating the CreateKbDTO object, you would then call the client.Knowledgebase.CreateAsync method, passing in the CreateKbDTO object as a parameter. This method would create the knowledge base and return a response that includes the ID of the newly created knowledge base.

Comment 12

ID: 741958 User: SSJA Badges: - Relative Date: 3 years, 3 months ago Absolute Date: Sun 11 Dec 2022 19:04 Selected Answer: BD Upvotes: 1

B & D is the correct answers.

Comment 13

ID: 677779 User: taer Badges: - Relative Date: 3 years, 5 months ago Absolute Date: Sat 24 Sep 2022 13:19 Selected Answer: BD Upvotes: 1

It should be BD.

Comment 14

ID: 636067 User: RamonKaus Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Sun 24 Jul 2022 15:50 Selected Answer: BD Upvotes: 2

I agree with Eltooth & czmiel

code from MS:
var createOp = await client.Knowledgebase.CreateAsync(createKbDto);

https://docs.microsoft.com/en-us/azure/cognitive-services/qnamaker/quickstarts/quickstart-sdk?tabs=v1%2Cversion-1&pivots=programming-language-csharp#create-a-knowledge-base

Comment 15

ID: 633084 User: Eltooth Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Mon 18 Jul 2022 18:05 Selected Answer: BD Upvotes: 1

B and D are correct answers.

https://docs.microsoft.com/en-us/azure/cognitive-services/qnamaker/quickstarts/quickstart-sdk?tabs=v1%2Cversion-1&pivots=programming-language-csharp#create-a-knowledge-base

Comment 16

ID: 465204 User: torekx Badges: - Relative Date: 4 years, 4 months ago Absolute Date: Wed 20 Oct 2021 15:53 Selected Answer: - Upvotes: 1

should be bcd options based on
https://docs.microsoft.com/en-us/azure/cognitive-services/qnamaker/quickstarts/quickstart-sdk?tabs=v1%2Cversion-1&pivots=programming-language-csharp#create-a-knowledge-base

Comment 17

ID: 446425 User: GMKanon Badges: - Relative Date: 4 years, 5 months ago Absolute Date: Fri 17 Sep 2021 09:26 Selected Answer: - Upvotes: 2

Should be BC rather.

88. AI-102 Topic 3 Question 10

Sequence
232
Discussion ID
75998
Source URL
https://www.examtopics.com/discussions/microsoft/view/75998-exam-ai-102-topic-3-question-10-discussion/
Posted By
jekko
Posted At
May 21, 2022, 11:22 a.m.

Question

You are building a conversational language understanding model.
You need to enable active learning.
What should you do?

  • A. Add show-all-intents=true to the prediction endpoint query.
  • B. Enable speech priming.
  • C. Add log=true to the prediction endpoint query.
  • D. Enable sentiment analysis.

Suggested Answer

C

Answer Description Click to expand


Community Answer Votes

Comments 8 comments Click to expand

Comment 1

ID: 633092 User: Eltooth Badges: Highly Voted Relative Date: 3 years, 7 months ago Absolute Date: Mon 18 Jul 2022 18:41 Selected Answer: C Upvotes: 13

C is the correct answer.

"To enable active learning, you must log user queries. This is accomplished by calling the endpoint query with the log=true query string parameter and value."

https://docs.microsoft.com/en-us/azure/cognitive-services/LUIS/how-to/improve-application#log-user-queries-to-enable-active-learning

Comment 2

ID: 1279594 User: famco Badges: Most Recent Relative Date: 1 year, 6 months ago Absolute Date: Fri 06 Sep 2024 16:06 Selected Answer: - Upvotes: 2

Really?!! log=true is that how active learning is enabled? That's a very bad choice and asking that in a certification exam shows courage LOL

Comment 3

ID: 1229243 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 12 Jun 2024 16:24 Selected Answer: C Upvotes: 1

C is answer.

Comment 4

ID: 1217554 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 16:04 Selected Answer: C Upvotes: 1

Add log=true

Comment 5

ID: 1156141 User: anto69 Badges: - Relative Date: 2 years ago Absolute Date: Thu 22 Feb 2024 06:19 Selected Answer: C Upvotes: 1

C is correct

Comment 6

ID: 837888 User: marti_tremblay000 Badges: - Relative Date: 2 years, 12 months ago Absolute Date: Mon 13 Mar 2023 12:39 Selected Answer: - Upvotes: 1

Log user queries to enable active learning
To enable active learning, you must log user queries. This is accomplished by calling the endpoint query with the log=true query string parameter and value.
https://learn.microsoft.com/en-us/azure/cognitive-services/LUIS/how-to/improve-application

Comment 7

ID: 816448 User: Marilena96 Badges: - Relative Date: 3 years ago Absolute Date: Tue 21 Feb 2023 11:31 Selected Answer: - Upvotes: 1

To enable active learning in a conversational language understanding model, you should add show-all-intents=true to the prediction endpoint query. This will allow you to see all the intents that the model is predicting, including the None intent.[0] This information can be used to improve the model by adding more training data for the None intent or other intents that are not being predicted accurately.

Comment 8

ID: 604779 User: jekko Badges: - Relative Date: 3 years, 9 months ago Absolute Date: Sat 21 May 2022 11:22 Selected Answer: - Upvotes: 1

Correct.

Reference: https://docs.microsoft.com/en-us/azure/cognitive-services/LUIS/how-to/improve-application

89. AI-102 Topic 3 Question 14

Sequence
233
Discussion ID
77617
Source URL
https://www.examtopics.com/discussions/microsoft/view/77617-exam-ai-102-topic-3-question-14-discussion/
Posted By
Eltooth
Posted At
July 19, 2022, 6:59 a.m.

Question

You are developing a method for an application that uses the Translator API.
The method will receive the content of a webpage, and then translate the content into Greek (el). The result will also contain a transliteration that uses the Roman alphabet.
You need to create the URI for the call to the Translator API.
You have the following URI.
https://api.cognitive.microsofttranslator.com/translate?api-version=3.0
Which three additional query parameters should you include in the URI? Each correct answer presents part of the solution.
NOTE: Each correct selection is worth one point.

  • A. toScript=Cyrl
  • B. from=el
  • C. textType=html
  • D. to=el
  • E. textType=plain
  • F. toScript=Latn

Suggested Answer

CDF

Answer Description Click to expand


Community Answer Votes

Comments 9 comments Click to expand

Comment 1

ID: 633352 User: Eltooth Badges: Highly Voted Relative Date: 3 years, 7 months ago Absolute Date: Tue 19 Jul 2022 06:59 Selected Answer: CDF Upvotes: 8

C, D and F are correct answers.
textType=html
to=el
toScript=Latn

https://docs.microsoft.com/en-us/azure/cognitive-services/translator/reference/v3-0-translate#translate-with-transliteration

Comment 2

ID: 936179 User: Pixelmate Badges: Highly Voted Relative Date: 2 years, 8 months ago Absolute Date: Wed 28 Jun 2023 07:38 Selected Answer: - Upvotes: 6

THIS WAS ON EXAM 28/06

Comment 3

ID: 1279607 User: famco Badges: Most Recent Relative Date: 1 year, 6 months ago Absolute Date: Fri 06 Sep 2024 16:21 Selected Answer: - Upvotes: 5

wow, microsoft wants me to know webpage is html, roman=latn.
Although I never understood why microsoft expects people to remember the .net class names or method names or api parameter names. But this knowledge is beyond anything. I'm guessing the people creating the questions are not from this area. They just look at the documentation and think what can I ask so that the people cannot remember what is the right answer.

Comment 4

ID: 1229239 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 12 Jun 2024 16:22 Selected Answer: CDF Upvotes: 1

I'll bump it up, so CDF is the answer.

Comment 5

ID: 1217544 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 16:01 Selected Answer: CDF Upvotes: 1

textType=html
to=el
toScript=Latn

Comment 6

ID: 1217356 User: funny_penguin Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 11:39 Selected Answer: CDF Upvotes: 1

correct answer

Comment 7

ID: 1132994 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 27 Jan 2024 03:51 Selected Answer: CDF Upvotes: 4

https://api.cognitive.microsofttranslator.com/translate?api-version=3.0&textType=html&to=el&toScript=Latn

Comment 8

ID: 726496 User: Aztek0403 Badges: - Relative Date: 3 years, 3 months ago Absolute Date: Fri 25 Nov 2022 09:18 Selected Answer: CDF Upvotes: 1

Correct!

Comment 9

ID: 636068 User: RamonKaus Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Sun 24 Jul 2022 15:55 Selected Answer: CDF Upvotes: 2

Agreed

90. AI-102 Topic 3 Question 23

Sequence
234
Discussion ID
85823
Source URL
https://www.examtopics.com/discussions/microsoft/view/85823-exam-ai-102-topic-3-question-23-discussion/
Posted By
David_ml
Posted At
Oct. 18, 2022, 7:52 p.m.

Question

You build a Conversational Language Understanding model by using the Language Services portal.
You export the model as a JSON file as shown in the following sample.
image
To what does the Weather.Historic entity correspond in the utterance?

  • A. by month
  • B. chicago
  • C. rain
  • D. location

Suggested Answer

A

Answer Description Click to expand


Community Answer Votes

Comments 9 comments Click to expand

Comment 1

ID: 1279662 User: famco Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Fri 06 Sep 2024 17:13 Selected Answer: - Upvotes: 1

It should be "last year". Does the question creators get extra time to make these questions because english is not their native language?

Comment 2

ID: 1263100 User: anto69 Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Fri 09 Aug 2024 17:05 Selected Answer: A Upvotes: 1

by month, basic math

Comment 3

ID: 1217522 User: reiwanotora Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 15:47 Selected Answer: A Upvotes: 2

23 to 30 is by month.

Comment 4

ID: 1217358 User: funny_penguin Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 11:39 Selected Answer: A Upvotes: 3

on exam, by month

Comment 5

ID: 1158440 User: anto69 Badges: - Relative Date: 2 years ago Absolute Date: Sun 25 Feb 2024 08:13 Selected Answer: A Upvotes: 1

"by month" seems correct

Comment 6

ID: 1133044 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 27 Jan 2024 04:52 Selected Answer: A Upvotes: 2

by month means to check history in the past by each month, no doubt answer is A

Comment 7

ID: 1130214 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Wed 24 Jan 2024 05:21 Selected Answer: A Upvotes: 1

No doubt, A checked by different AI: chatgpt, claude 2 and Google bard

Comment 8

ID: 1054589 User: chenglim Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Thu 26 Oct 2023 15:05 Selected Answer: A Upvotes: 2

correct answer

Comment 9

ID: 698417 User: David_ml Badges: - Relative Date: 3 years, 4 months ago Absolute Date: Tue 18 Oct 2022 19:52 Selected Answer: A Upvotes: 2

correct

91. AI-102 Topic 5 Question 35

Sequence
235
Discussion ID
111871
Source URL
https://www.examtopics.com/discussions/microsoft/view/111871-exam-ai-102-topic-5-question-35-discussion/
Posted By
973b658
Posted At
June 11, 2023, 7:30 a.m.

Question

You are building a chatbot.

You need to configure the bot to guide users through a product setup process.

Which type of dialog should you use?

  • A. component
  • B. action
  • C. waterfall
  • D. adaptive

Suggested Answer

C

Answer Description Click to expand


Community Answer Votes

Comments 8 comments Click to expand

Comment 1

ID: 1279790 User: famco Badges: Highly Voted Relative Date: 1 year, 6 months ago Absolute Date: Fri 06 Sep 2024 22:15 Selected Answer: - Upvotes: 5

bot should be out of syllabus

Comment 2

ID: 1235218 User: HaraTadahisa Badges: Most Recent Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 08:51 Selected Answer: C Upvotes: 1

I say this answer is C.

Comment 3

ID: 1229924 User: hatanaoki Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 17:17 Selected Answer: C Upvotes: 1

C is the correct answer for this.
J'aime les gros seins.

Comment 4

ID: 1221644 User: fuck_india Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 30 May 2024 16:13 Selected Answer: C Upvotes: 1

It must be C.
product setup process = Waterfall

Comment 5

ID: 1049996 User: sl_mslconsulting Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Sun 22 Oct 2023 02:35 Selected Answer: A Upvotes: 2

Setting up a product might not be a straight linear process in my view. You might need component dialog to coordinate among multiple waterfalls dialogs which by design work within the context of a component dialog. https://learn.microsoft.com/en-us/azure/bot-service/bot-builder-concept-waterfall-dialogs?view=azure-bot-service-4.0

Comment 6

ID: 979385 User: james2033 Badges: - Relative Date: 2 years, 7 months ago Absolute Date: Sat 12 Aug 2023 15:40 Selected Answer: C Upvotes: 2

Quote "waterfall dialog: Defines a sequence of steps, allowing your bot to guide a user through a linear process. These are typically designed to work within the context of a component dialog." at https://learn.microsoft.com/en-us/azure/bot-service/bot-builder-concept-dialog?view=azure-bot-service-4.0#dialog-types

Comment 6.1

ID: 1211882 User: abelarda Badges: - Relative Date: 1 year, 10 months ago Absolute Date: Wed 15 May 2024 12:41 Selected Answer: - Upvotes: 1

Shouldn't it be component to work in a "context of a component dialog"?

Comment 7

ID: 920436 User: 973b658 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Sun 11 Jun 2023 07:30 Selected Answer: C Upvotes: 1

C
Like #22 Topic5

92. AI-102 Topic 5 Question 27

Sequence
240
Discussion ID
102921
Source URL
https://www.examtopics.com/discussions/microsoft/view/102921-exam-ai-102-topic-5-question-27-discussion/
Posted By
marti_tremblay000
Posted At
March 17, 2023, 12:42 p.m.

Question

Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.

After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.

You have a chatbot that uses question answering in Azure Cognitive Service for Language.

Users report that the responses of the chatbot lack formality when answering spurious questions.

You need to ensure that the chatbot provides formal responses to spurious questions.

Solution: From Language Studio, you modify the question and answer pairs for the custom intents, and then retrain and republish the model.

Does this meet the goal?

  • A. Yes
  • B. No

Suggested Answer

B

Answer Description Click to expand


Community Answer Votes

Comments 7 comments Click to expand

Comment 1

ID: 934472 User: EliteAllen Badges: Highly Voted Relative Date: 2 years, 8 months ago Absolute Date: Mon 26 Jun 2023 14:59 Selected Answer: B Upvotes: 8

No, this solution does not meet the goal.

The formality of responses to spurious questions is not controlled by modifying the question and answer pairs for the custom intents in Language Studio. These pairs are used to train the model to understand and respond to specific intents, not to control the tone or formality of the responses.

To ensure that the chatbot provides formal responses to spurious questions, you would need to adjust the chatbot's response templates or scripts, not the question and answer pairs for the custom intents. This might involve programming the chatbot to use more formal language in its responses, or to respond to unrecognized or spurious inputs with a standard, formal message.

Comment 2

ID: 1271526 User: Moneybing Badges: Most Recent Relative Date: 1 year, 6 months ago Absolute Date: Sat 24 Aug 2024 06:51 Selected Answer: A Upvotes: 1

copilot says YES.

"By modifying the question and answer pairs for the custom intents in Language Studio, you can improve the chatbot’s responses to spurious questions. Retraining and republishing the model with the updated data will ensure that the chatbot provides more formal and accurate answers."

Comment 3

ID: 1140083 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sun 04 Feb 2024 13:38 Selected Answer: B Upvotes: 1

The qna_chitchat_professional.tsv is specifically designed to provide responses with a formal tone compared to other chitchat sources. It contains a curated set of question-and-answer pairs that are structured to reflect a professional and formal manner of communication. This makes it more suitable for environments or user interactions where formal responses are preferred, ensuring that the chatbot's replies to spurious or off-topic questions maintain the desired level of professionalism.

Comment 4

ID: 931978 User: Drummer Badges: - Relative Date: 2 years, 8 months ago Absolute Date: Sat 24 Jun 2023 00:06 Selected Answer: - Upvotes: 3

No, this solution does not meet the goal. Modifying the question and answer pairs for custom intents in Language Studio and retraining and republishing the model would only affect the responses of the chatbot to specific intents. B NO

Comment 5

ID: 920395 User: ziggy1117 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Sun 11 Jun 2023 05:59 Selected Answer: B Upvotes: 4

Spurious questions are questions that are not relevant to the conversation or are not meant to be answered. In other words, these are chit-chat. So to make it professional, choose a professional chitchat

Comment 6

ID: 912657 User: kail85 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Fri 02 Jun 2023 10:53 Selected Answer: B Upvotes: 2

you should use the Azure Cognitive Service for Language's Language Understanding (LUIS) feature to identify the intent of the user's input and filter out spurious questions or provide formal fallback responses. You can achieve this by training your LUIS model to recognize spurious questions as a separate intent and handle them accordingly in your chatbot's logic.

Comment 7

ID: 841939 User: marti_tremblay000 Badges: - Relative Date: 2 years, 12 months ago Absolute Date: Fri 17 Mar 2023 12:42 Selected Answer: A Upvotes: 3

ChatGPT answer :
Yes, modifying the question and answer pairs for the custom intents in Language Studio, and then retraining and republishing the model can meet the goal of ensuring that the chatbot provides formal responses to spurious questions.

The question and answer pairs in Language Studio are used to train the machine learning model for the chatbot. By modifying these pairs to provide more formal responses to spurious questions, and then retraining and republishing the model, the chatbot will have a better understanding of the user's input and will be able to provide more formal responses.

93. AI-102 Topic 5 Question 1

Sequence
241
Discussion ID
61688
Source URL
https://www.examtopics.com/discussions/microsoft/view/61688-exam-ai-102-topic-5-question-1-discussion/
Posted By
jayf
Posted At
Sept. 8, 2021, 7:59 a.m.

Question

You build a bot by using the Microsoft Bot Framework SDK and the Azure Bot Service.
You plan to deploy the bot to Azure.
You register the bot by using the Bot Channels Registration service.
Which two values are required to complete the deployment? Each correct answer presents part of the solution.
NOTE: Each correct selection is worth one point.

  • A. botId
  • B. tenantId
  • C. appId
  • D. objectId
  • E. appSecret

Suggested Answer

CE

Answer Description Click to expand


Community Answer Votes

Comments 10 comments Click to expand

Comment 1

ID: 1270539 User: Moneybing Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Thu 22 Aug 2024 08:49 Selected Answer: CE Upvotes: 1

copilot says CE

To deploy your bot to Azure, you’ll need the following two values:

App ID: This is a unique identifier for your bot. It’s generated when you register your bot with the Bot Channels Registration service. The App ID is used to authenticate your bot when it communicates with channels.
App Password/Secret: This is a secret key associated with your bot. It’s also generated during the registration process. The App Password is used for secure communication between your bot and the channels.

Comment 2

ID: 1229903 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 17:03 Selected Answer: CE Upvotes: 2

CE is answer.

Comment 3

ID: 1222695 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sat 01 Jun 2024 16:43 Selected Answer: CE Upvotes: 1

C and E are right answer.

Comment 4

ID: 1139776 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sun 04 Feb 2024 05:30 Selected Answer: CE Upvotes: 2

The answer is correct

Comment 5

ID: 633585 User: Eltooth Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Tue 19 Jul 2022 15:11 Selected Answer: CE Upvotes: 4

C and E are correct answers.

https://docs.microsoft.com/en-gb/learn/modules/design-bot-conversation-flow/5-deploy-bot

Comment 5.1

ID: 633586 User: Eltooth Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Tue 19 Jul 2022 15:12 Selected Answer: - Upvotes: 1

Create the Azure resources required to support your bot
Your will need to create an Azure application registration to give your bot an identity it can use to access resources, and a bot application service to host the bot.

Register an Azure app
You can create the application registration by using the az ad app create Azure command-line interface (CLI) command, specifying a display name and password for your app identity. This command registers the app and returns its registration information, including a unique application ID that you will need in the following step.

Create a bot application service
Your bot requires a Bot Channels Registration resource, along with associated application service and application service plan. To create these resources, you can use the Azure resource deployment templates provided with the Bot Framework SDK template you used to create your bot. Just run the az deployment group create command, referencing the deployment template and specifying your bot application registration's ID (from the az ad app create command output) and the password you specified.

Comment 6

ID: 559404 User: reachmymind Badges: - Relative Date: 4 years ago Absolute Date: Wed 02 Mar 2022 13:28 Selected Answer: - Upvotes: 4

appId
appSecret

az deployment sub create --template-file "<path-to-template-with-new-rg.json>" --location <region-location-name> --parameters appType="MultiTenant" appId="<app-id-from-previous-step>" appSecret="<password-from-previous-step>" botId="<id or bot-app-service-name>" botSku=F0 newAppServicePlanName="<new-service-plan-name>" newWebAppName="<bot-app-service-name>" groupName="<new-group-name>" groupLocation="<region-location-name>" newAppServicePlanLocation="<region-location-name>" --name "<bot-app-service-name>"

https://docs.microsoft.com/en-us/azure/bot-service/bot-builder-deploy-az-cli?view=azure-bot-service-4.0&tabs=csharp%2Cmultitenant

Comment 6.1

ID: 599401 User: JTWang Badges: - Relative Date: 3 years, 10 months ago Absolute Date: Tue 10 May 2022 06:56 Selected Answer: - Upvotes: 1

The bot channels registration registers your web service as a bot with the Bot Framework, provided you have a Microsoft App Id and App password (client secret).
https://github.com/OfficeDev/Microsoft-Teams-Samples/blob/main/samples/tab-adaptive-cards/nodejs/Wiki/azure-bot-channels-registration.md#create-the-bot-channels-registration

Comment 7

ID: 441244 User: jayf Badges: - Relative Date: 4 years, 6 months ago Absolute Date: Wed 08 Sep 2021 07:59 Selected Answer: - Upvotes: 3

Answer should be correct

Comment 7.1

ID: 473587 User: Mll1975 Badges: - Relative Date: 4 years, 4 months ago Absolute Date: Sat 06 Nov 2021 20:01 Selected Answer: - Upvotes: 2

Agree
https://github.com/MicrosoftDocs/bot-docs/blob/live/articles/bot-service-quickstart-registration.md#update-the-bot

94. AI-102 Topic 3 Question 42

Sequence
242
Discussion ID
111851
Source URL
https://www.examtopics.com/discussions/microsoft/view/111851-exam-ai-102-topic-3-question-42-discussion/
Posted By
973b658
Posted At
June 11, 2023, 1:15 a.m.

Question

DRAG DROP
-

You are building a transcription service for technical podcasts.

Testing reveals that the service fails to transcribe technical terms accurately.

You need to improve the accuracy of the service.

Which five actions should you perform in sequence? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.

image

Suggested Answer

image
Answer Description Click to expand


Comments 5 comments Click to expand

Comment 1

ID: 1266142 User: anto69 Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Thu 15 Aug 2024 04:34 Selected Answer: - Upvotes: 2

Given answer is correct

Comment 2

ID: 1248506 User: krzkrzkra Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Mon 15 Jul 2024 20:08 Selected Answer: - Upvotes: 1

1. Create Custom Speech project
2. Create speech-to-text model
3. Upload training datasets
4. Train model
5. Deploy model

Comment 3

ID: 1185104 User: f2c587e Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Fri 29 Mar 2024 01:10 Selected Answer: - Upvotes: 2

1. Create a Custom Voice Project
2. Create a speech-to-text model
3. Upload Training Datasets
4. Training Model
5. Implementation model

Comment 4

ID: 1185103 User: f2c587e Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Fri 29 Mar 2024 01:08 Selected Answer: - Upvotes: 1

According to the answer, then data should not be uploaded to train the model? Seriously? So how do you plan to train yourself if they're supposed to be technical words. I agree with zellck

Comment 5

ID: 920306 User: 973b658 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Sun 11 Jun 2023 01:15 Selected Answer: - Upvotes: 3

It is true.

95. AI-102 Topic 3 Question 53

Sequence
243
Discussion ID
112145
Source URL
https://www.examtopics.com/discussions/microsoft/view/112145-exam-ai-102-topic-3-question-53-discussion/
Posted By
973b658
Posted At
June 14, 2023, 9:41 a.m.

Question

DRAG DROP
-

You are building a customer support chatbot.

You need to configure the bot to identify the following:

• Code names for internal product development
• Messages that include credit card numbers

The solution must minimize development effort.

Which Azure Cognitive Service for Language feature should you use for each requirement? To answer, drag the appropriate features to the correct requirements. Each feature may be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view content.

NOTE: Each correct selection is worth one point.

image

Suggested Answer

image
Answer Description Click to expand


Comments 7 comments Click to expand

Comment 1

ID: 1266147 User: anto69 Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Thu 15 Aug 2024 04:42 Selected Answer: - Upvotes: 2

1. Custom NER
2. PII detection

Comment 2

ID: 1243294 User: anto69 Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 06 Jul 2024 11:28 Selected Answer: - Upvotes: 1

NER + PII

Comment 2.1

ID: 1266148 User: anto69 Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Thu 15 Aug 2024 04:42 Selected Answer: - Upvotes: 1

It's custom NER

Comment 3

ID: 1234518 User: HaraTadahisa Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Fri 21 Jun 2024 16:59 Selected Answer: - Upvotes: 2

1. Custom NER
2. PII

Comment 4

ID: 1215783 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 22 May 2024 15:55 Selected Answer: - Upvotes: 1

NER and PII.

Comment 5

ID: 925048 User: Tin_Tin Badges: - Relative Date: 2 years, 8 months ago Absolute Date: Fri 16 Jun 2023 11:55 Selected Answer: - Upvotes: 2

seems correct.
https://learn.microsoft.com/en-us/azure/cognitive-services/language-service/overview

Comment 6

ID: 922859 User: 973b658 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Wed 14 Jun 2023 09:41 Selected Answer: - Upvotes: 2

It is true.

96. AI-102 Topic 5 Question 15

Sequence
244
Discussion ID
77633
Source URL
https://www.examtopics.com/discussions/microsoft/view/77633-exam-ai-102-topic-5-question-15-discussion/
Posted By
Eltooth
Posted At
July 20, 2022, 12:36 p.m.

Question

DRAG DROP -
You have a chatbot that uses a QnA Maker application.
You enable active learning for the knowledge base used by the QnA Maker application.
You need to integrate user input into the model.
Which four actions should you perform in sequence? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.
Select and Place:
image

Suggested Answer

image
Answer Description Click to expand

Step 1: For the knowledge base, select Show active learning suggestions.
In order to see the suggested questions, on the Edit knowledge base page, select View Options, then select Show active learning suggestions.
Step 2: Approve and reject suggestions.
Each QnA pair suggests the new question alternatives with a check mark, , to accept the question or an x to reject the suggestions. Select the check mark to
ג"
add the question.
Step 3: Save and train the knowledge base.
Select Save and Train to save the changes to the knowledge base.
Step 4: Publish the knowledge base.
Select Publish to allow the changes to be available from the GenerateAnswer API.
When 5 or more similar queries are clustered, every 30 minutes, QnA Maker suggests the alternate questions for you to accept or reject.
Reference:
https://docs.microsoft.com/en-us/azure/cognitive-services/qnamaker/how-to/improve-knowledge-base

Comments 5 comments Click to expand

Comment 1

ID: 633983 User: Eltooth Badges: Highly Voted Relative Date: 3 years, 7 months ago Absolute Date: Wed 20 Jul 2022 12:36 Selected Answer: - Upvotes: 13

Answer provided is correct.

(Turn on active learning) - Done
1. Show active learning suggestions
2. Accept the question or reject the suggestions
3. Save and Train
4. Publish

https://docs.microsoft.com/en-us/azure/cognitive-services/qnamaker/how-to/improve-knowledge-base#view-suggested-questions

Comment 1.1

ID: 648166 User: ninjia Badges: - Relative Date: 3 years, 6 months ago Absolute Date: Wed 17 Aug 2022 19:09 Selected Answer: - Upvotes: 1

Agreed.

Comment 1.2

ID: 1222684 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sat 01 Jun 2024 16:09 Selected Answer: - Upvotes: 1

Your spelling mistake confuses the whole thing. The correct answer is this.

1. For the knowledge base, select Show active learning suggestions.
2. Approve and reject the suggestions.
3. Save and train knowledge base.
4. Publish knowledge base.

Comment 2

ID: 1266169 User: anto69 Badges: Most Recent Relative Date: 1 year, 6 months ago Absolute Date: Thu 15 Aug 2024 05:36 Selected Answer: - Upvotes: 2

1. From knowledge base, select show active learning suggestions.
2. Approve and reject the suggestions.
3. Save and train knowledge base.
4. Publish knowledge base.

Comment 3

ID: 1139893 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sun 04 Feb 2024 09:06 Selected Answer: - Upvotes: 1

1. Show active learning suggestions
2. Accept the question or reject the suggestions
3. Save and Train
4. Publish

97. AI-102 Topic 5 Question 23

Sequence
245
Discussion ID
111023
Source URL
https://www.examtopics.com/discussions/microsoft/view/111023-exam-ai-102-topic-5-question-23-discussion/
Posted By
kail85
Posted At
June 4, 2023, 1:26 a.m.

Question

DRAG DROP
-

You build a bot by using the Microsoft Bot Framework SDK.

You need to test the bot interactively on a local machine.

Which three actions should you perform in sequence? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.

NOTE: More than one order of answer choices is correct. You will receive credit for any of the correct orders you select.

image

Suggested Answer

image
Answer Description Click to expand


Comments 5 comments Click to expand

Comment 1

ID: 913957 User: kail85 Badges: Highly Voted Relative Date: 2 years, 9 months ago Absolute Date: Sun 04 Jun 2023 01:26 Selected Answer: - Upvotes: 7

Answer is correct.

Build and run the bot
Open the Bot Framework Emulator
Connect to the bot endpoint

Comment 2

ID: 1266170 User: anto69 Badges: Most Recent Relative Date: 1 year, 6 months ago Absolute Date: Thu 15 Aug 2024 05:38 Selected Answer: - Upvotes: 1

1. Build and run the bot
2. Open Bot Framework Emulator
3. Connect to bot endpoint

Comment 3

ID: 1247928 User: krzkrzkra Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sun 14 Jul 2024 20:21 Selected Answer: - Upvotes: 1

1. Build and run the bot
2. Open Bot Framework Emulator
3. Connect to bot endpoint

Comment 4

ID: 1227812 User: reiwanotora Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Mon 10 Jun 2024 12:50 Selected Answer: - Upvotes: 1

1. Build and run the bot
2. Open Bot Framework Emulator
3. Connect to bot endpoint

Comment 5

ID: 1139933 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sun 04 Feb 2024 10:21 Selected Answer: - Upvotes: 4

Given answer is correct and below is why:

1: Build and run the bot: This step involves compiling the bot's code and starting it on your local machine to make it ready to accept connections.
2: Open the Bot Framework Emulator: The Emulator is a desktop application that allows developers to test and debug their bots on localhost or remotely through a tunnel.
3: Connect to the bot endpoint: Within the Bot Framework Emulator, you would connect to the bot's endpoint, usually something like http://localhost:3978/api/messages, to start interacting with your bot.

98. AI-102 Topic 5 Question 48

Sequence
246
Discussion ID
112152
Source URL
https://www.examtopics.com/discussions/microsoft/view/112152-exam-ai-102-topic-5-question-48-discussion/
Posted By
973b658
Posted At
June 14, 2023, 10:20 a.m.

Question

HOTSPOT
-

You have a chatbot.

You need to test the bot by using the Bot Framework Emulator. The solution must ensure that you are prompted for credentials when you sign in to the bot.

Which three settings should you configure? To answer, select the appropriate settings in the answer area.

NOTE: Each correct selection is worth one point.

image

Suggested Answer

image
Answer Description Click to expand


Comments 5 comments Click to expand

Comment 1

ID: 1266173 User: anto69 Badges: - Relative Date: 1 year, 6 months ago Absolute Date: Thu 15 Aug 2024 05:41 Selected Answer: - Upvotes: 1

1. Enter the local path to ngrok.
2. Enable Run ngrok when the Emulator starts up.
3. Enable Use version 1.0 authentication tokens.

Comment 2

ID: 1145894 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 10 Feb 2024 02:58 Selected Answer: - Upvotes: 2

What is ngrok?

ngrok is a reverse proxy tool that creates a secure tunnel from a public endpoint to a locally running web service. It is widely used for testing local development versions of web applications, APIs, and bots, allowing them to be accessed over the internet. This is particularly useful for testing functionalities that require public accessibility, such as OAuth callbacks during authentication processes. By using ngrok, developers can simulate a production environment on their local machine, making it easier to test and debug features like sign-in flows that rely on external authentication services reaching the bot.

Comment 3

ID: 929236 User: Tin_Tin Badges: - Relative Date: 2 years, 8 months ago Absolute Date: Wed 21 Jun 2023 10:56 Selected Answer: - Upvotes: 1

The answer seems correct.
https://learn.microsoft.com/en-us/azure/bot-service/bot-service-debug-emulator?view=azure-bot-service-4.0&tabs=csharp

Comment 4

ID: 922885 User: 973b658 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Wed 14 Jun 2023 10:20 Selected Answer: - Upvotes: 1

No.
1st is OK.
2nd is OK.
3rd is "Use a sing-in verification code for OAuthCard"

Comment 4.1

ID: 929239 User: Tin_Tin Badges: - Relative Date: 2 years, 8 months ago Absolute Date: Wed 21 Jun 2023 10:58 Selected Answer: - Upvotes: 4

3rd should be enable "Use version 1.0 authentication tokens".
"Use a sing-in verification code for OAuthCard" is the setting for "Using a sign-in verification code"

99. AI-102 Topic 3 Question 39

Sequence
249
Discussion ID
110101
Source URL
https://www.examtopics.com/discussions/microsoft/view/110101-exam-ai-102-topic-3-question-39-discussion/
Posted By
sheldon73
Posted At
May 24, 2023, 10:39 a.m.

Question

You train a Conversational Language Understanding model to understand the natural language input of users.

You need to evaluate the accuracy of the model before deploying it.

What are two methods you can use? Each correct answer presents a complete solution.

NOTE: Each correct selection is worth one point.

  • A. From the language authoring REST endpoint, retrieve the model evaluation summary.
  • B. From Language Studio, enable Active Learning, and then validate the utterances logged for review.
  • C. From Language Studio, select Model performance.
  • D. From the Azure portal, enable log collection in Log Analytics, and then analyze the logs.

Suggested Answer

AC

Answer Description Click to expand


Community Answer Votes

Comments 13 comments Click to expand

Comment 1

ID: 920308 User: ziggy1117 Badges: Highly Voted Relative Date: 2 years, 9 months ago Absolute Date: Sun 11 Jun 2023 01:22 Selected Answer: AC Upvotes: 13

agree with SmallFire:
Active Learning cannot be initiated prior to the deployment of the model. The primary purpose of the 'Active Learning' feature is to leverage actual user interaction data to enhance the model's understanding capabilities. This is a continuous learning and optimization process that takes place after the model has been deployed and put into actual use.
so the answer is AC.

Comment 2

ID: 1265814 User: anto69 Badges: Most Recent Relative Date: 1 year, 6 months ago Absolute Date: Wed 14 Aug 2024 16:05 Selected Answer: AC Upvotes: 1

A-C
ChatGPT still think that B is better than A, but this time I'm not agree

Comment 3

ID: 1235328 User: reiwanotora Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 12:37 Selected Answer: AC Upvotes: 2

REST endpoint
select Model performance

Comment 4

ID: 1235195 User: HaraTadahisa Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 08:25 Selected Answer: AC Upvotes: 1

I say this answer is A and C.

Comment 5

ID: 1229223 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 12 Jun 2024 16:16 Selected Answer: AC Upvotes: 1

AC is answer.

Comment 6

ID: 1215796 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 22 May 2024 16:16 Selected Answer: AC Upvotes: 2

REST and model performance.

Comment 7

ID: 1133134 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 27 Jan 2024 08:35 Selected Answer: AC Upvotes: 3

A. From the language authoring REST endpoint, retrieve the model evaluation summary.
This summary typically includes metrics like precision, recall, and accuracy, which are crucial for evaluating the effectiveness of a language understanding model.
C. From Language Studio, select Model performance.
In Language Studio, the Model performance section typically provides detailed analytics about the model's performance, including various metrics and possibly confusion matrices.

Comment 8

ID: 1043622 User: sl_mslconsulting Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Sat 14 Oct 2023 18:52 Selected Answer: AC Upvotes: 1

If you look closely, you can see that the Model performance feature in Language Studio is about evaluating the performance of the model using test data as opposed to active learning which is using the real data from users’ interactions, which you couldn’t do prior to the deployment. My only pet peeve is that rest endpoint never works for me - always giving me the 404 no matter what.

Comment 9

ID: 1006120 User: jangotango Badges: - Relative Date: 2 years, 6 months ago Absolute Date: Wed 13 Sep 2023 03:25 Selected Answer: - Upvotes: 2

AC are correct - https://learn.microsoft.com/en-us/rest/api/language/2022-10-01-preview/text-analysis-authoring/get-model-evaluation-summary?tabs=HTTP

Comment 10

ID: 923293 User: EliteAllen Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Wed 14 Jun 2023 16:26 Selected Answer: BC Upvotes: 2

B. Active Learning in Language Studio is a feature that helps improve the performance of your model by suggesting utterances for you to review and label. This can help you evaluate the accuracy of your model by seeing how it performs on these suggested utterances.

C. The Model performance feature in Language Studio provides a detailed report on the performance of your model, including precision, recall, and F1 score. This can give you a good indication of the accuracy of your model.

Option A is incorrect because the language authoring REST endpoint does not provide a model evaluation summary.

Comment 10.1

ID: 986323 User: hawkzey Badges: - Relative Date: 2 years, 6 months ago Absolute Date: Mon 21 Aug 2023 10:29 Selected Answer: - Upvotes: 2

B is not correct as you cant not do active learning on a model that is not yet deployed

Comment 11

ID: 910814 User: SmallFire Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Wed 31 May 2023 07:12 Selected Answer: - Upvotes: 4

Active Learning cannot be initiated prior to the deployment of the model. The primary purpose of the 'Active Learning' feature is to leverage actual user interaction data to enhance the model's understanding capabilities. This is a continuous learning and optimization process that takes place after the model has been deployed and put into actual use.
so the answer is AC.

Comment 12

ID: 905651 User: sheldon73 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Wed 24 May 2023 10:39 Selected Answer: BC Upvotes: 1

Google Bard Answer : Sure, here are two methods you can use to evaluate the accuracy of a Conversational Language Understanding model before deploying it:

From Language Studio, select Model performance. This will show you a summary of the model's performance, including the F1 score, precision, and recall.
From Language Studio, enable Active Learning, and then validate the utterances logged for review. This will allow you to manually review utterances that the model has misclassified, and then retrain the model with the corrected data.
Here are the correct answers to your question:

C. From Language Studio, select Model performance.
B. From Language Studio, enable Active Learning, and then validate the utterances logged for review.

100. AI-102 Topic 5 Question 44

Sequence
250
Discussion ID
112148
Source URL
https://www.examtopics.com/discussions/microsoft/view/112148-exam-ai-102-topic-5-question-44-discussion/
Posted By
973b658
Posted At
June 14, 2023, 9:58 a.m.

Question

You have a Speech resource and a bot that was built by using the Microsoft Bot Framework Composer.

You need to add support for speech-based channels to the bot.

Which three actions should you perform? Each correct answer presents part of the solution.

NOTE: Each correct selection is worth one point.

  • A. Configure the language and voice settings for the Speech resource.
  • B. Add the endpoint and key of the Speech resource to the bot.
  • C. Add language understanding to dialogs.
  • D. Add Orchestrator to the bot.
  • E. Add Speech to the bot responses.
  • F. Remove the setSpeak configuration.

Suggested Answer

ABE

Answer Description Click to expand


Community Answer Votes

Comments 10 comments Click to expand

Comment 1

ID: 1265482 User: Shariq Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Wed 14 Aug 2024 04:44 Selected Answer: ABE Upvotes: 1

F is incorrect. Removing setSpeak configuration would not help in adding speech capabilities; in fact, it could hinder the bot's ability to produce speech responses effectively.

Comment 2

ID: 1229861 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 15:51 Selected Answer: ABE Upvotes: 1

We need to add support for speech-based channels to the bot.
So, Configure the language and voice settings, Add the endpoint and key of the Speech resource to the bot, Add Speech to the bot responses.
Simple. ABE is answer.

Comment 3

ID: 1214224 User: JamesKJoker Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Mon 20 May 2024 10:46 Selected Answer: - Upvotes: 2

A. Configure the language and voice settings for the Speech resource. You need to configure the Speech resource to match the language and voice settings of the bot.
B. Add the endpoint and key of the Speech resource to the bot. The bot needs to know where to connect to the Speech resource to process speech.
C. Add language understanding to dialogs. Language understanding allows the bot to understand the user's intent from spoken language.
Here's why the other options are incorrect:
D. Add Orchestrator to the bot. Orchestrator is a component of Azure Logic Apps and is not directly related to speech-based channels in a bot.
E. Add Speech to the bot responses. While the bot will respond with speech, you don't need to explicitly add speech to the bot responses. It's handled by the Speech resource.
F. Remove the setSpeak configuration. The setSpeak configuration is likely related to speech functionality and should not be removed.

Gemini Pro 1.5

Comment 4

ID: 1145888 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 10 Feb 2024 02:45 Selected Answer: - Upvotes: 3

A. Configure the language and voice settings for the Speech resource.
This step is crucial for ensuring that the Speech service can accurately recognize and synthesize speech in the desired language and voice.

B. Add the endpoint and key of the Speech resource to the bot.
To use the Azure Speech service with your bot, you need to authenticate your requests.

E. Add Speech to the bot responses.
To enhance the bot’s interactions over speech-based channels, you should include speech-specific responses.

Comment 5

ID: 1133299 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 27 Jan 2024 12:42 Selected Answer: ABE Upvotes: 1

A. Configure the language and voice settings for the Speech resource.
This step is important for ensuring that the Speech resource is set up to handle the specific languages and voice types that your bot will use.

B. Add the endpoint and key of the Speech resource to the bot.
This involves adding the endpoint URL and the key for your Speech resource to the bot's configuration.

E. Add Speech to the bot responses.
In the Bot Framework Composer, this typically involves specifying the speech output for each dialog or message.

Comment 6

ID: 1050177 User: sl_mslconsulting Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Sun 22 Oct 2023 05:31 Selected Answer: AEF Upvotes: 1

Check when you need to remove setSpeak https://learn.microsoft.com/en-us/composer/concept-speech?tabs=v2x#add-speech-components-to-your-bot-responses

Comment 6.1

ID: 1050178 User: sl_mslconsulting Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Sun 22 Oct 2023 05:33 Selected Answer: - Upvotes: 1

Also learn how to connect a bot to a channel : https://learn.microsoft.com/en-us/azure/bot-service/bot-service-channel-connect-directlinespeech?view=azure-bot-service-4.0

Comment 7

ID: 934984 User: Tin_Tin Badges: - Relative Date: 2 years, 8 months ago Absolute Date: Tue 27 Jun 2023 05:00 Selected Answer: ABE Upvotes: 1

should be ABE

Comment 8

ID: 934488 User: EliteAllen Badges: - Relative Date: 2 years, 8 months ago Absolute Date: Mon 26 Jun 2023 15:11 Selected Answer: ABE Upvotes: 3

A. Configure the language and voice settings for the Speech resource.
This is necessary to ensure that the speech services can correctly interpret and generate speech in the desired language and voice.

B. Add the endpoint and key of the Speech resource to the bot.
This allows the bot to use the Speech resource for speech-to-text and text-to-speech capabilities.

E. Add Speech to the bot responses.
This is necessary to enable the bot to generate spoken responses to user input.

Comment 9

ID: 922870 User: 973b658 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Wed 14 Jun 2023 09:58 Selected Answer: ABE Upvotes: 3

ABE is.

101. AI-102 Topic 5 Question 33

Sequence
251
Discussion ID
111870
Source URL
https://www.examtopics.com/discussions/microsoft/view/111870-exam-ai-102-topic-5-question-33-discussion/
Posted By
973b658
Posted At
June 11, 2023, 7:24 a.m.

Question

You have a Conversational Language Understanding model.

You export the model as a JSON file. The following is a sample of the file.

image

What represents the Weather.Historic entity in the sample utterance?

  • A. last year
  • B. by month
  • C. amount of
  • D. average

Suggested Answer

B

Answer Description Click to expand


Community Answer Votes

Comments 6 comments Click to expand

Comment 1

ID: 1265114 User: anto69 Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Tue 13 Aug 2024 12:56 Selected Answer: - Upvotes: 1

Repeated question: "by month" again

Comment 2

ID: 1229893 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 16:58 Selected Answer: B Upvotes: 2

redundancy

Comment 2.1

ID: 1244988 User: Toby86 Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Tue 09 Jul 2024 18:26 Selected Answer: - Upvotes: 3

I don't think the other question had "last year" as an option. Gotta be a bit more careful here since the correct answer "by month" and wrong answer "by year" are both time related and could be historical.

Comment 3

ID: 1193691 User: jv1 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Thu 11 Apr 2024 12:37 Selected Answer: - Upvotes: 3

The entity Weather.Historic in the sample utterance is represented by the string that starts at character position 23 and ends at character position 30. In the given text "average amount of rain by month in Chicago last year", the substring "last year" corresponds to those positions. Therefore, the entity Weather.Historic represents "last year".

Comment 4

ID: 979391 User: james2033 Badges: - Relative Date: 2 years, 7 months ago Absolute Date: Sat 12 Aug 2023 15:50 Selected Answer: B Upvotes: 3

"average amount of rain by month in Chicago last year" . The question has mistake, should start position - end position are 23 - 31.

Comment 5

ID: 920430 User: 973b658 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Sun 11 Jun 2023 07:24 Selected Answer: B Upvotes: 3

B.
Same #22Topic 3

102. AI-102 Topic 1 Question 8

Sequence
252
Discussion ID
54790
Source URL
https://www.examtopics.com/discussions/microsoft/view/54790-exam-ai-102-topic-1-question-8-discussion/
Posted By
motu
Posted At
June 7, 2021, 12:38 p.m.

Question

HOTSPOT -
You plan to deploy a containerized version of an Azure Cognitive Services service that will be used for text analysis.
You configure https://contoso.cognitiveservices.azure.com as the endpoint URI for the service, and you pull the latest version of the Text Analytics
Sentiment Analysis container.
You need to run the container on an Azure virtual machine by using Docker.
How should you complete the command? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Hot Area:
image

Suggested Answer

image
Answer Description Click to expand

Box 1: mcr.microsoft.com/azure-cognitive-services/textanalytics/sentiment
To run the Sentiment Analysis v3 container, execute the following docker run command. docker run --rm -it -p 5000:5000 --memory 8g --cpus 1 \ mcr.microsoft.com/azure-cognitive-services/textanalytics/sentiment \
Eula=accept \
Billing={ENDPOINT_URI} \
ApiKey={API_KEY} is the endpoint for accessing the Text Analytics API. https://.cognitiveservices.azure.com
Box 2: https://contoso.cognitiveservices.azure.com
{ENDPOINT_URI} is the endpoint for accessing the Text Analytics API: https://.cognitiveservices.a The endpoint for accessing the Text

Analytics API. zure.com -
Reference:
https://docs.microsoft.com/en-us/azure/cognitive-services/text-analytics/how-tos/text-analytics-how-to-install-containers?tabs=sentiment

Comments 6 comments Click to expand

Comment 1

ID: 389695 User: kolakone Badges: Highly Voted Relative Date: 3 years, 8 months ago Absolute Date: Fri 24 Jun 2022 16:42 Selected Answer: - Upvotes: 35

Seems correct.
For first answer, you "pull the latest version of the Text Analytics
Sentiment Analysis container", hence the "sentiment" in the endpoint.
For billing, other options are other services, hence the provided answer.

Comment 2

ID: 979132 User: james2033 Badges: Most Recent Relative Date: 1 year, 7 months ago Absolute Date: Mon 12 Aug 2024 07:13 Selected Answer: - Upvotes: 4

docker run --rm -it -p 5000:5000 --memory 8g --cpus 1 \
mcr.microsoft.com/azure-cognitive-services/textanalytics/sentiment \
Eula=accept \
Billing=http://contoso.cognitiveservices.azure.com \
ApiKey=xxxxxxxxxxxxxxxxx

See https://learn.microsoft.com/en-us/azure/ai-services/language-service/sentiment-opinion-mining/how-to/use-containers#run-the-container-with-docker-run

Comment 3

ID: 632413 User: AusAv Badges: - Relative Date: 2 years, 7 months ago Absolute Date: Mon 17 Jul 2023 02:11 Selected Answer: - Upvotes: 1

Example here: https://docs.microsoft.com/en-us/azure/cognitive-services/language-service/sentiment-opinion-mining/how-to/use-containers#run-the-container-with-docker-run

Comment 4

ID: 631336 User: Eltooth Badges: - Relative Date: 2 years, 8 months ago Absolute Date: Fri 14 Jul 2023 13:53 Selected Answer: - Upvotes: 1

mcr/.../sentiment
contoso.cognitiveservices

Comment 5

ID: 516117 User: sumanshu Badges: - Relative Date: 3 years, 2 months ago Absolute Date: Tue 03 Jan 2023 23:53 Selected Answer: - Upvotes: 2

As per question, we have to do the sentiment analysis, So URL should be appended with /sentiment and in Billing we have to mention the provided URL , So given answer is correct

Comment 6

ID: 376697 User: motu Badges: - Relative Date: 3 years, 9 months ago Absolute Date: Tue 07 Jun 2022 12:38 Selected Answer: - Upvotes: 1

https://docs.microsoft.com/en-us/azure/cognitive-services/anomaly-detector/anomaly-detector-container-howto

103. AI-102 Topic 5 Question 3

Sequence
253
Discussion ID
91978
Source URL
https://www.examtopics.com/discussions/microsoft/view/91978-exam-ai-102-topic-5-question-3-discussion/
Posted By
slcheng
Posted At
Dec. 18, 2022, 8:42 a.m.

Question

You are building a multilingual chatbot.
You need to send a different answer for positive and negative messages.
Which two Language service APIs should you use? Each correct answer presents part of the solution.
NOTE: Each correct selection is worth one point.

  • A. Linked entities from a well-known knowledge base
  • B. Sentiment Analysis
  • C. Key Phrases
  • D. Detect Language
  • E. Named Entity Recognition

Suggested Answer

BD

Answer Description Click to expand


Community Answer Votes

Comments 7 comments Click to expand

Comment 1

ID: 1057097 User: Lion007 Badges: Highly Voted Relative Date: 2 years, 4 months ago Absolute Date: Sun 29 Oct 2023 20:19 Selected Answer: BD Upvotes: 5

B. Sentiment Analysis <-- different answer for positive and negative messages
D. Detect Language <-- multilingual chatbot

Comment 2

ID: 1262553 User: anto69 Badges: Most Recent Relative Date: 1 year, 7 months ago Absolute Date: Thu 08 Aug 2024 16:59 Selected Answer: BD Upvotes: 1

BD is the correct answer without any doubt

Comment 3

ID: 1229904 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 17:04 Selected Answer: BD Upvotes: 1

BD is answer.

Comment 4

ID: 1222676 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sat 01 Jun 2024 16:01 Selected Answer: BD Upvotes: 1

B and D are correct answer.

Comment 5

ID: 771019 User: am20 Badges: - Relative Date: 3 years, 2 months ago Absolute Date: Tue 10 Jan 2023 04:10 Selected Answer: - Upvotes: 1

for sure B but not sure why D? although other options are not useful either!

Comment 5.1

ID: 813703 User: AzureJobsTillRetire Badges: - Relative Date: 3 years ago Absolute Date: Sun 19 Feb 2023 04:40 Selected Answer: - Upvotes: 6

You are building a "multilingual" chatbot, and that is why D (Detect Language) is correct

Comment 6

ID: 748670 User: slcheng Badges: - Relative Date: 3 years, 2 months ago Absolute Date: Sun 18 Dec 2022 08:42 Selected Answer: BD Upvotes: 2

Agreed with B & D

104. AI-102 Topic 5 Question 7

Sequence
254
Discussion ID
57089
Source URL
https://www.examtopics.com/discussions/microsoft/view/57089-exam-ai-102-topic-5-question-7-discussion/
Posted By
TheB
Posted At
July 4, 2021, 3:52 p.m.

Question

You build a conversational bot named bot1.
You need to configure the bot to use a QnA Maker application.
From the Azure Portal, where can you find the information required by bot1 to connect to the QnA Maker application?

  • A. Access control (IAM)
  • B. Properties
  • C. Keys and Endpoint
  • D. Identity

Suggested Answer

C

Answer Description Click to expand


Community Answer Votes

Comments 7 comments Click to expand

Comment 1

ID: 433137 User: olowoyinka Badges: Highly Voted Relative Date: 4 years, 6 months ago Absolute Date: Fri 27 Aug 2021 15:55 Selected Answer: - Upvotes: 13

Correct 👍

Comment 2

ID: 398383 User: TheB Badges: Highly Voted Relative Date: 4 years, 8 months ago Absolute Date: Sun 04 Jul 2021 15:52 Selected Answer: - Upvotes: 5

Given answer is correct!

Comment 3

ID: 1262554 User: anto69 Badges: Most Recent Relative Date: 1 year, 7 months ago Absolute Date: Thu 08 Aug 2024 17:00 Selected Answer: C Upvotes: 1

Should be C: keys and endpoints

Comment 4

ID: 1229901 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 17:02 Selected Answer: C Upvotes: 1

C is answer.

Comment 5

ID: 1222700 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sat 01 Jun 2024 16:47 Selected Answer: C Upvotes: 1

Will this question still be asked on June 2, 2024?

Comment 6

ID: 1139862 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sun 04 Feb 2024 08:22 Selected Answer: C Upvotes: 2

C. Keys and Endpoint

When configuring a conversational bot named bot1 to use the QnA Maker application, you need to obtain specific information from the QnA Maker service to establish the connection. This includes the service's endpoint address and the key used for authentication. You can find these details in the "Keys and Endpoint" section of the QnA Maker service in the Azure portal. Here, you will find the endpoint address and key for API calls, crucial for the bot to access and retrieve information from the QnA Maker service securely. These details enable your bot to interact safely with the QnA Maker service and retrieve answers based on your knowledge base content.

Comment 7

ID: 633599 User: Eltooth Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Tue 19 Jul 2022 15:30 Selected Answer: C Upvotes: 3

C is the correct answer : Keys and Endpoint.

105. AI-102 Topic 5 Question 5

Sequence
255
Discussion ID
56546
Source URL
https://www.examtopics.com/discussions/microsoft/view/56546-exam-ai-102-topic-5-question-5-discussion/
Posted By
azurelearner666
Posted At
June 30, 2021, 10:32 p.m.

Question

You are building a bot on a local computer by using the Microsoft Bot Framework. The bot will use an existing Language Understanding model.
You need to translate the Language Understanding model locally by using the Bot Framework CLI.
What should you do first?

  • A. From the Language Understanding portal, clone the model.
  • B. Export the model as an .lu file.
  • C. Create a new Speech service.
  • D. Create a new Language Understanding service.

Suggested Answer

B

Answer Description Click to expand


Community Answer Votes

Comments 7 comments Click to expand

Comment 1

ID: 1262225 User: moonlightc Badges: Highly Voted Relative Date: 1 year, 7 months ago Absolute Date: Wed 07 Aug 2024 22:50 Selected Answer: - Upvotes: 7

Save your time. Bot Framework is not part of AI-102 anymore

https://trainingsupport.microsoft.com/en-us/mcp/forum/all/ai-102-practice-exam-questions/05c54cce-cd0d-42ae-8be2-b76fb41c60fa

Comment 2

ID: 445146 User: Heegbaryee Badges: Highly Voted Relative Date: 4 years, 5 months ago Absolute Date: Wed 15 Sep 2021 12:35 Selected Answer: - Upvotes: 6

Correct

Comment 3

ID: 1229902 User: reigenchimpo Badges: Most Recent Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 17:03 Selected Answer: B Upvotes: 1

B is answer.

Comment 4

ID: 1222703 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sat 01 Jun 2024 16:55 Selected Answer: B Upvotes: 1

B is right answer.

Comment 5

ID: 1139857 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sun 04 Feb 2024 08:09 Selected Answer: B Upvotes: 6

"B. Export the model as an .lu file.

When you need to translate a Language Understanding (LUIS) model using the Microsoft Bot Framework CLI locally, the first step is to export the existing LUIS model as a .lu file. The .lu file format (Language Understanding file format) allows you to represent intents, entities, and other model components in a text-based form, enabling you to process, modify, or translate them using the Bot Framework CLI. This is a crucial step for working with and integrating LUIS models in a local environment."

Comment 6

ID: 652648 User: Nebary Badges: - Relative Date: 3 years, 6 months ago Absolute Date: Sat 27 Aug 2022 19:36 Selected Answer: B Upvotes: 1

Correct

Comment 7

ID: 395190 User: azurelearner666 Badges: - Relative Date: 4 years, 8 months ago Absolute Date: Wed 30 Jun 2021 22:32 Selected Answer: - Upvotes: 3

correct

106. AI-102 Topic 3 Question 15

Sequence
256
Discussion ID
75236
Source URL
https://www.examtopics.com/discussions/microsoft/view/75236-exam-ai-102-topic-3-question-15-discussion/
Posted By
PHD_CHENG
Posted At
May 6, 2022, 4:04 p.m.

Question

You have a chatbot that was built by using the Microsoft Bot Framework.
You need to debug the chatbot endpoint remotely.
Which two tools should you install on a local computer? Each correct answer presents part of the solution.
NOTE: Each correct selection is worth one point.

  • A. Fiddler
  • B. Bot Framework Composer
  • C. Bot Framework Emulator
  • D. Bot Framework CLI
  • E. ngrok
  • F. nginx

Suggested Answer

CE

Answer Description Click to expand


Community Answer Votes

Comments 6 comments Click to expand

Comment 1

ID: 1132996 User: evangelist Badges: Highly Voted Relative Date: 2 years, 1 month ago Absolute Date: Sat 27 Jan 2024 03:53 Selected Answer: CE Upvotes: 9

C. Bot Framework Emulator: This is an essential tool for debugging Microsoft Bot Framework bots. It allows you to test and debug your bots on your local machine by emulating the Bot Framework's channels and activities. It can be very helpful in a local development environment but is less suited for remote debugging.
E. ngrok: ngrok is a tool that creates a secure tunnel to your localhost. This is very useful for remote debugging because it allows you to expose your local development server to the internet, which is necessary for testing and debugging interactions with services like the Microsoft Bot Framework.

Comment 2

ID: 1261839 User: moonlightc Badges: Most Recent Relative Date: 1 year, 7 months ago Absolute Date: Tue 06 Aug 2024 23:58 Selected Answer: - Upvotes: 3

Bot framework is no longer part of the exam

Comment 3

ID: 1229238 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 12 Jun 2024 16:22 Selected Answer: CE Upvotes: 1

CE is answer.

Comment 4

ID: 1217559 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 16:08 Selected Answer: CE Upvotes: 1

Bot Framework Emulator takedajuku ngrok are right answer.

Comment 5

ID: 633354 User: Eltooth Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Tue 19 Jul 2022 07:00 Selected Answer: CE Upvotes: 2

C and E are correct answers.

Comment 6

ID: 597756 User: PHD_CHENG Badges: - Relative Date: 3 years, 10 months ago Absolute Date: Fri 06 May 2022 16:04 Selected Answer: CE Upvotes: 2

Answer is correct.
https://docs.microsoft.com/en-us/azure/bot-service/bot-service-debug-emulator?view=azure-bot-service-4.0&tabs=csharp

107. AI-102 Topic 5 Question 42

Sequence
258
Discussion ID
112146
Source URL
https://www.examtopics.com/discussions/microsoft/view/112146-exam-ai-102-topic-5-question-42-discussion/
Posted By
973b658
Posted At
June 14, 2023, 9:51 a.m.

Question

HOTSPOT
-

You have a chatbot.

You need to ensure that the bot conversation resets if a user fails to respond for 10 minutes.

How should you complete the code? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

image

Suggested Answer

image
Answer Description Click to expand


Comments 3 comments Click to expand

Comment 1

ID: 1133258 User: evangelist Badges: - Relative Date: 1 year, 7 months ago Absolute Date: Sat 27 Jul 2024 10:59 Selected Answer: - Upvotes: 4

In summary, send_activities is the standard method for sending messages to users. on_send_activities is not typically used for sending messages directly. send_trace_activity is for debugging, and update_activity is for modifying existing messages.

Comment 2

ID: 929134 User: Tin_Tin Badges: - Relative Date: 2 years, 2 months ago Absolute Date: Thu 21 Dec 2023 09:50 Selected Answer: - Upvotes: 2

The answer seems correct.
https://learn.microsoft.com/en-us/azure/bot-service/bot-builder-howto-expire-conversation?view=azure-bot-service-4.0&tabs=python

Comment 3

ID: 922865 User: 973b658 Badges: - Relative Date: 2 years, 2 months ago Absolute Date: Thu 14 Dec 2023 10:51 Selected Answer: - Upvotes: 2

It is true.

108. AI-102 Topic 5 Question 16

Sequence
260
Discussion ID
77634
Source URL
https://www.examtopics.com/discussions/microsoft/view/77634-exam-ai-102-topic-5-question-16-discussion/
Posted By
Eltooth
Posted At
July 20, 2022, 12:47 p.m.

Question

You need to enable speech capabilities for a chatbot.
Which three actions should you perform? Each correct answer presents part of the solution.
NOTE: Each correct selection is worth one point.

  • A. Enable WebSockets for the chatbot app.
  • B. Create a Speech service.
  • C. Register a Direct Line Speech channel.
  • D. Register a Cortana channel.
  • E. Enable CORS for the chatbot app.
  • F. Create a Language Understanding service.

Suggested Answer

ABC

Answer Description Click to expand


Community Answer Votes

Comments 9 comments Click to expand

Comment 1

ID: 633987 User: Eltooth Badges: Highly Voted Relative Date: 3 years, 7 months ago Absolute Date: Wed 20 Jul 2022 12:47 Selected Answer: ABC Upvotes: 14

A, B and C are correct answers in order shown below.

B. Create a Speech service
https://docs.microsoft.com/en-us/azure/cognitive-services/speech-service/tutorial-voice-enable-your-bot-speech-sdk#create-a-speech-service-resource

A. Enable WebSockets for the chatbot app
https://docs.microsoft.com/en-us/azure/cognitive-services/speech-service/tutorial-voice-enable-your-bot-speech-sdk#enable-web-sockets

C. Register a Direct Line Speech channel
https://docs.microsoft.com/en-us/azure/cognitive-services/speech-service/tutorial-voice-enable-your-bot-speech-sdk#register-the-direct-line-speech-channel

Comment 2

ID: 1248583 User: SAMBIT Badges: Most Recent Relative Date: 1 year, 7 months ago Absolute Date: Mon 15 Jul 2024 23:40 Selected Answer: - Upvotes: 1

Do create some to really get to the answer. A is not a choice. CORS is a must.

Comment 3

ID: 1229898 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 17:01 Selected Answer: ABC Upvotes: 2

ABC is answer. difficult.

Comment 4

ID: 1222682 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sat 01 Jun 2024 16:06 Selected Answer: ABC Upvotes: 1

ABC is justice! Sunshine Ikezaki!

Comment 5

ID: 1139894 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sun 04 Feb 2024 09:10 Selected Answer: ABC Upvotes: 1

Please do not choose BCE,

To enable speech capabilities for a chatbot, the correct actions include:

A. Enable WebSockets for the chatbot app: Necessary for the bot to communicate with the Direct Line Speech channel using web sockets.
B. Create a Speech service: Required to provide speech-to-text and text-to-speech capabilities for the chatbot.
C. Register a Direct Line Speech channel: Essential for connecting the chatbot with the Speech service to enable speech interactions.
For detailed steps, you can refer to the official Azure documentation: Tutorial: Voice-enable your bot.
https://learn.microsoft.com/en-us/azure/ai-services/speech-service/tutorial-voice-enable-your-bot-speech-sdk#enable-web-sockets

Comment 6

ID: 879400 User: Pffffff Badges: - Relative Date: 2 years, 10 months ago Absolute Date: Mon 24 Apr 2023 15:54 Selected Answer: ABC Upvotes: 3

said ChatGPT

Comment 7

ID: 840817 User: marti_tremblay000 Badges: - Relative Date: 2 years, 12 months ago Absolute Date: Thu 16 Mar 2023 11:37 Selected Answer: BCE Upvotes: 1

According to ChapGPT, the correct answers are BCE :
The three actions that need to be performed to enable speech capabilities for an Azure chatbot are:

B. Create a Speech service: A Speech service is needed to process the audio input and output of the chatbot. Azure provides the Speech service that can be created from the Azure portal.

C. Register a Direct Line Speech channel: A Direct Line Speech channel needs to be registered in the Azure portal for the chatbot app. The Direct Line Speech channel enables the chatbot to receive audio input and provide audio output to the user.

E. Enable CORS for the chatbot app: Cross-Origin Resource Sharing (CORS) needs to be enabled for the chatbot app to allow the browser to access resources from a different domain. This is required when using the Direct Line Speech channel.

Therefore, the correct actions are B, C, and E.

Comment 8

ID: 766960 User: Pyguy Badges: - Relative Date: 3 years, 2 months ago Absolute Date: Thu 05 Jan 2023 19:57 Selected Answer: BCE Upvotes: 2

instead of WebSockets you should enable CORS . Cross-Origin Resource Sharing , will allow your bot app to communicate with the Speech service.. Nothing to do websockets here..

Comment 8.1

ID: 816078 User: AzureJobsTillRetire Badges: - Relative Date: 3 years ago Absolute Date: Tue 21 Feb 2023 02:13 Selected Answer: - Upvotes: 2

Can you please share your reference?

Eltooth has provided link as below.
Enable web sockets
You need to make a small configuration change so that your bot can communicate with the Direct Line Speech channel by using web sockets.
https://docs.microsoft.com/en-us/azure/cognitive-services/speech-service/tutorial-voice-enable-your-bot-speech-sdk#enable-web-sockets

109. AI-102 Topic 3 Question 16

Sequence
262
Discussion ID
77618
Source URL
https://www.examtopics.com/discussions/microsoft/view/77618-exam-ai-102-topic-3-question-16-discussion/
Posted By
Eltooth
Posted At
July 19, 2022, 7:10 a.m.

Question

DRAG DROP -
You are building a retail chatbot that will use a QnA Maker service.
You upload an internal support document to train the model. The document contains the following question: "What is your warranty period?"
Users report that the chatbot returns the default QnA Maker answer when they ask the following question: "How long is the warranty coverage?"
The chatbot returns the correct answer when the users ask the following question: 'What is your warranty period?"
Both questions should return the same answer.
You need to increase the accuracy of the chatbot responses.
Which three actions should you perform in sequence? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.
Select and Place:
image

Suggested Answer

image
Answer Description Click to expand

Step 1: Add alternative phrasing to the question and answer (QnA) pair.
Add alternate questions to an existing QnA pair to improve the likelihood of a match to a user query.
Step 2: Retrain the model.
Periodically select Save and train after making edits to avoid losing changes.

Step 3: Republish the model -
Note: A knowledge base consists of question and answer (QnA) pairs. Each pair has one answer and a pair contains all the information associated with that answer.
Reference:
https://docs.microsoft.com/en-us/azure/cognitive-services/qnamaker/how-to/edit-knowledge-base

Comments 7 comments Click to expand

Comment 1

ID: 1133004 User: evangelist Badges: Highly Voted Relative Date: 2 years, 1 month ago Absolute Date: Sat 27 Jan 2024 03:59 Selected Answer: - Upvotes: 5

1: Add alternate phrasings as follow-up prompts or as additional questions in the QnA pair:
2: Train the QnA Maker model:
3: Publish the updated QnA Maker model:

Comment 2

ID: 1248503 User: krzkrzkra Badges: Most Recent Relative Date: 1 year, 7 months ago Absolute Date: Mon 15 Jul 2024 20:03 Selected Answer: - Upvotes: 2

1. Add alternative phrasing to the QnA pair.
2. Retrain model.
3. Republish model.

Comment 3

ID: 1217542 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 15:59 Selected Answer: - Upvotes: 2

Add alternative QnA
Retrain
Republish

Comment 4

ID: 1156932 User: anto69 Badges: - Relative Date: 2 years ago Absolute Date: Fri 23 Feb 2024 06:49 Selected Answer: - Upvotes: 1

correct answer

Comment 5

ID: 1102032 User: Gvalli Badges: - Relative Date: 2 years, 2 months ago Absolute Date: Thu 21 Dec 2023 00:29 Selected Answer: - Upvotes: 4

A modified version of this was in the exam today.

Comment 6

ID: 635757 User: RamonKaus Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Sun 24 Jul 2022 00:08 Selected Answer: - Upvotes: 4

Correct

Comment 7

ID: 633359 User: Eltooth Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Tue 19 Jul 2022 07:10 Selected Answer: - Upvotes: 4

Answer looks correct.

Add alternative phrasing
Retrain
Republish

110. AI-102 Topic 3 Question 59

Sequence
263
Discussion ID
122496
Source URL
https://www.examtopics.com/discussions/microsoft/view/122496-exam-ai-102-topic-3-question-59-discussion/
Posted By
SCyrus
Posted At
Oct. 5, 2023, 11:49 a.m.

Question

HOTSPOT
-

You are developing a streaming Speech to Text solution that will use the Speech SDK and MP3 encoding.

You need to develop a method to convert speech to text for streaming MP3 data.

How should you complete the code? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

image

Suggested Answer

image
Answer Description Click to expand


Comments 8 comments Click to expand

Comment 1

ID: 1139676 User: evangelist Badges: Highly Voted Relative Date: 2 years, 1 month ago Absolute Date: Sun 04 Feb 2024 01:06 Selected Answer: - Upvotes: 15

first one has to configure the format as MP3, so "AudioStreamFormat" option is chosen. Second, since it is speech to text, the SpeechRecognition option is needed.
The answer is CORRECT

Comment 2

ID: 1247891 User: krzkrzkra Badges: Most Recent Relative Date: 1 year, 8 months ago Absolute Date: Sun 14 Jul 2024 19:52 Selected Answer: - Upvotes: 4

1. AudioStreamFormat
2. SpeechRecognizer

Comment 3

ID: 1235648 User: JacobZ Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sun 23 Jun 2024 01:44 Selected Answer: - Upvotes: 1

Same as question 18 in topic 1

Comment 4

ID: 1225499 User: NagaoShingo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 06 Jun 2024 14:43 Selected Answer: - Upvotes: 2

1. AudioStreamFormat
2. SpeechRecognizer

Comment 5

ID: 1215780 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 22 May 2024 15:52 Selected Answer: - Upvotes: 2

audio_format = speechsdk.audio.AudioStreamFormat(compressed_stream_format=speechsdk.AudioStreamContainerFormat.MP3)
recognizer = speechsdk.SpeechRecognizer(speech_config, audio_config)

Comment 6

ID: 1117434 User: hkbnjos Badges: - Relative Date: 2 years, 2 months ago Absolute Date: Tue 09 Jan 2024 12:31 Selected Answer: - Upvotes: 1

this question is same as Topic 1 Q18 but with different answer, which is correct?

Comment 6.1

ID: 1170723 User: Mehe323 Badges: - Relative Date: 2 years ago Absolute Date: Mon 11 Mar 2024 02:21 Selected Answer: - Upvotes: 3

The answers are the same. Both chosen are AudioStreamFormat and SpeechRecognizer, but in question 18 there is an additional method for AudioStreamFormat (GetCompressedFormat).

Comment 7

ID: 1025510 User: SCyrus Badges: - Relative Date: 2 years, 5 months ago Absolute Date: Thu 05 Oct 2023 11:49 Selected Answer: - Upvotes: 4

Answer correct
https://github.com/Azure-Samples/cognitive-services-speech-sdk/blob/master/samples/python/console/speech_sample.py

111. AI-102 Topic 3 Question 67

Sequence
264
Discussion ID
135017
Source URL
https://www.examtopics.com/discussions/microsoft/view/135017-exam-ai-102-topic-3-question-67-discussion/
Posted By
Harry300
Posted At
March 1, 2024, 7:16 p.m.

Question

HOTSPOT -

You are developing a text processing solution.

You develop the following method.

image

You call the method by using the following code.

get_key_phrases(text_analytics_client, "the cat sat on the mat")

For each of the following statements, select Yes if the statement is true. Otherwise, select No.

NOTE: Each correct selection is worth one point.

image

Suggested Answer

image
Answer Description Click to expand


Comments 7 comments Click to expand

Comment 1

ID: 1163868 User: audlindr Badges: Highly Voted Relative Date: 2 years ago Absolute Date: Sat 02 Mar 2024 03:26 Selected Answer: - Upvotes: 20

Should be YNN

Refer: https://learn.microsoft.com/en-us/azure/ai-services/language-service/key-phrase-extraction/quickstart?pivots=programming-language-csharp

Comment 2

ID: 1215757 User: takaimomoGcup Badges: Highly Voted Relative Date: 1 year, 9 months ago Absolute Date: Wed 22 May 2024 15:27 Selected Answer: - Upvotes: 6

YNN is right answer.

Comment 3

ID: 1247893 User: krzkrzkra Badges: Most Recent Relative Date: 1 year, 8 months ago Absolute Date: Sun 14 Jul 2024 19:54 Selected Answer: - Upvotes: 5

Should be YNN

Comment 4

ID: 1184115 User: varinder82 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Wed 27 Mar 2024 14:10 Selected Answer: - Upvotes: 4

Final Answer :
YNN

Comment 5

ID: 1183548 User: f2c587e Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Tue 26 Mar 2024 20:13 Selected Answer: - Upvotes: 2

No se evidencia en el codigo que se muestre la confiabilidad de ninguna manera. YNN.

Comment 6

ID: 1179860 User: anto69 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Fri 22 Mar 2024 06:25 Selected Answer: - Upvotes: 2

Y-N-N for sure

Comment 7

ID: 1163667 User: Harry300 Badges: - Relative Date: 2 years ago Absolute Date: Fri 01 Mar 2024 19:16 Selected Answer: - Upvotes: 3

Should by YNN. Seriously, where should the confidence level come from. keyphase is a string

112. AI-102 Topic 4 Question 8

Sequence
266
Discussion ID
60582
Source URL
https://www.examtopics.com/discussions/microsoft/view/60582-exam-ai-102-topic-4-question-8-discussion/
Posted By
SuperPetey
Posted At
Aug. 25, 2021, 8:21 a.m.

Question

You are developing a solution to generate a word cloud based on the reviews of a company's products.
Which Text Analytics REST API endpoint should you use?

  • A. keyPhrases
  • B. sentiment
  • C. languages
  • D. entities/recognition/general

Suggested Answer

A

Answer Description Click to expand


Community Answer Votes

Comments 18 comments Click to expand

Comment 1

ID: 473572 User: Mll1975 Badges: Highly Voted Relative Date: 4 years, 4 months ago Absolute Date: Sat 06 Nov 2021 18:51 Selected Answer: - Upvotes: 23

I also thought it was B, but now I think given answer is correct:
Extracted from here
https://docs.microsoft.com/en-us/azure/cognitive-services/language-service/key-phrase-extraction/tutorials/integrate-power-bi#create-the-word-cloud
The key phrases provide us with the important words from our customer comments, not just the most common words. Also, word sizing in the resulting cloud isn't skewed by the frequent use of a word in a relatively small number of comments.

Comment 2

ID: 538142 User: TamHas Badges: Highly Voted Relative Date: 4 years, 1 month ago Absolute Date: Tue 01 Feb 2022 19:48 Selected Answer: A Upvotes: 8

correct

Comment 3

ID: 1247903 User: krzkrzkra Badges: Most Recent Relative Date: 1 year, 8 months ago Absolute Date: Sun 14 Jul 2024 20:00 Selected Answer: A Upvotes: 1

Selected Answer: A

Comment 4

ID: 1229845 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 15:33 Selected Answer: A Upvotes: 1

In my opinion, this is only option A.

Comment 5

ID: 1218707 User: reiwanotora Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sun 26 May 2024 05:23 Selected Answer: A Upvotes: 1

It must be A.

Comment 6

ID: 1143848 User: PCRamirez Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Wed 07 Feb 2024 22:56 Selected Answer: - Upvotes: 1

According to Windows Copilot:
The best choice for this scenario is **A. keyPhrases**. This endpoint extracts the main points or topics from a text document, which can be used to generate a word cloud that shows the most frequent or relevant terms in the reviews. The other endpoints are not suitable for this task because:
- B. sentiment analyzes the positive, negative, or neutral tone of a text document, which is not relevant for generating a word cloud.
- C. languages detects the language of a text document, which is not relevant for generating a word cloud.
- D. entities/recognition/general identifies and categorizes named entities in a text document, such as person names, locations, organizations, etc. This might not capture the main points or topics of the reviews, and might include irrelevant or sensitive information in the word cloud.

Comment 7

ID: 1075977 User: sismer Badges: - Relative Date: 2 years, 3 months ago Absolute Date: Tue 21 Nov 2023 03:16 Selected Answer: B Upvotes: 1

I think the correct answer is B:

Sentiment Analysis and opinion mining:
Sentiment analysis and opinion mining are preconfigured features that help you find out what people think of your brand or topic by mining text for clues about positive or negative sentiment, and can associate them with specific aspects of the text.

https://learn.microsoft.com/en-us/azure/ai-services/language-service/overview#sentiment-analysis-and-opinion-mining

Comment 7.1

ID: 1085198 User: tdctdc Badges: - Relative Date: 2 years, 3 months ago Absolute Date: Fri 01 Dec 2023 13:18 Selected Answer: - Upvotes: 2

Word Cloud has nothing to do with sentiment analysis.

Comment 8

ID: 936206 User: Pixelmate Badges: - Relative Date: 2 years, 8 months ago Absolute Date: Wed 28 Jun 2023 07:56 Selected Answer: - Upvotes: 3

This was on exam 28/06

Comment 9

ID: 912571 User: kail85 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Fri 02 Jun 2023 08:07 Selected Answer: - Upvotes: 1

A. keyPhrases
The keyPhrases endpoint extracts key phrases from the given text, which can be useful for creating a word cloud. Key phrases represent the main points or topics discussed in the text, making them ideal for visualizing the most prominent themes in the reviews.

Comment 10

ID: 748653 User: slcheng Badges: - Relative Date: 3 years, 2 months ago Absolute Date: Sun 18 Dec 2022 08:17 Selected Answer: A Upvotes: 1

Agreed with A

Comment 11

ID: 643859 User: ExamGuruBhai Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Sun 07 Aug 2022 23:01 Selected Answer: A Upvotes: 1

correct

Comment 12

ID: 633463 User: Eltooth Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Tue 19 Jul 2022 11:15 Selected Answer: A Upvotes: 1

A is correct answer.

Comment 13

ID: 600345 User: ManAtWorkAtNight Badges: - Relative Date: 3 years, 10 months ago Absolute Date: Thu 12 May 2022 00:39 Selected Answer: - Upvotes: 3

Answer is A. keyPhrases
Word Cloud is a set of most frequently appeared words(actually an image)
https://docs.microsoft.com/en-us/rest/api/cognitiveservices-textanalytics/3.0/key-phrases/key-phrases

Comment 14

ID: 598921 User: PHD_CHENG Badges: - Relative Date: 3 years, 10 months ago Absolute Date: Mon 09 May 2022 09:22 Selected Answer: A Upvotes: 2

I support "A" as answer as question is asking to develop a "Word cloud based" solution by reviewing of Company's product. (initially, I was thinking about "B" as an answer but I changed my mind.)

Comment 15

ID: 535842 User: prabhjot Badges: - Relative Date: 4 years, 1 month ago Absolute Date: Sun 30 Jan 2022 01:24 Selected Answer: - Upvotes: 1

Sentiment is correct one

Comment 16

ID: 431187 User: SuperPetey Badges: - Relative Date: 4 years, 6 months ago Absolute Date: Wed 25 Aug 2021 08:21 Selected Answer: - Upvotes: 3

Correct Answer should be "B":sentiment analysis because it is analyzing customer reviews. See this blog post from MSFT that uses sentiment analysis to create a word cloud application for the same purpose: https://devblogs.microsoft.com/premier-developer/creating-words-cloud-for-sentiment-analysis-with-azure-cognitive-services-text-analytics/

Comment 16.1

ID: 432552 User: astralyt Badges: - Relative Date: 4 years, 6 months ago Absolute Date: Thu 26 Aug 2021 22:14 Selected Answer: - Upvotes: 21

It doesn't say analyzing it. Word clouds comprise of key phrases. So I think the answer is right.

113. AI-102 Topic 5 Question 12

Sequence
268
Discussion ID
62325
Source URL
https://www.examtopics.com/discussions/microsoft/view/62325-exam-ai-102-topic-5-question-12-discussion/
Posted By
Diem
Posted At
Sept. 18, 2021, 2:21 p.m.

Question

HOTSPOT -
You are building a bot and that will use Language Understanding.
You have a LUDown file that contains the following content.
image
Use the drop-down menus to select the answer choice that completes each statement based on the information presented in the graphic.
NOTE: Each correct selection is worth one point.
Hot Area:
image

Suggested Answer

image
Answer Description Click to expand

Reference:
https://github.com/solliancenet/tech-immersion-data-ai/blob/master/ai-exp1/README.md

Comments 7 comments Click to expand

Comment 1

ID: 463294 User: Zoul Badges: Highly Voted Relative Date: 4 years, 4 months ago Absolute Date: Sat 16 Oct 2021 23:39 Selected Answer: - Upvotes: 23

Answer is correct. From the link provided:
SelectItem is the intent, and each item below it are example utterances that capture ways users can express this intent. Entities in .lu files are denoted using {<entityName>=<labeled value>} notation. Taking from our sample code once again, you can find the bottom left entity within the following utterance: choose the {DirectionalReference=bottom left}.

Comment 2

ID: 447066 User: Diem Badges: Highly Voted Relative Date: 4 years, 5 months ago Absolute Date: Sat 18 Sep 2021 14:21 Selected Answer: - Upvotes: 7

Based on the reference, it should be entity for the second question

Comment 3

ID: 1247924 User: krzkrzkra Badges: Most Recent Relative Date: 1 year, 8 months ago Absolute Date: Sun 14 Jul 2024 20:19 Selected Answer: - Upvotes: 3

1. intent
2. utterance

Comment 4

ID: 637908 User: ParkXD Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Wed 27 Jul 2022 09:51 Selected Answer: - Upvotes: 1

Answer is correct.
The LUIS model begins with categories of user intentions called intents. Each intent needs examples of user utterances (something a user should say). Each utterance can provide a variety of data that needs to be extracted with entities.

Comment 5

ID: 633609 User: Eltooth Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Tue 19 Jul 2022 15:53 Selected Answer: - Upvotes: 2

Intent
Utterance.

Comment 6

ID: 620250 User: sdokmak Badges: - Relative Date: 3 years, 8 months ago Absolute Date: Wed 22 Jun 2022 10:49 Selected Answer: - Upvotes: 2

a trick stolen from previous questions
(U) (S) A, Utterance Statement
(E)(n)gland, Entity Noun
(I)(v)ory Coast, Intent Verb

Comment 7

ID: 599392 User: PHD_CHENG Badges: - Relative Date: 3 years, 10 months ago Absolute Date: Tue 10 May 2022 05:41 Selected Answer: - Upvotes: 1

Answer is correct

114. AI-102 Topic 5 Question 17

Sequence
269
Discussion ID
80793
Source URL
https://www.examtopics.com/discussions/microsoft/view/80793-exam-ai-102-topic-5-question-17-discussion/
Posted By
Davard
Posted At
Sept. 7, 2022, 4:36 a.m.

Question

You use the Microsoft Bot Framework Composer to build a chatbot that enables users to purchase items.
You need to ensure that the users can cancel in-progress transactions. The solution must minimize development effort.
What should you add to the bot?

  • A. a language generator
  • B. a custom event
  • C. a dialog trigger
  • D. a conversation activity

Suggested Answer

C

Answer Description Click to expand


Community Answer Votes

Comments 12 comments Click to expand

Comment 1

ID: 661871 User: Davard Badges: Highly Voted Relative Date: 3 years, 6 months ago Absolute Date: Wed 07 Sep 2022 04:36 Selected Answer: - Upvotes: 15

It seems like it should be C: a dialog trigger. Anyone else?

Comment 1.1

ID: 682667 User: Tickxit Badges: - Relative Date: 3 years, 5 months ago Absolute Date: Thu 29 Sep 2022 14:23 Selected Answer: - Upvotes: 6

I agree, an interruption occurs when a trigger is fired. Context of adaptive dialogs.

Comment 2

ID: 715845 User: ArchMelody Badges: Highly Voted Relative Date: 3 years, 4 months ago Absolute Date: Fri 11 Nov 2022 08:57 Selected Answer: C Upvotes: 5

C seems to be the correct answer for me as well for the aforementioned reasons.

Comment 3

ID: 1247926 User: krzkrzkra Badges: Most Recent Relative Date: 1 year, 8 months ago Absolute Date: Sun 14 Jul 2024 20:21 Selected Answer: C Upvotes: 1

Selected Answer: C

Comment 4

ID: 1229896 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 17:00 Selected Answer: C Upvotes: 1

C is answer.

Comment 5

ID: 1222680 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sat 01 Jun 2024 16:05 Selected Answer: C Upvotes: 1

I often make mistakes. This is C!

Comment 6

ID: 1139895 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sun 04 Feb 2024 09:12 Selected Answer: C Upvotes: 2

The answer is C:
To ensure that users can cancel in-progress transactions with minimal development effort, you should add C. a dialog trigger to the bot. A dialog trigger allows the bot to respond to specific conditions or intents, such as a user's request to cancel a transaction, by invoking predefined dialogs that handle these scenarios efficiently.

Comment 7

ID: 1053333 User: propanther Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Wed 25 Oct 2023 02:37 Selected Answer: C Upvotes: 3

You should use dialog triggers to:
Take actions immediately when the dialog starts, even before the recognizer is called.
Take actions when a "cancel" signal is detected.
Take actions on messages received or sent.
Evaluate the content of the incoming activity.

https://learn.microsoft.com/en-us/composer/concept-events-and-triggers?tabs=v2x#dialog-events-triggers

Comment 8

ID: 1049799 User: sl_mslconsulting Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Sat 21 Oct 2023 22:15 Selected Answer: D Upvotes: 2

Tiggers are not a good fit for handling scenarios like these let along there are all 4 types of triggers for dialogs. Follow the link provided by the answer and read the code carefully and you will understand why it’s D.

Comment 8.1

ID: 1049801 User: sl_mslconsulting Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Sat 21 Oct 2023 22:16 Selected Answer: - Upvotes: 2

*only 4 types… Dialog started (Begin dialog event)
Dialog cancelled (Cancel dialog event)
Error occurred (Error event)
Re-prompt for input (Reprompt dialog event). How are they appropriate in handling scenarios like this?

Comment 9

ID: 934460 User: EliteAllen Badges: - Relative Date: 2 years, 8 months ago Absolute Date: Mon 26 Jun 2023 14:50 Selected Answer: C Upvotes: 3

A dialog trigger in Bot Framework Composer allows you to define the conditions under which a specific dialog is started. For example, you can create a dialog trigger that listens for the user to say something like "cancel" or "stop", and then initiate a dialog that handles the cancellation of the in-progress transaction.

This approach doesn't require you to manually code the cancellation logic, and it leverages the built-in capabilities of the Bot Framework Composer, thus minimizing development effort.

Comment 10

ID: 698833 User: DiegoGonL Badges: - Relative Date: 3 years, 4 months ago Absolute Date: Wed 19 Oct 2022 10:42 Selected Answer: C Upvotes: 3

As Tickxit said: an interruption occurs when a trigger is fired in the context of adaptive dialogs.

115. AI-102 Topic 5 Question 24

Sequence
270
Discussion ID
102798
Source URL
https://www.examtopics.com/discussions/microsoft/view/102798-exam-ai-102-topic-5-question-24-discussion/
Posted By
marti_tremblay000
Posted At
March 16, 2023, 11:40 a.m.

Question

You create a bot by using the Microsoft Bot Framework SDK.

You need to configure the bot to respond to events by using custom text responses.

What should you use?

  • A. a dialog
  • B. an activity handler
  • C. an adaptive card
  • D. a skill

Suggested Answer

B

Answer Description Click to expand


Community Answer Votes

Comments 6 comments Click to expand

Comment 1

ID: 840818 User: marti_tremblay000 Badges: Highly Voted Relative Date: 2 years, 12 months ago Absolute Date: Thu 16 Mar 2023 11:40 Selected Answer: B Upvotes: 9

B is the correct answer as explained by ChatGPT :
B. An activity handler is the correct choice for configuring the bot to respond to events by using custom text responses.

An activity handler is a class in the Bot Framework SDK that processes incoming activities (e.g., messages, events, etc.) from the user and generates outgoing activities (e.g., replies). By overriding the OnMessageActivityAsync method of the activity handler, you can provide custom logic for responding to user messages.

Comment 2

ID: 1247929 User: krzkrzkra Badges: Most Recent Relative Date: 1 year, 8 months ago Absolute Date: Sun 14 Jul 2024 20:22 Selected Answer: B Upvotes: 1

Selected Answer: B

Comment 3

ID: 1234240 User: cxw23 Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Fri 21 Jun 2024 09:24 Selected Answer: - Upvotes: 3

Azure Bot Service has been removed from the exam since oct 2023.

Comment 4

ID: 1229894 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 16:59 Selected Answer: B Upvotes: 1

B is answer.

Comment 5

ID: 1139934 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sun 04 Feb 2024 10:23 Selected Answer: B Upvotes: 1

To configure the bot to respond to events using custom text responses, you should use B. an activity handler. Activity handlers allow you to manage and respond to different types of activities and events in a bot's conversation, such as messages from users, with customized logic and responses.

Comment 6

ID: 937832 User: Tin_Tin Badges: - Relative Date: 2 years, 8 months ago Absolute Date: Thu 29 Jun 2023 11:13 Selected Answer: B Upvotes: 3

correct
https://learn.microsoft.com/en-us/azure/bot-service/bot-activity-handler-concept?view=azure-bot-service-4.0&tabs=csharp

116. AI-102 Topic 5 Question 39

Sequence
271
Discussion ID
106151
Source URL
https://www.examtopics.com/discussions/microsoft/view/106151-exam-ai-102-topic-5-question-39-discussion/
Posted By
tzuyichao
Posted At
April 14, 2023, 2:50 a.m.

Question

HOTSPOT
-

You are building a chatbot.

You need to use the Content Moderator service to identify messages that contain sexually explicit language.

Which section in the response from the service will contain the category score, and which category will be assigned to the message? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

image

Suggested Answer

image
Answer Description Click to expand


Comments 5 comments Click to expand

Comment 1

ID: 1145884 User: evangelist Badges: Highly Voted Relative Date: 2 years, 1 month ago Absolute Date: Sat 10 Feb 2024 02:31 Selected Answer: - Upvotes: 8

In the Content Moderator service response, the classification section can include different category scores that represent various types of content. Here's a summary of what each category generally signifies:

Category 1: Typically represents sexually explicit content. This includes images, text, or videos that contain nudity or sexual activities.

Category 2: Usually denotes sexually suggestive content. While not explicit, this content has sexual undertones or risqué material that might not be suitable for all audiences.

Category 3: Often related to offensive content. This can cover a broad range of materials, from language that might be considered profane or derogatory, to content that includes hate speech, violence, or other types of content that could be offensive or disturbing to some viewers.

Comment 2

ID: 869867 User: tzuyichao Badges: Highly Voted Relative Date: 2 years, 11 months ago Absolute Date: Fri 14 Apr 2023 02:50 Selected Answer: - Upvotes: 6

Correct

Ref: https://learn.microsoft.com/en-us/azure/cognitive-services/content-moderator/text-moderation-api#classification

Comment 3

ID: 1247930 User: krzkrzkra Badges: Most Recent Relative Date: 1 year, 8 months ago Absolute Date: Sun 14 Jul 2024 20:24 Selected Answer: - Upvotes: 1

1. Classification
2. 1

Comment 4

ID: 1133178 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 27 Jan 2024 09:45 Selected Answer: - Upvotes: 4

Category1 refers to potential presence of language that may be considered sexually explicit or adult in certain situations.
Category2 refers to potential presence of language that may be considered sexually suggestive or mature in certain situations.
Category3 refers to potential presence of language that may be considered offensive in certain situations.

Comment 5

ID: 1102010 User: Gvalli Badges: - Relative Date: 2 years, 2 months ago Absolute Date: Thu 21 Dec 2023 00:11 Selected Answer: - Upvotes: 4

This was in the exam today.

117. AI-102 Topic 5 Question 46

Sequence
272
Discussion ID
112150
Source URL
https://www.examtopics.com/discussions/microsoft/view/112150-exam-ai-102-topic-5-question-46-discussion/
Posted By
973b658
Posted At
June 14, 2023, 10:13 a.m.

Question

HOTSPOT
-

You have a bot that was built by using the Microsoft Bot Framework composer as shown in the following exhibit.

image

Use the drop-down menus to select the answer choice that completes each statement based on the information presented in the graphic.

NOTE: Each correct selection is worth one point.

image

Suggested Answer

image
Answer Description Click to expand


Comments 6 comments Click to expand

Comment 1

ID: 1247933 User: krzkrzkra Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sun 14 Jul 2024 20:26 Selected Answer: - Upvotes: 1

1. identity New York as city entity
2. Language Understanding intent recognised

Comment 2

ID: 1225525 User: NagaoShingo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 06 Jun 2024 15:10 Selected Answer: - Upvotes: 2

1. identity New York as city entity
2. Language Understanding intent recognized

Comment 3

ID: 1145891 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 10 Feb 2024 02:54 Selected Answer: - Upvotes: 1

given answer is correct

Comment 4

ID: 1133305 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 27 Jan 2024 12:47 Selected Answer: - Upvotes: 1

Connect with the Bot Framework Emulator:

Use the Bot Framework Emulator to connect to your bot running on localhost. The default URL is typically http://localhost:3978/api/messages.
Test Your Bot: Interact with your bot via the Emulator to see the responses and the trace activities.

Comment 5

ID: 922877 User: 973b658 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Wed 14 Jun 2023 10:13 Selected Answer: - Upvotes: 1

It is true.

Comment 5.1

ID: 1235855 User: anjanc Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sun 23 Jun 2024 16:44 Selected Answer: - Upvotes: 1

again he wrote!

118. AI-102 Topic 3 Question 29

Sequence
273
Discussion ID
83483
Source URL
https://www.examtopics.com/discussions/microsoft/view/83483-exam-ai-102-topic-3-question-29-discussion/
Posted By
Davard
Posted At
Sept. 25, 2022, 2:05 a.m.

Question

HOTSPOT -
You are developing an application that includes language translation.
The application will translate text retrieved by using a function named get_text_to_be_translated. The text can be in one of many languages. The content of the text must remain within the Americas Azure geography.
You need to develop code to translate the text to a single language.
How should you complete the code? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Hot Area:
image

Suggested Answer

image
Answer Description Click to expand

Box 1: ("api-nam.cognitive.microsofttranslator.com")
Geography USA: api-nam.cognitive.microsofttranslator.com
Datacenters: East US, South Central US, West Central US, and West US 2
Box 2: "/translate?to=en"
Must specify the language which it is being translated to. The 'to' parameter is required
Reference:
https://docs.microsoft.com/en-us/azure/cognitive-services/translator/reference/v3-0-reference https://docs.microsoft.com/en-us/azure/cognitive-services/translator/reference/v3-0-translate

Comments 7 comments Click to expand

Comment 1

ID: 1247876 User: krzkrzkra Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sun 14 Jul 2024 19:38 Selected Answer: - Upvotes: 1

1.api-nam.cognitive.microsofttranslator.com
2. /translate?to=en

Comment 2

ID: 1217524 User: reiwanotora Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 15:48 Selected Answer: - Upvotes: 2

api-nam.cognitive.microsofttranslator.com
translate?to=en

Comment 3

ID: 1130249 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Wed 24 Jan 2024 06:31 Selected Answer: - Upvotes: 4

"api-apc.cognitive.microsofttranslator.com" refers to the Asia Pacific endpoint. Using this endpoint will route requests to Azure data centers located in Asia Pacific for processing.

"api-nam.cognitive.microsofttranslator.com" refers to the North America endpoint. Using this endpoint will route requests to Azure data centers located in North America for processing.

So in summary:

"api-apc.cognitive.microsofttranslator.com" - Asia Pacific endpoint
"api-nam.cognitive.microsofttranslator.com" - North America endpoint

Comment 4

ID: 1102033 User: Gvalli Badges: - Relative Date: 2 years, 2 months ago Absolute Date: Thu 21 Dec 2023 00:30 Selected Answer: - Upvotes: 2

A modified version of this was in the exam today.

Comment 5

ID: 987215 User: ddogg Badges: - Relative Date: 2 years, 6 months ago Absolute Date: Tue 22 Aug 2023 10:08 Selected Answer: - Upvotes: 2

I feel like the second one should be /detect?to=en since the input can be multiple unspecified languages.

Comment 6

ID: 920304 User: ziggy1117 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Sun 11 Jun 2023 01:09 Selected Answer: - Upvotes: 1

answer is correct

Comment 7

ID: 678295 User: Davard Badges: - Relative Date: 3 years, 5 months ago Absolute Date: Sun 25 Sep 2022 02:05 Selected Answer: - Upvotes: 4

The answer seems correct to me. api.nam would keep it within the US.

119. AI-102 Topic 3 Question 34

Sequence
274
Discussion ID
102572
Source URL
https://www.examtopics.com/discussions/microsoft/view/102572-exam-ai-102-topic-3-question-34-discussion/
Posted By
marti_tremblay000
Posted At
March 14, 2023, 12:38 p.m.

Question

HOTSPOT
-

You are building content for a video training solution.

You need to create narration to accompany the video content. The solution must use Custom Neural Voice.

What should you use to create a custom neural voice, and which service should you use to generate the narration? To answer, select the appropriate options in the answer area.

NOTE: Each correct answer is worth one point.

image

Suggested Answer

image
Answer Description Click to expand


Comments 5 comments Click to expand

Comment 1

ID: 1139002 User: evangelist Badges: Highly Voted Relative Date: 2 years, 1 month ago Absolute Date: Sat 03 Feb 2024 05:39 Selected Answer: - Upvotes: 6

For Creating a Custom Neural Voice:
Use: The Speech Studio Portal
Explanation: The Speech Studio portal is part of Microsoft Azure's Cognitive Services and is specifically designed for speech-related applications, including the creation of Custom Neural Voices.

For Generating Narration:
Use: Text-to-Speech (TTS) Service
Explanation: Once you have created a Custom Neural Voice model in the Speech Studio portal, you can use Azure's Text-to-Speech service to generate narration. The Text-to-Speech service converts written text into spoken words in a natural-sounding voice, leveraging the custom voice model you've created.

Comment 2

ID: 1247877 User: krzkrzkra Badges: Most Recent Relative Date: 1 year, 8 months ago Absolute Date: Sun 14 Jul 2024 19:39 Selected Answer: - Upvotes: 1

1. Speech Studio portal
2. Text-to-speech

Comment 3

ID: 1217512 User: reiwanotora Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 15:37 Selected Answer: - Upvotes: 2

The Speech Studio portal
Text-to-speech

Comment 4

ID: 1162258 User: anto69 Badges: - Relative Date: 2 years ago Absolute Date: Thu 29 Feb 2024 07:11 Selected Answer: - Upvotes: 2

Absolutely correct

Comment 5

ID: 838787 User: marti_tremblay000 Badges: - Relative Date: 2 years, 12 months ago Absolute Date: Tue 14 Mar 2023 12:38 Selected Answer: - Upvotes: 2

The answer is correct. Speech Studio and text to speech.

120. AI-102 Topic 3 Question 45

Sequence
275
Discussion ID
108985
Source URL
https://www.examtopics.com/discussions/microsoft/view/108985-exam-ai-102-topic-3-question-45-discussion/
Posted By
WinzigWeich
Posted At
May 11, 2023, 6:47 p.m.

Question

HOTSPOT
-

You are building a text-to-speech app that will use a custom neural voice.

You need to create an SSML file for the app. The solution must ensure that the voice profile meets the following requirements:

• Expresses a calm tone
• Imitates the voice of a young adult female

How should you complete the code? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

image

Suggested Answer

image
Answer Description Click to expand


Comments 7 comments Click to expand

Comment 1

ID: 895235 User: WinzigWeich Badges: Highly Voted Relative Date: 2 years, 10 months ago Absolute Date: Thu 11 May 2023 18:47 Selected Answer: - Upvotes: 9

correct https://learn.microsoft.com/en-us/azure/cognitive-services/speech-service/speech-synthesis-markup-voice

Comment 2

ID: 1247883 User: krzkrzkra Badges: Most Recent Relative Date: 1 year, 8 months ago Absolute Date: Sun 14 Jul 2024 19:42 Selected Answer: - Upvotes: 1

1. role
2. style

Comment 3

ID: 1247882 User: krzkrzkra Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sun 14 Jul 2024 19:42 Selected Answer: - Upvotes: 1

1. role
2. style

Comment 4

ID: 1225506 User: NagaoShingo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 06 Jun 2024 14:46 Selected Answer: - Upvotes: 1

1. role
2. style

Comment 5

ID: 1217360 User: funny_penguin Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 11:42 Selected Answer: - Upvotes: 2

on exam, role and style.

Comment 6

ID: 1215797 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 22 May 2024 16:17 Selected Answer: - Upvotes: 1

role and style

Comment 7

ID: 984582 User: james2033 Badges: - Relative Date: 2 years, 6 months ago Absolute Date: Fri 18 Aug 2023 16:42 Selected Answer: - Upvotes: 2

role

style

see role at https://learn.microsoft.com/en-us/azure/ai-services/speech-service/speech-synthesis-markup-voice#role-example . May styles at https://learn.microsoft.com/en-us/azure/ai-services/speech-service/speech-synthesis-markup-voice#speaking-styles-and-roles

121. AI-102 Topic 3 Question 56

Sequence
276
Discussion ID
123659
Source URL
https://www.examtopics.com/discussions/microsoft/view/123659-exam-ai-102-topic-3-question-56-discussion/
Posted By
sl_mslconsulting
Posted At
Oct. 15, 2023, 3:11 p.m.

Question

You have the following C# function.

image

You call the function by using the following code.

image

Which output will you receive?

  • A. The quick -
    The lazy
  • B. the quick brown fox jumps over the lazy dog
  • C. jumps over the
  • D. quick brown fox
    lazy dog

Suggested Answer

D

Answer Description Click to expand


Community Answer Votes

Comments 9 comments Click to expand

Comment 1

ID: 1107477 User: MelMac Badges: Highly Voted Relative Date: 2 years, 2 months ago Absolute Date: Thu 28 Dec 2023 06:59 Selected Answer: D Upvotes: 6

I tried it out. D is correct. Key Phrases:
quick brown fox
lazy dog

Comment 2

ID: 1247889 User: krzkrzkra Badges: Most Recent Relative Date: 1 year, 8 months ago Absolute Date: Sun 14 Jul 2024 19:46 Selected Answer: D Upvotes: 1

Selected Answer: D

Comment 3

ID: 1235143 User: LM12 Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 07:36 Selected Answer: - Upvotes: 2

was on exam 20.06.24

Comment 4

ID: 1229217 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 12 Jun 2024 16:14 Selected Answer: D Upvotes: 1

D is answer.

Comment 5

ID: 1220848 User: hatanaoki Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 29 May 2024 13:16 Selected Answer: D Upvotes: 1

quick brown fox lazy dog

Comment 6

ID: 1215775 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 22 May 2024 15:47 Selected Answer: D Upvotes: 2

Key Phrases are "quick brown fox" and "lazy dog".

Comment 7

ID: 1139687 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sun 04 Feb 2024 01:35 Selected Answer: - Upvotes: 1

because the method is to extract "key Phrases" so article "the" will not be extracted.

Comment 8

ID: 1044161 User: sl_mslconsulting Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Sun 15 Oct 2023 15:11 Selected Answer: D Upvotes: 3

Verified the answer using a simple console program

Comment 8.1

ID: 1044336 User: sl_mslconsulting Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Sun 15 Oct 2023 19:14 Selected Answer: - Upvotes: 4

Tried a couple examples and seems it will pick up the pattern: adjective + noun

122. AI-102 Topic 5 Question 2

Sequence
277
Discussion ID
56834
Source URL
https://www.examtopics.com/discussions/microsoft/view/56834-exam-ai-102-topic-5-question-2-discussion/
Posted By
Duch003
Posted At
July 1, 2021, 9:30 p.m.

Question

HOTSPOT -
You are building a chatbot by using the Microsoft Bot Framework Composer.
You have the dialog design shown in the following exhibit.
image
For each of the following statements, select Yes if the statement is true. Otherwise, select No.
NOTE: Each correct selection is worth one point.
Hot Area:
image

Suggested Answer

image
Answer Description Click to expand

Box 1: No -
User.name is a property.

Box 2: Yes -

Box 3: Yes -
The coalesce() function evaluates a list of expressions and returns the first non-null (or non-empty for string) expression.
Reference:
https://docs.microsoft.com/en-us/composer/concept-language-generation https://docs.microsoft.com/en-us/azure/data-explorer/kusto/query/coalescefunction

Comments 15 comments Click to expand

Comment 1

ID: 396294 User: Duch003 Badges: Highly Voted Relative Date: 4 years, 8 months ago Absolute Date: Thu 01 Jul 2021 21:30 Selected Answer: - Upvotes: 15

I am not sure if the correct answer for third is yes - the fields are showing examples, not the actual values typed in the boxes.

Comment 1.1

ID: 450593 User: TuongNN Badges: - Relative Date: 4 years, 5 months ago Absolute Date: Fri 24 Sep 2021 02:39 Selected Answer: - Upvotes: 7

That's what I thinking about..

Comment 2

ID: 398391 User: TheB Badges: Highly Voted Relative Date: 4 years, 8 months ago Absolute Date: Sun 04 Jul 2021 15:57 Selected Answer: - Upvotes: 10

The answer is correct.

Comment 3

ID: 1246799 User: krzkrzkra Badges: Most Recent Relative Date: 1 year, 8 months ago Absolute Date: Fri 12 Jul 2024 15:50 Selected Answer: - Upvotes: 3

NYY is the answer.

Comment 4

ID: 1234506 User: HaraTadahisa Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Fri 21 Jun 2024 16:53 Selected Answer: - Upvotes: 1

My answer is that
No
Yes
Yes

Comment 5

ID: 1230857 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sat 15 Jun 2024 10:55 Selected Answer: - Upvotes: 1

No
Yes
Yes

Comment 6

ID: 1222702 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sat 01 Jun 2024 16:54 Selected Answer: - Upvotes: 1

No Yes Yes

Comment 7

ID: 1180272 User: Murtuza Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Fri 22 Mar 2024 20:18 Selected Answer: - Upvotes: 2

The third answer in this question is YES
The chatbot indeed attempts to extract relevant information from the user’s input. In this case, it looks for non-null values associated with the userName or personName entities and assigns those values to the user.name property. This process allows the chatbot to personalize its interactions and provide a more tailored experience for the user.

Comment 8

ID: 936208 User: Pixelmate Badges: - Relative Date: 2 years, 8 months ago Absolute Date: Wed 28 Jun 2023 08:00 Selected Answer: - Upvotes: 6

this was on exam 28/06

Comment 9

ID: 592742 User: 2ez4Zane Badges: - Relative Date: 3 years, 10 months ago Absolute Date: Wed 27 Apr 2022 02:57 Selected Answer: - Upvotes: 3

The answer is correct. The 3rd one is yes
The coalesce() function evaluates a list of expressions and returns the first non-null (or non-empty for string) expression.

Comment 9.1

ID: 610479 User: Moody_L Badges: - Relative Date: 3 years, 9 months ago Absolute Date: Thu 02 Jun 2022 09:37 Selected Answer: - Upvotes: 1

Disagreed. The coaleasce() function uses comma to separate out a list of expressions. However, in the question, the separator is a period.
```coalesce(tolong("not a number"), tolong("42"), 33) == 42```
Ref: https://docs.microsoft.com/en-us/azure/data-explorer/kusto/query/coalescefunction

Comment 9.1.1

ID: 646958 User: ninjia Badges: - Relative Date: 3 years, 6 months ago Absolute Date: Mon 15 Aug 2022 02:09 Selected Answer: - Upvotes: 4

I would agree with 2ez4Zane. Your link is a very good reference. I would consider the period in the question is a typo.

Comment 9.1.1.1

ID: 1224698 User: demonite Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 05 Jun 2024 13:58 Selected Answer: - Upvotes: 1

I'd agree, perhaps userName is just missing the dot

Comment 10

ID: 513554 User: RamonRW Badges: - Relative Date: 4 years, 2 months ago Absolute Date: Thu 30 Dec 2021 18:30 Selected Answer: - Upvotes: 2

Answers 1 and 2, I are correct in my eyes. But with the third answer, I am not sure. Especially it is referring to personName and I do not see that property anywhere. Usually that's a hint, if an answer should be wrong. What do you think or do I miss something here?

Comment 10.1

ID: 513557 User: RamonRW Badges: - Relative Date: 4 years, 2 months ago Absolute Date: Thu 30 Dec 2021 18:33 Selected Answer: - Upvotes: 1

I missed something indeed. There is a coalesce function for username and personname. I am still not sure, if the correct answer should actually be no.

123. AI-102 Topic 3 Question 20

Sequence
283
Discussion ID
76239
Source URL
https://www.examtopics.com/discussions/microsoft/view/76239-exam-ai-102-topic-3-question-20-discussion/
Posted By
Isidro
Posted At
May 24, 2022, 9:54 a.m.

Question

You are training a Language Understanding model for a user support system.
You create the first intent named GetContactDetails and add 200 examples.
You need to decrease the likelihood of a false positive.
What should you do?

  • A. Enable active learning.
  • B. Add a machine learned entity.
  • C. Add additional examples to the GetContactDetails intent.
  • D. Add examples to the None intent.

Suggested Answer

D

Answer Description Click to expand


Community Answer Votes

Comments 19 comments Click to expand

Comment 1

ID: 606556 User: Isidro Badges: Highly Voted Relative Date: 3 years, 9 months ago Absolute Date: Tue 24 May 2022 09:54 Selected Answer: - Upvotes: 33

I would say is D) as per the following: https://docs.microsoft.com/en-us/azure/cognitive-services/language-service/conversational-language-understanding/concepts/none-intent#adding-examples-to-the-none-intent

Comment 1.1

ID: 663122 User: mk1967 Badges: - Relative Date: 3 years, 6 months ago Absolute Date: Thu 08 Sep 2022 06:57 Selected Answer: - Upvotes: 5

Agreed, as stated in the link:
"You should also consider adding false positive examples to the None intent."

Comment 2

ID: 1133040 User: evangelist Badges: Highly Voted Relative Date: 2 years, 1 month ago Absolute Date: Sat 27 Jan 2024 04:50 Selected Answer: - Upvotes: 9

False positive means =>
The model needs examples of what it should not classify as "GetContactDetails," which is the role of the "None" intent.
Therefore, the most effective approach is to add a diverse range of examples to the "None" intent, covering various phrases and queries that are outside the scope of "GetContactDetails." This helps create a clear boundary for the model, reducing the likelihood of it mistakenly classifying unrelated inputs as belonging to the "GetContactDetails" intent.

Comment 3

ID: 1235201 User: HaraTadahisa Badges: Most Recent Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 08:27 Selected Answer: D Upvotes: 1

I say this answer is D.

Comment 4

ID: 1229237 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 12 Jun 2024 16:21 Selected Answer: D Upvotes: 1

D is answer.

Comment 5

ID: 1224485 User: anto69 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 05 Jun 2024 05:11 Selected Answer: D Upvotes: 1

For me and ChatGPT: d

Comment 6

ID: 1217552 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 16:04 Selected Answer: D Upvotes: 1

It must be D.

Comment 7

ID: 1130202 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Wed 24 Jan 2024 05:08 Selected Answer: D Upvotes: 3

The correct option to decrease the likelihood of a false positive in the Language Understanding model is to add additional None intent examples.

Option D is correct. By adding more varied examples that do not map to a valid intent to the None intent, the model can better learn when an utterance does not apply and avoid falsely matching invalid queries to a valid intent.

Options A, B, and C may improve the model in certain ways, but they do not directly address reducing false positives. Only adding additional out-of-scope examples to the None intent will help the model better distinguish when new utterances do not match any existing intent's patterns.

So out of the options, adding examples to the None intent is the way to decrease the likelihood of false positives.

Comment 8

ID: 984516 User: james2033 Badges: - Relative Date: 2 years, 6 months ago Absolute Date: Fri 18 Aug 2023 15:15 Selected Answer: D Upvotes: 1

200 sample data. --> much false positive. --> Increase number of training data. --> Add example to the None intent, not active learning in this context.

Comment 9

ID: 921788 User: Drummer Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Tue 13 Jun 2023 01:00 Selected Answer: - Upvotes: 1

A. Enable active learning.

By enabling active learning, the model can actively request feedback from users when it encounters uncertain or ambiguous queries. This feedback loop helps improve the model's understanding and reduces false positives by incorporating user input into its training process. Option A (Enable active learning) is the correct choice to decrease the likelihood of false positives.

Comment 10

ID: 920753 User: ziggy1117 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Sun 11 Jun 2023 16:13 Selected Answer: D Upvotes: 3

You should also consider adding false positive examples to the None intent. For example, in a flight booking project it is likely that the utterance "I want to buy a book" could be confused with a Book Flight intent. Adding "I want to buy a book" or "I love reading books" as None training utterances helps alter the predictions of those types of utterances towards the None intent instead of Book Flight.

https://learn.microsoft.com/en-us/azure/cognitive-services/language-service/conversational-language-understanding/concepts/none-intent#adding-examples-to-the-none-intent

Comment 11

ID: 918759 User: ziggy1117 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Fri 09 Jun 2023 01:44 Selected Answer: - Upvotes: 1

You should also consider adding false positive examples to the None intent. For example, in a flight booking project it is likely that the utterance "I want to buy a book" could be confused with a Book Flight intent. Adding "I want to buy a book" or "I love reading books" as None training utterances helps alter the predictions of those types of utterances towards the None intent instead of Book Flight.

https://learn.microsoft.com/en-us/azure/cognitive-services/language-service/conversational-language-understanding/concepts/none-intent#adding-examples-to-the-none-intent

Comment 12

ID: 817673 User: Marilena96 Badges: - Relative Date: 3 years ago Absolute Date: Wed 22 Feb 2023 10:30 Selected Answer: - Upvotes: 2

To decrease the likelihood of a false positive, you can add additional examples to the GetContactDetails intent. This will help the model to better understand the intent and reduce the likelihood of false positive predictions.

Comment 12.1

ID: 899829 User: Rob77 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Wed 17 May 2023 09:56 Selected Answer: - Upvotes: 1

Nope, 20-30 examples per intent is recommended. See https://learn.microsoft.com/en-us/azure/cognitive-services/LUIS/concepts/application-design#create-example-utterances-for-each-intent

Comment 13

ID: 779290 User: ap1234pa Badges: - Relative Date: 3 years, 1 month ago Absolute Date: Tue 17 Jan 2023 20:52 Selected Answer: D Upvotes: 4

As explained in MS Document
"false positive" = None intent

Comment 14

ID: 778387 User: ap1234pa Badges: - Relative Date: 3 years, 1 month ago Absolute Date: Tue 17 Jan 2023 00:50 Selected Answer: D Upvotes: 2

Add examples to "None" intent

Comment 15

ID: 742076 User: SSJA Badges: - Relative Date: 3 years, 3 months ago Absolute Date: Sun 11 Dec 2022 21:33 Selected Answer: D Upvotes: 3

Answer is D.
Reference - https://learn.microsoft.com/en-us/azure/cognitive-services/LUIS/concepts/application-design

Comment 16

ID: 705586 User: GigaCaster Badges: - Relative Date: 3 years, 4 months ago Absolute Date: Thu 27 Oct 2022 15:21 Selected Answer: A Upvotes: 3

My reasoning for active learning is that it will be better in the long run whereas None intent is but a temporary fix to the current issue.

Comment 17

ID: 668948 User: Internal_Koala Badges: - Relative Date: 3 years, 5 months ago Absolute Date: Wed 14 Sep 2022 13:20 Selected Answer: D Upvotes: 2

As described by Isidro and mk1967.

124. AI-102 Topic 3 Question 33

Sequence
284
Discussion ID
102571
Source URL
https://www.examtopics.com/discussions/microsoft/view/102571-exam-ai-102-topic-3-question-33-discussion/
Posted By
marti_tremblay000
Posted At
March 14, 2023, 12:35 p.m.

Question

You have an Azure Cognitive Services model named Model1 that identifies the intent of text input.

You develop an app in C# named App1.

You need to configure App1 to use Model1.

Which package should you add to App1?

  • A. Universal.Microsoft.CognitiveServices.Speech
  • B. SpeechServicesToolkit
  • C. Azure.AI.Language.Conversations
  • D. Xamarin.Cognitive.Speech

Suggested Answer

C

Answer Description Click to expand


Community Answer Votes

Comments 8 comments Click to expand

Comment 1

ID: 1139035 User: evangelist Badges: Highly Voted Relative Date: 2 years, 1 month ago Absolute Date: Sat 03 Feb 2024 08:02 Selected Answer: C Upvotes: 8

Azure Conversational Language Understanding is part of Azure Cognitive Services and is designed to understand the intent of text input. The Azure.AI.Language.Conversations package provides C# developers with the client libraries needed to interact with these services, enabling you to send text to models and receive the results of intent and entity recognition.
Explanation of other options:

A. Universal.Microsoft.CognitiveServices.Speech is primarily used for speech services such as speech-to-text conversion, and not specifically for text intent recognition.

Comment 2

ID: 1235198 User: HaraTadahisa Badges: Most Recent Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 08:25 Selected Answer: C Upvotes: 1

I say this answer is C.

Comment 3

ID: 1229227 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 12 Jun 2024 16:17 Selected Answer: C Upvotes: 1

C is answer.

Comment 4

ID: 1217513 User: reiwanotora Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 15:40 Selected Answer: C Upvotes: 1

C is right answer.
Xamarin is done.

Comment 5

ID: 906646 User: EliteAllen Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Thu 25 May 2023 14:27 Selected Answer: C Upvotes: 2

C. Azure.AI.Language.Conversations

The Azure.AI.Language.Conversations package is part of the Azure SDK for .NET and is used for interacting with Azure Cognitive Services' Language Understanding (LUIS) models, which are used to identify the intent of text input. This makes it the correct choice for integrating Model1 into App1.

Comment 6

ID: 879188 User: Pffffff Badges: - Relative Date: 2 years, 10 months ago Absolute Date: Mon 24 Apr 2023 11:33 Selected Answer: C Upvotes: 4

ChatGPT: The package you should add to App1 to use Model1 is C. Azure.AI.Language.Conversations.

Azure.AI.Language.Conversations is a package that provides the Language Understanding (LUIS) service, which can be used to identify the intent of text input. This package contains classes for authenticating with the LUIS service and sending text to the service to obtain intent and entity information.

Comment 7

ID: 840868 User: marti_tremblay000 Badges: - Relative Date: 2 years, 12 months ago Absolute Date: Thu 16 Mar 2023 12:13 Selected Answer: - Upvotes: 2

ChatGPT says the correct answer isn't even listed :
To configure App1 to use Model1, you should add the Azure.AI.TextAnalytics package to the project. This package provides the necessary libraries and functionality to integrate with Azure Cognitive Services text analytics models such as Model1.

Therefore, the correct answer is not listed among the options. The correct answer is Azure.AI.TextAnalytics.

Comment 8

ID: 838786 User: marti_tremblay000 Badges: - Relative Date: 2 years, 12 months ago Absolute Date: Tue 14 Mar 2023 12:35 Selected Answer: C Upvotes: 3

The question is about the intent of text input. It has nothing to do with speech.
Therefore the Azure.AI.Language.Conversations is the answer :
Conversation Analysis is a cloud-based conversational AI service that applies custom machine-learning intelligence to a user's conversational, natural language text to predict overall meaning, and pull out relevant, detailed information.
Reference : https://learn.microsoft.com/en-us/samples/azure/azure-sdk-for-net/azureailanguageconversations-samples/

125. AI-102 Topic 3 Question 41

Sequence
285
Discussion ID
111850
Source URL
https://www.examtopics.com/discussions/microsoft/view/111850-exam-ai-102-topic-3-question-41-discussion/
Posted By
973b658
Posted At
June 11, 2023, 1:08 a.m.

Question

You have an Azure subscription that contains an Azure Cognitive Service for Language resource.

You need to identify the URL of the REST interface for the Language service.

Which blade should you use in the Azure portal?

  • A. Identity
  • B. Keys and Endpoint
  • C. Networking
  • D. Properties

Suggested Answer

B

Answer Description Click to expand


Community Answer Votes

Comments 6 comments Click to expand

Comment 1

ID: 920303 User: 973b658 Badges: Highly Voted Relative Date: 2 years, 9 months ago Absolute Date: Sun 11 Jun 2023 01:08 Selected Answer: B Upvotes: 9

It is B.

Comment 2

ID: 1235194 User: HaraTadahisa Badges: Most Recent Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 08:24 Selected Answer: B Upvotes: 1

I say this answer is B.

Comment 3

ID: 1229221 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 12 Jun 2024 16:16 Selected Answer: B Upvotes: 1

B is answer.

Comment 4

ID: 1215792 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 22 May 2024 16:11 Selected Answer: B Upvotes: 1

B is right answer.

Comment 5

ID: 1139138 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 03 Feb 2024 10:38 Selected Answer: B Upvotes: 3

B. Keys and Endpoint blade in the Azure portal.

This blade provides the endpoint URL needed to access the Cognitive Services API, along with the keys required for authentication. The endpoint URL is essential for making API calls to the service, including those for Language features such as sentiment analysis, key phrase extraction, named entity recognition, and more.

Comment 6

ID: 1055010 User: chenglim Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Fri 27 Oct 2023 02:41 Selected Answer: B Upvotes: 1

ans correct

126. AI-102 Topic 5 Question 29

Sequence
286
Discussion ID
102800
Source URL
https://www.examtopics.com/discussions/microsoft/view/102800-exam-ai-102-topic-5-question-29-discussion/
Posted By
marti_tremblay000
Posted At
March 16, 2023, 12:10 p.m.

Question

You create five bots by using Microsoft Bot Framework Composer.

You need to make a single bot available to users that combines the bots. The solution must support dynamic routing to the bots based on user input.

Which three actions should you perform? Each correct answer presents part of the solution.

NOTE: Each correct selection is worth one point.

  • A. Create a composer extension.
  • B. Change the Recognizer/Dispatch type.
  • C. Create an Orchestrator model.
  • D. Enable WebSockets.
  • E. Create a custom recognizer JSON file.
  • F. Install the Orchestrator package.

Suggested Answer

BCF

Answer Description Click to expand


Community Answer Votes

Comments 6 comments Click to expand

Comment 1

ID: 1140081 User: evangelist Badges: Highly Voted Relative Date: 2 years, 1 month ago Absolute Date: Sun 04 Feb 2024 13:32 Selected Answer: BCF Upvotes: 6

B. Change the Recognizer/Dispatch type: Adjusting the recognizer type to one that supports dispatching allows the bot to understand and route user input to the appropriate sub-bot.
================================================
C. Create an Orchestrator model: The Orchestrator is an AI-based component that helps in making decisions on routing the conversation to the most appropriate bot based on the user's intent.
================================================
F. Install the Orchestrator package: This is necessary to use the Orchestrator model within your bot, enabling it to dynamically route conversations to the correct bot based on the context and content of the user input.

Comment 2

ID: 1235332 User: reiwanotora Badges: Most Recent Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 12:45 Selected Answer: BCF Upvotes: 2

Recognizer/Dispatch type
double Orchestrator

Comment 3

ID: 1235220 User: HaraTadahisa Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 08:53 Selected Answer: BCF Upvotes: 2

I say this answer is BCF.

Comment 4

ID: 1222672 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sat 01 Jun 2024 15:50 Selected Answer: BCE Upvotes: 1

Memorize dispatch and Orchestrator.

Comment 5

ID: 1222671 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sat 01 Jun 2024 15:50 Selected Answer: - Upvotes: 1

Memorize dispatch and Orchestrator.

Comment 6

ID: 840864 User: marti_tremblay000 Badges: - Relative Date: 2 years, 12 months ago Absolute Date: Thu 16 Mar 2023 12:10 Selected Answer: BCF Upvotes: 4

ChatGPT answers :
To make a single bot available to users that combines the bots and supports dynamic routing to the bots based on user input, the following three actions should be performed:

B. Change the Recognizer/Dispatch type: The Recognizer/Dispatch type should be changed to enable the bot to recognize user input and dispatch it to the appropriate bot.

C. Create an Orchestrator model: An Orchestrator model should be created to handle the routing of user input to the appropriate bot.

F. Install the Orchestrator package: The Orchestrator package should be installed to provide the bot with the necessary functionality to route user input to the appropriate bot.

Therefore, options B, C, and F are the correct answers.

127. AI-102 Topic 5 Question 34

Sequence
287
Discussion ID
111027
Source URL
https://www.examtopics.com/discussions/microsoft/view/111027-exam-ai-102-topic-5-question-34-discussion/
Posted By
kail85
Posted At
June 4, 2023, 1:43 a.m.

Question

You are building a chatbot by using Microsoft Bot Framework Composer.

You need to configure the chatbot to present a list of available options. The solution must ensure that an image is provided for each option.

Which two features should you use? Each correct answer presents part of the solution.

NOTE: Each correct selection is worth one point.

  • A. an entity
  • B. an Azure function
  • C. an utterance
  • D. an adaptive card
  • E. a dialog

Suggested Answer

DE

Answer Description Click to expand


Community Answer Votes

Comments 5 comments Click to expand

Comment 1

ID: 913965 User: kail85 Badges: Highly Voted Relative Date: 2 years, 9 months ago Absolute Date: Sun 04 Jun 2023 01:43 Selected Answer: DE Upvotes: 11

To configure the chatbot to present a list of available options with images, you should use the following features:

D. An adaptive card: Adaptive cards are a flexible, customizable way to present rich, interactive content within a chatbot conversation. You can use adaptive cards to display a list of available options, along with images and other interactive elements, such as buttons or inputs.

E. A dialog: Dialogs are used to manage and organize the conversation flow in your chatbot. You can create a dialog that is responsible for presenting the list of available options using an adaptive card. The dialog can be triggered when needed, and it can handle the user's selection of an option.

Comment 2

ID: 1235219 User: HaraTadahisa Badges: Most Recent Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 08:52 Selected Answer: DE Upvotes: 1

I say this answer is D and E.

Comment 3

ID: 1229892 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 16:56 Selected Answer: DE Upvotes: 1

DE is answer.

Comment 4

ID: 1140681 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Mon 05 Feb 2024 06:17 Selected Answer: DE Upvotes: 2

only card enables showing image in an option

Comment 5

ID: 979392 User: james2033 Badges: - Relative Date: 2 years, 7 months ago Absolute Date: Sat 12 Aug 2023 15:53 Selected Answer: DE Upvotes: 2

Quote “Creating Adaptive Cards attachments, as seen above.” at https://learn.microsoft.com/en-us/composer/introduction?tabs=v2x#language-generation . Dialog https://learn.microsoft.com/en-us/composer/concept-dialog?tabs=v2x

128. AI-102 Topic 5 Question 36

Sequence
288
Discussion ID
107302
Source URL
https://www.examtopics.com/discussions/microsoft/view/107302-exam-ai-102-topic-5-question-36-discussion/
Posted By
Codia
Posted At
April 24, 2023, 1:37 p.m.

Question

You have a chatbot that was built by using Microsoft Bot Framework and deployed to Azure.

You need to configure the bot to support voice interactions. The solution must support multiple client apps.

Which type of channel should you use?

  • A. Cortana
  • B. Microsoft Teams
  • C. Direct Line Speech

Suggested Answer

C

Answer Description Click to expand


Community Answer Votes

Comments 6 comments Click to expand

Comment 1

ID: 1235217 User: HaraTadahisa Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 08:51 Selected Answer: C Upvotes: 1

I say this answer is C

Comment 2

ID: 1229866 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 15:54 Selected Answer: C Upvotes: 1

C is answer.

Comment 3

ID: 1133167 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 27 Jan 2024 09:38 Selected Answer: C Upvotes: 3

Designed for Voice Interactions: ***Direct Line Speech*** is specifically designed to enable voice interactions for bots. It provides an integrated set of capabilities to convert speech to text (STT), perform natural language understanding, and then convert text back to speech (TTS). This is essential for creating a seamless voice interaction experience.

***Cortana*** is tailored for interactions with Microsoft's virtual assistant and is limited to the Cortana ecosystem.
***Microsoft Teams*** is focused on integrating bots within the Teams platform for collaboration and communication, but it's not specifically designed for voice-first experiences.

Comment 4

ID: 1050009 User: sl_mslconsulting Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Sun 22 Oct 2023 02:56 Selected Answer: C Upvotes: 2

Quoted from the link down below. Using DirectLine, not only can you define your own UI with your own features, but you can connect multiple applications to a registered Bot simultaneously. https://blog.botframework.com/2017/06/26/speech-to-text/

Comment 5

ID: 979384 User: james2033 Badges: - Relative Date: 2 years, 7 months ago Absolute Date: Sat 12 Aug 2023 15:38 Selected Answer: C Upvotes: 3

https://learn.microsoft.com/en-us/azure/bot-service/directline-speech-bot?view=azure-bot-service-4.0

Comment 6

ID: 879267 User: Codia Badges: - Relative Date: 2 years, 10 months ago Absolute Date: Mon 24 Apr 2023 13:37 Selected Answer: - Upvotes: 4

Direct Line Speech is a robust, end-to-end solution for creating a flexible, extensible voice assistant. It is powered by the Bot Framework and its Direct Line Speech channel, that is optimized for voice-in, voice-out interaction with bots.
https://learn.microsoft.com/en-us/azure/cognitive-services/speech-service/direct-line-speech

129. AI-102 Topic 3 Question 57

Sequence
291
Discussion ID
124687
Source URL
https://www.examtopics.com/discussions/microsoft/view/124687-exam-ai-102-topic-3-question-57-discussion/
Posted By
chenglim
Posted At
Oct. 27, 2023, 3:45 a.m.

Question

You have the following Python method.

image

You need to deploy an Azure resource to the East US Azure region. The resource will be used to perform sentiment analysis.

How should you call the method?

  • A. create_resource("res1", "TextAnalytics", "Standard", "East US")
  • B. create_resource("res1", "ContentModerator", "S0", "eastus")
  • C. create_resource("res1", "ContentModerator", "Standard", "East US")
  • D. create_resource("res1", "TextAnalytics", "S0", "eastus")

Suggested Answer

D

Answer Description Click to expand


Community Answer Votes

Comments 6 comments Click to expand

Comment 1

ID: 1235188 User: HaraTadahisa Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 08:22 Selected Answer: D Upvotes: 1

ComputerVision, F0.
TextAnalysis, S0.

Comment 2

ID: 1231287 User: PeteColag Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sun 16 Jun 2024 11:49 Selected Answer: - Upvotes: 2

The correct answer can be found by a process of elimination:

1. Need to provide reference to correct service -> Text Analytics
2. Need to provide reference to a correct service tier: S0 or F0
3. Need to provide reference to a correct name for a computer region:
United States:

East US: eastus
East US 2: eastus2
Central US: centralus
North Central US: northcentralus
South Central US: southcentralus
West US: westus
West US 2: westus2

D is the only answer that captures all these correctly.

Comment 3

ID: 1229216 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 12 Jun 2024 16:13 Selected Answer: D Upvotes: 1

redundant question.
TextAnalytics, S0, eastus

Comment 4

ID: 1215773 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 22 May 2024 15:40 Selected Answer: D Upvotes: 1

D is right answer.

Comment 5

ID: 1139682 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sun 04 Feb 2024 01:24 Selected Answer: - Upvotes: 2

In Azure services, "S0" represents a pricing tier or service level. For Azure Cognitive Services, including Text Analytics, the pricing tier determines the service's performance, request rate limits, features, and costs. Each pricing tier offers different service capabilities and quotas, and they come with varying price points.

Comment 6

ID: 1055043 User: chenglim Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Fri 27 Oct 2023 03:45 Selected Answer: D Upvotes: 1

correct answer

130. AI-102 Topic 3 Question 62

Sequence
292
Discussion ID
135626
Source URL
https://www.examtopics.com/discussions/microsoft/view/135626-exam-ai-102-topic-3-question-62-discussion/
Posted By
GHill1982
Posted At
March 10, 2024, 7:32 a.m.

Question

You are building a chatbot.

You need to ensure that the bot will recognize the names of your company’s products and codenames. The solution must minimize development effort.

Which Azure Cognitive Service for Language service should you include in the solution?

  • A. custom text classification
  • B. entity linking
  • C. custom Named Entity Recognition (NER)
  • D. key phrase extraction

Suggested Answer

C

Answer Description Click to expand


Community Answer Votes

Comments 6 comments Click to expand

Comment 1

ID: 1235186 User: HaraTadahisa Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Sat 22 Jun 2024 08:20 Selected Answer: C Upvotes: 1

I say this answer is C.

Comment 2

ID: 1229213 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 12 Jun 2024 16:12 Selected Answer: C Upvotes: 1

C is answer.

Comment 3

ID: 1226713 User: anto69 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sat 08 Jun 2024 13:37 Selected Answer: C Upvotes: 1

C for me and ChatGPT too

Comment 4

ID: 1215738 User: takaimomoGcup Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 22 May 2024 15:08 Selected Answer: C Upvotes: 1

NER is right answer.

Comment 5

ID: 1177767 User: Murtuza Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Tue 19 Mar 2024 22:51 Selected Answer: - Upvotes: 2

Named Entity Recognition (NER) identifies and categorizes specific names or terms in text, such as names of people, organizations, places, and more. By customizing NER, you can tailor it to recognize your company-specific entities, making it an efficient solution for your chatbot’s needs45.

So, the correct answer is C. custom Named Entity Recognition (NER).

Comment 6

ID: 1170079 User: GHill1982 Badges: - Relative Date: 2 years ago Absolute Date: Sun 10 Mar 2024 07:32 Selected Answer: C Upvotes: 3

C. custom Named Entity Recognition (NER)
Custom NER allows you to train a model to recognize specific terms relevant to your business, such as product names and codenames. https://learn.microsoft.com/en-us/azure/ai-services/computer-vision/concept-shelf-analysis

131. AI-102 Topic 5 Question 28

Sequence
297
Discussion ID
102922
Source URL
https://www.examtopics.com/discussions/microsoft/view/102922-exam-ai-102-topic-5-question-28-discussion/
Posted By
marti_tremblay000
Posted At
March 17, 2023, 12:43 p.m.

Question

Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.

After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.

You have a chatbot that uses question answering in Azure Cognitive Service for Language.

Users report that the responses of the chatbot lack formality when answering spurious questions.

You need to ensure that the chatbot provides formal responses to spurious questions.

Solution: From Language Studio, you change the chitchat source to qna_chitchat_professional.tsv, and then retrain and republish the model.

Does this meet the goal?

  • A. Yes
  • B. No

Suggested Answer

A

Answer Description Click to expand


Community Answer Votes

Comments 4 comments Click to expand

Comment 1

ID: 1234482 User: HaraTadahisa Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Fri 21 Jun 2024 16:41 Selected Answer: A Upvotes: 1

Focus on "qna_chitchat_professional.tsv". HipHop!

Comment 2

ID: 1140082 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sun 04 Feb 2024 13:37 Selected Answer: A Upvotes: 2

The qna_chitchat_professional.tsv is specifically designed to provide responses with a formal tone compared to other chitchat sources. It contains a curated set of question-and-answer pairs that are structured to reflect a professional and formal manner of communication. This makes it more suitable for environments or user interactions where formal responses are preferred, ensuring that the chatbot's replies to spurious or off-topic questions maintain the desired level of professionalism.

Comment 3

ID: 984776 User: james2033 Badges: - Relative Date: 2 years, 6 months ago Absolute Date: Fri 18 Aug 2023 22:47 Selected Answer: A Upvotes: 4

Formal response, it is Personality = Professional, it is file qna_chitchat_professional.tsv . https://learn.microsoft.com/en-us/azure/ai-services/language-service/question-answering/concepts/best-practices#choosing-a-personality

Comment 4

ID: 841941 User: marti_tremblay000 Badges: - Relative Date: 2 years, 12 months ago Absolute Date: Fri 17 Mar 2023 12:43 Selected Answer: A Upvotes: 2

ChatGPT answer :
Yes, changing the chitchat source to "qna_chitchat_professional.tsv" and retraining and republishing the model can meet the goal of ensuring that the chatbot provides formal responses to spurious questions.

The "qna_chitchat_professional.tsv" file is a source file for professional chitchat, which includes conversational responses for formal topics, such as business and finance. This file is designed to provide more formal responses to chitchat questions, so changing the source file to this file and retraining the model can improve the formality of responses to spurious questions.

132. AI-102 Topic 5 Question 30

Sequence
298
Discussion ID
111868
Source URL
https://www.examtopics.com/discussions/microsoft/view/111868-exam-ai-102-topic-5-question-30-discussion/
Posted By
973b658
Posted At
June 11, 2023, 7:17 a.m.

Question

Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.

After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.

You are building a chatbot that will use question answering in Azure Cognitive Service for Language.

You have a PDF named Doc1.pdf that contains a product catalogue and a price list.

You upload Doc1.pdf and train the model.

During testing, users report that the chatbot responds correctly to the following question: What is the price of ?

The chatbot fails to respond to the following question: How much does cost?

You need to ensure that the chatbot responds correctly to both questions.

Solution: From Language Studio, you add alternative phrasing to the question and answer pair, and then retrain and republish the model.

Does this meet the goal?

  • A. Yes
  • B. No

Suggested Answer

A

Answer Description Click to expand


Community Answer Votes

Comments 6 comments Click to expand

Comment 1

ID: 1234481 User: HaraTadahisa Badges: - Relative Date: 1 year, 8 months ago Absolute Date: Fri 21 Jun 2024 16:39 Selected Answer: A Upvotes: 2

"From Language Studio, you add alternative phrasing to the question and answer pair, and then retrain and republish the model." is correct solution. HipHop!

Comment 2

ID: 1221650 User: fuck_india Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 30 May 2024 16:18 Selected Answer: A Upvotes: 2

A is right answer.

Comment 3

ID: 1140112 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sun 04 Feb 2024 14:03 Selected Answer: A Upvotes: 2

Yes
adding alternative phrasing to the question and answer pair in Azure Cognitive Service for Language, then retraining and republishing the model, meets the goal. This approach allows the model to understand and respond to variations in how questions are asked, improving its ability to correctly answer questions with similar meanings but different wordings.

Comment 4

ID: 1102029 User: Gvalli Badges: - Relative Date: 2 years, 2 months ago Absolute Date: Thu 21 Dec 2023 00:26 Selected Answer: - Upvotes: 2

A modified version of this was in the exam today.

Comment 5

ID: 979396 User: james2033 Badges: - Relative Date: 2 years, 7 months ago Absolute Date: Sat 12 Aug 2023 16:04 Selected Answer: A Upvotes: 3

"Question: What is the price of Microsoft Stock?"

"How much does a Microsoft share cost?"

at https://learn.microsoft.com/en-us/azure/ai-services/language-service/question-answering/concepts/best-practices#when-should-you-add-alternate-questions-to-question-and-answer-pairs

Comment 6

ID: 920428 User: 973b658 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Sun 11 Jun 2023 07:17 Selected Answer: A Upvotes: 4

A from ChatGPT 3.5.

133. AI-102 Topic 5 Question 40

Sequence
303
Discussion ID
111874
Source URL
https://www.examtopics.com/discussions/microsoft/view/111874-exam-ai-102-topic-5-question-40-discussion/
Posted By
973b658
Posted At
June 11, 2023, 7:59 a.m.

Question

You are building a chatbot for a travel agent. The bot will ask users for a destination and must repeat the question until a valid input is received, or the user closes the conversation.

Which type of dialog should you use?

  • A. prompt
  • B. input
  • C. adaptive
  • D. QnA Maker

Suggested Answer

A

Answer Description Click to expand


Community Answer Votes

Comments 6 comments Click to expand

Comment 1

ID: 1229864 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 15:53 Selected Answer: A Upvotes: 1

A is answer.

Comment 2

ID: 1221642 User: reiwanotora Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 30 May 2024 16:11 Selected Answer: A Upvotes: 1

A is right answer.

Comment 3

ID: 1143929 User: PCRamirez Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Thu 08 Feb 2024 01:55 Selected Answer: - Upvotes: 2

According to Windows Copilot:
I understand your reasoning, but option A is not the best choice for this scenario. A prompt dialog can validate the user’s input, but it cannot handle complex conversational flows, such as interruptions, validations, and conditional logic. For example, if the user wants to change their destination, or ask for more information, or cancel the conversation, a prompt dialog cannot adapt to these situations. A prompt dialog can only ask the user for a single piece of information and return it to the parent dialog.

An adaptive dialog, on the other hand, can dynamically adjust to the user’s input and context. It can handle multiple inputs, interruptions, validations, and conditional logic. It can also use language generation templates to create natural and varied responses. An adaptive dialog can provide a better user experience and a more robust chatbot for a travel agent.

I hope this clarifies why option C is the correct answer.

Comment 4

ID: 1133230 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 27 Jan 2024 11:27 Selected Answer: A Upvotes: 3

In this scenario, where a chatbot needs to ask users for a destination and must repeat the question until a valid input is received, the most appropriate type of dialog to use is a "prompt" dialog. Here's why:

Prompt Dialog: The prompt dialog is specifically designed to obtain user input and can be configured to validate this input and reprompt the user if necessary. In the Microsoft Bot Framework, prompts are used to handle various types of user input such as text, numbers, dates, and confirmations. For your requirement, you can use a text prompt that keeps asking the user for a destination until a valid one is provided or the user exits the conversation.

Comment 5

ID: 929116 User: Tin_Tin Badges: - Relative Date: 2 years, 8 months ago Absolute Date: Wed 21 Jun 2023 08:28 Selected Answer: - Upvotes: 3

answer seems correct.
https://learn.microsoft.com/en-us/training/modules/create-bot-with-bot-framework-composer/4-dialogs

Comment 6

ID: 920448 User: 973b658 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Sun 11 Jun 2023 07:59 Selected Answer: A Upvotes: 3

A. prompt

134. AI-102 Topic 5 Question 41

Sequence
304
Discussion ID
111875
Source URL
https://www.examtopics.com/discussions/microsoft/view/111875-exam-ai-102-topic-5-question-41-discussion/
Posted By
973b658
Posted At
June 11, 2023, 8:01 a.m.

Question

You are building a chatbot.

You need to configure the chatbot to query a knowledge base.

Which dialog class should you use?

  • A. QnAMakerDialog
  • B. AdaptiveDialog
  • C. SkillDialog
  • D. ComponentDialog

Suggested Answer

A

Answer Description Click to expand


Community Answer Votes

Comments 4 comments Click to expand

Comment 1

ID: 1229863 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 15:53 Selected Answer: A Upvotes: 1

A is answer.

Comment 2

ID: 1221643 User: reiwanotora Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 30 May 2024 16:11 Selected Answer: A Upvotes: 1

A is correct answer.

Comment 3

ID: 1133252 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 27 Jan 2024 11:53 Selected Answer: A Upvotes: 1

Here's why:

QnAMakerDialog: This class is specifically designed for integration with QnA Maker, a cloud-based API service offered by Microsoft Azure that creates a question and answer layer over your data. QnAMakerDialog simplifies the process of querying a QnA Maker knowledge base and handling the responses within your chatbot. It manages the interaction pattern with the QnA Maker service, sending user queries and receiving answers that are then presented to the user.

Comment 4

ID: 920451 User: 973b658 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Sun 11 Jun 2023 08:01 Selected Answer: A Upvotes: 2

A. QnAMakerDialog

135. AI-102 Topic 5 Question 43

Sequence
305
Discussion ID
112147
Source URL
https://www.examtopics.com/discussions/microsoft/view/112147-exam-ai-102-topic-5-question-43-discussion/
Posted By
973b658
Posted At
June 14, 2023, 9:52 a.m.

Question

You develop a Conversational Language Understanding model by using Language Studio.

During testing, users receive incorrect responses to requests that do NOT relate to the capabilities of the model.

You need to ensure that the model identifies spurious requests.

What should you do?

  • A. Enable active learning.
  • B. Add entities.
  • C. Add examples to the None intent.
  • D. Add examples to the custom intents.

Suggested Answer

C

Answer Description Click to expand


Community Answer Votes

Comments 6 comments Click to expand

Comment 1

ID: 1229862 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 15:52 Selected Answer: C Upvotes: 1

C is answer.

Comment 2

ID: 1226717 User: anto69 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sat 08 Jun 2024 13:50 Selected Answer: - Upvotes: 1

ChatGPT says C but I've doubt on A

Comment 3

ID: 1133260 User: evangelist Badges: - Relative Date: 2 years, 1 month ago Absolute Date: Sat 27 Jan 2024 12:01 Selected Answer: C Upvotes: 3

In a Conversational Language Understanding model, especially when using a tool like Language Studio, the "None" intent plays a crucial role in handling queries or requests that do not pertain to the defined capabilities of the model. Here's why adding examples to the None intent is the right approach:

None Intent: The None intent is designed to capture user inputs that fall outside of the other specified intents in your model. By training the None intent with examples of irrelevant or unrelated requests, you teach the model to recognize and appropriately handle spurious queries that do not align with its capabilities. This helps in reducing incorrect responses by ensuring that the model does not force these out-of-scope queries into one of the custom intents.

Comment 4

ID: 1091252 User: Gvalli Badges: - Relative Date: 2 years, 3 months ago Absolute Date: Fri 08 Dec 2023 19:14 Selected Answer: C Upvotes: 1

C - None intent.
Because of 'NOT' related utterances, and 'NOT' related utterances do not add value, hence 'Not active learning'.
Keyword - 'NOT' related.
During testing, users receive incorrect responses to requests that do NOT relate to the capabilities of the model.

Comment 5

ID: 922867 User: 973b658 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Wed 14 Jun 2023 09:52 Selected Answer: C Upvotes: 3

C.
#22 topic3

Comment 5.1

ID: 998630 User: M25 Badges: - Relative Date: 2 years, 6 months ago Absolute Date: Mon 04 Sep 2023 17:07 Selected Answer: - Upvotes: 1

Now #19 topic 3

136. AI-102 Topic 5 Question 51

Sequence
306
Discussion ID
125138
Source URL
https://www.examtopics.com/discussions/microsoft/view/125138-exam-ai-102-topic-5-question-51-discussion/
Posted By
_LAW_
Posted At
Nov. 2, 2023, 12:03 a.m.

Question

You are building a chatbot by using the Microsoft Bot Framework SDK. The bot will be used to accept food orders from customers and allow the customers to customize each food item.

You need to configure the bot to ask the user for additional input based on the type of item ordered. The solution must minimize development effort.

Which two types of dialogs should you use? Each correct answer presents part of the solution.

NOTE: Each correct selection is worth one point.

  • A. adaptive
  • B. action
  • C. waterfall
  • D. prompt
  • E. input

Suggested Answer

CD

Answer Description Click to expand


Community Answer Votes

Comments 5 comments Click to expand

Comment 1

ID: 1133331 User: evangelist Badges: Highly Voted Relative Date: 2 years, 1 month ago Absolute Date: Sat 27 Jan 2024 13:17 Selected Answer: CD Upvotes: 9

=======Waterfall Dialogs===========:
Guide users through a predefined, linear sequence of steps.
Each step prompts the user for specific input, and the bot moves to the next step based on the user's response. Useful for structured conversations where the bot needs to collect information in a specific order.
Ideal for gathering customization options for food items in a structured way.
========Prompt Dialogs===========:
Simple dialogs that present a prompt to the user and wait for their response.
Can be used within waterfall dialogs or as standalone prompts.
Offer flexibility in how you ask questions and collect user input.

Can be incorporated within waterfall dialogs to ask for customization details for each item.

Comment 2

ID: 1229909 User: hatanaoki Badges: Most Recent Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 17:11 Selected Answer: CD Upvotes: 1

CD is the correct answer for this.

Comment 3

ID: 1229854 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Thu 13 Jun 2024 15:43 Selected Answer: CD Upvotes: 1

CD is answer.

Comment 4

ID: 1218943 User: reiwanotora Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Sun 26 May 2024 13:05 Selected Answer: CD Upvotes: 2

C and D are right answer.

Comment 5

ID: 1060111 User: _LAW_ Badges: - Relative Date: 2 years, 4 months ago Absolute Date: Thu 02 Nov 2023 00:03 Selected Answer: CD Upvotes: 4

Correct

137. AI-102 Topic 3 Question 13

Sequence
307
Discussion ID
77612
Source URL
https://www.examtopics.com/discussions/microsoft/view/77612-exam-ai-102-topic-3-question-13-discussion/
Posted By
Eltooth
Posted At
July 18, 2022, 7:10 p.m.

Question

You need to upload speech samples to a Speech Studio project for use in training.
How should you upload the samples?

  • A. Combine the speech samples into a single audio file in the .wma format and upload the file.
  • B. Upload a .zip file that contains a collection of audio files in the .wav format and a corresponding text transcript file.
  • C. Upload individual audio files in the FLAC format and manually upload a corresponding transcript in Microsoft Word format.
  • D. Upload individual audio files in the .wma format.

Suggested Answer

B

Answer Description Click to expand


Community Answer Votes

Comments 10 comments Click to expand

Comment 1

ID: 1132989 User: evangelist Badges: Highly Voted Relative Date: 2 years, 1 month ago Absolute Date: Sat 27 Jan 2024 03:46 Selected Answer: B Upvotes: 8

, the best option is B. Upload a .zip file that contains a collection of audio files in the .wav format and a corresponding text transcript file. This method provides a balance of audio quality (with .wav files) and organization (having audio and transcripts together), which is essential for efficient and accurate training of speech recognition models.

Comment 2

ID: 1102031 User: Gvalli Badges: Highly Voted Relative Date: 2 years, 2 months ago Absolute Date: Thu 21 Dec 2023 00:28 Selected Answer: - Upvotes: 6

A modified version of this was in the exam today.

Comment 3

ID: 1229241 User: reigenchimpo Badges: Most Recent Relative Date: 1 year, 9 months ago Absolute Date: Wed 12 Jun 2024 16:23 Selected Answer: B Upvotes: 1

B is answer.

Comment 4

ID: 1224484 User: anto69 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 05 Jun 2024 05:09 Selected Answer: B Upvotes: 1

For me and ChatGPT: B

Comment 5

ID: 1217546 User: nanaw770 Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 16:01 Selected Answer: B Upvotes: 1

wav zip

Comment 6

ID: 1217355 User: funny_penguin Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Fri 24 May 2024 11:38 Selected Answer: B Upvotes: 1

B is correct

Comment 7

ID: 1094133 User: orionduo Badges: - Relative Date: 2 years, 3 months ago Absolute Date: Tue 12 Dec 2023 06:13 Selected Answer: B Upvotes: 2

Answer is correct

Comment 8

ID: 1024595 User: ManvaIT Badges: - Relative Date: 2 years, 5 months ago Absolute Date: Wed 04 Oct 2023 11:01 Selected Answer: - Upvotes: 3

B is the Answer. Got this in Oct2023 exam

Comment 8.1

ID: 1026285 User: JDKJDKJDK Badges: - Relative Date: 2 years, 5 months ago Absolute Date: Fri 06 Oct 2023 07:53 Selected Answer: - Upvotes: 3

just want to confirm, theres no labs included right?

Comment 9

ID: 633102 User: Eltooth Badges: - Relative Date: 3 years, 7 months ago Absolute Date: Mon 18 Jul 2022 19:10 Selected Answer: B Upvotes: 1

B is correct answer.

https://docs.microsoft.com/en-us/azure/cognitive-services/speech-service/how-to-custom-speech-test-and-train

138. AI-102 Topic 3 Question 26

Sequence
308
Discussion ID
88019
Source URL
https://www.examtopics.com/discussions/microsoft/view/88019-exam-ai-102-topic-3-question-26-discussion/
Posted By
halfway
Posted At
Nov. 20, 2022, 3:06 a.m.

Question

SIMULATION -
You need to configure bot12345678 support the French (FR-FR) language.
Export the bot to C:\Resources\Bot\Bot1.zip.
To complete this task, use the Microsoft Bot Framework Composer.

Suggested Answer

image
Answer Description Click to expand


Comments 3 comments Click to expand

Comment 1

ID: 1229248 User: reigenchimpo Badges: - Relative Date: 1 year, 9 months ago Absolute Date: Wed 12 Jun 2024 16:27 Selected Answer: - Upvotes: 3

I saw in the Microsoft question and answer that this question will not be asked, Mr. Big Tits.

Comment 2

ID: 918784 User: ziggy1117 Badges: - Relative Date: 2 years, 9 months ago Absolute Date: Fri 09 Jun 2023 02:33 Selected Answer: - Upvotes: 4

Select the Configure page from the left and then select the Localization tab. Select Manage bot language to choose your bot's languages.

Comment 3

ID: 722337 User: halfway Badges: - Relative Date: 3 years, 3 months ago Absolute Date: Sun 20 Nov 2022 03:06 Selected Answer: - Upvotes: 3

The answer is incorrect. The instructions in the following link should be used to add multi-language support to a bot:
https://learn.microsoft.com/en-us/composer/how-to-use-multiple-language?tabs=v2x#update-language-settings

139. AI-102 Topic 3 Question 73

Sequence
309
Discussion ID
135148
Source URL
https://www.examtopics.com/discussions/microsoft/view/135148-exam-ai-102-topic-3-question-73-discussion/
Posted By
Razvan_C
Posted At
March 4, 2024, 6:44 a.m.

Question

HOTSPOT
-

You are building an app that will automatically translate speech from English to French, German, and Spanish by using Azure AI service.

You need to define the output languages and configure the Azure AI Speech service.

How should you complete the code? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

image

Suggested Answer

image
Answer Description Click to expand


Comments 5 comments Click to expand

Comment 1

ID: 1188988 User: anto69 Badges: Highly Voted Relative Date: 1 year, 11 months ago Absolute Date: Thu 04 Apr 2024 02:19 Selected Answer: - Upvotes: 6

Duplicate

Comment 2

ID: 1229205 User: reigenchimpo Badges: Most Recent Relative Date: 1 year, 9 months ago Absolute Date: Wed 12 Jun 2024 16:07 Selected Answer: - Upvotes: 4

1. fr, de, es
2. TranslationRecognizer

Comment 3

ID: 1177395 User: rober13 Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Tue 19 Mar 2024 15:11 Selected Answer: - Upvotes: 4

Same Question 68

Comment 3.1

ID: 1185583 User: f2c587e Badges: - Relative Date: 1 year, 11 months ago Absolute Date: Fri 29 Mar 2024 17:48 Selected Answer: - Upvotes: 2

The answer options for this question are slightly different from question 68 on topic 3

Comment 4

ID: 1165330 User: Razvan_C Badges: - Relative Date: 2 years ago Absolute Date: Mon 04 Mar 2024 06:44 Selected Answer: - Upvotes: 4

Answer seems to be correct.

140. AI-102 Topic 5 Question 13

Sequence
310
Discussion ID
62268
Source URL
https://www.examtopics.com/discussions/microsoft/view/62268-exam-ai-102-topic-5-question-13-discussion/
Posted By
SnowCheetah
Posted At
Sept. 17, 2021, 10:29 a.m.

Question

HOTSPOT -
You are designing a conversation flow to be used in a chatbot.
You need to test the conversation flow by using the Microsoft Bot Framework Emulator.
How should you complete the .chat file? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Hot Area:
image

Suggested Answer

image
Answer Description Click to expand

Reference:
https://docs.microsoft.com/en-us/azure/bot-service/bot-builder-howto-add-media-attachments?view=azure-bot-service-4.0&tabs=csharp

Comments 5 comments Click to expand

Comment 1

ID: 451675 User: Diem Badges: Highly Voted Relative Date: 3 years, 5 months ago Absolute Date: Mon 26 Sep 2022 07:36 Selected Answer: - Upvotes: 14

The given answer is correct. Second box is Carousel which includes multiple attachments.

Comment 2

ID: 920334 User: ziggy1117 Badges: Most Recent Relative Date: 1 year, 9 months ago Absolute Date: Tue 11 Jun 2024 02:23 Selected Answer: - Upvotes: 1

answer is correct:

In the context of Azure Bot Service, a carousel refers to a visual component that displays a set of cards or items in a horizontally scrollable format. It is commonly used to present multiple options or pieces of information to the user in a visually appealing manner.

A carousel typically consists of multiple cards, where each card represents a specific item, option, or piece of content. Each card within the carousel can contain text, images, buttons, or other interactive elements. Users can scroll through the carousel horizontally to view and interact with different cards.

Comment 3

ID: 446480 User: GMKanon Badges: - Relative Date: 3 years, 5 months ago Absolute Date: Sat 17 Sep 2022 10:50 Selected Answer: - Upvotes: 2

Second box should be thumbnail

Comment 4

ID: 446466 User: SnowCheetah Badges: - Relative Date: 3 years, 5 months ago Absolute Date: Sat 17 Sep 2022 10:29 Selected Answer: - Upvotes: 3

I am not sure on the second answer is correct or not

1. Typing is correct ( it's answer indicate on [delay] ) https://docs.microsoft.com/en-us/azure/bot-service/bot-builder-howto-send-messages?view=azure-bot-service-4.0&tabs=csharp

2. Since attachment is attach with 2 images ==> it cannot be thumbnail, which can only a simple card as a proper response. however I am not sure adaptive card can be selected as well for this choice.

3. Adaptive card is correct ( in context after user send selection, bot is sending detail of selected item ) https://docs.microsoft.com/en-us/azure/bot-service/bot-builder-howto-add-media-attachments?view=azure-bot-service-4.0&tabs=csharp

Comment 4.1

ID: 534224 User: timmayy54 Badges: - Relative Date: 3 years, 1 month ago Absolute Date: Sat 28 Jan 2023 01:36 Selected Answer: - Upvotes: 1

Fully Agree, for 2. having 2 attachments makes Carousel more right than Adaptive, plus Adaptive is last answer and double using the same one is quite rare.

141. AI-102 Topic 1 Question 65

Sequence
320
Discussion ID
135060
Source URL
https://www.examtopics.com/discussions/microsoft/view/135060-exam-ai-102-topic-1-question-65-discussion/
Posted By
audlindr
Posted At
March 2, 2024, 6:45 p.m.

Question

HOTSPOT
-

You plan to deploy a containerized version of an Azure Cognitive Services service that will be used for sentiment analysis.

You configure https://contoso.cognitiveservices.azure.com as the endpoint URI for the service.

You need to run the container on an Azure virtual machine by using Docker.

How should you complete the command? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

image

Suggested Answer

image
Answer Description Click to expand


Comments 3 comments Click to expand

Comment 1

ID: 1165023 User: carcasgon Badges: Highly Voted Relative Date: 2 years ago Absolute Date: Sun 03 Mar 2024 20:10 Selected Answer: - Upvotes: 16

Correct. https://learn.microsoft.com/en-us/azure/ai-services/language-service/sentiment-opinion-mining/how-to/use-containers

Comment 2

ID: 1214011 User: dragongoseki Badges: Highly Voted Relative Date: 1 year, 9 months ago Absolute Date: Mon 20 May 2024 00:36 Selected Answer: - Upvotes: 8

I looked similar question before.

Comment 3

ID: 1194124 User: michaelmorar Badges: Most Recent Relative Date: 1 year, 11 months ago Absolute Date: Fri 12 Apr 2024 05:45 Selected Answer: - Upvotes: 3

I agree with the answer. The first parameter indicates the service type, and the second is specifically named "billing" so should point to our service instance.