@xdev-asia/xdev-knowledge-mcp 1.0.41 → 1.0.42
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/content/series/luyen-thi/luyen-thi-aws-ai-practitioner/chapters/01-domain-1-fundamentals-ai-ml/lessons/01-bai-1-ai-ml-deep-learning-concepts.md +287 -0
- package/content/series/luyen-thi/luyen-thi-aws-ai-practitioner/chapters/01-domain-1-fundamentals-ai-ml/lessons/02-bai-2-ml-lifecycle-aws-services.md +258 -0
- package/content/series/luyen-thi/luyen-thi-aws-ai-practitioner/chapters/02-domain-2-fundamentals-generative-ai/lessons/03-bai-3-generative-ai-foundation-models.md +218 -0
- package/content/series/luyen-thi/luyen-thi-aws-ai-practitioner/chapters/02-domain-2-fundamentals-generative-ai/lessons/04-bai-4-llm-transformers-multimodal.md +232 -0
- package/content/series/luyen-thi/luyen-thi-aws-ai-practitioner/chapters/03-domain-3-applications-foundation-models/lessons/05-bai-5-prompt-engineering-techniques.md +254 -0
- package/content/series/luyen-thi/luyen-thi-aws-ai-practitioner/chapters/03-domain-3-applications-foundation-models/lessons/06-bai-6-rag-vector-databases-knowledge-bases.md +244 -0
- package/content/series/luyen-thi/luyen-thi-aws-ai-practitioner/chapters/03-domain-3-applications-foundation-models/lessons/07-bai-7-fine-tuning-model-customization.md +247 -0
- package/content/series/luyen-thi/luyen-thi-aws-ai-practitioner/chapters/03-domain-3-applications-foundation-models/lessons/08-bai-8-amazon-bedrock-deep-dive.md +276 -0
- package/content/series/luyen-thi/luyen-thi-aws-ai-practitioner/chapters/04-domain-4-responsible-ai/lessons/09-bai-9-responsible-ai-fairness-bias-transparency.md +224 -0
- package/content/series/luyen-thi/luyen-thi-aws-ai-practitioner/chapters/04-domain-4-responsible-ai/lessons/10-bai-10-aws-responsible-ai-tools.md +252 -0
- package/content/series/luyen-thi/luyen-thi-aws-ai-practitioner/chapters/05-domain-5-security-compliance/lessons/11-bai-11-ai-security-data-privacy-compliance.md +279 -0
- package/content/series/luyen-thi/luyen-thi-aws-ai-practitioner/chapters/05-domain-5-security-compliance/lessons/12-bai-12-exam-strategy-cheat-sheet.md +229 -0
- package/content/series/luyen-thi/luyen-thi-aws-ai-practitioner/index.md +257 -0
- package/content/series/luyen-thi/luyen-thi-aws-ml-specialty/index.md +240 -0
- package/content/series/luyen-thi/luyen-thi-gcp-ml-engineer/index.md +225 -0
- package/data/categories.json +16 -4
- package/data/quizzes.json +764 -0
- package/package.json +1 -1
|
@@ -0,0 +1,764 @@
|
|
|
1
|
+
[
|
|
2
|
+
{
|
|
3
|
+
"id": "aws-ai-practitioner",
|
|
4
|
+
"title": "AWS Certified AI Practitioner (AIF-C01)",
|
|
5
|
+
"slug": "aws-ai-practitioner",
|
|
6
|
+
"description": "Practice exam for AWS Certified AI Practitioner — 20 questions covering all 5 domains",
|
|
7
|
+
"icon": "award",
|
|
8
|
+
"provider": "AWS",
|
|
9
|
+
"level": "Foundational",
|
|
10
|
+
"duration_minutes": 30,
|
|
11
|
+
"passing_score": 70,
|
|
12
|
+
"questions_count": 20,
|
|
13
|
+
"tags": [
|
|
14
|
+
"AWS",
|
|
15
|
+
"AI",
|
|
16
|
+
"Cloud",
|
|
17
|
+
"Bedrock",
|
|
18
|
+
"GenAI"
|
|
19
|
+
],
|
|
20
|
+
"series_slug": "luyen-thi-aws-ai-practitioner",
|
|
21
|
+
"domains": [
|
|
22
|
+
{
|
|
23
|
+
"name": "Domain 1: Fundamentals of AI and ML",
|
|
24
|
+
"weight": 20,
|
|
25
|
+
"lessons": [
|
|
26
|
+
{
|
|
27
|
+
"title": "Bài 1: AI, ML & Deep Learning Concepts",
|
|
28
|
+
"slug": "01-bai-1-ai-ml-deep-learning-concepts"
|
|
29
|
+
},
|
|
30
|
+
{
|
|
31
|
+
"title": "Bài 2: ML Lifecycle & AWS AI Services",
|
|
32
|
+
"slug": "02-bai-2-ml-lifecycle-aws-services"
|
|
33
|
+
}
|
|
34
|
+
]
|
|
35
|
+
},
|
|
36
|
+
{
|
|
37
|
+
"name": "Domain 2: Fundamentals of Generative AI",
|
|
38
|
+
"weight": 24,
|
|
39
|
+
"lessons": [
|
|
40
|
+
{
|
|
41
|
+
"title": "Bài 3: Generative AI & Foundation Models",
|
|
42
|
+
"slug": "03-bai-3-generative-ai-foundation-models"
|
|
43
|
+
},
|
|
44
|
+
{
|
|
45
|
+
"title": "Bài 4: LLMs, Transformers & Multi-modal",
|
|
46
|
+
"slug": "04-bai-4-llm-transformers-multimodal"
|
|
47
|
+
}
|
|
48
|
+
]
|
|
49
|
+
},
|
|
50
|
+
{
|
|
51
|
+
"name": "Domain 3: Applications of Foundation Models",
|
|
52
|
+
"weight": 28,
|
|
53
|
+
"lessons": [
|
|
54
|
+
{
|
|
55
|
+
"title": "Bài 5: Prompt Engineering",
|
|
56
|
+
"slug": "05-bai-5-prompt-engineering-techniques"
|
|
57
|
+
},
|
|
58
|
+
{
|
|
59
|
+
"title": "Bài 6: RAG & Knowledge Bases",
|
|
60
|
+
"slug": "06-bai-6-rag-vector-databases-knowledge-bases"
|
|
61
|
+
},
|
|
62
|
+
{
|
|
63
|
+
"title": "Bài 7: Fine-tuning & Model Customization",
|
|
64
|
+
"slug": "07-bai-7-fine-tuning-model-customization"
|
|
65
|
+
},
|
|
66
|
+
{
|
|
67
|
+
"title": "Bài 8: Amazon Bedrock Deep Dive",
|
|
68
|
+
"slug": "08-bai-8-amazon-bedrock-deep-dive"
|
|
69
|
+
}
|
|
70
|
+
]
|
|
71
|
+
},
|
|
72
|
+
{
|
|
73
|
+
"name": "Domain 4: Guidelines for Responsible AI",
|
|
74
|
+
"weight": 14,
|
|
75
|
+
"lessons": [
|
|
76
|
+
{
|
|
77
|
+
"title": "Bài 9: Responsible AI — Fairness & Bias",
|
|
78
|
+
"slug": "09-bai-9-responsible-ai-fairness-bias-transparency"
|
|
79
|
+
},
|
|
80
|
+
{
|
|
81
|
+
"title": "Bài 10: AWS Responsible AI Tools",
|
|
82
|
+
"slug": "10-bai-10-aws-responsible-ai-tools"
|
|
83
|
+
}
|
|
84
|
+
]
|
|
85
|
+
},
|
|
86
|
+
{
|
|
87
|
+
"name": "Domain 5: Security, Compliance & Governance",
|
|
88
|
+
"weight": 14,
|
|
89
|
+
"lessons": [
|
|
90
|
+
{
|
|
91
|
+
"title": "Bài 11: AI Security & Data Privacy",
|
|
92
|
+
"slug": "11-bai-11-ai-security-data-privacy-compliance"
|
|
93
|
+
},
|
|
94
|
+
{
|
|
95
|
+
"title": "Bài 12: Exam Strategy & Cheat Sheet",
|
|
96
|
+
"slug": "12-bai-12-exam-strategy-cheat-sheet"
|
|
97
|
+
}
|
|
98
|
+
]
|
|
99
|
+
}
|
|
100
|
+
],
|
|
101
|
+
"questions": [
|
|
102
|
+
{
|
|
103
|
+
"id": 1,
|
|
104
|
+
"domain": "Domain 2: Fundamentals of Generative AI",
|
|
105
|
+
"question": "What is a Foundation Model?",
|
|
106
|
+
"options": [
|
|
107
|
+
"A model designed for only one specific task",
|
|
108
|
+
"A large AI model pre-trained on broad data that can be adapted to many downstream tasks",
|
|
109
|
+
"A model that only processes structured tabular data",
|
|
110
|
+
"A model trained entirely using Reinforcement Learning"
|
|
111
|
+
],
|
|
112
|
+
"correct": 1,
|
|
113
|
+
"explanation": "A Foundation Model is a large AI model pre-trained on vast, diverse datasets. It can be adapted to many downstream tasks through fine-tuning, RAG, or prompt engineering."
|
|
114
|
+
},
|
|
115
|
+
{
|
|
116
|
+
"id": 2,
|
|
117
|
+
"domain": "Domain 3: Applications of Foundation Models",
|
|
118
|
+
"question": "What is the PRIMARY purpose of Amazon Bedrock?",
|
|
119
|
+
"options": [
|
|
120
|
+
"Managing relational databases",
|
|
121
|
+
"Deploying containers on the cloud",
|
|
122
|
+
"Accessing and using Foundation Models from multiple providers through a single API",
|
|
123
|
+
"Monitoring cloud costs"
|
|
124
|
+
],
|
|
125
|
+
"correct": 2,
|
|
126
|
+
"explanation": "Amazon Bedrock is a fully managed service that provides access to Foundation Models from multiple providers (Anthropic, Meta, Amazon, Mistral, etc.) through a single API for building generative AI applications."
|
|
127
|
+
},
|
|
128
|
+
{
|
|
129
|
+
"id": 3,
|
|
130
|
+
"domain": "Domain 1: Fundamentals of AI and ML",
|
|
131
|
+
"question": "How does Supervised Learning differ from Unsupervised Learning?",
|
|
132
|
+
"options": [
|
|
133
|
+
"Supervised Learning does not require any data",
|
|
134
|
+
"Supervised Learning uses labeled data to train the model",
|
|
135
|
+
"Unsupervised Learning always produces more accurate results",
|
|
136
|
+
"Supervised Learning can only be used for classification tasks"
|
|
137
|
+
],
|
|
138
|
+
"correct": 1,
|
|
139
|
+
"explanation": "Supervised Learning uses labeled data (input-output pairs) to train models for classification or regression, while Unsupervised Learning discovers hidden patterns in unlabeled data (e.g., clustering)."
|
|
140
|
+
},
|
|
141
|
+
{
|
|
142
|
+
"id": 4,
|
|
143
|
+
"domain": "Domain 3: Applications of Foundation Models",
|
|
144
|
+
"question": "What problem does RAG (Retrieval-Augmented Generation) solve for LLMs?",
|
|
145
|
+
"options": [
|
|
146
|
+
"It increases inference speed",
|
|
147
|
+
"It reduces training costs",
|
|
148
|
+
"It reduces hallucination by grounding responses in external knowledge sources",
|
|
149
|
+
"It increases the context window size"
|
|
150
|
+
],
|
|
151
|
+
"correct": 2,
|
|
152
|
+
"explanation": "RAG combines retrieval of relevant external data with generation, helping LLMs produce more accurate, fact-based answers by grounding responses in retrieved documents rather than relying solely on training knowledge."
|
|
153
|
+
},
|
|
154
|
+
{
|
|
155
|
+
"id": 5,
|
|
156
|
+
"domain": "Domain 2: Fundamentals of Generative AI",
|
|
157
|
+
"question": "A customer support chatbot gives inconsistent and overly creative answers to factual questions. Which inference parameter should be adjusted?",
|
|
158
|
+
"options": [
|
|
159
|
+
"Increase temperature to 1.0",
|
|
160
|
+
"Decrease temperature closer to 0",
|
|
161
|
+
"Increase max tokens",
|
|
162
|
+
"Increase top-k to 500"
|
|
163
|
+
],
|
|
164
|
+
"correct": 1,
|
|
165
|
+
"explanation": "Lower temperature values (closer to 0) make the model more deterministic and focused, producing consistent and factual responses. Higher temperature values increase randomness and creativity."
|
|
166
|
+
},
|
|
167
|
+
{
|
|
168
|
+
"id": 6,
|
|
169
|
+
"domain": "Domain 3: Applications of Foundation Models",
|
|
170
|
+
"question": "Which prompting technique is MOST effective for improving a model's accuracy on complex mathematical reasoning tasks?",
|
|
171
|
+
"options": [
|
|
172
|
+
"Zero-shot prompting",
|
|
173
|
+
"Negative prompting",
|
|
174
|
+
"Chain-of-Thought (CoT) prompting",
|
|
175
|
+
"System prompting"
|
|
176
|
+
],
|
|
177
|
+
"correct": 2,
|
|
178
|
+
"explanation": "Chain-of-Thought prompting instructs the model to reason step by step before giving a final answer. This significantly improves accuracy on math, logic, and multi-step reasoning tasks."
|
|
179
|
+
},
|
|
180
|
+
{
|
|
181
|
+
"id": 7,
|
|
182
|
+
"domain": "Domain 3: Applications of Foundation Models",
|
|
183
|
+
"question": "A company wants to build a Q&A assistant that answers questions from internal documents stored in Amazon S3. The documents are updated weekly. Which approach is MOST suitable?",
|
|
184
|
+
"options": [
|
|
185
|
+
"Fine-tune a foundation model on the documents",
|
|
186
|
+
"Use RAG with Amazon Bedrock Knowledge Bases",
|
|
187
|
+
"Pre-train a custom model from scratch",
|
|
188
|
+
"Use zero-shot prompting with a large context window"
|
|
189
|
+
],
|
|
190
|
+
"correct": 1,
|
|
191
|
+
"explanation": "Amazon Bedrock Knowledge Bases provides managed RAG — it automatically chunks, embeds, and indexes S3 documents, retrieves relevant information per query, and stays current via auto-sync without model retraining."
|
|
192
|
+
},
|
|
193
|
+
{
|
|
194
|
+
"id": 8,
|
|
195
|
+
"domain": "Domain 2: Fundamentals of Generative AI",
|
|
196
|
+
"question": "Which Transformer architecture type is BEST suited for text generation tasks such as chatbots and content creation?",
|
|
197
|
+
"options": [
|
|
198
|
+
"Encoder-only (e.g., BERT)",
|
|
199
|
+
"Decoder-only (e.g., GPT, Claude)",
|
|
200
|
+
"Encoder-Decoder (e.g., T5)",
|
|
201
|
+
"Convolutional Neural Network (CNN)"
|
|
202
|
+
],
|
|
203
|
+
"correct": 1,
|
|
204
|
+
"explanation": "Decoder-only architectures (like GPT, Claude, Llama) generate text autoregressively one token at a time and are the basis for most modern chatbots and text generators."
|
|
205
|
+
},
|
|
206
|
+
{
|
|
207
|
+
"id": 9,
|
|
208
|
+
"domain": "Domain 3: Applications of Foundation Models",
|
|
209
|
+
"question": "A retail company wants to build an AI assistant that can check inventory, process returns, and answer product questions from their catalog. Which Amazon Bedrock feature should they use?",
|
|
210
|
+
"options": [
|
|
211
|
+
"Bedrock Guardrails",
|
|
212
|
+
"Bedrock Knowledge Bases only",
|
|
213
|
+
"Bedrock Agents with Action Groups and Knowledge Bases",
|
|
214
|
+
"Bedrock Model Evaluation"
|
|
215
|
+
],
|
|
216
|
+
"correct": 2,
|
|
217
|
+
"explanation": "Bedrock Agents can orchestrate multi-step tasks by calling APIs (action groups for inventory/returns) and retrieving information (knowledge bases for product catalog) — combining reasoning with actions."
|
|
218
|
+
},
|
|
219
|
+
{
|
|
220
|
+
"id": 10,
|
|
221
|
+
"domain": "Domain 3: Applications of Foundation Models",
|
|
222
|
+
"question": "Which technique allows fine-tuning a large language model while updating only a small fraction of the model's parameters?",
|
|
223
|
+
"options": [
|
|
224
|
+
"Full fine-tuning",
|
|
225
|
+
"LoRA (Low-Rank Adaptation)",
|
|
226
|
+
"Continued pre-training",
|
|
227
|
+
"RLHF (Reinforcement Learning from Human Feedback)"
|
|
228
|
+
],
|
|
229
|
+
"correct": 1,
|
|
230
|
+
"explanation": "LoRA is a Parameter-Efficient Fine-Tuning (PEFT) technique that adds small trainable adapter matrices while freezing the original model weights — typically updating less than 1% of total parameters, reducing cost significantly."
|
|
231
|
+
},
|
|
232
|
+
{
|
|
233
|
+
"id": 11,
|
|
234
|
+
"domain": "Domain 4: Guidelines for Responsible AI",
|
|
235
|
+
"question": "What is the PRIMARY purpose of Amazon Bedrock Guardrails?",
|
|
236
|
+
"options": [
|
|
237
|
+
"Accelerating model inference",
|
|
238
|
+
"Implementing safety controls such as content filtering, denied topics, and PII detection for AI applications",
|
|
239
|
+
"Compressing model size for deployment",
|
|
240
|
+
"Managing billing and costs"
|
|
241
|
+
],
|
|
242
|
+
"correct": 1,
|
|
243
|
+
"explanation": "Bedrock Guardrails implement safety controls including content filters (hate, violence, sexual), denied topics, word filters, PII detection/redaction, and contextual grounding checks — applied to both model inputs and outputs."
|
|
244
|
+
},
|
|
245
|
+
{
|
|
246
|
+
"id": 12,
|
|
247
|
+
"domain": "Domain 4: Guidelines for Responsible AI",
|
|
248
|
+
"question": "A hiring AI system consistently ranks male candidates higher than equally qualified female candidates. What is the MOST likely cause?",
|
|
249
|
+
"options": [
|
|
250
|
+
"Measurement bias in data collection",
|
|
251
|
+
"Selection bias in the training data reflecting historical hiring patterns",
|
|
252
|
+
"The model's architecture is too complex",
|
|
253
|
+
"The inference temperature is set too high"
|
|
254
|
+
],
|
|
255
|
+
"correct": 1,
|
|
256
|
+
"explanation": "If training data contained historical hiring decisions that favored male candidates, the model would learn and reproduce that selection bias — the training data didn't represent the qualified population fairly."
|
|
257
|
+
},
|
|
258
|
+
{
|
|
259
|
+
"id": 13,
|
|
260
|
+
"domain": "Domain 4: Guidelines for Responsible AI",
|
|
261
|
+
"question": "Which AWS service can detect bias in ML model predictions and provide per-prediction explainability using SHAP values?",
|
|
262
|
+
"options": [
|
|
263
|
+
"Amazon Rekognition",
|
|
264
|
+
"Amazon SageMaker Clarify",
|
|
265
|
+
"Amazon Bedrock Guardrails",
|
|
266
|
+
"Amazon Comprehend"
|
|
267
|
+
],
|
|
268
|
+
"correct": 1,
|
|
269
|
+
"explanation": "SageMaker Clarify provides pre-training bias detection (data analysis), post-training bias detection (prediction analysis across demographic groups), and model explainability through SHAP values."
|
|
270
|
+
},
|
|
271
|
+
{
|
|
272
|
+
"id": 14,
|
|
273
|
+
"domain": "Domain 4: Guidelines for Responsible AI",
|
|
274
|
+
"question": "A document processing application needs human review when AI-extracted data has low confidence. Which AWS service provides this human-in-the-loop capability?",
|
|
275
|
+
"options": [
|
|
276
|
+
"Amazon SageMaker Ground Truth",
|
|
277
|
+
"Amazon Augmented AI (A2I)",
|
|
278
|
+
"Amazon Mechanical Turk directly",
|
|
279
|
+
"Amazon Bedrock Agents"
|
|
280
|
+
],
|
|
281
|
+
"correct": 1,
|
|
282
|
+
"explanation": "Amazon A2I provides human-in-the-loop workflows with built-in integration for Amazon Textract and Rekognition. It automatically triggers human review when AI confidence falls below a defined threshold."
|
|
283
|
+
},
|
|
284
|
+
{
|
|
285
|
+
"id": 15,
|
|
286
|
+
"domain": "Domain 5: Security, Compliance & Governance",
|
|
287
|
+
"question": "A financial services company wants to ensure Amazon Bedrock API calls do NOT traverse the public internet. What should they configure?",
|
|
288
|
+
"options": [
|
|
289
|
+
"AWS Direct Connect only",
|
|
290
|
+
"VPC endpoint (AWS PrivateLink) for Amazon Bedrock",
|
|
291
|
+
"A VPN connection",
|
|
292
|
+
"Amazon CloudFront distribution"
|
|
293
|
+
],
|
|
294
|
+
"correct": 1,
|
|
295
|
+
"explanation": "A VPC interface endpoint (AWS PrivateLink) for Amazon Bedrock allows private connectivity from within a VPC without any traffic going through the public internet."
|
|
296
|
+
},
|
|
297
|
+
{
|
|
298
|
+
"id": 16,
|
|
299
|
+
"domain": "Domain 5: Security, Compliance & Governance",
|
|
300
|
+
"question": "According to the AWS Shared Responsibility Model, who is responsible for ensuring ML training data does not contain bias?",
|
|
301
|
+
"options": [
|
|
302
|
+
"AWS",
|
|
303
|
+
"The foundation model provider",
|
|
304
|
+
"The customer",
|
|
305
|
+
"Both AWS and the customer equally"
|
|
306
|
+
],
|
|
307
|
+
"correct": 2,
|
|
308
|
+
"explanation": "Under the Shared Responsibility Model, customers are responsible for 'security IN the cloud' — including training data quality, bias detection, model selection, IAM, and ethical AI practices."
|
|
309
|
+
},
|
|
310
|
+
{
|
|
311
|
+
"id": 17,
|
|
312
|
+
"domain": "Domain 5: Security, Compliance & Governance",
|
|
313
|
+
"question": "A chatbot must NEVER reveal customer credit card numbers in responses. Which approach provides the STRONGEST guarantee?",
|
|
314
|
+
"options": [
|
|
315
|
+
"Add 'never output credit card numbers' to the system prompt",
|
|
316
|
+
"Fine-tune the model to avoid outputting PII",
|
|
317
|
+
"Use Amazon Bedrock Guardrails with PII filters set to BLOCK",
|
|
318
|
+
"Remove credit card numbers from the knowledge base"
|
|
319
|
+
],
|
|
320
|
+
"correct": 2,
|
|
321
|
+
"explanation": "Bedrock Guardrails with PII filters provide programmatic detection and blocking of credit card numbers — this cannot be bypassed by prompt injection, unlike system prompts which are soft constraints."
|
|
322
|
+
},
|
|
323
|
+
{
|
|
324
|
+
"id": 18,
|
|
325
|
+
"domain": "Domain 5: Security, Compliance & Governance",
|
|
326
|
+
"question": "A company needs to discover which Amazon S3 buckets contain personally identifiable information (PII) before using the data for ML training. Which AWS service should they use?",
|
|
327
|
+
"options": [
|
|
328
|
+
"Amazon Comprehend",
|
|
329
|
+
"Amazon Macie",
|
|
330
|
+
"Amazon Inspector",
|
|
331
|
+
"AWS Config"
|
|
332
|
+
],
|
|
333
|
+
"correct": 1,
|
|
334
|
+
"explanation": "Amazon Macie uses ML to automatically discover and classify sensitive data (including PII) stored in Amazon S3 buckets. Comprehend detects PII in text at runtime, but Macie is designed for S3-level data discovery."
|
|
335
|
+
},
|
|
336
|
+
{
|
|
337
|
+
"id": 19,
|
|
338
|
+
"domain": "Domain 3: Applications of Foundation Models",
|
|
339
|
+
"question": "A company wants to process 50,000 customer reviews overnight using a foundation model for sentiment analysis. Which Amazon Bedrock pricing model is MOST cost-effective?",
|
|
340
|
+
"options": [
|
|
341
|
+
"On-Demand pricing",
|
|
342
|
+
"Provisioned Throughput",
|
|
343
|
+
"Batch Inference",
|
|
344
|
+
"Free tier"
|
|
345
|
+
],
|
|
346
|
+
"correct": 2,
|
|
347
|
+
"explanation": "Batch Inference is designed for large-scale, non-real-time workloads and offers up to 50% cost savings compared to on-demand pricing. Ideal for processing large datasets overnight."
|
|
348
|
+
},
|
|
349
|
+
{
|
|
350
|
+
"id": 20,
|
|
351
|
+
"domain": "Domain 3: Applications of Foundation Models",
|
|
352
|
+
"question": "A non-technical marketing team wants to experiment with generative AI applications without an AWS account or coding skills. Which AWS service should they use?",
|
|
353
|
+
"options": [
|
|
354
|
+
"Amazon SageMaker Canvas",
|
|
355
|
+
"Amazon Bedrock Console",
|
|
356
|
+
"Amazon PartyRock",
|
|
357
|
+
"Amazon Q Business"
|
|
358
|
+
],
|
|
359
|
+
"correct": 2,
|
|
360
|
+
"explanation": "Amazon PartyRock is a free, no-code playground for generative AI that requires no AWS account. Users can build and share GenAI apps with drag-and-drop — ideal for experimentation and learning."
|
|
361
|
+
}
|
|
362
|
+
]
|
|
363
|
+
},
|
|
364
|
+
{
|
|
365
|
+
"id": "aws-ml-specialty",
|
|
366
|
+
"title": "AWS Certified Machine Learning - Specialty",
|
|
367
|
+
"slug": "aws-ml-specialty",
|
|
368
|
+
"description": "Luyện thi chứng chỉ AWS ML Specialty — build, train, deploy ML trên AWS",
|
|
369
|
+
"icon": "award",
|
|
370
|
+
"provider": "AWS",
|
|
371
|
+
"level": "Chuyên gia",
|
|
372
|
+
"duration_minutes": 180,
|
|
373
|
+
"passing_score": 75,
|
|
374
|
+
"questions_count": 15,
|
|
375
|
+
"tags": [
|
|
376
|
+
"AWS",
|
|
377
|
+
"ML",
|
|
378
|
+
"SageMaker"
|
|
379
|
+
],
|
|
380
|
+
"series_slug": "luyen-thi-aws-ml-specialty",
|
|
381
|
+
"questions": [
|
|
382
|
+
{
|
|
383
|
+
"id": 1,
|
|
384
|
+
"question": "SageMaker built-in algorithm nào phù hợp nhất cho bài toán phát hiện bất thường (anomaly detection)?",
|
|
385
|
+
"options": [
|
|
386
|
+
"XGBoost",
|
|
387
|
+
"Random Cut Forest",
|
|
388
|
+
"BlazingText",
|
|
389
|
+
"DeepAR"
|
|
390
|
+
],
|
|
391
|
+
"correct": 1,
|
|
392
|
+
"explanation": "Random Cut Forest (RCF) là thuật toán unsupervised trong SageMaker, chuyên detect anomaly trong dữ liệu streaming hoặc time series."
|
|
393
|
+
},
|
|
394
|
+
{
|
|
395
|
+
"id": 2,
|
|
396
|
+
"question": "Feature Store trong SageMaker dùng để làm gì?",
|
|
397
|
+
"options": [
|
|
398
|
+
"Lưu trữ mô hình đã train",
|
|
399
|
+
"Quản lý và chia sẻ features giữa các team ML, đảm bảo consistency",
|
|
400
|
+
"Giám sát endpoint inference",
|
|
401
|
+
"Quản lý IAM policies"
|
|
402
|
+
],
|
|
403
|
+
"correct": 1,
|
|
404
|
+
"explanation": "SageMaker Feature Store là kho lưu trữ features centralized, giúp các team ML chia sẻ features, tránh duplicate work, và đảm bảo tính nhất quán giữa training và inference."
|
|
405
|
+
},
|
|
406
|
+
{
|
|
407
|
+
"id": 3,
|
|
408
|
+
"question": "SageMaker sử dụng mode nào để train trên nhiều instance cùng lúc?",
|
|
409
|
+
"options": [
|
|
410
|
+
"Pipe mode",
|
|
411
|
+
"Distributed training mode",
|
|
412
|
+
"File mode",
|
|
413
|
+
"Batch mode"
|
|
414
|
+
],
|
|
415
|
+
"correct": 1,
|
|
416
|
+
"explanation": "SageMaker hỗ trợ distributed training cho phép chia workload training ra nhiều instances (data parallelism hoặc model parallelism) để tăng tốc."
|
|
417
|
+
},
|
|
418
|
+
{
|
|
419
|
+
"id": 4,
|
|
420
|
+
"question": "SageMaker Model Monitor phát hiện loại drift nào?",
|
|
421
|
+
"options": [
|
|
422
|
+
"Chỉ concept drift",
|
|
423
|
+
"Data quality, model quality, bias drift, và feature attribution drift",
|
|
424
|
+
"Chỉ data drift",
|
|
425
|
+
"Chỉ bias drift"
|
|
426
|
+
],
|
|
427
|
+
"correct": 1,
|
|
428
|
+
"explanation": "SageMaker Model Monitor phát hiện 4 loại: Data Quality (thay đổi schema/statistics), Model Quality (độ chính xác giảm), Bias Drift (bias thay đổi), Feature Attribution Drift (feature importance thay đổi)."
|
|
429
|
+
},
|
|
430
|
+
{
|
|
431
|
+
"id": 5,
|
|
432
|
+
"question": "Khi nào nên dùng SageMaker Inference Pipeline?",
|
|
433
|
+
"options": [
|
|
434
|
+
"Khi cần chạy batch transform",
|
|
435
|
+
"Khi cần chain nhiều bước xử lý (preprocessing → model → postprocessing) trong một endpoint",
|
|
436
|
+
"Khi cần train nhiều model",
|
|
437
|
+
"Khi cần A/B testing"
|
|
438
|
+
],
|
|
439
|
+
"correct": 1,
|
|
440
|
+
"explanation": "Inference Pipeline cho phép chain tối đa 15 containers trong một endpoint — ví dụ: data preprocessing → feature engineering → model prediction → postprocessing."
|
|
441
|
+
},
|
|
442
|
+
{
|
|
443
|
+
"id": 6,
|
|
444
|
+
"question": "BlazingText trong SageMaker được dùng cho tác vụ nào?",
|
|
445
|
+
"options": [
|
|
446
|
+
"Object detection",
|
|
447
|
+
"Word2Vec và text classification",
|
|
448
|
+
"Time series forecasting",
|
|
449
|
+
"Recommender systems"
|
|
450
|
+
],
|
|
451
|
+
"correct": 1,
|
|
452
|
+
"explanation": "BlazingText là implementation siêu nhanh của Word2Vec và nhận diện text classification. Nó hỗ trợ training trên multi-GPU với tốc độ rất cao."
|
|
453
|
+
},
|
|
454
|
+
{
|
|
455
|
+
"id": 7,
|
|
456
|
+
"question": "SageMaker Ground Truth được sử dụng để?",
|
|
457
|
+
"options": [
|
|
458
|
+
"Deploy model lên production",
|
|
459
|
+
"Tạo labeled datasets với hỗ trợ của human annotators và active learning",
|
|
460
|
+
"Tối ưu hyperparameter",
|
|
461
|
+
"Giám sát chi phí training"
|
|
462
|
+
],
|
|
463
|
+
"correct": 1,
|
|
464
|
+
"explanation": "Ground Truth là dịch vụ data labeling, kết hợp human annotators (Amazon Mechanical Turk, private team, hoặc vendors) với active learning để giảm chi phí labeling."
|
|
465
|
+
},
|
|
466
|
+
{
|
|
467
|
+
"id": 8,
|
|
468
|
+
"question": "Elastic Inference trong SageMaker dùng để?",
|
|
469
|
+
"options": [
|
|
470
|
+
"Tăng dung lượng storage",
|
|
471
|
+
"Gắn GPU fractional vào instance để giảm chi phí inference",
|
|
472
|
+
"Tự động scale số lượng model",
|
|
473
|
+
"Nén model để deploy nhanh"
|
|
474
|
+
],
|
|
475
|
+
"correct": 1,
|
|
476
|
+
"explanation": "Elastic Inference cho phép gắn GPU acceleration với chi phí thấp vào SageMaker endpoints hoặc notebook instances — chỉ trả tiền cho GPU resource thực sự dùng."
|
|
477
|
+
},
|
|
478
|
+
{
|
|
479
|
+
"id": 9,
|
|
480
|
+
"question": "Chiến lược nào giúp xử lý dữ liệu mất cân bằng (imbalanced dataset)?",
|
|
481
|
+
"options": [
|
|
482
|
+
"Chỉ dùng accuracy làm metric",
|
|
483
|
+
"SMOTE (oversampling), undersampling, class weights, hoặc ensemble methods",
|
|
484
|
+
"Tăng learning rate",
|
|
485
|
+
"Giảm số epoch training"
|
|
486
|
+
],
|
|
487
|
+
"correct": 1,
|
|
488
|
+
"explanation": "Imbalanced data cần kỹ thuật đặc biệt: SMOTE tạo thêm sample cho class thiểu số, undersampling giảm class đa số, hoặc điều chỉnh class weights trong loss function."
|
|
489
|
+
},
|
|
490
|
+
{
|
|
491
|
+
"id": 10,
|
|
492
|
+
"question": "SageMaker Clarify dùng để?",
|
|
493
|
+
"options": [
|
|
494
|
+
"Tối ưu hyperparameter",
|
|
495
|
+
"Phát hiện bias trong dữ liệu và mô hình, giải thích dự đoán (explainability)",
|
|
496
|
+
"Quản lý experiment",
|
|
497
|
+
"Xây dựng data pipeline"
|
|
498
|
+
],
|
|
499
|
+
"correct": 1,
|
|
500
|
+
"explanation": "SageMaker Clarify giúp phát hiện bias trong data và model, cung cấp feature importance (SHAP values), hỗ trợ Responsible AI và regulatory compliance."
|
|
501
|
+
},
|
|
502
|
+
{
|
|
503
|
+
"id": 11,
|
|
504
|
+
"question": "DeepAR trong SageMaker được dùng cho bài toán nào?",
|
|
505
|
+
"options": [
|
|
506
|
+
"Image classification",
|
|
507
|
+
"Dự báo chuỗi thời gian (time series forecasting)",
|
|
508
|
+
"Text summarization",
|
|
509
|
+
"Object detection"
|
|
510
|
+
],
|
|
511
|
+
"correct": 1,
|
|
512
|
+
"explanation": "DeepAR là thuật toán RNN-based cho time series forecasting, đặc biệt hiệu quả khi có nhiều chuỗi thời gian liên quan (cold-start problem)."
|
|
513
|
+
},
|
|
514
|
+
{
|
|
515
|
+
"id": 12,
|
|
516
|
+
"question": "Multi-Model Endpoint trong SageMaker có ưu điểm gì?",
|
|
517
|
+
"options": [
|
|
518
|
+
"Chỉ hỗ trợ GPU instances",
|
|
519
|
+
"Host nhiều model trên cùng một endpoint, giảm chi phí khi có nhiều model ít traffic",
|
|
520
|
+
"Tăng tốc training",
|
|
521
|
+
"Chỉ support PyTorch"
|
|
522
|
+
],
|
|
523
|
+
"correct": 1,
|
|
524
|
+
"explanation": "Multi-Model Endpoint cho phép host hàng trăm model trên cùng endpoint, load model on-demand — tiết kiệm chi phí rất lớn so với mỗi model một endpoint riêng."
|
|
525
|
+
},
|
|
526
|
+
{
|
|
527
|
+
"id": 13,
|
|
528
|
+
"question": "Khi nào nên dùng SageMaker Batch Transform thay vì Real-time Endpoint?",
|
|
529
|
+
"options": [
|
|
530
|
+
"Khi cần inference nhanh, real-time",
|
|
531
|
+
"Khi cần xử lý inference cho dataset lớn không cần response ngay lập tức",
|
|
532
|
+
"Khi cần A/B testing",
|
|
533
|
+
"Khi cần model auto-scaling"
|
|
534
|
+
],
|
|
535
|
+
"correct": 1,
|
|
536
|
+
"explanation": "Batch Transform phù hợp khi cần inference lượng lớn dữ liệu, không cần response real-time — ví dụ: nightly scoring, preprocessing dataset lớn."
|
|
537
|
+
},
|
|
538
|
+
{
|
|
539
|
+
"id": 14,
|
|
540
|
+
"question": "SageMaker Autopilot là gì?",
|
|
541
|
+
"options": [
|
|
542
|
+
"Tool deploy model tự động",
|
|
543
|
+
"AutoML — tự động phân tích data, thử nhiều algorithms, và chọn model tốt nhất",
|
|
544
|
+
"Tool giám sát endpoint",
|
|
545
|
+
"Framework training distributed"
|
|
546
|
+
],
|
|
547
|
+
"correct": 1,
|
|
548
|
+
"explanation": "SageMaker Autopilot là giải pháp AutoML, tự động phân tích data, feature engineering, thử nhiều algorithms/hyperparameters, và đề xuất model tốt nhất — kèm notebook giải thích."
|
|
549
|
+
},
|
|
550
|
+
{
|
|
551
|
+
"id": 15,
|
|
552
|
+
"question": "Pipe Mode trong SageMaker training có lợi ích gì?",
|
|
553
|
+
"options": [
|
|
554
|
+
"Tăng kích thước model",
|
|
555
|
+
"Stream dữ liệu trực tiếp từ S3 vào training container, không cần download toàn bộ trước",
|
|
556
|
+
"Giảm thời gian deploy",
|
|
557
|
+
"Tự động chọn algorithm"
|
|
558
|
+
],
|
|
559
|
+
"correct": 1,
|
|
560
|
+
"explanation": "Pipe Mode stream dữ liệu từ S3 vào training container thay vì copy toàn bộ (File Mode) — giảm startup time và disk requirement, đặc biệt hiệu quả với dataset lớn."
|
|
561
|
+
}
|
|
562
|
+
]
|
|
563
|
+
},
|
|
564
|
+
{
|
|
565
|
+
"id": "gcp-ml-engineer",
|
|
566
|
+
"title": "Google Cloud Professional ML Engineer",
|
|
567
|
+
"slug": "gcp-ml-engineer",
|
|
568
|
+
"description": "Luyện thi chứng chỉ Google Cloud Professional Machine Learning Engineer",
|
|
569
|
+
"icon": "award",
|
|
570
|
+
"provider": "Google Cloud",
|
|
571
|
+
"level": "Chuyên nghiệp",
|
|
572
|
+
"duration_minutes": 120,
|
|
573
|
+
"passing_score": 70,
|
|
574
|
+
"questions_count": 15,
|
|
575
|
+
"tags": [
|
|
576
|
+
"GCP",
|
|
577
|
+
"ML",
|
|
578
|
+
"Vertex AI"
|
|
579
|
+
],
|
|
580
|
+
"series_slug": "luyen-thi-gcp-ml-engineer",
|
|
581
|
+
"questions": [
|
|
582
|
+
{
|
|
583
|
+
"id": 1,
|
|
584
|
+
"question": "Vertex AI Pipeline được xây dựng trên framework nào?",
|
|
585
|
+
"options": [
|
|
586
|
+
"Apache Spark",
|
|
587
|
+
"Kubeflow Pipelines / TFX",
|
|
588
|
+
"Apache Airflow",
|
|
589
|
+
"Jenkins"
|
|
590
|
+
],
|
|
591
|
+
"correct": 1,
|
|
592
|
+
"explanation": "Vertex AI Pipelines dựa trên Kubeflow Pipelines SDK và TFX (TensorFlow Extended), cho phép orchestrate ML workflow trên Google Cloud."
|
|
593
|
+
},
|
|
594
|
+
{
|
|
595
|
+
"id": 2,
|
|
596
|
+
"question": "BigQuery ML cho phép làm gì đặc biệt?",
|
|
597
|
+
"options": [
|
|
598
|
+
"Chỉ query dữ liệu",
|
|
599
|
+
"Train và deploy ML model trực tiếp bằng SQL trong BigQuery",
|
|
600
|
+
"Chỉ export dữ liệu sang CSV",
|
|
601
|
+
"Quản lý Kubernetes cluster"
|
|
602
|
+
],
|
|
603
|
+
"correct": 1,
|
|
604
|
+
"explanation": "BigQuery ML (BQML) cho phép data analysts train model ML bằng SQL quen thuộc ngay trong BigQuery — không cần viết Python hay setup infrastructure riêng."
|
|
605
|
+
},
|
|
606
|
+
{
|
|
607
|
+
"id": 3,
|
|
608
|
+
"question": "Vertex AI Feature Store khác gì so với lưu features trong database thông thường?",
|
|
609
|
+
"options": [
|
|
610
|
+
"Không có gì khác",
|
|
611
|
+
"Hỗ trợ serving features với low-latency, đảm bảo training-serving consistency, và feature versioning",
|
|
612
|
+
"Chỉ hỗ trợ structured data",
|
|
613
|
+
"Chỉ dùng được với TensorFlow"
|
|
614
|
+
],
|
|
615
|
+
"correct": 1,
|
|
616
|
+
"explanation": "Feature Store chuyên biệt cho ML: serving features online (low-latency) và offline (batch), đảm bảo features đồng nhất giữa training và serving, hỗ trợ time-travel và monitoring."
|
|
617
|
+
},
|
|
618
|
+
{
|
|
619
|
+
"id": 4,
|
|
620
|
+
"question": "Khi nào nên dùng AutoML thay vì custom training trên Vertex AI?",
|
|
621
|
+
"options": [
|
|
622
|
+
"Khi cần kiểm soát hoàn toàn architecture",
|
|
623
|
+
"Khi team không có nhiều ML expertise hoặc cần baseline model nhanh",
|
|
624
|
+
"Khi dataset rất lớn (>1TB)",
|
|
625
|
+
"Khi cần distributed training"
|
|
626
|
+
],
|
|
627
|
+
"correct": 1,
|
|
628
|
+
"explanation": "AutoML phù hợp khi cần model nhanh, team có ít ML expertise, hoặc cần baseline. Custom training khi cần kiểm soát architecture, thuật toán đặc thù, hoặc tối ưu sâu."
|
|
629
|
+
},
|
|
630
|
+
{
|
|
631
|
+
"id": 5,
|
|
632
|
+
"question": "Vertex AI Experiments dùng để?",
|
|
633
|
+
"options": [
|
|
634
|
+
"Deploy model lên production",
|
|
635
|
+
"Track, compare và reproduce ML experiments (hyperparameters, metrics, artifacts)",
|
|
636
|
+
"Tạo dataset mới",
|
|
637
|
+
"Quản lý IAM"
|
|
638
|
+
],
|
|
639
|
+
"correct": 1,
|
|
640
|
+
"explanation": "Vertex AI Experiments cung cấp experiment tracking: log hyperparameters, metrics, model artifacts — cho phép compare nhiều runs và reproduce kết quả."
|
|
641
|
+
},
|
|
642
|
+
{
|
|
643
|
+
"id": 6,
|
|
644
|
+
"question": "TFX (TensorFlow Extended) bao gồm những component chính nào?",
|
|
645
|
+
"options": [
|
|
646
|
+
"Chỉ có ExampleGen và Trainer",
|
|
647
|
+
"ExampleGen, StatisticsGen, SchemaGen, ExampleValidator, Transform, Trainer, Evaluator, Pusher",
|
|
648
|
+
"Chỉ có Trainer và Serving",
|
|
649
|
+
"Chỉ có Transform và Evaluator"
|
|
650
|
+
],
|
|
651
|
+
"correct": 1,
|
|
652
|
+
"explanation": "TFX là end-to-end ML platform gồm: ExampleGen (ingest), StatisticsGen + SchemaGen + ExampleValidator (validate), Transform (feature eng), Trainer, Tuner, Evaluator, Pusher (deploy)."
|
|
653
|
+
},
|
|
654
|
+
{
|
|
655
|
+
"id": 7,
|
|
656
|
+
"question": "Vertex AI Model Monitoring kiểm tra điều gì?",
|
|
657
|
+
"options": [
|
|
658
|
+
"Chỉ monitor CPU/memory",
|
|
659
|
+
"Skew (training-serving) và drift (prediction data thay đổi theo thời gian)",
|
|
660
|
+
"Chỉ monitor latency",
|
|
661
|
+
"Chỉ monitor cost"
|
|
662
|
+
],
|
|
663
|
+
"correct": 1,
|
|
664
|
+
"explanation": "Model Monitoring phát hiện: training-serving skew (feature distribution khác nhau) và prediction drift (dữ liệu production drift khỏi baseline), trigger alert khi vượt threshold."
|
|
665
|
+
},
|
|
666
|
+
{
|
|
667
|
+
"id": 8,
|
|
668
|
+
"question": "Google Cloud AI Platform Prediction hỗ trợ chiến lược deploy nào?",
|
|
669
|
+
"options": [
|
|
670
|
+
"Chỉ single model deployment",
|
|
671
|
+
"Traffic splitting cho A/B testing và canary deployments",
|
|
672
|
+
"Chỉ batch prediction",
|
|
673
|
+
"Chỉ edge deployment"
|
|
674
|
+
],
|
|
675
|
+
"correct": 1,
|
|
676
|
+
"explanation": "Vertex AI Prediction hỗ trợ traffic splitting: có thể route % traffic sang model versions khác nhau — phục vụ A/B testing, canary release, và progressive rollout."
|
|
677
|
+
},
|
|
678
|
+
{
|
|
679
|
+
"id": 9,
|
|
680
|
+
"question": "Dataflow trong ML pipeline đóng vai trò gì?",
|
|
681
|
+
"options": [
|
|
682
|
+
"Training model",
|
|
683
|
+
"Xử lý dữ liệu quy mô lớn (batch & streaming) cho data preprocessing/feature engineering",
|
|
684
|
+
"Deploy model",
|
|
685
|
+
"Monitor model"
|
|
686
|
+
],
|
|
687
|
+
"correct": 1,
|
|
688
|
+
"explanation": "Dataflow (dựa trên Apache Beam) xử lý data ở scale lớn: ETL, feature engineering cho cả batch và streaming — bước tiền xử lý quan trọng trong ML pipeline."
|
|
689
|
+
},
|
|
690
|
+
{
|
|
691
|
+
"id": 10,
|
|
692
|
+
"question": "Vertex AI Matching Engine dùng cho bài toán nào?",
|
|
693
|
+
"options": [
|
|
694
|
+
"Training model",
|
|
695
|
+
"Tìm kiếm nearest neighbor (vector similarity search) ở quy mô lớn",
|
|
696
|
+
"Data labeling",
|
|
697
|
+
"Model serving thông thường"
|
|
698
|
+
],
|
|
699
|
+
"correct": 1,
|
|
700
|
+
"explanation": "Matching Engine là managed approximate nearest neighbor (ANN) service — dùng cho similarity search, recommendation, RAG retrieval ở quy mô tỷ vectors."
|
|
701
|
+
},
|
|
702
|
+
{
|
|
703
|
+
"id": 11,
|
|
704
|
+
"question": "Vertex AI Workbench khác gì Colab Enterprise?",
|
|
705
|
+
"options": [
|
|
706
|
+
"Giống hệt nhau",
|
|
707
|
+
"Workbench là JupyterLab managed instances cho ML production, Colab Enterprise cho collaboration và exploration",
|
|
708
|
+
"Workbench chỉ support R",
|
|
709
|
+
"Colab Enterprise chỉ dùng miễn phí"
|
|
710
|
+
],
|
|
711
|
+
"correct": 1,
|
|
712
|
+
"explanation": "Workbench cung cấp JupyterLab managed instances với tích hợp sâu vào GCP services (BigQuery, GCS) cho production ML. Colab Enterprise thiên về collaboration, sharing và exploration."
|
|
713
|
+
},
|
|
714
|
+
{
|
|
715
|
+
"id": 12,
|
|
716
|
+
"question": "Kỹ thuật nào giảm kích thước model để deploy trên edge devices?",
|
|
717
|
+
"options": [
|
|
718
|
+
"Tăng layers",
|
|
719
|
+
"Quantization, pruning, knowledge distillation",
|
|
720
|
+
"Tăng batch size",
|
|
721
|
+
"Dùng thêm GPU"
|
|
722
|
+
],
|
|
723
|
+
"correct": 1,
|
|
724
|
+
"explanation": "Model compression: Quantization (giảm precision: FP32→INT8), Pruning (loại bỏ weights/neurons không quan trọng), Knowledge Distillation (teacher model dạy student model nhỏ hơn)."
|
|
725
|
+
},
|
|
726
|
+
{
|
|
727
|
+
"id": 13,
|
|
728
|
+
"question": "Vertex AI GenAI Studio dùng để?",
|
|
729
|
+
"options": [
|
|
730
|
+
"Chỉ train model từ đầu",
|
|
731
|
+
"Prototyping, testing, và tuning Foundation Models (PaLM, Gemini) trên Google Cloud",
|
|
732
|
+
"Quản lý billing",
|
|
733
|
+
"Giám sát network"
|
|
734
|
+
],
|
|
735
|
+
"correct": 1,
|
|
736
|
+
"explanation": "GenAI Studio cung cấp UI và API để thử nghiệm Foundation Models, prompt design, tuning, và deploy — không cần ML expertise sâu."
|
|
737
|
+
},
|
|
738
|
+
{
|
|
739
|
+
"id": 14,
|
|
740
|
+
"question": "Khi data có nhiều missing values, chiến lược nào phù hợp?",
|
|
741
|
+
"options": [
|
|
742
|
+
"Luôn xoá rows có missing values",
|
|
743
|
+
"Tuỳ context: imputation (mean/median/mode, KNN, model-based), hoặc tạo indicator feature cho missingness",
|
|
744
|
+
"Luôn fill bằng 0",
|
|
745
|
+
"Bỏ qua và train trực tiếp"
|
|
746
|
+
],
|
|
747
|
+
"correct": 1,
|
|
748
|
+
"explanation": "Xử lý missing values tuỳ thuộc vào pattern (MCAR/MAR/MNAR): imputation thống kê (mean/median), model-based (KNN, MICE), hoặc thêm feature indicator. Xoá rows chỉ khi missing ít và MCAR."
|
|
749
|
+
},
|
|
750
|
+
{
|
|
751
|
+
"id": 15,
|
|
752
|
+
"question": "Continuous Training (CT) trong MLOps là gì?",
|
|
753
|
+
"options": [
|
|
754
|
+
"Train model chỉ một lần",
|
|
755
|
+
"Tự động retrain model khi phát hiện trigger (data drift, schedule, hoặc performance degradation)",
|
|
756
|
+
"Train model thủ công hàng tuần",
|
|
757
|
+
"Chỉ dùng cho deep learning"
|
|
758
|
+
],
|
|
759
|
+
"correct": 1,
|
|
760
|
+
"explanation": "Continuous Training tự động kích hoạt retrain pipeline khi: data mới đến (scheduled), data drift vượt threshold, hoặc model performance giảm — đảm bảo model luôn fresh."
|
|
761
|
+
}
|
|
762
|
+
]
|
|
763
|
+
}
|
|
764
|
+
]
|