@xdev-asia/xdev-knowledge-mcp 1.0.42 → 1.0.44

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (32) hide show
  1. package/content/pages/xoa-du-lieu-nguoi-dung.md +68 -0
  2. package/content/series/luyen-thi/luyen-thi-aws-ml-specialty/chapters/01-phan-1-data-engineering/lessons/01-bai-1-data-repositories-ingestion.md +198 -0
  3. package/content/series/luyen-thi/luyen-thi-aws-ml-specialty/chapters/01-phan-1-data-engineering/lessons/02-bai-2-data-transformation.md +183 -0
  4. package/content/series/luyen-thi/luyen-thi-aws-ml-specialty/chapters/01-phan-1-data-engineering/lessons/03-bai-3-data-analysis.md +159 -0
  5. package/content/series/luyen-thi/luyen-thi-aws-ml-specialty/chapters/02-phan-2-modeling/lessons/04-bai-4-sagemaker-built-in-algorithms.md +186 -0
  6. package/content/series/luyen-thi/luyen-thi-aws-ml-specialty/chapters/02-phan-2-modeling/lessons/05-bai-5-training-hyperparameter-tuning.md +159 -0
  7. package/content/series/luyen-thi/luyen-thi-aws-ml-specialty/chapters/02-phan-2-modeling/lessons/06-bai-6-model-evaluation.md +169 -0
  8. package/content/series/luyen-thi/luyen-thi-aws-ml-specialty/chapters/03-phan-3-implementation-operations/lessons/07-bai-7-model-deployment.md +193 -0
  9. package/content/series/luyen-thi/luyen-thi-aws-ml-specialty/chapters/03-phan-3-implementation-operations/lessons/08-bai-8-model-monitoring-mlops.md +184 -0
  10. package/content/series/luyen-thi/luyen-thi-aws-ml-specialty/chapters/03-phan-3-implementation-operations/lessons/09-bai-9-security-cost.md +166 -0
  11. package/content/series/luyen-thi/luyen-thi-aws-ml-specialty/chapters/04-phan-4-on-tap/lessons/10-bai-10-bai-toan-thuong-gap.md +181 -0
  12. package/content/series/luyen-thi/luyen-thi-aws-ml-specialty/chapters/04-phan-4-on-tap/lessons/11-bai-11-cheat-sheet.md +110 -0
  13. package/content/series/luyen-thi/luyen-thi-aws-ml-specialty/chapters/04-phan-4-on-tap/lessons/12-bai-12-chien-luoc-thi.md +113 -0
  14. package/content/series/luyen-thi/luyen-thi-aws-ml-specialty/index.md +1 -1
  15. package/content/series/luyen-thi/luyen-thi-cka/index.md +217 -0
  16. package/content/series/luyen-thi/luyen-thi-ckad/index.md +199 -0
  17. package/content/series/luyen-thi/luyen-thi-gcp-ml-engineer/chapters/01-phan-1-problem-framing/lessons/01-bai-1-framing-ml-problems.md +136 -0
  18. package/content/series/luyen-thi/luyen-thi-gcp-ml-engineer/chapters/01-phan-1-problem-framing/lessons/02-bai-2-gcp-ai-ml-ecosystem.md +160 -0
  19. package/content/series/luyen-thi/luyen-thi-gcp-ml-engineer/chapters/02-phan-2-data-engineering/lessons/03-bai-3-data-pipeline.md +174 -0
  20. package/content/series/luyen-thi/luyen-thi-gcp-ml-engineer/chapters/02-phan-2-data-engineering/lessons/04-bai-4-feature-engineering.md +156 -0
  21. package/content/series/luyen-thi/luyen-thi-gcp-ml-engineer/chapters/03-phan-3-model-development/lessons/05-bai-5-vertex-ai-training.md +155 -0
  22. package/content/series/luyen-thi/luyen-thi-gcp-ml-engineer/chapters/03-phan-3-model-development/lessons/06-bai-6-bigquery-ml-tensorflow.md +141 -0
  23. package/content/series/luyen-thi/luyen-thi-gcp-ml-engineer/chapters/04-phan-4-deployment-mlops/lessons/07-bai-7-model-deployment.md +134 -0
  24. package/content/series/luyen-thi/luyen-thi-gcp-ml-engineer/chapters/04-phan-4-deployment-mlops/lessons/08-bai-8-vertex-ai-pipelines-mlops.md +149 -0
  25. package/content/series/luyen-thi/luyen-thi-gcp-ml-engineer/chapters/05-phan-5-responsible-ai/lessons/09-bai-9-responsible-ai.md +128 -0
  26. package/content/series/luyen-thi/luyen-thi-gcp-ml-engineer/chapters/05-phan-5-responsible-ai/lessons/10-bai-10-cheat-sheet-chien-luoc-thi.md +108 -0
  27. package/content/series/luyen-thi/luyen-thi-gcp-ml-engineer/index.md +1 -1
  28. package/content/series/luyen-thi/luyen-thi-kcna/index.md +168 -0
  29. package/data/quizzes/aws-ai-practitioner.json +362 -0
  30. package/data/quizzes/aws-ml-specialty.json +200 -0
  31. package/data/quizzes/gcp-ml-engineer.json +200 -0
  32. package/package.json +1 -1
@@ -0,0 +1,134 @@
1
+ ---
2
+ id: 019c9619-lt03-l07
3
+ title: 'Bài 7: Model Deployment & Prediction'
4
+ slug: bai-7-model-deployment
5
+ description: >-
6
+ Vertex AI Endpoints: online, batch prediction.
7
+ Model versioning, traffic splitting. Edge deployment.
8
+ Scaling config, GPU allocation.
9
+ duration_minutes: 60
10
+ is_free: true
11
+ video_url: null
12
+ sort_order: 7
13
+ section_title: "Phần 4: Model Deployment & MLOps"
14
+ course:
15
+ id: 019c9619-lt03-7003-c003-lt0300000003
16
+ title: 'Luyện thi Google Cloud Professional Machine Learning Engineer'
17
+ slug: luyen-thi-gcp-ml-engineer
18
+ ---
19
+
20
+ <div style="text-align: center; margin: 2rem 0;">
21
+ <img src="/storage/uploads/2026/04/gcp-mle-bai7-deployment.png" alt="Vertex AI Model Deployment" style="max-width: 800px; width: 100%; border-radius: 12px;" />
22
+ <p><em>Vertex AI Deployment: Online Prediction, Batch Prediction, traffic splitting, và edge deployment</em></p>
23
+ </div>
24
+
25
+ <h2 id="deployment-types"><strong>1. Prediction Types on Vertex AI</strong></h2>
26
+
27
+ <table>
28
+ <thead><tr><th>Type</th><th>Latency</th><th>When to Use</th></tr></thead>
29
+ <tbody>
30
+ <tr><td><strong>Online Prediction</strong></td><td>Milliseconds (sync)</td><td>Real-time apps, user-facing APIs</td></tr>
31
+ <tr><td><strong>Batch Prediction</strong></td><td>Minutes/Hours (async)</td><td>Large datasets, scheduled scoring</td></tr>
32
+ <tr><td><strong>Streaming Prediction</strong></td><td>Near real-time</td><td>Pub/Sub events + Dataflow + Vertex AI</td></tr>
33
+ </tbody>
34
+ </table>
35
+
36
+ <h2 id="vertex-endpoints"><strong>2. Vertex AI Endpoints</strong></h2>
37
+
38
+ <pre><code class="language-text">Vertex AI Endpoint Architecture:
39
+
40
+ Client Request
41
+
42
+ Vertex AI Endpoint (load balancer)
43
+ ├── Model Version A (70% traffic)
44
+ │ └── Deployed Model (e.g., v1.0)
45
+ └── Model Version B (30% traffic) ← Canary/A-B test
46
+ └── Deployed Model (e.g., v1.1)
47
+ </code></pre>
48
+
49
+ <p>Mỗi Endpoint có thể có <strong>nhiều model versions</strong> với <strong>traffic splitting</strong> — dùng để A/B testing và canary deployments.</p>
50
+
51
+ <table>
52
+ <thead><tr><th>Feature</th><th>Details</th></tr></thead>
53
+ <tbody>
54
+ <tr><td><strong>Dedicated Endpoint</strong></td><td>Dedicated resources, lowest latency, higher cost</td></tr>
55
+ <tr><td><strong>Shared Endpoint</strong></td><td>Multi-tenant, lower cost, potential cold start</td></tr>
56
+ <tr><td><strong>Explanation</strong></td><td>Enable Vertex Explainability per deployed model</td></tr>
57
+ <tr><td><strong>Min/Max Replicas</strong></td><td>Autoscaling based on request rate</td></tr>
58
+ <tr><td><strong>GPU allocation</strong></td><td>Specify GPU type (NVIDIA T4, A100) per deployment</td></tr>
59
+ </tbody>
60
+ </table>
61
+
62
+ <blockquote>
63
+ <p><strong>Exam tip:</strong> Traffic splitting trong Vertex AI Endpoints là cách triển khai <strong>Canary deployment</strong> hoặc <strong>A/B testing</strong>. Câu hỏi "roll out new model version safely" → Traffic splitting (ví dụ: 90% old, 10% new).</p>
64
+ </blockquote>
65
+
66
+ <h2 id="batch-prediction"><strong>3. Batch Prediction</strong></h2>
67
+
68
+ <table>
69
+ <thead><tr><th>Property</th><th>Value</th></tr></thead>
70
+ <tbody>
71
+ <tr><td><strong>Input</strong></td><td>Cloud Storage (CSV, JSON, JSONL, TFRecords, Avro)</td></tr>
72
+ <tr><td><strong>Output</strong></td><td>Cloud Storage (predictions as JSON/CSV)</td></tr>
73
+ <tr><td><strong>No Endpoint needed</strong></td><td>Runs directly from Model Registry, no persistent endpoint</td></tr>
74
+ <tr><td><strong>Auto-scaling</strong></td><td>Scales to zero when done (cost-efficient)</td></tr>
75
+ <tr><td><strong>Accelerators</strong></td><td>Supports GPU/TPU for batch inference</td></tr>
76
+ </tbody>
77
+ </table>
78
+
79
+ <h2 id="model-versioning"><strong>4. Model Versioning & Registry</strong></h2>
80
+
81
+ <pre><code class="language-text">Vertex AI Model Registry:
82
+
83
+ Model: churn-predictor
84
+ ├── v1 (Logistic Regression) ← Champion in production
85
+ │ - Accuracy: 0.87
86
+ │ - Deployed to: endpoint/prod (70% traffic)
87
+
88
+ └── v2 (XGBoost) ← Challenger
89
+ - Accuracy: 0.91
90
+ - Deployed to: endpoint/prod (30% traffic)
91
+
92
+ After validation: promote v2 to Champion
93
+ </code></pre>
94
+
95
+ <h2 id="edge-deployment"><strong>5. Edge Deployment</strong></h2>
96
+
97
+ <table>
98
+ <thead><tr><th>Platform</th><th>Solution</th></tr></thead>
99
+ <tbody>
100
+ <tr><td>Mobile (Android/iOS)</td><td>TFLite + Vertex AI model export</td></tr>
101
+ <tr><td>Edge devices (IoT)</td><td>TFLite Micro / Edge TPU (Coral)</td></tr>
102
+ <tr><td>On-premise servers</td><td>TF Serving in Docker container</td></tr>
103
+ <tr><td>Kubernetes</td><td>KServe (formerly KFServing) on GKE</td></tr>
104
+ </tbody>
105
+ </table>
106
+
107
+ <h2 id="practice"><strong>6. Practice Questions</strong></h2>
108
+
109
+ <p><strong>Q1:</strong> A company needs to score 50 million customer records for churn risk. Results are needed within 2 hours but not in real time. Which Vertex AI prediction option is MOST cost-effective?</p>
110
+ <ul>
111
+ <li>A) Online Prediction with high replica count</li>
112
+ <li>B) Batch Prediction ✓</li>
113
+ <li>C) Streaming prediction via Dataflow</li>
114
+ <li>D) Deploy on dedicated GPU endpoint</li>
115
+ </ul>
116
+ <p><em>Explanation: Batch Prediction is designed for large-scale asynchronous scoring. It scales compute resources up during the job and back to zero when done, with no persistent endpoint cost. Online Prediction would be wasteful since real-time response isn't needed for batch scoring.</em></p>
117
+
118
+ <p><strong>Q2:</strong> A team is deploying a new model version. They want to gradually route 10% of production traffic to the new version while the old version handles 90%, allowing comparison of performance metrics before full rollout. Which Vertex AI feature enables this?</p>
119
+ <ul>
120
+ <li>A) Model Registry versioning</li>
121
+ <li>B) Traffic splitting on Vertex AI Endpoints ✓</li>
122
+ <li>C) Batch Prediction comparison</li>
123
+ <li>D) Vertex AI Experiments</li>
124
+ </ul>
125
+ <p><em>Explanation: Vertex AI Endpoints support deploying multiple model versions simultaneously with configurable traffic splits (e.g., 90%/10%). This enables canary deployments and A/B testing to compare live performance before committing to a full rollout.</em></p>
126
+
127
+ <p><strong>Q3:</strong> A retail company wants to detect product defects on a factory floor without network connectivity to cloud. Which deployment approach should they use?</p>
128
+ <ul>
129
+ <li>A) Vertex AI Online Prediction Endpoint</li>
130
+ <li>B) AutoML Edge Model deployed to device using TFLite ✓</li>
131
+ <li>C) BigQuery ML batch prediction</li>
132
+ <li>D) TF Serving on Cloud Run</li>
133
+ </ul>
134
+ <p><em>Explanation: Edge deployment with TFLite (or AutoML Edge Model) runs inference locally on the device without network connectivity. TFLite supports on-device inference for computer vision models, suitable for factory floor equipment with no internet access.</em></p>
@@ -0,0 +1,149 @@
1
+ ---
2
+ id: 019c9619-lt03-l08
3
+ title: 'Bài 8: Vertex AI Pipelines & MLOps'
4
+ slug: bai-8-vertex-ai-pipelines-mlops
5
+ description: >-
6
+ Vertex AI Pipelines (Kubeflow Pipelines SDK).
7
+ Model Registry, Experiments, Metadata Store.
8
+ Vertex AI Model Monitoring: skew, drift detection.
9
+ CI/CD cho ML: Cloud Build + Vertex AI.
10
+ duration_minutes: 60
11
+ is_free: true
12
+ video_url: null
13
+ sort_order: 8
14
+ section_title: "Phần 4: Model Deployment & MLOps"
15
+ course:
16
+ id: 019c9619-lt03-7003-c003-lt0300000003
17
+ title: 'Luyện thi Google Cloud Professional Machine Learning Engineer'
18
+ slug: luyen-thi-gcp-ml-engineer
19
+ ---
20
+
21
+ <div style="text-align: center; margin: 2rem 0;">
22
+ <img src="/storage/uploads/2026/04/gcp-mle-bai8-mlops-cicd.png" alt="Vertex AI Pipelines & MLOps" style="max-width: 800px; width: 100%; border-radius: 12px;" />
23
+ <p><em>Vertex AI MLOps: Pipelines, CI/CD, Model Registry, và monitoring cho production ML</em></p>
24
+ </div>
25
+
26
+ <h2 id="mlops-maturity"><strong>1. MLOps Maturity Levels</strong></h2>
27
+
28
+ <table>
29
+ <thead><tr><th>Level</th><th>Description</th><th>Automation</th></tr></thead>
30
+ <tbody>
31
+ <tr><td><strong>Level 0</strong></td><td>Manual process, scripts only</td><td>None</td></tr>
32
+ <tr><td><strong>Level 1</strong></td><td>ML pipeline automation, continuous training</td><td>Training pipeline</td></tr>
33
+ <tr><td><strong>Level 2</strong></td><td>Full CI/CD for ML, automated retraining triggers</td><td>Everything</td></tr>
34
+ </tbody>
35
+ </table>
36
+
37
+ <h2 id="vertex-pipelines"><strong>2. Vertex AI Pipelines</strong></h2>
38
+
39
+ <p>Vertex AI Pipelines là managed execution environment cho <strong>Kubeflow Pipelines (KFP)</strong>. Pipeline được định nghĩa bằng Python SDK và compile thành YAML.</p>
40
+
41
+ <pre><code class="language-text">Vertex AI Pipeline Structure:
42
+
43
+ @component (preprocess_data)
44
+
45
+ @component (train_model)
46
+
47
+ @component (evaluate_model)
48
+ ↓ (if accuracy > threshold)
49
+ @component (deploy_model)
50
+
51
+ Each component = isolated Docker container
52
+ Artifacts (data, models) stored in Cloud Storage
53
+ Metadata tracked in Vertex ML Metadata Store
54
+ </code></pre>
55
+
56
+ <table>
57
+ <thead><tr><th>Pipeline SDK</th><th>Notes</th></tr></thead>
58
+ <tbody>
59
+ <tr><td><strong>Kubeflow Pipelines SDK v2</strong></td><td>Primary SDK for Vertex AI Pipelines</td></tr>
60
+ <tr><td><strong>TFX</strong></td><td>TensorFlow-specific pipeline components</td></tr>
61
+ <tr><td><strong>Google Cloud Pipeline Components</strong></td><td>Pre-built components cho Vertex AI services</td></tr>
62
+ </tbody>
63
+ </table>
64
+
65
+ <h2 id="model-monitoring"><strong>3. Vertex AI Model Monitoring</strong></h2>
66
+
67
+ <table>
68
+ <thead><tr><th>Monitoring Type</th><th>What It Detects</th></tr></thead>
69
+ <tbody>
70
+ <tr><td><strong>Feature Skew Monitoring</strong></td><td>Serving feature distribution ≠ training baseline</td></tr>
71
+ <tr><td><strong>Feature Drift Monitoring</strong></td><td>Serving feature distribution changes over time</td></tr>
72
+ <tr><td><strong>Prediction Drift</strong></td><td>Model output distribution changes (indirect label drift)</td></tr>
73
+ </tbody>
74
+ </table>
75
+
76
+ <pre><code class="language-text">Model Monitoring Workflow:
77
+
78
+ Training Data Baseline (BigQuery/GCS)
79
+ ↓ (establish distribution)
80
+ Deploy to Endpoint with Monitoring enabled
81
+ ↓ (collect serving requests)
82
+ Periodic Analysis (hourly/daily)
83
+ ↓ (compare distributions)
84
+ Alert if skew/drift > threshold
85
+
86
+ Retrain trigger → new Pipeline run
87
+ </code></pre>
88
+
89
+ <h2 id="experiments-metadata"><strong>4. Vertex AI Experiments & Metadata</strong></h2>
90
+
91
+ <table>
92
+ <thead><tr><th>Component</th><th>Purpose</th></tr></thead>
93
+ <tbody>
94
+ <tr><td><strong>Vertex AI Experiments</strong></td><td>Track hyperparameters, metrics, artifacts across runs</td></tr>
95
+ <tr><td><strong>ML Metadata Store</strong></td><td>Track lineage: data → model → endpoint</td></tr>
96
+ <tr><td><strong>Vertex AI TensorBoard</strong></td><td>Visualize training metrics (loss, accuracy curves)</td></tr>
97
+ </tbody>
98
+ </table>
99
+
100
+ <h2 id="cicd-ml"><strong>5. CI/CD for ML on GCP</strong></h2>
101
+
102
+ <pre><code class="language-text">ML CI/CD Pipeline on GCP:
103
+
104
+ Code Push to Cloud Source Repositories
105
+
106
+ Cloud Build trigger (CI)
107
+ ├── Unit tests for ML components
108
+ ├── Data validation tests
109
+ └── Build Docker image → push to Artifact Registry
110
+
111
+ Vertex AI Pipeline trigger (CD/CT)
112
+ ├── Data preprocessing
113
+ ├── Model training
114
+ ├── Model evaluation
115
+ └── Conditional deployment → Vertex AI Endpoint
116
+ </code></pre>
117
+
118
+ <blockquote>
119
+ <p><strong>Exam tip:</strong> CI/CD cho ML = Cloud Build (code testing + Docker build) + Vertex AI Pipelines (training + deployment orchestration). Cloud Source Repositories là GCP's Git hosting. Artifact Registry thay thế Container Registry để lưu Docker images.</p>
120
+ </blockquote>
121
+
122
+ <h2 id="practice"><strong>6. Practice Questions</strong></h2>
123
+
124
+ <p><strong>Q1:</strong> A production ML model's prediction distribution has shifted significantly over 3 weeks, but ground truth labels are not yet available to measure accuracy directly. Which Vertex AI monitoring type detects this?</p>
125
+ <ul>
126
+ <li>A) Feature Skew Monitoring</li>
127
+ <li>B) Prediction Drift Monitoring ✓</li>
128
+ <li>C) Training data validation</li>
129
+ <li>D) Vertex AI Experiments baseline comparison</li>
130
+ </ul>
131
+ <p><em>Explanation: Prediction Drift Monitoring tracks how the model's output distribution changes over time, serving as an indirect signal of model degradation even when ground truth labels are unavailable. Feature Skew compares serving vs training feature distributions (requires known training baseline).</em></p>
132
+
133
+ <p><strong>Q2:</strong> A team is building a Vertex AI Pipeline that includes data preprocessing, model training, and deployment. They need to track all inputs, outputs, and model artifacts for auditability and reproducibility. Which service stores this lineage information?</p>
134
+ <ul>
135
+ <li>A) Cloud Logging</li>
136
+ <li>B) Vertex AI ML Metadata Store ✓</li>
137
+ <li>C) Cloud Storage versioning</li>
138
+ <li>D) Vertex AI Experiments dashboard</li>
139
+ </ul>
140
+ <p><em>Explanation: Vertex AI ML Metadata Store (also called Vertex ML Metadata) automatically tracks lineage: which datasets produced which models, which models were deployed to which endpoints, including hyperparameters and evaluation metrics — enabling full provenance tracking.</em></p>
141
+
142
+ <p><strong>Q3:</strong> A company wants to automatically retrain their ML model whenever new training data is available in Cloud Storage. The retraining should run a Vertex AI Pipeline and deploy if metrics pass thresholds. Which GCP service should trigger the pipeline?</p>
143
+ <ul>
144
+ <li>A) Vertex AI Schedules</li>
145
+ <li>B) Cloud Storage notifications + Cloud Functions/Eventarc → Vertex AI Pipelines ✓</li>
146
+ <li>C) BigQuery scheduled queries</li>
147
+ <li>D) Cloud Scheduler alone</li>
148
+ </ul>
149
+ <p><em>Explanation: Cloud Storage object finalize notifications can trigger Cloud Functions or Eventarc, which then programmatically start a Vertex AI Pipeline run. This creates event-driven continuous training (MLOps Level 1). Cloud Scheduler triggers on time, not on data availability.</em></p>
@@ -0,0 +1,128 @@
1
+ ---
2
+ id: 019c9619-lt03-l09
3
+ title: 'Bài 9: Responsible AI & Security'
4
+ slug: bai-9-responsible-ai
5
+ description: >-
6
+ Google Responsible AI principles. Vertex AI Explainability (SHAP, IG).
7
+ Fairness indicators. Privacy: differential privacy, federated learning.
8
+ IAM, VPC-SC, CMEK cho ML workloads.
9
+ duration_minutes: 50
10
+ is_free: true
11
+ video_url: null
12
+ sort_order: 9
13
+ section_title: "Phần 5: Responsible AI & Ôn tập"
14
+ course:
15
+ id: 019c9619-lt03-7003-c003-lt0300000003
16
+ title: 'Luyện thi Google Cloud Professional Machine Learning Engineer'
17
+ slug: luyen-thi-gcp-ml-engineer
18
+ ---
19
+
20
+ <h2 id="responsible-ai"><strong>1. Google's Responsible AI Principles</strong></h2>
21
+
22
+ <table>
23
+ <thead><tr><th>Principle</th><th>Key Requirement</th></tr></thead>
24
+ <tbody>
25
+ <tr><td><strong>Socially Beneficial</strong></td><td>Benefits society and individuals</td></tr>
26
+ <tr><td><strong>Avoid Unfair Bias</strong></td><td>Test fairness across demographic groups</td></tr>
27
+ <tr><td><strong>Safety</strong></td><td>Test across diverse scenarios, continuous evaluation</td></tr>
28
+ <tr><td><strong>Accountable</strong></td><td>Appropriate human oversight and control</td></tr>
29
+ <tr><td><strong>Privacy Preserving</strong></td><td>Protect training data privacy</td></tr>
30
+ <tr><td><strong>Scientific Excellence</strong></td><td>Rigorous research standards</td></tr>
31
+ <tr><td><strong>Available for Beneficial Uses</strong></td><td>Primary benefit criteria</td></tr>
32
+ </tbody>
33
+ </table>
34
+
35
+ <h2 id="explainability"><strong>2. Vertex AI Explainability</strong></h2>
36
+
37
+ <p>Vertex AI Explainability cung cấp feature attribution scores — giải thích tại sao model đưa ra prediction nào đó.</p>
38
+
39
+ <table>
40
+ <thead><tr><th>Method</th><th>For</th><th>How</th></tr></thead>
41
+ <tbody>
42
+ <tr><td><strong>SHAP (Shapley Values)</strong></td><td>Tabular models</td><td>Game theory: contribution của mỗi feature</td></tr>
43
+ <tr><td><strong>Integrated Gradients (IG)</strong></td><td>Neural networks (image, text)</td><td>Gradient accumulation from baseline to input</td></tr>
44
+ <tr><td><strong>XRAI</strong></td><td>Image models</td><td>Pixel-region attribution (better UX than IG)</td></tr>
45
+ <tr><td><strong>Sampled Shapley</strong></td><td>Large tabular datasets</td><td>Approximate SHAP, faster</td></tr>
46
+ </tbody>
47
+ </table>
48
+
49
+ <blockquote>
50
+ <p><strong>Exam tip:</strong> "Explain why a loan was denied" → SHAP for tabular models. "Highlight which image regions drove classification" → Integrated Gradients or XRAI. Vertex AI Explainability phải được enable lúc deploy endpoint.</p>
51
+ </blockquote>
52
+
53
+ <h2 id="fairness"><strong>3. Fairness & Bias Detection</strong></h2>
54
+
55
+ <table>
56
+ <thead><tr><th>Tool/Concept</th><th>Description</th></tr></thead>
57
+ <tbody>
58
+ <tr><td><strong>Fairness Indicators</strong></td><td>GCP tool: evaluate model fairness metrics across demographic slices</td></tr>
59
+ <tr><td><strong>What-If Tool</strong></td><td>Interactive exploration of model behavior, counterfactuals</td></tr>
60
+ <tr><td><strong>Demographic parity</strong></td><td>Model predicts same rate across demographic groups</td></tr>
61
+ <tr><td><strong>Equal opportunity</strong></td><td>Same recall/TPR across groups</td></tr>
62
+ <tr><td><strong>Data slice evaluation</strong></td><td>Evaluate metrics per gender, race, age in TFX Evaluator</td></tr>
63
+ </tbody>
64
+ </table>
65
+
66
+ <h2 id="privacy"><strong>4. Privacy Techniques</strong></h2>
67
+
68
+ <table>
69
+ <thead><tr><th>Technique</th><th>Description</th></tr></thead>
70
+ <tbody>
71
+ <tr><td><strong>Differential Privacy</strong></td><td>Add statistical noise to training data/model, prevents individual data re-identification</td></tr>
72
+ <tr><td><strong>Federated Learning</strong></td><td>Train on distributed data without centralizing raw data — model updates only</td></tr>
73
+ <tr><td><strong>Data Anonymization</strong></td><td>Remove PII before training (Cloud DLP API)</td></tr>
74
+ </tbody>
75
+ </table>
76
+
77
+ <h2 id="security"><strong>5. Security Controls for ML Workloads</strong></h2>
78
+
79
+ <table>
80
+ <thead><tr><th>Control</th><th>Purpose</th></tr></thead>
81
+ <tbody>
82
+ <tr><td><strong>IAM roles</strong></td><td>Least-privilege access for ML service accounts</td></tr>
83
+ <tr><td><strong>VPC Service Controls (VPC-SC)</strong></td><td>Security perimeter: prevent data exfiltration from BigQuery, GCS</td></tr>
84
+ <tr><td><strong>CMEK (Customer-Managed Encryption Keys)</strong></td><td>Control encryption keys via Cloud KMS</td></tr>
85
+ <tr><td><strong>Private IP for Vertex AI</strong></td><td>Training and endpoints use private networking</td></tr>
86
+ <tr><td><strong>Cloud Audit Logs</strong></td><td>Who accessed what data, when (Data Access + Admin Activity)</td></tr>
87
+ </tbody>
88
+ </table>
89
+
90
+ <pre><code class="language-text">VPC Service Controls Perimeter:
91
+
92
+ ┌────── Security Perimeter ─────────┐
93
+ │ BigQuery │ Cloud Storage │
94
+ │ Vertex AI │ Cloud KMS │
95
+ │ Dataflow │ Secret Manager │
96
+ └──────────────────────────────────┘
97
+ │ (no exfiltration outside perimeter)
98
+ ✗ Unauthorized access blocked
99
+ </code></pre>
100
+
101
+ <h2 id="practice"><strong>6. Practice Questions</strong></h2>
102
+
103
+ <p><strong>Q1:</strong> A financial services company deployed a loan approval ML model. Regulators require the company to explain why specific loan applications were denied. Which Vertex AI feature provides per-prediction feature importance scores for tabular models?</p>
104
+ <ul>
105
+ <li>A) Vertex AI Experiments</li>
106
+ <li>B) Vertex AI Explainability with SHAP ✓</li>
107
+ <li>C) Vertex AI Model Monitoring</li>
108
+ <li>D) Fairness Indicators</li>
109
+ </ul>
110
+ <p><em>Explanation: Vertex AI Explainability with Shapley Values (SHAP) assigns an importance score to each feature for each individual prediction, explaining why a specific loan was denied by attributing the model's decision to specific input features like credit_score, income, debt_ratio.</em></p>
111
+
112
+ <p><strong>Q2:</strong> A healthcare company needs to train ML models on patient data distributed across multiple hospitals. Data privacy regulations prohibit centralizing raw patient records. Which privacy-preserving ML approach should they use?</p>
113
+ <ul>
114
+ <li>A) Differential Privacy with central training</li>
115
+ <li>B) Federated Learning ✓</li>
116
+ <li>C) Data anonymization + BigQuery ML</li>
117
+ <li>D) Cloud DLP de-identification</li>
118
+ </ul>
119
+ <p><em>Explanation: Federated Learning trains models on distributed data without moving raw data to a central location. Each hospital trains locally on its own data; only model updates (gradients) are shared and aggregated. Raw patient records never leave the hospital's environment.</em></p>
120
+
121
+ <p><strong>Q3:</strong> A company processes sensitive financial data in BigQuery for ML training. They need to prevent data from being moved outside an approved security boundary to unauthorized GCP projects. Which GCP feature should they implement?</p>
122
+ <ul>
123
+ <li>A) Cloud KMS CMEK encryption</li>
124
+ <li>B) VPC Service Controls (VPC-SC) perimeter ✓</li>
125
+ <li>C) IAM role deny policies</li>
126
+ <li>D) Cloud Armor WAF</li>
127
+ </ul>
128
+ <p><em>Explanation: VPC Service Controls creates a security perimeter around GCP services (BigQuery, Cloud Storage, Vertex AI). It prevents data exfiltration by blocking requests that would move data outside the defined perimeter, even from authenticated users. CMEK provides encryption control but doesn't prevent exfiltration.</em></p>
@@ -0,0 +1,108 @@
1
+ ---
2
+ id: 019c9619-lt03-l10
3
+ title: 'Bài 10: Cheat Sheet & Chiến lược thi GCP MLE'
4
+ slug: bai-10-cheat-sheet-chien-luoc-thi
5
+ description: >-
6
+ Bảng tổng hợp toàn khoá GCP Professional Machine Learning Engineer.
7
+ GCP service reference, evaluation metrics, domain weights, và chiến lược thi.
8
+ duration_minutes: 40
9
+ is_free: true
10
+ video_url: null
11
+ sort_order: 10
12
+ section_title: "Phần 5: Responsible AI & Ôn tập"
13
+ course:
14
+ id: 019c9619-lt03-7003-c003-lt0300000003
15
+ title: 'Luyện thi Google Cloud Professional Machine Learning Engineer'
16
+ slug: luyen-thi-gcp-ml-engineer
17
+ ---
18
+
19
+ <h2 id="exam-structure"><strong>1. Cấu Trúc Đề Thi GCP Professional ML Engineer</strong></h2>
20
+
21
+ <table>
22
+ <thead><tr><th>Item</th><th>Details</th></tr></thead>
23
+ <tbody>
24
+ <tr><td><strong>Total Questions</strong></td><td>60 câu</td></tr>
25
+ <tr><td><strong>Time Limit</strong></td><td>120 phút (2 giờ)</td></tr>
26
+ <tr><td><strong>Passing Score</strong></td><td>~70% (Google không công bố chính xác)</td></tr>
27
+ <tr><td><strong>Format</strong></td><td>Multiple choice, multiple select</td></tr>
28
+ <tr><td><strong>Validity</strong></td><td>2 năm</td></tr>
29
+ <tr><td><strong>Level</strong></td><td>Professional (intermediate to advanced)</td></tr>
30
+ </tbody>
31
+ </table>
32
+
33
+ <h2 id="domain-weights"><strong>2. Domain Weights</strong></h2>
34
+
35
+ <table>
36
+ <thead><tr><th>Domain</th><th>Weight</th></tr></thead>
37
+ <tbody>
38
+ <tr><td>1. Architecting low-code ML solutions</td><td>~10%</td></tr>
39
+ <tr><td>2. Collaborate within and across teams to manage data and models</td><td>~20%</td></tr>
40
+ <tr><td>3. Scale prototypes into ML models</td><td>~20%</td></tr>
41
+ <tr><td>4. Serve and scale models</td><td>~20%</td></tr>
42
+ <tr><td>5. Automate & orchestrate ML pipelines</td><td>~20%</td></tr>
43
+ <tr><td>6. Monitor ML solutions</td><td>~10%</td></tr>
44
+ </tbody>
45
+ </table>
46
+
47
+ <h2 id="service-cheat-sheet"><strong>3. GCP ML Services Cheat Sheet</strong></h2>
48
+
49
+ <table>
50
+ <thead><tr><th>Task</th><th>GCP Service</th></tr></thead>
51
+ <tbody>
52
+ <tr><td>No-code image classification</td><td>Vertex AI AutoML Image</td></tr>
53
+ <tr><td>SQL-based ML in data warehouse</td><td>BigQuery ML</td></tr>
54
+ <tr><td>Custom TensorFlow/PyTorch training</td><td>Vertex AI Custom Training</td></tr>
55
+ <tr><td>Hyperparameter optimization</td><td>Vertex AI Hyperparameter Tuning (Bayesian)</td></tr>
56
+ <tr><td>Feature consistency training/serving</td><td>Vertex AI Feature Store</td></tr>
57
+ <tr><td>ML workflow orchestration (pipelines)</td><td>Vertex AI Pipelines (KFP)</td></tr>
58
+ <tr><td>Experiment tracking</td><td>Vertex AI Experiments</td></tr>
59
+ <tr><td>Model versioning</td><td>Vertex AI Model Registry</td></tr>
60
+ <tr><td>A/B testing model versions</td><td>Vertex AI Endpoints traffic splitting</td></tr>
61
+ <tr><td>Monitor feature skew/drift</td><td>Vertex AI Model Monitoring</td></tr>
62
+ <tr><td>Explain model predictions</td><td>Vertex AI Explainability (SHAP, IG)</td></tr>
63
+ <tr><td>Real-time event ingestion</td><td>Pub/Sub</td></tr>
64
+ <tr><td>Batch + streaming ETL (unified)</td><td>Dataflow (Apache Beam)</td></tr>
65
+ <tr><td>Spark/Hadoop workloads</td><td>Dataproc</td></tr>
66
+ <tr><td>ML pipeline orchestration (multi-service)</td><td>Cloud Composer (Airflow)</td></tr>
67
+ <tr><td>Natural language analysis (no training)</td><td>Cloud Natural Language API</td></tr>
68
+ <tr><td>Document extraction</td><td>Document AI</td></tr>
69
+ <tr><td>Speech to text</td><td>Cloud Speech-to-Text API</td></tr>
70
+ <tr><td>Prevent data exfiltration</td><td>VPC Service Controls</td></tr>
71
+ <tr><td>Customer-managed encryption</td><td>Cloud KMS (CMEK)</td></tr>
72
+ </tbody>
73
+ </table>
74
+
75
+ <h2 id="traps"><strong>4. Common Exam Traps</strong></h2>
76
+
77
+ <table>
78
+ <thead><tr><th>Trap</th><th>Correct Answer</th></tr></thead>
79
+ <tbody>
80
+ <tr><td>"No ML expertise, image classification"</td><td>AutoML Image (not custom training)</td></tr>
81
+ <tr><td>"Train on data already in BigQuery"</td><td>BigQuery ML (not Vertex AI)</td></tr>
82
+ <tr><td>"Features differ at training vs serving"</td><td>Vertex AI Feature Store (not re-training)</td></tr>
83
+ <tr><td>"Trigger retraining when data arrives"</td><td>GCS notification → Eventarc → Vertex AI Pipeline</td></tr>
84
+ <tr><td>"Explain why model rejected application"</td><td>Vertex AI Explainability (SHAP)</td></tr>
85
+ <tr><td>"Train on distributed hospital data"</td><td>Federated Learning</td></tr>
86
+ <tr><td>"Prevent BigQuery data exfiltration"</td><td>VPC Service Controls</td></tr>
87
+ <tr><td>"Compare model performance across runs"</td><td>Vertex AI Experiments</td></tr>
88
+ </tbody>
89
+ </table>
90
+
91
+ <blockquote>
92
+ <p><strong>Exam tip:</strong> GCP Professional ML Engineer thường hỏi về architecture decisions, không phải API syntax. Key question patterns: "which service BEST fits the requirement", "what is the FIRST step", "which approach requires the LEAST operational overhead". Luôn ưu tiên managed services của GCP khi câu hỏi có "minimal management" hoặc "serverless".</p>
93
+ </blockquote>
94
+
95
+ <h2 id="study-plan"><strong>5. Kế Hoạch Ôn Tập</strong></h2>
96
+
97
+ <table>
98
+ <thead><tr><th>Ngày</th><th>Focus</th></tr></thead>
99
+ <tbody>
100
+ <tr><td>Day 1</td><td>Vertex AI full platform: Training, Pipelines, Endpoints, Monitoring</td></tr>
101
+ <tr><td>Day 2</td><td>Data engineering: Pub/Sub, Dataflow, Dataproc, Cloud Composer</td></tr>
102
+ <tr><td>Day 3</td><td>BigQuery ML + Feature Engineering + Feature Store</td></tr>
103
+ <tr><td>Day 4</td><td>Responsible AI: Explainability, Fairness, Privacy, Security</td></tr>
104
+ <tr><td>Day 5</td><td>Practice exam 1 — identify weak areas</td></tr>
105
+ <tr><td>Day 6</td><td>Review weak areas + Practice exam 2</td></tr>
106
+ <tr><td>Day 7</td><td>Cheat sheet review only</td></tr>
107
+ </tbody>
108
+ </table>
@@ -6,7 +6,7 @@ description: >-
6
6
  Lộ trình ôn tập toàn diện cho kỳ thi Google Cloud Professional Machine Learning
7
7
  Engineer. Vertex AI, BigQuery ML, TFX pipeline, MLOps trên GCP.
8
8
 
9
- featured_image: null
9
+ featured_image: images/blog/gcp-ml-engineer-series-banner.png
10
10
  level: advanced
11
11
  duration_hours: 35
12
12
  lesson_count: 10