decision_agent 0.3.0 → 1.0.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (115) hide show
  1. checksums.yaml +4 -4
  2. data/README.md +272 -7
  3. data/lib/decision_agent/agent.rb +72 -1
  4. data/lib/decision_agent/context.rb +1 -0
  5. data/lib/decision_agent/data_enrichment/cache/memory_adapter.rb +86 -0
  6. data/lib/decision_agent/data_enrichment/cache_adapter.rb +49 -0
  7. data/lib/decision_agent/data_enrichment/circuit_breaker.rb +135 -0
  8. data/lib/decision_agent/data_enrichment/client.rb +220 -0
  9. data/lib/decision_agent/data_enrichment/config.rb +78 -0
  10. data/lib/decision_agent/data_enrichment/errors.rb +36 -0
  11. data/lib/decision_agent/decision.rb +102 -2
  12. data/lib/decision_agent/dmn/feel/evaluator.rb +28 -6
  13. data/lib/decision_agent/dsl/condition_evaluator.rb +982 -839
  14. data/lib/decision_agent/dsl/schema_validator.rb +51 -13
  15. data/lib/decision_agent/evaluators/dmn_evaluator.rb +106 -19
  16. data/lib/decision_agent/evaluators/json_rule_evaluator.rb +69 -9
  17. data/lib/decision_agent/explainability/condition_trace.rb +83 -0
  18. data/lib/decision_agent/explainability/explainability_result.rb +52 -0
  19. data/lib/decision_agent/explainability/rule_trace.rb +39 -0
  20. data/lib/decision_agent/explainability/trace_collector.rb +24 -0
  21. data/lib/decision_agent/monitoring/alert_manager.rb +5 -1
  22. data/lib/decision_agent/simulation/errors.rb +18 -0
  23. data/lib/decision_agent/simulation/impact_analyzer.rb +498 -0
  24. data/lib/decision_agent/simulation/monte_carlo_simulator.rb +635 -0
  25. data/lib/decision_agent/simulation/replay_engine.rb +486 -0
  26. data/lib/decision_agent/simulation/scenario_engine.rb +318 -0
  27. data/lib/decision_agent/simulation/scenario_library.rb +163 -0
  28. data/lib/decision_agent/simulation/shadow_test_engine.rb +287 -0
  29. data/lib/decision_agent/simulation/what_if_analyzer.rb +1002 -0
  30. data/lib/decision_agent/simulation.rb +17 -0
  31. data/lib/decision_agent/version.rb +1 -1
  32. data/lib/decision_agent/versioning/activerecord_adapter.rb +23 -8
  33. data/lib/decision_agent/web/public/app.js +119 -0
  34. data/lib/decision_agent/web/public/index.html +49 -0
  35. data/lib/decision_agent/web/public/simulation.html +130 -0
  36. data/lib/decision_agent/web/public/simulation_impact.html +478 -0
  37. data/lib/decision_agent/web/public/simulation_replay.html +551 -0
  38. data/lib/decision_agent/web/public/simulation_shadow.html +546 -0
  39. data/lib/decision_agent/web/public/simulation_whatif.html +532 -0
  40. data/lib/decision_agent/web/public/styles.css +65 -0
  41. data/lib/decision_agent/web/server.rb +594 -23
  42. data/lib/decision_agent.rb +60 -2
  43. metadata +53 -73
  44. data/spec/ab_testing/ab_test_assignment_spec.rb +0 -253
  45. data/spec/ab_testing/ab_test_manager_spec.rb +0 -612
  46. data/spec/ab_testing/ab_test_spec.rb +0 -270
  47. data/spec/ab_testing/ab_testing_agent_spec.rb +0 -655
  48. data/spec/ab_testing/storage/adapter_spec.rb +0 -64
  49. data/spec/ab_testing/storage/memory_adapter_spec.rb +0 -485
  50. data/spec/activerecord_thread_safety_spec.rb +0 -553
  51. data/spec/advanced_operators_spec.rb +0 -3150
  52. data/spec/agent_spec.rb +0 -289
  53. data/spec/api_contract_spec.rb +0 -430
  54. data/spec/audit_adapters_spec.rb +0 -92
  55. data/spec/auth/access_audit_logger_spec.rb +0 -394
  56. data/spec/auth/authenticator_spec.rb +0 -112
  57. data/spec/auth/password_reset_spec.rb +0 -294
  58. data/spec/auth/permission_checker_spec.rb +0 -207
  59. data/spec/auth/permission_spec.rb +0 -73
  60. data/spec/auth/rbac_adapter_spec.rb +0 -778
  61. data/spec/auth/rbac_config_spec.rb +0 -82
  62. data/spec/auth/role_spec.rb +0 -51
  63. data/spec/auth/session_manager_spec.rb +0 -172
  64. data/spec/auth/session_spec.rb +0 -112
  65. data/spec/auth/user_spec.rb +0 -130
  66. data/spec/comprehensive_edge_cases_spec.rb +0 -1777
  67. data/spec/context_spec.rb +0 -127
  68. data/spec/decision_agent_spec.rb +0 -96
  69. data/spec/decision_spec.rb +0 -423
  70. data/spec/dmn/decision_graph_spec.rb +0 -282
  71. data/spec/dmn/decision_tree_spec.rb +0 -203
  72. data/spec/dmn/feel/errors_spec.rb +0 -18
  73. data/spec/dmn/feel/functions_spec.rb +0 -400
  74. data/spec/dmn/feel/simple_parser_spec.rb +0 -274
  75. data/spec/dmn/feel/types_spec.rb +0 -176
  76. data/spec/dmn/feel_parser_spec.rb +0 -489
  77. data/spec/dmn/hit_policy_spec.rb +0 -202
  78. data/spec/dmn/integration_spec.rb +0 -226
  79. data/spec/dsl/condition_evaluator_spec.rb +0 -774
  80. data/spec/dsl_validation_spec.rb +0 -648
  81. data/spec/edge_cases_spec.rb +0 -353
  82. data/spec/evaluation_spec.rb +0 -364
  83. data/spec/evaluation_validator_spec.rb +0 -165
  84. data/spec/examples/feedback_aware_evaluator_spec.rb +0 -460
  85. data/spec/examples.txt +0 -1909
  86. data/spec/fixtures/dmn/complex_decision.dmn +0 -81
  87. data/spec/fixtures/dmn/invalid_structure.dmn +0 -31
  88. data/spec/fixtures/dmn/simple_decision.dmn +0 -40
  89. data/spec/issue_verification_spec.rb +0 -759
  90. data/spec/json_rule_evaluator_spec.rb +0 -587
  91. data/spec/monitoring/alert_manager_spec.rb +0 -378
  92. data/spec/monitoring/metrics_collector_spec.rb +0 -501
  93. data/spec/monitoring/monitored_agent_spec.rb +0 -225
  94. data/spec/monitoring/prometheus_exporter_spec.rb +0 -242
  95. data/spec/monitoring/storage/activerecord_adapter_spec.rb +0 -498
  96. data/spec/monitoring/storage/base_adapter_spec.rb +0 -61
  97. data/spec/monitoring/storage/memory_adapter_spec.rb +0 -247
  98. data/spec/performance_optimizations_spec.rb +0 -493
  99. data/spec/replay_edge_cases_spec.rb +0 -699
  100. data/spec/replay_spec.rb +0 -210
  101. data/spec/rfc8785_canonicalization_spec.rb +0 -215
  102. data/spec/scoring_spec.rb +0 -225
  103. data/spec/spec_helper.rb +0 -60
  104. data/spec/testing/batch_test_importer_spec.rb +0 -693
  105. data/spec/testing/batch_test_runner_spec.rb +0 -307
  106. data/spec/testing/test_coverage_analyzer_spec.rb +0 -292
  107. data/spec/testing/test_result_comparator_spec.rb +0 -392
  108. data/spec/testing/test_scenario_spec.rb +0 -113
  109. data/spec/thread_safety_spec.rb +0 -490
  110. data/spec/thread_safety_spec.rb.broken +0 -878
  111. data/spec/versioning/adapter_spec.rb +0 -156
  112. data/spec/versioning_spec.rb +0 -1030
  113. data/spec/web/middleware/auth_middleware_spec.rb +0 -133
  114. data/spec/web/middleware/permission_middleware_spec.rb +0 -247
  115. data/spec/web_ui_rack_spec.rb +0 -2134
@@ -1,392 +0,0 @@
1
- require "spec_helper"
2
- require "tempfile"
3
-
4
- RSpec.describe DecisionAgent::Testing::TestResultComparator do
5
- let(:comparator) { DecisionAgent::Testing::TestResultComparator.new }
6
-
7
- describe "#compare" do
8
- let(:scenarios) do
9
- [
10
- DecisionAgent::Testing::TestScenario.new(
11
- id: "test_1",
12
- context: { user_id: 123 },
13
- expected_decision: "approve",
14
- expected_confidence: 0.95
15
- ),
16
- DecisionAgent::Testing::TestScenario.new(
17
- id: "test_2",
18
- context: { user_id: 456 },
19
- expected_decision: "reject",
20
- expected_confidence: 0.80
21
- )
22
- ]
23
- end
24
-
25
- let(:results) do
26
- [
27
- DecisionAgent::Testing::TestResult.new(
28
- scenario_id: "test_1",
29
- decision: "approve",
30
- confidence: 0.95
31
- ),
32
- DecisionAgent::Testing::TestResult.new(
33
- scenario_id: "test_2",
34
- decision: "reject",
35
- confidence: 0.80
36
- )
37
- ]
38
- end
39
-
40
- it "compares results with expected outcomes" do
41
- summary = comparator.compare(results, scenarios)
42
-
43
- expect(summary[:total]).to eq(2)
44
- expect(summary[:matches]).to eq(2)
45
- expect(summary[:mismatches]).to eq(0)
46
- expect(summary[:accuracy_rate]).to eq(1.0)
47
- end
48
-
49
- it "identifies mismatches" do
50
- mismatched_results = [
51
- DecisionAgent::Testing::TestResult.new(
52
- scenario_id: "test_1",
53
- decision: "reject", # Wrong decision
54
- confidence: 0.95
55
- ),
56
- DecisionAgent::Testing::TestResult.new(
57
- scenario_id: "test_2",
58
- decision: "reject",
59
- confidence: 0.50 # Wrong confidence
60
- )
61
- ]
62
-
63
- summary = comparator.compare(mismatched_results, scenarios)
64
-
65
- expect(summary[:matches]).to eq(0)
66
- expect(summary[:mismatches]).to eq(2)
67
- expect(summary[:accuracy_rate]).to eq(0.0)
68
- expect(summary[:mismatches_detail].size).to eq(2)
69
- end
70
-
71
- it "handles confidence tolerance" do
72
- comparator_with_tolerance = DecisionAgent::Testing::TestResultComparator.new(
73
- confidence_tolerance: 0.1
74
- )
75
-
76
- results_with_tolerance = [
77
- DecisionAgent::Testing::TestResult.new(
78
- scenario_id: "test_1",
79
- decision: "approve",
80
- confidence: 0.96 # Within 0.1 tolerance of 0.95
81
- )
82
- ]
83
-
84
- scenarios_single = [scenarios[0]]
85
- summary = comparator_with_tolerance.compare(results_with_tolerance, scenarios_single)
86
-
87
- expect(summary[:matches]).to eq(1)
88
- expect(summary[:confidence_accuracy]).to eq(1.0)
89
- end
90
-
91
- it "handles fuzzy matching" do
92
- comparator_fuzzy = DecisionAgent::Testing::TestResultComparator.new(fuzzy_match: true)
93
-
94
- scenarios_fuzzy = [
95
- DecisionAgent::Testing::TestScenario.new(
96
- id: "test_1",
97
- context: { user_id: 123 },
98
- expected_decision: "APPROVE", # Uppercase
99
- expected_confidence: 0.95
100
- )
101
- ]
102
-
103
- results_fuzzy = [
104
- DecisionAgent::Testing::TestResult.new(
105
- scenario_id: "test_1",
106
- decision: "approve", # Lowercase - should match with fuzzy
107
- confidence: 0.95
108
- )
109
- ]
110
-
111
- summary = comparator_fuzzy.compare(results_fuzzy, scenarios_fuzzy)
112
- expect(summary[:matches]).to eq(1)
113
- end
114
-
115
- it "handles fuzzy matching with whitespace" do
116
- comparator_fuzzy = DecisionAgent::Testing::TestResultComparator.new(fuzzy_match: true)
117
-
118
- scenarios_fuzzy = [
119
- DecisionAgent::Testing::TestScenario.new(
120
- id: "test_1",
121
- context: { user_id: 123 },
122
- expected_decision: " approve ", # With spaces
123
- expected_confidence: 0.95
124
- )
125
- ]
126
-
127
- results_fuzzy = [
128
- DecisionAgent::Testing::TestResult.new(
129
- scenario_id: "test_1",
130
- decision: "approve", # Without spaces - should match with fuzzy
131
- confidence: 0.95
132
- )
133
- ]
134
-
135
- summary = comparator_fuzzy.compare(results_fuzzy, scenarios_fuzzy)
136
- expect(summary[:matches]).to eq(1)
137
- end
138
-
139
- it "handles nil expected confidence" do
140
- scenarios_nil_conf = [
141
- DecisionAgent::Testing::TestScenario.new(
142
- id: "test_1",
143
- context: { user_id: 123 },
144
- expected_decision: "approve",
145
- expected_confidence: nil
146
- )
147
- ]
148
-
149
- results_nil_conf = [
150
- DecisionAgent::Testing::TestResult.new(
151
- scenario_id: "test_1",
152
- decision: "approve",
153
- confidence: 0.95
154
- )
155
- ]
156
-
157
- summary = comparator.compare(results_nil_conf, scenarios_nil_conf)
158
- expect(summary[:matches]).to eq(1)
159
- end
160
-
161
- it "handles nil actual confidence when expected is present" do
162
- scenarios_with_conf = [
163
- DecisionAgent::Testing::TestScenario.new(
164
- id: "test_1",
165
- context: { user_id: 123 },
166
- expected_decision: "approve",
167
- expected_confidence: 0.95
168
- )
169
- ]
170
-
171
- results_no_conf = [
172
- DecisionAgent::Testing::TestResult.new(
173
- scenario_id: "test_1",
174
- decision: "approve",
175
- confidence: nil
176
- )
177
- ]
178
-
179
- summary = comparator.compare(results_no_conf, scenarios_with_conf)
180
- expect(summary[:matches]).to eq(0)
181
- expect(summary[:mismatches]).to eq(1)
182
- end
183
-
184
- it "handles missing results for scenarios" do
185
- scenarios_missing = [
186
- DecisionAgent::Testing::TestScenario.new(
187
- id: "test_1",
188
- context: { user_id: 123 },
189
- expected_decision: "approve",
190
- expected_confidence: 0.95
191
- ),
192
- DecisionAgent::Testing::TestScenario.new(
193
- id: "test_2",
194
- context: { user_id: 456 },
195
- expected_decision: "reject",
196
- expected_confidence: 0.80
197
- )
198
- ]
199
-
200
- # Only provide result for test_1
201
- results_missing = [
202
- DecisionAgent::Testing::TestResult.new(
203
- scenario_id: "test_1",
204
- decision: "approve",
205
- confidence: 0.95
206
- )
207
- ]
208
-
209
- summary = comparator.compare(results_missing, scenarios_missing)
210
- # Should only compare test_1 since test_2 has no result
211
- expect(summary[:total]).to eq(1)
212
- end
213
-
214
- it "handles confidence outside tolerance" do
215
- comparator_strict = DecisionAgent::Testing::TestResultComparator.new(
216
- confidence_tolerance: 0.01
217
- )
218
-
219
- scenarios_strict = [
220
- DecisionAgent::Testing::TestScenario.new(
221
- id: "test_1",
222
- context: { user_id: 123 },
223
- expected_decision: "approve",
224
- expected_confidence: 0.95
225
- )
226
- ]
227
-
228
- results_outside = [
229
- DecisionAgent::Testing::TestResult.new(
230
- scenario_id: "test_1",
231
- decision: "approve",
232
- confidence: 0.98 # Outside 0.01 tolerance
233
- )
234
- ]
235
-
236
- summary = comparator_strict.compare(results_outside, scenarios_strict)
237
- expect(summary[:matches]).to eq(0)
238
- expect(summary[:confidence_accuracy]).to eq(0.0)
239
- end
240
-
241
- it "handles missing expected results" do
242
- scenarios_no_expected = [
243
- DecisionAgent::Testing::TestScenario.new(
244
- id: "test_1",
245
- context: { user_id: 123 }
246
- # No expected_decision
247
- )
248
- ]
249
-
250
- summary = comparator.compare(results, scenarios_no_expected)
251
-
252
- # Should not compare scenarios without expected results
253
- expect(summary[:total]).to eq(0)
254
- end
255
-
256
- it "handles failed test results" do
257
- failed_results = [
258
- DecisionAgent::Testing::TestResult.new(
259
- scenario_id: "test_1",
260
- error: StandardError.new("Test failed")
261
- )
262
- ]
263
-
264
- # Only compare scenarios that have expected results
265
- scenarios_with_expected = scenarios.select(&:expected_result?)
266
- summary = comparator.compare(failed_results, scenarios_with_expected)
267
-
268
- expect(summary[:mismatches]).to eq(1)
269
- expect(comparator.comparison_results[0].match).to be false
270
- end
271
- end
272
-
273
- describe "#generate_summary" do
274
- it "returns empty summary when no comparisons" do
275
- summary = comparator.generate_summary
276
-
277
- expect(summary[:total]).to eq(0)
278
- expect(summary[:matches]).to eq(0)
279
- expect(summary[:accuracy_rate]).to eq(0.0)
280
- end
281
- end
282
-
283
- describe "#export_csv" do
284
- it "exports comparison results to CSV" do
285
- scenarios = [
286
- DecisionAgent::Testing::TestScenario.new(
287
- id: "test_1",
288
- context: { user_id: 123 },
289
- expected_decision: "approve",
290
- expected_confidence: 0.95
291
- )
292
- ]
293
-
294
- results = [
295
- DecisionAgent::Testing::TestResult.new(
296
- scenario_id: "test_1",
297
- decision: "approve",
298
- confidence: 0.95
299
- )
300
- ]
301
-
302
- comparator.compare(results, scenarios)
303
-
304
- file = Tempfile.new(["comparison", ".csv"])
305
- comparator.export_csv(file.path)
306
-
307
- content = File.read(file.path)
308
- expect(content).to include("scenario_id")
309
- expect(content).to include("test_1")
310
- expect(content).to include("true") # match
311
-
312
- file.unlink
313
- end
314
- end
315
-
316
- describe "#export_json" do
317
- it "exports comparison results to JSON" do
318
- scenarios = [
319
- DecisionAgent::Testing::TestScenario.new(
320
- id: "test_1",
321
- context: { user_id: 123 },
322
- expected_decision: "approve",
323
- expected_confidence: 0.95
324
- )
325
- ]
326
-
327
- results = [
328
- DecisionAgent::Testing::TestResult.new(
329
- scenario_id: "test_1",
330
- decision: "approve",
331
- confidence: 0.95
332
- )
333
- ]
334
-
335
- comparator.compare(results, scenarios)
336
-
337
- file = Tempfile.new(["comparison", ".json"])
338
- comparator.export_json(file.path)
339
-
340
- content = JSON.parse(File.read(file.path))
341
- expect(content).to have_key("summary")
342
- expect(content).to have_key("results")
343
- expect(content["summary"]["total"]).to eq(1)
344
-
345
- file.unlink
346
- end
347
-
348
- it "handles empty comparison results" do
349
- file = Tempfile.new(["comparison", ".csv"])
350
- comparator.export_csv(file.path)
351
-
352
- content = File.read(file.path)
353
- expect(content).to include("scenario_id")
354
-
355
- file.unlink
356
- end
357
- end
358
-
359
- describe "ComparisonResult" do
360
- let(:comparison_result) do
361
- DecisionAgent::Testing::ComparisonResult.new(
362
- scenario_id: "test_1",
363
- match: true,
364
- decision_match: true,
365
- confidence_match: true,
366
- differences: [],
367
- actual: { decision: "approve", confidence: 0.95 },
368
- expected: { decision: "approve", confidence: 0.95 }
369
- )
370
- end
371
-
372
- it "creates a comparison result" do
373
- expect(comparison_result.scenario_id).to eq("test_1")
374
- expect(comparison_result.match).to be true
375
- expect(comparison_result.decision_match).to be true
376
- expect(comparison_result.confidence_match).to be true
377
- end
378
-
379
- it "converts to hash" do
380
- hash = comparison_result.to_h
381
-
382
- expect(hash[:scenario_id]).to eq("test_1")
383
- expect(hash[:match]).to be true
384
- expect(hash[:actual][:decision]).to eq("approve")
385
- expect(hash[:expected][:decision]).to eq("approve")
386
- end
387
-
388
- it "freezes the comparison result" do
389
- expect(comparison_result.frozen?).to be true
390
- end
391
- end
392
- end
@@ -1,113 +0,0 @@
1
- require "spec_helper"
2
-
3
- RSpec.describe DecisionAgent::Testing::TestScenario do
4
- describe "#initialize" do
5
- it "creates a test scenario with required fields" do
6
- scenario = DecisionAgent::Testing::TestScenario.new(
7
- id: "test_1",
8
- context: { user_id: 123, amount: 1000 }
9
- )
10
-
11
- expect(scenario.id).to eq("test_1")
12
- expect(scenario.context).to eq({ user_id: 123, amount: 1000 })
13
- expect(scenario.expected_decision).to be_nil
14
- expect(scenario.expected_confidence).to be_nil
15
- end
16
-
17
- it "creates a test scenario with expected results" do
18
- scenario = DecisionAgent::Testing::TestScenario.new(
19
- id: "test_2",
20
- context: { user_id: 456 },
21
- expected_decision: "approve",
22
- expected_confidence: 0.95
23
- )
24
-
25
- expect(scenario.expected_decision).to eq("approve")
26
- expect(scenario.expected_confidence).to eq(0.95)
27
- end
28
-
29
- it "freezes the scenario for immutability" do
30
- scenario = DecisionAgent::Testing::TestScenario.new(
31
- id: "test_3",
32
- context: { key: "value" }
33
- )
34
-
35
- expect(scenario.frozen?).to be true
36
- end
37
- end
38
-
39
- describe "#expected_result?" do
40
- it "returns true when expected_decision is set" do
41
- scenario = DecisionAgent::Testing::TestScenario.new(
42
- id: "test_1",
43
- context: { key: "value" },
44
- expected_decision: "approve"
45
- )
46
-
47
- expect(scenario.expected_result?).to be true
48
- end
49
-
50
- it "returns false when expected_decision is nil" do
51
- scenario = DecisionAgent::Testing::TestScenario.new(
52
- id: "test_1",
53
- context: { key: "value" }
54
- )
55
-
56
- expect(scenario.expected_result?).to be false
57
- end
58
- end
59
-
60
- describe "#to_h" do
61
- it "converts scenario to hash" do
62
- scenario = DecisionAgent::Testing::TestScenario.new(
63
- id: "test_1",
64
- context: { user_id: 123 },
65
- expected_decision: "approve",
66
- expected_confidence: 0.9,
67
- metadata: { source: "csv" }
68
- )
69
-
70
- hash = scenario.to_h
71
-
72
- expect(hash).to eq({
73
- id: "test_1",
74
- context: { user_id: 123 },
75
- expected_decision: "approve",
76
- expected_confidence: 0.9,
77
- metadata: { source: "csv" }
78
- })
79
- end
80
- end
81
-
82
- describe "#==" do
83
- it "returns true for equal scenarios" do
84
- scenario1 = DecisionAgent::Testing::TestScenario.new(
85
- id: "test_1",
86
- context: { user_id: 123 },
87
- expected_decision: "approve"
88
- )
89
-
90
- scenario2 = DecisionAgent::Testing::TestScenario.new(
91
- id: "test_1",
92
- context: { user_id: 123 },
93
- expected_decision: "approve"
94
- )
95
-
96
- expect(scenario1).to eq(scenario2)
97
- end
98
-
99
- it "returns false for different scenarios" do
100
- scenario1 = DecisionAgent::Testing::TestScenario.new(
101
- id: "test_1",
102
- context: { user_id: 123 }
103
- )
104
-
105
- scenario2 = DecisionAgent::Testing::TestScenario.new(
106
- id: "test_2",
107
- context: { user_id: 123 }
108
- )
109
-
110
- expect(scenario1).not_to eq(scenario2)
111
- end
112
- end
113
- end