plainstamp 0.1.0 → 0.3.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md
CHANGED
|
@@ -9,14 +9,30 @@ The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.1.0/),
|
|
|
9
9
|
### Planned next
|
|
10
10
|
|
|
11
11
|
- Add EU member-state AI Act implementation specifics where they diverge from the regulation (Germany, France, Spain, Italy, Netherlands first).
|
|
12
|
-
- Add sector-specific rules: FDA Software-as-a-Medical-Device AI guidance, FINRA chatbot disclosure, healthcare HIPAA-adjacent AI rules.
|
|
13
12
|
- Add a third watcher source (Cal Leg Info first; EUR-Lex if a usable feed surfaces).
|
|
14
|
-
- Cloudflare Workers deployment of the MCP server for free-tier hosted access.
|
|
15
|
-
- Get plainstamp listed on MCP registries (Anthropic registry, mcp-market, MCP Hive).
|
|
13
|
+
- Cloudflare Workers deployment of the MCP server for free-tier hosted access — gates the smithery.ai / pulsemcp.com / official MCP registry submissions, all of which require a hosted MCP endpoint or GitHub-verified namespace ownership.
|
|
16
14
|
|
|
17
15
|
### Distribution
|
|
18
16
|
|
|
19
|
-
Distribution is **npm-only**. Source remains in the operating organization's private repository; there is no public source repository host. Contact channel for issues, accuracy reports, security reports, and contribution proposals is **helpfulbutton140@agentmail.to** (see `
|
|
17
|
+
Distribution is **npm-only**. Source remains in the operating organization's private repository; there is no public source repository host. Contact channel for issues, accuracy reports, security reports, and contribution proposals is **helpfulbutton140@agentmail.to** (see `docs/CONTRIBUTING.md`, `docs/SECURITY.md`).
|
|
18
|
+
|
|
19
|
+
## [0.3.0] — 2026-05-08
|
|
20
|
+
|
|
21
|
+
### Added
|
|
22
|
+
|
|
23
|
+
- HHS Section 1557 — Patient Care Decision Support Tools nondiscrimination (45 CFR § 92.210, May 6, 2024 final rule). Covered entities (most healthcare providers receiving federal financial assistance, many health insurers, HHS-administered programs) must identify uses of AI/ML clinical decision-support tools and make reasonable efforts to mitigate algorithmic discrimination. Compliance deadline May 1, 2025 — now in effect and enforceable. Use case `healthcare`.
|
|
24
|
+
- Second SEO guide: `docs/guides/colorado-ai-act-sb-24-205-builder-guide.md` — long-form coverage of Colorado's comprehensive AI Act, the high-risk AI system definition, deployer/developer obligations, the consumer-disclosure components, the June 30, 2026 deadline, and how SB 24-205 stacks with parallel state and federal AI rules. Targets the high-traffic Colorado-compliance search vertical (deadline pressure + uncertainty about scope).
|
|
25
|
+
- Rule count 19 → 20. Tests still 51/51 passing.
|
|
26
|
+
|
|
27
|
+
## [0.2.0] — 2026-05-08
|
|
28
|
+
|
|
29
|
+
### Added
|
|
30
|
+
|
|
31
|
+
- FINRA Regulatory Notice 24-09 — AI in customer communications. Member-firm obligations under FINRA Rules 2210 (communications), 2090 (KYC), 2111 (suitability), 3110 (supervision), 4511 (records), 3220 (gifts) all apply to AI-driven customer communications and recommendations; firms remain responsible for third-party AI vendor outputs. Use case `financial-services`. Issued 2024-06-27.
|
|
32
|
+
- New SEO-leaning guide: `docs/guides/eu-ai-act-article-50-chatbot-disclosure.md` — long-form builder-focused guide on Article 50 disclosure requirements, the August 2026 application date, the Omnibus VII provisional agreement, and how the rule stacks with GDPR Article 22 and EU Member-State implementations. Ships in the npm package and renders on the npm package page (which is well-indexed).
|
|
33
|
+
- Package `files` array now includes `docs/guides` so SEO-leaning content ships with the published artifact.
|
|
34
|
+
- Keywords expanded: `gdpr`, `finra`, `cfpb`, `eeoc`, `regtech` added to support discovery via npm search and search-engine indexing of the npm package page.
|
|
35
|
+
- Rule count 18 → 19. Tests still 51/51 passing.
|
|
20
36
|
|
|
21
37
|
## [0.1.0] — 2026-05-08
|
|
22
38
|
|
|
@@ -0,0 +1,216 @@
|
|
|
1
|
+
# Colorado AI Act (SB 24-205): a builder's guide
|
|
2
|
+
|
|
3
|
+
> **Informational only — not legal advice.** Verify against the cited
|
|
4
|
+
> regulator-published text and consult counsel for production deployments.
|
|
5
|
+
> See `AI-DISCLOSURE.md` in this package.
|
|
6
|
+
|
|
7
|
+
If your AI product is sold to or used by people in Colorado and any of
|
|
8
|
+
its decisions could affect a person's access to housing, employment,
|
|
9
|
+
education, healthcare, financial services, government services, legal
|
|
10
|
+
services, or essential goods and services, **the Colorado AI Act
|
|
11
|
+
applies to you**. The rule is one of the strictest comprehensive AI
|
|
12
|
+
laws in the U.S. and its consumer-disclosure obligation goes into
|
|
13
|
+
effect **June 30, 2026** after a delay from the original February 2026
|
|
14
|
+
date. This guide walks through what the rule requires, what it does
|
|
15
|
+
*not* require, and what to ship before the deadline.
|
|
16
|
+
|
|
17
|
+
## What SB 24-205 actually does
|
|
18
|
+
|
|
19
|
+
Colorado SB 24-205 (codified at Colorado Revised Statutes § 6-1-1701
|
|
20
|
+
et seq.) creates obligations for two parties:
|
|
21
|
+
|
|
22
|
+
- **Developers** of high-risk AI systems — entities that build and
|
|
23
|
+
deploy a high-risk AI system or substantially modify one.
|
|
24
|
+
- **Deployers** of high-risk AI systems — entities that use a high-
|
|
25
|
+
risk AI system in their operations affecting Colorado consumers.
|
|
26
|
+
|
|
27
|
+
A "high-risk AI system" is one that, when deployed, makes or is a
|
|
28
|
+
substantial factor in making a "consequential decision" — defined to
|
|
29
|
+
include decisions affecting access to or cost of:
|
|
30
|
+
|
|
31
|
+
- Educational opportunities
|
|
32
|
+
- Employment or employment opportunities
|
|
33
|
+
- Financial or lending services
|
|
34
|
+
- Essential government services
|
|
35
|
+
- Healthcare services
|
|
36
|
+
- Housing
|
|
37
|
+
- Insurance
|
|
38
|
+
- Legal services
|
|
39
|
+
|
|
40
|
+
The Act layers two distinct sets of obligations: substantive (avoid
|
|
41
|
+
algorithmic discrimination) and procedural (impact assessments, risk
|
|
42
|
+
management, regulator notifications, consumer notices).
|
|
43
|
+
|
|
44
|
+
## The consumer-disclosure obligation — what to ship
|
|
45
|
+
|
|
46
|
+
The consumer-facing piece — the part most builders need to ship — has
|
|
47
|
+
three components:
|
|
48
|
+
|
|
49
|
+
### 1. Pre-decision disclosure (deployer obligation)
|
|
50
|
+
|
|
51
|
+
Before a high-risk AI system makes a consequential decision about a
|
|
52
|
+
consumer, the deployer must give the consumer:
|
|
53
|
+
|
|
54
|
+
- A statement disclosing that a high-risk AI system has been used
|
|
55
|
+
in the consequential decision-making process.
|
|
56
|
+
- A description of the high-risk AI system, its purpose, and how
|
|
57
|
+
it has been used.
|
|
58
|
+
- The nature of the consequential decision.
|
|
59
|
+
- Contact information for the deployer.
|
|
60
|
+
- A description of any human components of the decision-making
|
|
61
|
+
process and how the AI system contributes to the decision.
|
|
62
|
+
- A description of the consumer's rights under SB 24-205,
|
|
63
|
+
including the right to opt out of the processing of personal
|
|
64
|
+
data for profiling that produces legal or similarly significant
|
|
65
|
+
effects (under the Colorado Privacy Act), the right to correct
|
|
66
|
+
incorrect personal data, and the right to appeal an adverse
|
|
67
|
+
consequential decision.
|
|
68
|
+
|
|
69
|
+
### 2. Adverse-decision notice (deployer obligation)
|
|
70
|
+
|
|
71
|
+
If the high-risk AI system contributes to an adverse consequential
|
|
72
|
+
decision, the deployer must additionally disclose to the consumer:
|
|
73
|
+
|
|
74
|
+
- The principal reason(s) for the adverse decision.
|
|
75
|
+
- The degree to which the AI system contributed to the decision.
|
|
76
|
+
- The type of data processed by the AI system in making the
|
|
77
|
+
decision and the source of that data.
|
|
78
|
+
- The right to correct incorrect personal data, the right to
|
|
79
|
+
appeal the adverse decision, and the right to opt out of
|
|
80
|
+
profiling.
|
|
81
|
+
|
|
82
|
+
### 3. Public-facing statement (developer + deployer)
|
|
83
|
+
|
|
84
|
+
Both developers and deployers must publish a public statement summarizing:
|
|
85
|
+
|
|
86
|
+
- The types of high-risk AI systems they currently develop /
|
|
87
|
+
deploy.
|
|
88
|
+
- How the entity manages known or reasonably foreseeable risks of
|
|
89
|
+
algorithmic discrimination.
|
|
90
|
+
- The most recent date the public statement was updated.
|
|
91
|
+
|
|
92
|
+
## What SB 24-205 does *not* require
|
|
93
|
+
|
|
94
|
+
Common misconceptions worth clearing up:
|
|
95
|
+
|
|
96
|
+
- **It is not a CCPA-style right of deletion**. SB 24-205 layers on
|
|
97
|
+
the existing Colorado Privacy Act for personal-data rights; it
|
|
98
|
+
doesn't create new general-purpose data rights.
|
|
99
|
+
- **It does not require pre-approval or registration of every AI
|
|
100
|
+
system** with a Colorado regulator. Developers must notify the
|
|
101
|
+
Colorado Attorney General within 90 days of discovering that a
|
|
102
|
+
high-risk AI system has caused or is reasonably likely to have
|
|
103
|
+
caused algorithmic discrimination, but routine deployment doesn't
|
|
104
|
+
require pre-clearance.
|
|
105
|
+
- **It does not apply to most generative AI consumer products**
|
|
106
|
+
unless a specific deployment of that product is itself a high-
|
|
107
|
+
risk AI system making consequential decisions. A general-purpose
|
|
108
|
+
LLM helping a user write an email is not a high-risk AI system;
|
|
109
|
+
the same LLM scoring resumes for an employer is.
|
|
110
|
+
|
|
111
|
+
## The deadlines
|
|
112
|
+
|
|
113
|
+
- **June 30, 2026** — consumer-disclosure obligations apply to
|
|
114
|
+
deployers (delayed from the original February 2026 date).
|
|
115
|
+
- **Public statement and risk-management obligations apply on
|
|
116
|
+
the same date.**
|
|
117
|
+
- **Algorithmic-discrimination notification to the Attorney General**
|
|
118
|
+
applies on the same date.
|
|
119
|
+
|
|
120
|
+
## How SB 24-205 stacks with other AI rules
|
|
121
|
+
|
|
122
|
+
Colorado SB 24-205 is part of a comprehensive U.S.-state AI regime
|
|
123
|
+
that's emerging unevenly across jurisdictions. Builders deploying
|
|
124
|
+
across multiple states need to layer obligations:
|
|
125
|
+
|
|
126
|
+
- **California**: AB 2013 (training-data transparency, effective
|
|
127
|
+
2026-01-01); B&P § 17941 (bot disclosure); SB 942 (AI provenance);
|
|
128
|
+
the California Privacy Protection Agency's automated-decision-
|
|
129
|
+
making rulemaking.
|
|
130
|
+
- **Illinois**: HB 3773 amending the Illinois Human Rights Act
|
|
131
|
+
(employment AI, effective 2026-01-01).
|
|
132
|
+
- **Texas**: TRAIGA (HB 149, effective 2026-01-01) — government-
|
|
133
|
+
agency and healthcare-provider AI disclosure obligations.
|
|
134
|
+
- **Utah**: SB 149 + SB 226 — GenAI disclosure in regulated
|
|
135
|
+
occupations.
|
|
136
|
+
- **New York City**: Local Law 144 — AEDT bias audits for
|
|
137
|
+
employment AI.
|
|
138
|
+
- **Maryland**: Labor & Employment § 3-717 — facial recognition
|
|
139
|
+
in interviews requires written consent.
|
|
140
|
+
- **Federal**: EEOC technical assistance on Title VII selection
|
|
141
|
+
procedures; CFPB Circular 2023-03 on AI adverse-action notices;
|
|
142
|
+
HHS Section 1557 on patient-care decision support tools; FINRA
|
|
143
|
+
Regulatory Notice 24-09 on AI in member-firm communications.
|
|
144
|
+
- **EU**: AI Act Articles 50(1) and 50(2); GDPR Article 22 on
|
|
145
|
+
automated decisions.
|
|
146
|
+
|
|
147
|
+
A consumer-facing AI product operating across these jurisdictions
|
|
148
|
+
needs disclosure copy for each — and the disclosures often differ in
|
|
149
|
+
content, timing, and format. That's the maintenance problem
|
|
150
|
+
`plainstamp` exists to solve.
|
|
151
|
+
|
|
152
|
+
## How plainstamp helps
|
|
153
|
+
|
|
154
|
+
`plainstamp` ships an `us-co-sb24-205-consumer-disclosure` rule that
|
|
155
|
+
returns the live disclosure-element checklist for SB 24-205, ready-
|
|
156
|
+
to-paste plain-language and formal-language templates, citation back
|
|
157
|
+
to the Colorado Office of Legislative Legal Services source URL, and
|
|
158
|
+
a `last_verified` date.
|
|
159
|
+
|
|
160
|
+
Typical lookup for a deployer notifying a Colorado employment-AI user
|
|
161
|
+
before a hiring decision:
|
|
162
|
+
|
|
163
|
+
```bash
|
|
164
|
+
npx plainstamp lookup --jurisdiction us-co \
|
|
165
|
+
--channel email-transactional \
|
|
166
|
+
--use-case employment-decisions
|
|
167
|
+
```
|
|
168
|
+
|
|
169
|
+
This returns the SB 24-205 consumer-disclosure rule. To pick up the
|
|
170
|
+
parallel federal-floor obligation (EEOC technical assistance) and the
|
|
171
|
+
parallel state-employment rules in other states the deployer
|
|
172
|
+
operates in, query each jurisdiction in turn. plainstamp's
|
|
173
|
+
parent-jurisdiction inheritance rule means a `us-co` query also
|
|
174
|
+
matches federal-level `us` rules.
|
|
175
|
+
|
|
176
|
+
For the public-facing statement (developer or deployer) and the
|
|
177
|
+
internal-governance items (impact assessments, risk-management
|
|
178
|
+
program), consult Colorado Attorney General published guidance
|
|
179
|
+
directly — those are above plainstamp's scope (which covers per-
|
|
180
|
+
interaction or per-decision disclosure text, not corporate
|
|
181
|
+
governance program documentation).
|
|
182
|
+
|
|
183
|
+
## The minimum viable Colorado disclosure
|
|
184
|
+
|
|
185
|
+
If you ship one thing this quarter, ship the pre-decision disclosure:
|
|
186
|
+
|
|
187
|
+
1. A clear statement that a high-risk AI system is being used in
|
|
188
|
+
the consequential decision.
|
|
189
|
+
2. A description of the AI system's purpose and role in the
|
|
190
|
+
decision.
|
|
191
|
+
3. A description of any human components of the decision.
|
|
192
|
+
4. Contact information for the deployer.
|
|
193
|
+
5. A summary of the consumer's appeal, correction, and opt-out
|
|
194
|
+
rights, with a path to exercise them.
|
|
195
|
+
|
|
196
|
+
If your AI system can produce adverse outcomes (denials, rejections,
|
|
197
|
+
adverse employment actions, etc.), also ship the adverse-decision
|
|
198
|
+
notice with principal reasons, the AI's contribution, and data-source
|
|
199
|
+
disclosure.
|
|
200
|
+
|
|
201
|
+
## Source-of-truth links
|
|
202
|
+
|
|
203
|
+
- **Colorado SB 24-205 — full text and legislative history**
|
|
204
|
+
([leg.colorado.gov](https://leg.colorado.gov/bills/sb24-205))
|
|
205
|
+
- **Colorado Attorney General — AI Act guidance and rulemaking**
|
|
206
|
+
([coag.gov](https://coag.gov/))
|
|
207
|
+
- **Colorado Privacy Act**, into which SB 24-205 connects for personal-
|
|
208
|
+
data rights ([leg.colorado.gov](https://leg.colorado.gov/))
|
|
209
|
+
|
|
210
|
+
`plainstamp` is maintained by an autonomous AI agent operating under
|
|
211
|
+
KS Elevated Solutions LLC. Accuracy reports, rule-update suggestions,
|
|
212
|
+
and security disclosures: [helpfulbutton140@agentmail.to](mailto:helpfulbutton140@agentmail.to).
|
|
213
|
+
|
|
214
|
+
---
|
|
215
|
+
|
|
216
|
+
[`← Back to plainstamp on npm`](https://www.npmjs.com/package/plainstamp)
|
|
@@ -0,0 +1,161 @@
|
|
|
1
|
+
# EU AI Act Article 50: a builder's guide to chatbot disclosure
|
|
2
|
+
|
|
3
|
+
> **Informational only — not legal advice.** Verify against the cited
|
|
4
|
+
> regulator-published text and consult counsel for production deployments.
|
|
5
|
+
> See `AI-DISCLOSURE.md` in this package.
|
|
6
|
+
|
|
7
|
+
If your product talks to people in the EU and an AI is doing the talking,
|
|
8
|
+
Article 50 of the EU AI Act applies to you. This guide covers what the
|
|
9
|
+
rule actually says, when it applies, what counts as compliance, and the
|
|
10
|
+
deadline pressure most teams aren't tracking yet.
|
|
11
|
+
|
|
12
|
+
## What Article 50 actually requires
|
|
13
|
+
|
|
14
|
+
Article 50(1) of Regulation (EU) 2024/1689 says:
|
|
15
|
+
|
|
16
|
+
> Providers of AI systems intended to interact directly with natural
|
|
17
|
+
> persons must design and develop them in such a way that the natural
|
|
18
|
+
> persons concerned are informed that they are interacting with an AI
|
|
19
|
+
> system.
|
|
20
|
+
|
|
21
|
+
There is one exception: if the fact that the user is talking to an AI
|
|
22
|
+
is "obvious from the point of view of a reasonably well-informed person
|
|
23
|
+
taking into account the circumstances and the context of use," the
|
|
24
|
+
disclosure is not required. The bar for "obvious" is high — a chat
|
|
25
|
+
window labeled "AI Assistant" probably qualifies; a chat window
|
|
26
|
+
labeled "Customer Support" does not, even if the bot sounds robotic.
|
|
27
|
+
|
|
28
|
+
Article 50(2) layers a separate obligation: any AI-generated synthetic
|
|
29
|
+
audio, image, video, or text must be marked as artificially generated
|
|
30
|
+
or manipulated, in a machine-readable format. The text-content
|
|
31
|
+
sub-clause has narrow exemptions (assistive editing, no substantive
|
|
32
|
+
change, etc.) that we cover later.
|
|
33
|
+
|
|
34
|
+
## Who is the "provider"
|
|
35
|
+
|
|
36
|
+
The Act distinguishes **providers** (who develop or place the AI system
|
|
37
|
+
on the market) from **deployers** (who use it). Article 50 falls
|
|
38
|
+
primarily on providers — but the deployer obligations under Article 50(4)
|
|
39
|
+
on emotion-recognition / biometric systems and on deepfakes still apply
|
|
40
|
+
where relevant.
|
|
41
|
+
|
|
42
|
+
For a typical SaaS chatbot: the company that builds the chatbot model
|
|
43
|
+
or wraps an LLM into a product is the provider. The customer that
|
|
44
|
+
embeds the chatbot on their site is a deployer. Both have obligations
|
|
45
|
+
under different Article 50 paragraphs.
|
|
46
|
+
|
|
47
|
+
## When the obligation kicks in
|
|
48
|
+
|
|
49
|
+
Article 50 applies as soon as a natural person begins interacting with
|
|
50
|
+
the AI system. Practically, this means **the disclosure must appear at
|
|
51
|
+
the start of the conversation**, before the AI has produced any
|
|
52
|
+
substantive output that a user might rely on.
|
|
53
|
+
|
|
54
|
+
A persistent banner reading "You are chatting with an AI assistant" at
|
|
55
|
+
the top of the chat surface satisfies this for most chat UIs. A
|
|
56
|
+
voice-channel disclosure must be spoken at session start. A
|
|
57
|
+
video-avatar disclosure typically combines a spoken introduction with a
|
|
58
|
+
visible on-screen indicator.
|
|
59
|
+
|
|
60
|
+
## The "machine-readable" requirement (Art. 50(2))
|
|
61
|
+
|
|
62
|
+
For AI-generated synthetic content, the marking must be machine-
|
|
63
|
+
readable. The Act doesn't mandate a specific technical standard, but
|
|
64
|
+
the European Commission has signaled that watermarking schemes
|
|
65
|
+
compliant with C2PA, the SynthID variants, and similar provenance
|
|
66
|
+
metadata will be acceptable. As of 2026, the Commission is finalizing
|
|
67
|
+
implementing acts that will narrow the technical options.
|
|
68
|
+
|
|
69
|
+
If you're producing AI-generated images, audio, or video at scale,
|
|
70
|
+
adopt a watermarking standard now — retrofitting watermarks across an
|
|
71
|
+
existing content corpus is materially harder than baking them into the
|
|
72
|
+
generation pipeline.
|
|
73
|
+
|
|
74
|
+
## Penalties and timing
|
|
75
|
+
|
|
76
|
+
Article 50 obligations apply from **August 2, 2026**. Penalties under
|
|
77
|
+
Article 99 of the Act for Article 50 violations can reach **€15 million
|
|
78
|
+
or 3% of global annual turnover, whichever is higher**.
|
|
79
|
+
|
|
80
|
+
A separate provisional agreement under the EU's Omnibus VII package
|
|
81
|
+
(provisional agreement 2026-05-07) reduced the transparency-solutions
|
|
82
|
+
grace period from 6 months to 3 months, moving the practical
|
|
83
|
+
compliance deadline for Article 50(2) machine-readable marking
|
|
84
|
+
implementations to **December 2, 2026**. Re-verify against the final
|
|
85
|
+
adopted text — Omnibus VII's provisional agreement may shift before
|
|
86
|
+
formal adoption.
|
|
87
|
+
|
|
88
|
+
## How Article 50 stacks with other EU rules
|
|
89
|
+
|
|
90
|
+
Article 50 doesn't operate in isolation. Builders should also check:
|
|
91
|
+
|
|
92
|
+
- **GDPR Article 22** — if the AI conversation feeds into an automated
|
|
93
|
+
decision producing legal or similarly significant effects, the
|
|
94
|
+
data-subject rights to human intervention, point-of-view expression,
|
|
95
|
+
and contestation apply on top of the chatbot disclosure.
|
|
96
|
+
- **GDPR Articles 13(2)(f) and 14(2)(g)** — when personal data is
|
|
97
|
+
collected during the AI interaction, the controller must inform the
|
|
98
|
+
data subject about the existence of automated decision-making, the
|
|
99
|
+
logic involved, and the envisaged consequences.
|
|
100
|
+
- **EU AI Act Article 13** — high-risk AI systems have separate
|
|
101
|
+
transparency obligations to deployers (instructions for use,
|
|
102
|
+
expected outputs, characteristics and limitations).
|
|
103
|
+
- **Digital Services Act (Regulation (EU) 2022/2065)** — provider
|
|
104
|
+
obligations around content moderation transparency layer over the
|
|
105
|
+
chatbot disclosure when the AI is moderating user-generated content.
|
|
106
|
+
- **Member-state implementations** — Germany's BDSG, France's Loi
|
|
107
|
+
Informatique et Libertés, and Spain's AESIA framework all add
|
|
108
|
+
national-level safeguards on top of the EU regulation. Verify the
|
|
109
|
+
rules of every Member State your service operates in.
|
|
110
|
+
|
|
111
|
+
## How plainstamp helps
|
|
112
|
+
|
|
113
|
+
`plainstamp` ships with `eu-ai-act-art50-chatbot` and
|
|
114
|
+
`eu-ai-act-art50-genai-content` rules that surface the live text of
|
|
115
|
+
Article 50, the required disclosure elements, and ready-to-paste plain-
|
|
116
|
+
language and formal-language disclosure templates. Each rule cites the
|
|
117
|
+
EUR-Lex source URL and carries a `last_verified` date so you know
|
|
118
|
+
whether the text you're reading is current.
|
|
119
|
+
|
|
120
|
+
A typical lookup:
|
|
121
|
+
|
|
122
|
+
```bash
|
|
123
|
+
npx plainstamp lookup --jurisdiction eu \
|
|
124
|
+
--channel live-chat \
|
|
125
|
+
--use-case b2c-customer-support
|
|
126
|
+
```
|
|
127
|
+
|
|
128
|
+
returns the rule, the disclosure-element checklist, and template text
|
|
129
|
+
you can drop into your chat surface. For deployers running across
|
|
130
|
+
multiple jurisdictions, the same query against `us-ca`, `us-co`,
|
|
131
|
+
`us-il`, `us-tx`, `us-ut`, etc. will surface the parallel state-level
|
|
132
|
+
obligations that often layer on top.
|
|
133
|
+
|
|
134
|
+
## The minimum viable Article 50 disclosure
|
|
135
|
+
|
|
136
|
+
If you ship one thing this week, ship a chat-surface header that
|
|
137
|
+
includes:
|
|
138
|
+
|
|
139
|
+
1. A clear statement that the user is interacting with an AI ("You
|
|
140
|
+
are chatting with an AI assistant").
|
|
141
|
+
2. A path to escalate to a human (where applicable to your service
|
|
142
|
+
model and required by sectoral rules — e.g., financial-services
|
|
143
|
+
rules in many jurisdictions require an escalation path).
|
|
144
|
+
3. A link to your privacy notice covering AI data use.
|
|
145
|
+
|
|
146
|
+
Then, if you process AI-generated synthetic media, prioritize
|
|
147
|
+
machine-readable marking for the Art. 50(2) deadline.
|
|
148
|
+
|
|
149
|
+
## Source-of-truth links
|
|
150
|
+
|
|
151
|
+
- **Regulation (EU) 2024/1689 — full text on EUR-Lex** ([eur-lex.europa.eu](https://eur-lex.europa.eu/legal-content/EN/TXT/?uri=CELEX%3A32024R1689))
|
|
152
|
+
- **GDPR Article 22 on EUR-Lex** ([eur-lex.europa.eu](https://eur-lex.europa.eu/legal-content/EN/TXT/?uri=CELEX%3A32016R0679))
|
|
153
|
+
- **EDPB Guidelines on Automated Decision-Making (WP251rev.01)** — apply alongside Art. 22 obligations.
|
|
154
|
+
|
|
155
|
+
`plainstamp` is maintained by an autonomous AI agent operating under
|
|
156
|
+
KS Elevated Solutions LLC. Accuracy reports, rule-update suggestions,
|
|
157
|
+
and security disclosures: [helpfulbutton140@agentmail.to](mailto:helpfulbutton140@agentmail.to).
|
|
158
|
+
|
|
159
|
+
---
|
|
160
|
+
|
|
161
|
+
[`← Back to plainstamp on npm`](https://www.npmjs.com/package/plainstamp)
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "plainstamp",
|
|
3
|
-
"version": "0.
|
|
3
|
+
"version": "0.3.0",
|
|
4
4
|
"description": "AI disclosure compliance assistant — generates legally-grounded AI disclosure text per (jurisdiction × channel × use-case) and tracks regulatory updates. Operated by an autonomous AI agent under KS Elevated Solutions LLC.",
|
|
5
5
|
"type": "module",
|
|
6
6
|
"license": "MIT",
|
|
@@ -17,6 +17,7 @@
|
|
|
17
17
|
"files": [
|
|
18
18
|
"dist",
|
|
19
19
|
"rules",
|
|
20
|
+
"docs/guides",
|
|
20
21
|
"README.md",
|
|
21
22
|
"AI-DISCLOSURE.md",
|
|
22
23
|
"CHANGELOG.md",
|
|
@@ -53,7 +54,12 @@
|
|
|
53
54
|
"compliance",
|
|
54
55
|
"ccpa",
|
|
55
56
|
"eu-ai-act",
|
|
57
|
+
"gdpr",
|
|
56
58
|
"ftc",
|
|
59
|
+
"finra",
|
|
60
|
+
"cfpb",
|
|
61
|
+
"eeoc",
|
|
62
|
+
"regtech",
|
|
57
63
|
"agent",
|
|
58
64
|
"autonomous-ai"
|
|
59
65
|
]
|
package/rules/seed.json
CHANGED
|
@@ -812,6 +812,100 @@
|
|
|
812
812
|
"formal": "Notice of Adverse Action under the Equal Credit Opportunity Act (15 U.S.C. § 1691(d)) and Regulation B (12 CFR § 1002.9), as further interpreted by CFPB Circular 2023-03 in the context of artificial-intelligence and machine-learning credit decisions: The application identified by reference number [REF] has been adversely acted upon. The specific principal reasons that most adversely affected the decision in this case, as identified by the creditor's review of the AI/ML model output, are: (1) [reason]; (2) [reason]; (3) [reason]. The applicant may request a written statement of the specific reasons within 60 days of this notice; the creditor will provide such statement within 30 days of receipt of the request. Federal law prohibits creditors from discriminating against credit applicants on prohibited bases enumerated in 15 U.S.C. § 1691(a). The federal agency administering compliance with the ECOA concerning this creditor is [agency, address]."
|
|
813
813
|
},
|
|
814
814
|
"notes": "CFPB Circular 2023-03 makes explicit a position the CFPB had taken in supervisory guidance for years: the technological complexity of an AI/ML model is not a defense for failing to provide ECOA-compliant adverse-action reasons. Creditors must identify the specific factors that affected THIS APPLICANT'S decision — not generic factors that influence the model in general. Practical implications for AI-credit fintechs: (1) the model itself must be explainable to a level that supports per-applicant reason codes — if the model cannot do this, the model cannot be deployed for credit decisions; (2) the reason codes must be checked for accuracy, not just plausibility — using post-hoc SHAP / LIME explanations as the source of reason codes is acceptable IF the creditor has validated that those explanations actually reflect what drove the decision in each case; (3) generic or boilerplate codes ('credit application incomplete', 'failed model threshold') are insufficient — the codes must point to applicant-specific factors. ECOA's statutory penalties combined with ongoing CFPB enforcement priority make this a high-stakes obligation. Note: Regulation B's adverse-action requirements run in parallel with the FCRA's adverse-action requirements (15 U.S.C. § 1681m) when the decision was based in whole or in part on a consumer report — both sets of obligations apply to the same notice."
|
|
815
|
+
},
|
|
816
|
+
{
|
|
817
|
+
"id": "us-finra-rn-24-09-ai-customer-communications",
|
|
818
|
+
"jurisdiction": "us",
|
|
819
|
+
"channels": ["live-chat", "voice", "email-marketing", "ai-generated-content"],
|
|
820
|
+
"use_cases": ["financial-services"],
|
|
821
|
+
"severity": "mandatory",
|
|
822
|
+
"short_title": "FINRA Regulatory Notice 24-09 — AI in customer communications",
|
|
823
|
+
"summary": "FINRA Regulatory Notice 24-09 (June 27, 2024) addresses member firm use of generative artificial intelligence and other large language model technologies in their securities business. The Notice does not create new rules; it confirms that existing FINRA rules apply to AI-driven customer communications and reminds member firms of their obligations: (a) Rule 3110 — supervisory systems reasonably designed to achieve compliance with applicable rules apply to AI tools used by associated persons or in customer-facing roles; (b) Rule 2210 — communications with the public, including any communication generated by an AI tool, must be fair, balanced, not misleading, and (where applicable) supervised, principal-approved, or filed with FINRA; (c) Rule 2090 (Know Your Customer) and Rule 2111 (suitability) — AI-generated recommendations are subject to the same suitability and KYC obligations as human-generated ones; (d) Rule 4511 — books-and-records obligations apply to AI inputs and outputs that constitute communications with customers; (e) Rule 3220 — gifts and gratuities standards apply to AI-generated promotional materials. Member firms remain responsible for AI tool outputs even when the tool is provided by a third-party vendor. Notice 24-09 also flags risks including hallucination, bias, data privacy, and intellectual-property concerns; firms should address these in written supervisory procedures.",
|
|
824
|
+
"required_elements": [
|
|
825
|
+
{
|
|
826
|
+
"id": "ai-communication-supervision",
|
|
827
|
+
"description": "AI-generated communications with the public are subject to FINRA Rule 2210 standards (fair, balanced, not misleading) and the firm's existing principal-review / pre-approval / filing workflow as applicable to the communication type.",
|
|
828
|
+
"required": true,
|
|
829
|
+
"example": "All customer-facing communications generated by the AI assistant are reviewed by a qualified principal under FINRA Rule 2210 before delivery and retained per the firm's books-and-records policy under Rule 4511."
|
|
830
|
+
},
|
|
831
|
+
{
|
|
832
|
+
"id": "ai-recommendation-suitability",
|
|
833
|
+
"description": "AI-generated investment recommendations or advice are subject to FINRA Rule 2111 suitability obligations on the same terms as human-generated recommendations; firm WSPs must address how AI-generated recommendations are reviewed for suitability.",
|
|
834
|
+
"required": true,
|
|
835
|
+
"example": "Any investment recommendation generated by the AI tool for a customer account is subject to a Rule 2111 suitability review against the customer's investment profile under the firm's written supervisory procedures."
|
|
836
|
+
},
|
|
837
|
+
{
|
|
838
|
+
"id": "third-party-vendor-responsibility",
|
|
839
|
+
"description": "Firm responsibility for AI tool outputs persists when the tool is operated by a third-party vendor; vendor due diligence and oversight are part of the firm's Rule 3110 supervisory obligation.",
|
|
840
|
+
"required": true,
|
|
841
|
+
"example": "AI tools operated by third-party vendors are vetted, monitored, and supervised by the firm under FINRA Rule 3110; the firm remains responsible for any communications, recommendations, or records generated by those tools in connection with its securities business."
|
|
842
|
+
},
|
|
843
|
+
{
|
|
844
|
+
"id": "wsp-ai-coverage",
|
|
845
|
+
"description": "Written supervisory procedures address AI tool use, including risk areas of hallucination, bias, data privacy, and IP. (System / governance requirement, not per-message text.)",
|
|
846
|
+
"required": false
|
|
847
|
+
}
|
|
848
|
+
],
|
|
849
|
+
"citation": {
|
|
850
|
+
"statute": "FINRA Rules 2210, 2090, 2111, 3110, 4511, 3220 (existing); FINRA Regulatory Notice 24-09, 'FINRA Reminds Member Firms of Their Obligations When Using Generative Artificial Intelligence and Large Language Models' (June 27, 2024)",
|
|
851
|
+
"section": "Member-firm obligations when using AI in securities business",
|
|
852
|
+
"source_url": "https://www.finra.org/rules-guidance/notices/24-09",
|
|
853
|
+
"publisher": "Financial Industry Regulatory Authority"
|
|
854
|
+
},
|
|
855
|
+
"effective_date": "2024-06-27",
|
|
856
|
+
"last_verified": "2026-05-08",
|
|
857
|
+
"template": {
|
|
858
|
+
"plain": "Notice — Customer Communication via AI Tool: This message (or recommendation) was prepared with the assistance of an artificial-intelligence tool and is subject to the same review and supervision standards as any communication delivered by [Member Firm]. The communication is reviewed under FINRA Rule 2210 standards and, where applicable, has been reviewed by a qualified principal. Any investment recommendation in this communication remains subject to the firm's suitability analysis under FINRA Rule 2111 against your investment profile. If you have questions about this communication or the role of AI in producing it, contact [contact].",
|
|
859
|
+
"formal": "Notice under FINRA Regulatory Notice 24-09 and Rules 2210, 2090, 2111, 3110, 4511, and 3220: This communication was generated, in whole or in part, with the assistance of artificial-intelligence technology. The member firm has reviewed and supervised this communication under its written supervisory procedures consistent with FINRA Rule 3110, and the communication satisfies the standards of FINRA Rule 2210 governing communications with the public. Any investment recommendation contained herein has been evaluated for suitability under FINRA Rule 2111 against the customer's investment profile under FINRA Rule 2090. The firm retains records of this communication under FINRA Rule 4511. The member firm remains responsible for AI tool outputs whether the tool is internally operated or provided by a third-party vendor."
|
|
860
|
+
},
|
|
861
|
+
"notes": "FINRA Regulatory Notice 24-09 is reminder-and-clarification guidance — it does not create new rules. The binding obligations are the existing FINRA rules (2210, 2090, 2111, 3110, 4511, 3220), which apply by their existing terms to AI-driven communications, recommendations, and records. Member firms (broker-dealers and their associated persons) are bound; non-member firms are not directly bound by FINRA rules but may face parallel obligations under SEC rules (e.g., Rule 17a-4 books-and-records, Investment Advisers Act fiduciary duty for IA-registered firms) — this rule's `jurisdiction` is `us` because FINRA is a self-regulatory organization with national scope, not a single-state regulator. The 2023 SEC Staff Bulletin on conflicts of interest for AI/PDA-using broker-dealers and investment advisers (and the SEC's proposed PDA rule, Rel. No. 34-97990) layers additional obligations specifically around conflicts; firms with PDA / AI advisory tools should consult both. FINRA expects firms to update their WSPs to specifically address AI tool use; using AI without WSP coverage is an immediate Rule 3110 supervision deficiency. Firms should also be aware of state-level adverse-action and disclosure overlays (e.g., NYDFS's October 2024 cybersecurity / AI guidance for licensed entities)."
|
|
862
|
+
},
|
|
863
|
+
{
|
|
864
|
+
"id": "us-hhs-section-1557-pcdst-2024",
|
|
865
|
+
"jurisdiction": "us",
|
|
866
|
+
"channels": ["ai-generated-content", "about-page", "privacy-policy"],
|
|
867
|
+
"use_cases": ["healthcare"],
|
|
868
|
+
"severity": "mandatory",
|
|
869
|
+
"short_title": "HHS Section 1557 — Patient Care Decision Support Tools nondiscrimination (2024 final rule)",
|
|
870
|
+
"summary": "On May 6, 2024, the U.S. Department of Health and Human Services Office for Civil Rights published a final rule (89 Fed. Reg. 37522) implementing Section 1557 of the Affordable Care Act that imposes nondiscrimination obligations on covered entities' use of 'patient care decision support tools' (PCDSTs) — defined to include automated and non-automated tools, including artificial-intelligence and machine-learning-based clinical decision support. Covered entities (most healthcare providers receiving federal financial assistance, many health insurers, and HHS-administered health programs) must (a) make reasonable efforts to identify uses of PCDSTs in their health programs and activities that employ input variables or factors that measure race, color, national origin, sex, age, or disability; AND (b) make reasonable efforts to mitigate the risk of discrimination resulting from the tool's use. The compliance deadline for the PCDST nondiscrimination obligation was May 1, 2025; the obligation is now in effect and enforceable. Penalties for Section 1557 violations include loss of federal financial assistance, OCR-imposed corrective-action plans, and potential private-right-of-action claims for discrimination.",
|
|
871
|
+
"required_elements": [
|
|
872
|
+
{
|
|
873
|
+
"id": "pcdst-identification",
|
|
874
|
+
"description": "Reasonable efforts to identify uses of PCDSTs (including AI/ML clinical decision support tools) in the entity's health programs and activities.",
|
|
875
|
+
"required": false,
|
|
876
|
+
"example": "Internal inventory and documentation of all AI/ML clinical decision support tools deployed in patient care, with notation of input variables and use cases. (System / governance requirement; does not require per-patient disclosure.)"
|
|
877
|
+
},
|
|
878
|
+
{
|
|
879
|
+
"id": "pcdst-mitigation",
|
|
880
|
+
"description": "Reasonable efforts to mitigate the risk of discrimination resulting from PCDST use, including documentation of mitigation steps and ongoing monitoring.",
|
|
881
|
+
"required": false,
|
|
882
|
+
"example": "Documented mitigation procedures, periodic testing for adverse impact across protected classes, and a designated responsible person or office. (System / governance requirement.)"
|
|
883
|
+
},
|
|
884
|
+
{
|
|
885
|
+
"id": "patient-facing-pcdst-notice",
|
|
886
|
+
"description": "Patient-facing notice that AI/ML decision-support tools may inform clinical decisions, where the entity's notice-of-availability obligations under § 92.11 apply (translation requirements + civil rights coordinator + grievance procedures).",
|
|
887
|
+
"required": true,
|
|
888
|
+
"example": "Notice: Some clinical decisions in your care may be informed by automated decision-support tools, including artificial intelligence. You have the right to discuss any care decision with your provider. If you believe you have experienced discrimination on the basis of race, color, national origin, sex, age, or disability in connection with these tools or any other aspect of your care, contact our Civil Rights Coordinator at [contact] or file a complaint with the HHS Office for Civil Rights."
|
|
889
|
+
},
|
|
890
|
+
{
|
|
891
|
+
"id": "civil-rights-coordinator-designation",
|
|
892
|
+
"description": "Designation of a Civil Rights Coordinator responsible for the entity's Section 1557 compliance, including PCDST nondiscrimination obligations. (Governance, not per-patient text.)",
|
|
893
|
+
"required": false
|
|
894
|
+
}
|
|
895
|
+
],
|
|
896
|
+
"citation": {
|
|
897
|
+
"statute": "Section 1557 of the Patient Protection and Affordable Care Act (42 U.S.C. § 18116); 45 CFR Part 92, as amended by the May 6, 2024 final rule, 89 Fed. Reg. 37522",
|
|
898
|
+
"section": "45 CFR § 92.210 (Discrimination through the use of patient care decision support tools)",
|
|
899
|
+
"source_url": "https://www.federalregister.gov/documents/2024/05/06/2024-08711/nondiscrimination-in-health-programs-and-activities",
|
|
900
|
+
"publisher": "U.S. Department of Health and Human Services, Office for Civil Rights"
|
|
901
|
+
},
|
|
902
|
+
"effective_date": "2025-05-01",
|
|
903
|
+
"last_verified": "2026-05-08",
|
|
904
|
+
"template": {
|
|
905
|
+
"plain": "Notice — Use of Decision-Support Tools in Your Care: Some clinical decisions in your care may be informed by automated decision-support tools, including artificial-intelligence and machine-learning systems. These tools assist your healthcare team and do not replace the judgment of a licensed clinician. You have the right to discuss any care decision with your provider. If you believe you have experienced discrimination on the basis of race, color, national origin, sex, age, or disability in connection with these tools or any other aspect of your care, please contact our Civil Rights Coordinator at [contact] or file a complaint with the HHS Office for Civil Rights at https://www.hhs.gov/ocr/.",
|
|
906
|
+
"formal": "Notice under Section 1557 of the Patient Protection and Affordable Care Act (42 U.S.C. § 18116) and the implementing regulations at 45 CFR Part 92 (as amended by the May 6, 2024 final rule, 89 Fed. Reg. 37522): The covered entity uses one or more patient care decision support tools, including artificial-intelligence and machine-learning-based clinical decision support, in its health programs and activities. The covered entity has identified its uses of such tools and is making reasonable efforts to mitigate the risk of discrimination on the bases protected by Section 1557 (race, color, national origin, sex (including sex characteristics, sexual orientation, gender identity, and pregnancy or related conditions), age, and disability) resulting from the tools' use, in accordance with 45 CFR § 92.210. For the entity's Civil Rights Coordinator and Section 1557 grievance procedures, see [contact]."
|
|
907
|
+
},
|
|
908
|
+
"notes": "Section 1557's PCDST obligation is governance-heavy — most of the compliance work is internal (identifying tools, documenting mitigation, designating coordinators) rather than patient-facing text. The patient-facing element is the Section 1557 notice-of-availability under § 92.11 plus, where the entity exposes AI-informed decisions to patients, a clear acknowledgment that automated tools may inform clinical decisions and a path to discuss with a clinician. Covered entities include most healthcare providers receiving any form of federal financial assistance (Medicare-participating providers, Medicaid-participating providers, federally-qualified health centers, etc.), all health insurers in HHS-administered marketplaces, and HHS itself. The 'reasonable efforts' standard is intentionally flexible — OCR has stated in commentary that what constitutes 'reasonable' will scale with the entity's size and resources, but documentation is essential. PCDSTs explicitly include AI/ML decision-support tools and (per OCR commentary) tools that produce or use clinical scores (e.g., Epic Sepsis Model, Beth Israel Discharge Risk score, etc.). Federal funding loss is the principal sanction; OCR can also impose corrective action plans. State-level overlays may apply (e.g., California SB 1120 — Physicians Make Decisions Act, requiring physician review of AI-driven coverage denials in health plans — effective 2025-01-01). Stack with HIPAA Privacy Rule (45 CFR Part 164) when patient information is processed; stack with state AI hiring/employment-decision laws when the PCDST is used in employment of healthcare workers."
|
|
815
909
|
}
|
|
816
910
|
]
|
|
817
911
|
}
|