specweave 0.18.1 → 0.20.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CLAUDE.md +229 -1817
- package/README.md +68 -0
- package/bin/specweave.js +62 -6
- package/dist/plugins/specweave/lib/hooks/sync-living-docs.d.ts.map +1 -1
- package/dist/plugins/specweave/lib/hooks/sync-living-docs.js +3 -0
- package/dist/plugins/specweave/lib/hooks/sync-living-docs.js.map +1 -1
- package/dist/plugins/specweave/lib/hooks/update-ac-status.d.ts +21 -0
- package/dist/plugins/specweave/lib/hooks/update-ac-status.d.ts.map +1 -0
- package/dist/plugins/specweave/lib/hooks/update-ac-status.js +162 -0
- package/dist/plugins/specweave/lib/hooks/update-ac-status.js.map +1 -0
- package/dist/plugins/specweave-ado/lib/ado-spec-content-sync.d.ts.map +1 -1
- package/dist/plugins/specweave-ado/lib/ado-spec-content-sync.js +65 -6
- package/dist/plugins/specweave-ado/lib/ado-spec-content-sync.js.map +1 -1
- package/dist/plugins/specweave-github/lib/completion-calculator.d.ts +112 -0
- package/dist/plugins/specweave-github/lib/completion-calculator.d.ts.map +1 -0
- package/dist/plugins/specweave-github/lib/completion-calculator.js +301 -0
- package/dist/plugins/specweave-github/lib/completion-calculator.js.map +1 -0
- package/dist/plugins/specweave-github/lib/duplicate-detector.d.ts +3 -3
- package/dist/plugins/specweave-github/lib/duplicate-detector.js +3 -3
- package/dist/plugins/specweave-github/lib/epic-content-builder.d.ts +7 -0
- package/dist/plugins/specweave-github/lib/epic-content-builder.d.ts.map +1 -1
- package/dist/plugins/specweave-github/lib/epic-content-builder.js +42 -0
- package/dist/plugins/specweave-github/lib/epic-content-builder.js.map +1 -1
- package/dist/plugins/specweave-github/lib/github-client-v2.d.ts +14 -0
- package/dist/plugins/specweave-github/lib/github-client-v2.d.ts.map +1 -1
- package/dist/plugins/specweave-github/lib/github-client-v2.js +51 -0
- package/dist/plugins/specweave-github/lib/github-client-v2.js.map +1 -1
- package/dist/plugins/specweave-github/lib/github-epic-sync.js +1 -1
- package/dist/plugins/specweave-github/lib/github-epic-sync.js.map +1 -1
- package/dist/plugins/specweave-github/lib/github-feature-sync.d.ts +87 -0
- package/dist/plugins/specweave-github/lib/github-feature-sync.d.ts.map +1 -0
- package/dist/plugins/specweave-github/lib/github-feature-sync.js +412 -0
- package/dist/plugins/specweave-github/lib/github-feature-sync.js.map +1 -0
- package/dist/plugins/specweave-github/lib/github-spec-content-sync.d.ts.map +1 -1
- package/dist/plugins/specweave-github/lib/github-spec-content-sync.js +64 -13
- package/dist/plugins/specweave-github/lib/github-spec-content-sync.js.map +1 -1
- package/dist/plugins/specweave-github/lib/progress-comment-builder.d.ts +78 -0
- package/dist/plugins/specweave-github/lib/progress-comment-builder.d.ts.map +1 -0
- package/dist/plugins/specweave-github/lib/progress-comment-builder.js +237 -0
- package/dist/plugins/specweave-github/lib/progress-comment-builder.js.map +1 -0
- package/dist/plugins/specweave-github/lib/user-story-content-builder.d.ts +97 -0
- package/dist/plugins/specweave-github/lib/user-story-content-builder.d.ts.map +1 -0
- package/dist/plugins/specweave-github/lib/user-story-content-builder.js +301 -0
- package/dist/plugins/specweave-github/lib/user-story-content-builder.js.map +1 -0
- package/dist/plugins/specweave-github/lib/user-story-issue-builder.d.ts +83 -0
- package/dist/plugins/specweave-github/lib/user-story-issue-builder.d.ts.map +1 -0
- package/dist/plugins/specweave-github/lib/user-story-issue-builder.js +386 -0
- package/dist/plugins/specweave-github/lib/user-story-issue-builder.js.map +1 -0
- package/dist/plugins/specweave-jira/lib/enhanced-jira-sync.d.ts +8 -6
- package/dist/plugins/specweave-jira/lib/enhanced-jira-sync.d.ts.map +1 -1
- package/dist/plugins/specweave-jira/lib/enhanced-jira-sync.js +78 -117
- package/dist/plugins/specweave-jira/lib/enhanced-jira-sync.js.map +1 -1
- package/dist/plugins/specweave-kafka/lib/cli/kcat-wrapper.d.ts +57 -0
- package/dist/plugins/specweave-kafka/lib/cli/kcat-wrapper.d.ts.map +1 -0
- package/dist/plugins/specweave-kafka/lib/cli/kcat-wrapper.js +248 -0
- package/dist/plugins/specweave-kafka/lib/cli/kcat-wrapper.js.map +1 -0
- package/dist/plugins/specweave-kafka/lib/cli/types.d.ts +82 -0
- package/dist/plugins/specweave-kafka/lib/cli/types.d.ts.map +1 -0
- package/dist/plugins/specweave-kafka/lib/cli/types.js +13 -0
- package/dist/plugins/specweave-kafka/lib/cli/types.js.map +1 -0
- package/dist/plugins/specweave-kafka/lib/mcp/detector.d.ts +49 -0
- package/dist/plugins/specweave-kafka/lib/mcp/detector.d.ts.map +1 -0
- package/dist/plugins/specweave-kafka/lib/mcp/detector.js +316 -0
- package/dist/plugins/specweave-kafka/lib/mcp/detector.js.map +1 -0
- package/dist/plugins/specweave-kafka/lib/mcp/types.d.ts +70 -0
- package/dist/plugins/specweave-kafka/lib/mcp/types.d.ts.map +1 -0
- package/dist/plugins/specweave-kafka/lib/mcp/types.js +23 -0
- package/dist/plugins/specweave-kafka/lib/mcp/types.js.map +1 -0
- package/dist/plugins/specweave-kafka/lib/utils/partitioning.d.ts +85 -0
- package/dist/plugins/specweave-kafka/lib/utils/partitioning.d.ts.map +1 -0
- package/dist/plugins/specweave-kafka/lib/utils/partitioning.js +281 -0
- package/dist/plugins/specweave-kafka/lib/utils/partitioning.js.map +1 -0
- package/dist/plugins/specweave-kafka/lib/utils/sizing.d.ts +75 -0
- package/dist/plugins/specweave-kafka/lib/utils/sizing.d.ts.map +1 -0
- package/dist/plugins/specweave-kafka/lib/utils/sizing.js +238 -0
- package/dist/plugins/specweave-kafka/lib/utils/sizing.js.map +1 -0
- package/dist/src/cli/commands/import-docs.js +4 -4
- package/dist/src/cli/commands/import-docs.js.map +1 -1
- package/dist/src/cli/commands/init-multiproject.d.ts.map +1 -1
- package/dist/src/cli/commands/init-multiproject.js +17 -18
- package/dist/src/cli/commands/init-multiproject.js.map +1 -1
- package/dist/src/cli/commands/migrate-to-multiproject.d.ts.map +1 -1
- package/dist/src/cli/commands/migrate-to-multiproject.js +8 -4
- package/dist/src/cli/commands/migrate-to-multiproject.js.map +1 -1
- package/dist/src/cli/commands/switch-project.d.ts.map +1 -1
- package/dist/src/cli/commands/switch-project.js +9 -26
- package/dist/src/cli/commands/switch-project.js.map +1 -1
- package/dist/src/cli/commands/sync-spec-content.js +3 -0
- package/dist/src/cli/commands/sync-spec-content.js.map +1 -1
- package/dist/src/core/deduplication/command-deduplicator.d.ts +166 -0
- package/dist/src/core/deduplication/command-deduplicator.d.ts.map +1 -0
- package/dist/src/core/deduplication/command-deduplicator.js +254 -0
- package/dist/src/core/deduplication/command-deduplicator.js.map +1 -0
- package/dist/src/core/increment/active-increment-manager.d.ts +42 -15
- package/dist/src/core/increment/active-increment-manager.d.ts.map +1 -1
- package/dist/src/core/increment/active-increment-manager.js +113 -46
- package/dist/src/core/increment/active-increment-manager.js.map +1 -1
- package/dist/src/core/increment/conflict-resolver.d.ts +40 -0
- package/dist/src/core/increment/conflict-resolver.d.ts.map +1 -0
- package/dist/src/core/increment/conflict-resolver.js +219 -0
- package/dist/src/core/increment/conflict-resolver.js.map +1 -0
- package/dist/src/core/increment/discipline-checker.d.ts.map +1 -1
- package/dist/src/core/increment/discipline-checker.js +7 -1
- package/dist/src/core/increment/discipline-checker.js.map +1 -1
- package/dist/src/core/increment/duplicate-detector.d.ts +52 -0
- package/dist/src/core/increment/duplicate-detector.d.ts.map +1 -0
- package/dist/src/core/increment/duplicate-detector.js +276 -0
- package/dist/src/core/increment/duplicate-detector.js.map +1 -0
- package/dist/src/core/increment/increment-archiver.d.ts +90 -0
- package/dist/src/core/increment/increment-archiver.d.ts.map +1 -0
- package/dist/src/core/increment/increment-archiver.js +368 -0
- package/dist/src/core/increment/increment-archiver.js.map +1 -0
- package/dist/src/core/increment/increment-reopener.d.ts +165 -0
- package/dist/src/core/increment/increment-reopener.d.ts.map +1 -0
- package/dist/src/core/increment/increment-reopener.js +390 -0
- package/dist/src/core/increment/increment-reopener.js.map +1 -0
- package/dist/src/core/increment/metadata-manager.d.ts +26 -1
- package/dist/src/core/increment/metadata-manager.d.ts.map +1 -1
- package/dist/src/core/increment/metadata-manager.js +143 -5
- package/dist/src/core/increment/metadata-manager.js.map +1 -1
- package/dist/src/core/increment/recent-work-scanner.d.ts +121 -0
- package/dist/src/core/increment/recent-work-scanner.d.ts.map +1 -0
- package/dist/src/core/increment/recent-work-scanner.js +303 -0
- package/dist/src/core/increment/recent-work-scanner.js.map +1 -0
- package/dist/src/core/increment/types.d.ts +1 -0
- package/dist/src/core/increment/types.d.ts.map +1 -1
- package/dist/src/core/increment-utils.d.ts +112 -0
- package/dist/src/core/increment-utils.d.ts.map +1 -0
- package/dist/src/core/increment-utils.js +210 -0
- package/dist/src/core/increment-utils.js.map +1 -0
- package/dist/src/core/living-docs/ac-project-specific-generator.d.ts +65 -0
- package/dist/src/core/living-docs/ac-project-specific-generator.d.ts.map +1 -0
- package/dist/src/core/living-docs/ac-project-specific-generator.js +175 -0
- package/dist/src/core/living-docs/ac-project-specific-generator.js.map +1 -0
- package/dist/src/core/living-docs/feature-archiver.d.ts +130 -0
- package/dist/src/core/living-docs/feature-archiver.d.ts.map +1 -0
- package/dist/src/core/living-docs/feature-archiver.js +549 -0
- package/dist/src/core/living-docs/feature-archiver.js.map +1 -0
- package/dist/src/core/living-docs/feature-id-manager.d.ts +81 -0
- package/dist/src/core/living-docs/feature-id-manager.d.ts.map +1 -0
- package/dist/src/core/living-docs/feature-id-manager.js +339 -0
- package/dist/src/core/living-docs/feature-id-manager.js.map +1 -0
- package/dist/src/core/living-docs/hierarchy-mapper.d.ts +144 -83
- package/dist/src/core/living-docs/hierarchy-mapper.d.ts.map +1 -1
- package/dist/src/core/living-docs/hierarchy-mapper.js +488 -270
- package/dist/src/core/living-docs/hierarchy-mapper.js.map +1 -1
- package/dist/src/core/living-docs/index.d.ts +6 -0
- package/dist/src/core/living-docs/index.d.ts.map +1 -1
- package/dist/src/core/living-docs/index.js +6 -0
- package/dist/src/core/living-docs/index.js.map +1 -1
- package/dist/src/core/living-docs/project-detector.d.ts +6 -0
- package/dist/src/core/living-docs/project-detector.d.ts.map +1 -1
- package/dist/src/core/living-docs/project-detector.js +35 -1
- package/dist/src/core/living-docs/project-detector.js.map +1 -1
- package/dist/src/core/living-docs/spec-distributor.d.ts +100 -26
- package/dist/src/core/living-docs/spec-distributor.d.ts.map +1 -1
- package/dist/src/core/living-docs/spec-distributor.js +1275 -258
- package/dist/src/core/living-docs/spec-distributor.js.map +1 -1
- package/dist/src/core/living-docs/task-project-specific-generator.d.ts +109 -0
- package/dist/src/core/living-docs/task-project-specific-generator.d.ts.map +1 -0
- package/dist/src/core/living-docs/task-project-specific-generator.js +221 -0
- package/dist/src/core/living-docs/task-project-specific-generator.js.map +1 -0
- package/dist/src/core/living-docs/types.d.ts +143 -0
- package/dist/src/core/living-docs/types.d.ts.map +1 -1
- package/dist/src/core/project-manager.d.ts +2 -17
- package/dist/src/core/project-manager.d.ts.map +1 -1
- package/dist/src/core/project-manager.js +68 -48
- package/dist/src/core/project-manager.js.map +1 -1
- package/dist/src/core/spec-content-sync.d.ts +1 -1
- package/dist/src/core/spec-content-sync.d.ts.map +1 -1
- package/dist/src/core/sync/enhanced-content-builder.d.ts.map +1 -1
- package/dist/src/core/sync/enhanced-content-builder.js +2 -1
- package/dist/src/core/sync/enhanced-content-builder.js.map +1 -1
- package/dist/src/core/sync/performance-optimizer.d.ts +153 -0
- package/dist/src/core/sync/performance-optimizer.d.ts.map +1 -0
- package/dist/src/core/sync/performance-optimizer.js +220 -0
- package/dist/src/core/sync/performance-optimizer.js.map +1 -0
- package/dist/src/core/sync/retry-handler.d.ts +98 -0
- package/dist/src/core/sync/retry-handler.d.ts.map +1 -0
- package/dist/src/core/sync/retry-handler.js +196 -0
- package/dist/src/core/sync/retry-handler.js.map +1 -0
- package/dist/src/core/types/config.d.ts +94 -0
- package/dist/src/core/types/config.d.ts.map +1 -1
- package/dist/src/core/types/config.js +16 -0
- package/dist/src/core/types/config.js.map +1 -1
- package/dist/src/core/types/increment-metadata.d.ts +6 -0
- package/dist/src/core/types/increment-metadata.d.ts.map +1 -1
- package/dist/src/core/types/increment-metadata.js +10 -1
- package/dist/src/core/types/increment-metadata.js.map +1 -1
- package/dist/src/integrations/jira/jira-incremental-mapper.d.ts.map +1 -1
- package/dist/src/integrations/jira/jira-incremental-mapper.js +4 -8
- package/dist/src/integrations/jira/jira-incremental-mapper.js.map +1 -1
- package/dist/src/integrations/jira/jira-mapper.d.ts.map +1 -1
- package/dist/src/integrations/jira/jira-mapper.js +4 -8
- package/dist/src/integrations/jira/jira-mapper.js.map +1 -1
- package/package.json +1 -1
- package/plugins/specweave/COMMANDS.md +13 -4
- package/plugins/specweave/commands/specweave-abandon.md +22 -20
- package/plugins/specweave/commands/specweave-archive-features.md +121 -0
- package/plugins/specweave/commands/specweave-archive-increments.md +82 -0
- package/plugins/specweave/commands/specweave-archive.md +363 -0
- package/plugins/specweave/commands/specweave-backlog.md +211 -0
- package/plugins/specweave/commands/specweave-fix-duplicates.md +517 -0
- package/plugins/specweave/commands/specweave-increment.md +4 -3
- package/plugins/specweave/commands/specweave-progress.md +176 -27
- package/plugins/specweave/commands/specweave-reopen.md +391 -0
- package/plugins/specweave/commands/specweave-restore-feature.md +90 -0
- package/plugins/specweave/commands/specweave-restore.md +309 -0
- package/plugins/specweave/commands/specweave-resume.md +51 -23
- package/plugins/specweave/commands/specweave-status.md +41 -7
- package/plugins/specweave/commands/specweave-sync-specs.md +425 -0
- package/plugins/specweave/hooks/hooks.json +4 -0
- package/plugins/specweave/hooks/lib/sync-spec-content.sh +2 -2
- package/plugins/specweave/hooks/post-task-completion.sh +39 -0
- package/plugins/specweave/hooks/pre-command-deduplication.sh +83 -0
- package/plugins/specweave/hooks/user-prompt-submit.sh +1 -1
- package/plugins/specweave/lib/hooks/sync-living-docs.js +2 -0
- package/plugins/specweave/lib/hooks/sync-living-docs.ts +4 -0
- package/plugins/specweave/lib/hooks/update-ac-status.js +102 -0
- package/plugins/specweave/lib/hooks/update-ac-status.ts +192 -0
- package/plugins/specweave/skills/archive-increments/SKILL.md +198 -0
- package/plugins/specweave/skills/increment-planner/scripts/feature-utils.js +14 -0
- package/plugins/specweave/skills/smart-reopen-detector/SKILL.md +244 -0
- package/plugins/specweave-ado/lib/ado-spec-content-sync.js +49 -5
- package/plugins/specweave-ado/lib/ado-spec-content-sync.ts +72 -6
- package/plugins/specweave-confluent/.claude-plugin/plugin.json +23 -0
- package/plugins/specweave-confluent/README.md +375 -0
- package/plugins/specweave-confluent/agents/confluent-architect/AGENT.md +306 -0
- package/plugins/specweave-confluent/skills/confluent-kafka-connect/SKILL.md +453 -0
- package/plugins/specweave-confluent/skills/confluent-ksqldb/SKILL.md +470 -0
- package/plugins/specweave-confluent/skills/confluent-schema-registry/SKILL.md +316 -0
- package/plugins/specweave-github/agents/github-task-splitter/AGENT.md +2 -2
- package/plugins/specweave-github/agents/user-story-updater/AGENT.md +148 -0
- package/plugins/specweave-github/commands/specweave-github-cleanup-duplicates.md +1 -1
- package/plugins/specweave-github/commands/specweave-github-update-user-story.md +156 -0
- package/plugins/specweave-github/hooks/post-task-completion.sh +10 -9
- package/plugins/specweave-github/lib/completion-calculator.js +262 -0
- package/plugins/specweave-github/lib/completion-calculator.ts +434 -0
- package/plugins/specweave-github/lib/duplicate-detector.js +3 -3
- package/plugins/specweave-github/lib/duplicate-detector.ts +4 -4
- package/plugins/specweave-github/lib/epic-content-builder.js +38 -0
- package/plugins/specweave-github/lib/epic-content-builder.ts +59 -0
- package/plugins/specweave-github/lib/github-client-v2.js +49 -0
- package/plugins/specweave-github/lib/github-client-v2.ts +59 -0
- package/plugins/specweave-github/lib/github-epic-sync.ts +1 -1
- package/plugins/specweave-github/lib/github-feature-sync.js +381 -0
- package/plugins/specweave-github/lib/github-feature-sync.ts +568 -0
- package/plugins/specweave-github/lib/github-spec-content-sync.js +40 -10
- package/plugins/specweave-github/lib/github-spec-content-sync.ts +82 -14
- package/plugins/specweave-github/lib/progress-comment-builder.js +229 -0
- package/plugins/specweave-github/lib/progress-comment-builder.ts +324 -0
- package/plugins/specweave-github/lib/user-story-content-builder.js +299 -0
- package/plugins/specweave-github/lib/user-story-content-builder.ts +413 -0
- package/plugins/specweave-github/lib/user-story-issue-builder.js +344 -0
- package/plugins/specweave-github/lib/user-story-issue-builder.ts +543 -0
- package/plugins/specweave-github/skills/github-issue-standard/SKILL.md +189 -0
- package/plugins/specweave-jira/lib/enhanced-jira-sync.js +134 -0
- package/plugins/specweave-jira/lib/{enhanced-jira-sync.ts.disabled → enhanced-jira-sync.ts} +26 -52
- package/plugins/specweave-kafka/.claude-plugin/plugin.json +26 -0
- package/plugins/specweave-kafka/IMPLEMENTATION-COMPLETE.md +483 -0
- package/plugins/specweave-kafka/README.md +242 -0
- package/plugins/specweave-kafka/agents/kafka-architect/AGENT.md +235 -0
- package/plugins/specweave-kafka/agents/kafka-devops/AGENT.md +209 -0
- package/plugins/specweave-kafka/agents/kafka-observability/AGENT.md +266 -0
- package/plugins/specweave-kafka/commands/deploy.md +99 -0
- package/plugins/specweave-kafka/commands/dev-env.md +176 -0
- package/plugins/specweave-kafka/commands/mcp-configure.md +101 -0
- package/plugins/specweave-kafka/commands/monitor-setup.md +96 -0
- package/plugins/specweave-kafka/docker/kafka-local/docker-compose.yml +187 -0
- package/plugins/specweave-kafka/docker/redpanda/docker-compose.yml +199 -0
- package/plugins/specweave-kafka/docker/templates/consumer-nodejs.js +225 -0
- package/plugins/specweave-kafka/docker/templates/consumer-python.py +220 -0
- package/plugins/specweave-kafka/docker/templates/producer-nodejs.js +168 -0
- package/plugins/specweave-kafka/docker/templates/producer-python.py +167 -0
- package/plugins/specweave-kafka/lib/adapters/apache-kafka-adapter.js +438 -0
- package/plugins/specweave-kafka/lib/adapters/apache-kafka-adapter.ts +541 -0
- package/plugins/specweave-kafka/lib/adapters/platform-adapter.js +47 -0
- package/plugins/specweave-kafka/lib/adapters/platform-adapter.ts +343 -0
- package/plugins/specweave-kafka/lib/cli/kcat-wrapper.js +258 -0
- package/plugins/specweave-kafka/lib/cli/kcat-wrapper.ts +298 -0
- package/plugins/specweave-kafka/lib/cli/types.js +10 -0
- package/plugins/specweave-kafka/lib/cli/types.ts +92 -0
- package/plugins/specweave-kafka/lib/connectors/connector-catalog.js +305 -0
- package/plugins/specweave-kafka/lib/connectors/connector-catalog.ts +528 -0
- package/plugins/specweave-kafka/lib/documentation/diagram-generator.js +114 -0
- package/plugins/specweave-kafka/lib/documentation/diagram-generator.ts +195 -0
- package/plugins/specweave-kafka/lib/documentation/exporter.js +210 -0
- package/plugins/specweave-kafka/lib/documentation/exporter.ts +338 -0
- package/plugins/specweave-kafka/lib/documentation/schema-catalog-generator.js +60 -0
- package/plugins/specweave-kafka/lib/documentation/schema-catalog-generator.ts +130 -0
- package/plugins/specweave-kafka/lib/documentation/topology-generator.js +143 -0
- package/plugins/specweave-kafka/lib/documentation/topology-generator.ts +290 -0
- package/plugins/specweave-kafka/lib/mcp/detector.js +298 -0
- package/plugins/specweave-kafka/lib/mcp/detector.ts +352 -0
- package/plugins/specweave-kafka/lib/mcp/types.js +21 -0
- package/plugins/specweave-kafka/lib/mcp/types.ts +77 -0
- package/plugins/specweave-kafka/lib/multi-cluster/cluster-config-manager.js +193 -0
- package/plugins/specweave-kafka/lib/multi-cluster/cluster-config-manager.ts +362 -0
- package/plugins/specweave-kafka/lib/multi-cluster/cluster-switcher.js +188 -0
- package/plugins/specweave-kafka/lib/multi-cluster/cluster-switcher.ts +359 -0
- package/plugins/specweave-kafka/lib/multi-cluster/health-aggregator.js +195 -0
- package/plugins/specweave-kafka/lib/multi-cluster/health-aggregator.ts +380 -0
- package/plugins/specweave-kafka/lib/observability/opentelemetry-kafka.js +209 -0
- package/plugins/specweave-kafka/lib/observability/opentelemetry-kafka.ts +358 -0
- package/plugins/specweave-kafka/lib/patterns/advanced-ksqldb-patterns.js +354 -0
- package/plugins/specweave-kafka/lib/patterns/advanced-ksqldb-patterns.ts +563 -0
- package/plugins/specweave-kafka/lib/patterns/circuit-breaker-resilience.js +259 -0
- package/plugins/specweave-kafka/lib/patterns/circuit-breaker-resilience.ts +516 -0
- package/plugins/specweave-kafka/lib/patterns/dead-letter-queue.js +233 -0
- package/plugins/specweave-kafka/lib/patterns/dead-letter-queue.ts +423 -0
- package/plugins/specweave-kafka/lib/patterns/exactly-once-semantics.js +266 -0
- package/plugins/specweave-kafka/lib/patterns/exactly-once-semantics.ts +445 -0
- package/plugins/specweave-kafka/lib/patterns/flink-kafka-integration.js +312 -0
- package/plugins/specweave-kafka/lib/patterns/flink-kafka-integration.ts +561 -0
- package/plugins/specweave-kafka/lib/patterns/multi-dc-replication.js +289 -0
- package/plugins/specweave-kafka/lib/patterns/multi-dc-replication.ts +607 -0
- package/plugins/specweave-kafka/lib/patterns/rate-limiting-backpressure.js +264 -0
- package/plugins/specweave-kafka/lib/patterns/rate-limiting-backpressure.ts +498 -0
- package/plugins/specweave-kafka/lib/patterns/stream-processing-optimization.js +263 -0
- package/plugins/specweave-kafka/lib/patterns/stream-processing-optimization.ts +549 -0
- package/plugins/specweave-kafka/lib/patterns/tiered-storage-compaction.js +205 -0
- package/plugins/specweave-kafka/lib/patterns/tiered-storage-compaction.ts +399 -0
- package/plugins/specweave-kafka/lib/performance/performance-optimizer.js +249 -0
- package/plugins/specweave-kafka/lib/performance/performance-optimizer.ts +427 -0
- package/plugins/specweave-kafka/lib/security/kafka-security.js +252 -0
- package/plugins/specweave-kafka/lib/security/kafka-security.ts +494 -0
- package/plugins/specweave-kafka/lib/utils/capacity-planner.js +203 -0
- package/plugins/specweave-kafka/lib/utils/capacity-planner.ts +469 -0
- package/plugins/specweave-kafka/lib/utils/config-validator.js +419 -0
- package/plugins/specweave-kafka/lib/utils/config-validator.ts +564 -0
- package/plugins/specweave-kafka/lib/utils/partitioning.js +329 -0
- package/plugins/specweave-kafka/lib/utils/partitioning.ts +473 -0
- package/plugins/specweave-kafka/lib/utils/sizing.js +221 -0
- package/plugins/specweave-kafka/lib/utils/sizing.ts +374 -0
- package/plugins/specweave-kafka/monitoring/grafana/dashboards/kafka-broker-metrics.json +628 -0
- package/plugins/specweave-kafka/monitoring/grafana/dashboards/kafka-cluster-overview.json +564 -0
- package/plugins/specweave-kafka/monitoring/grafana/dashboards/kafka-consumer-lag.json +509 -0
- package/plugins/specweave-kafka/monitoring/grafana/dashboards/kafka-jvm-metrics.json +674 -0
- package/plugins/specweave-kafka/monitoring/grafana/dashboards/kafka-topic-metrics.json +578 -0
- package/plugins/specweave-kafka/monitoring/grafana/provisioning/dashboards/kafka.yml +17 -0
- package/plugins/specweave-kafka/monitoring/grafana/provisioning/datasources/prometheus.yml +17 -0
- package/plugins/specweave-kafka/monitoring/prometheus/kafka-alerts.yml +415 -0
- package/plugins/specweave-kafka/monitoring/prometheus/kafka-jmx-exporter.yml +256 -0
- package/plugins/specweave-kafka/package.json +41 -0
- package/plugins/specweave-kafka/skills/kafka-architecture/SKILL.md +647 -0
- package/plugins/specweave-kafka/skills/kafka-cli-tools/SKILL.md +433 -0
- package/plugins/specweave-kafka/skills/kafka-iac-deployment/SKILL.md +449 -0
- package/plugins/specweave-kafka/skills/kafka-kubernetes/SKILL.md +667 -0
- package/plugins/specweave-kafka/skills/kafka-mcp-integration/SKILL.md +273 -0
- package/plugins/specweave-kafka/skills/kafka-observability/SKILL.md +576 -0
- package/plugins/specweave-kafka/templates/config/broker-production.properties +254 -0
- package/plugins/specweave-kafka/templates/config/consumer-low-latency.properties +112 -0
- package/plugins/specweave-kafka/templates/config/producer-high-throughput.properties +120 -0
- package/plugins/specweave-kafka/templates/migration/mirrormaker2-config.properties +234 -0
- package/plugins/specweave-kafka/templates/monitoring/grafana/multi-cluster-dashboard.json +686 -0
- package/plugins/specweave-kafka/terraform/apache-kafka/main.tf +347 -0
- package/plugins/specweave-kafka/terraform/apache-kafka/outputs.tf +107 -0
- package/plugins/specweave-kafka/terraform/apache-kafka/templates/kafka-broker-init.sh.tpl +216 -0
- package/plugins/specweave-kafka/terraform/apache-kafka/variables.tf +156 -0
- package/plugins/specweave-kafka/terraform/aws-msk/main.tf +362 -0
- package/plugins/specweave-kafka/terraform/aws-msk/outputs.tf +93 -0
- package/plugins/specweave-kafka/terraform/aws-msk/templates/server.properties.tpl +32 -0
- package/plugins/specweave-kafka/terraform/aws-msk/variables.tf +235 -0
- package/plugins/specweave-kafka/terraform/azure-event-hubs/main.tf +281 -0
- package/plugins/specweave-kafka/terraform/azure-event-hubs/outputs.tf +118 -0
- package/plugins/specweave-kafka/terraform/azure-event-hubs/variables.tf +148 -0
- package/plugins/specweave-kafka/tsconfig.json +21 -0
- package/plugins/specweave-kafka-streams/.claude-plugin/plugin.json +23 -0
- package/plugins/specweave-kafka-streams/README.md +310 -0
- package/plugins/specweave-kafka-streams/skills/kafka-streams-topology/SKILL.md +539 -0
- package/plugins/specweave-n8n/.claude-plugin/plugin.json +22 -0
- package/plugins/specweave-n8n/README.md +354 -0
- package/plugins/specweave-n8n/skills/n8n-kafka-workflows/SKILL.md +504 -0
- package/plugins/specweave-release/commands/specweave-release-platform.md +1 -1
- package/plugins/specweave-release/hooks/post-task-completion.sh +2 -2
- package/src/templates/AGENTS.md.template +601 -7
- package/src/templates/CLAUDE.md.template +188 -88
- package/plugins/specweave-ado/commands/specweave-ado-sync-spec.md +0 -255
- package/plugins/specweave-github/commands/specweave-github-sync-epic.md +0 -248
- package/plugins/specweave-github/commands/specweave-github-sync-from.md +0 -147
- package/plugins/specweave-github/commands/specweave-github-sync-spec.md +0 -208
- package/plugins/specweave-github/commands/specweave-github-sync-tasks.md +0 -530
- package/plugins/specweave-jira/commands/specweave-jira-sync-epic.md +0 -267
- package/plugins/specweave-jira/commands/specweave-jira-sync-spec.md +0 -240
|
@@ -0,0 +1,298 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* kcat CLI Wrapper
|
|
3
|
+
*
|
|
4
|
+
* Type-safe TypeScript wrapper for kcat (kafkacat) CLI tool.
|
|
5
|
+
* Provides produce, consume, metadata, and query operations.
|
|
6
|
+
*/
|
|
7
|
+
|
|
8
|
+
import { exec } from 'child_process';
|
|
9
|
+
import { promisify } from 'util';
|
|
10
|
+
import {
|
|
11
|
+
KcatProduceOptions,
|
|
12
|
+
KcatConsumeOptions,
|
|
13
|
+
KcatMetadataOptions,
|
|
14
|
+
KcatQueryOptions,
|
|
15
|
+
ClusterMetadata,
|
|
16
|
+
ConsumerMessage,
|
|
17
|
+
CLIExecutionResult,
|
|
18
|
+
TopicMetadata,
|
|
19
|
+
PartitionMetadata
|
|
20
|
+
} from './types';
|
|
21
|
+
|
|
22
|
+
const execAsync = promisify(exec);
|
|
23
|
+
|
|
24
|
+
export class KcatWrapper {
|
|
25
|
+
/**
|
|
26
|
+
* Produce a message to Kafka topic
|
|
27
|
+
*/
|
|
28
|
+
async produce(message: string, options: KcatProduceOptions): Promise<CLIExecutionResult> {
|
|
29
|
+
const args: string[] = [
|
|
30
|
+
'kcat',
|
|
31
|
+
'-P', // Producer mode
|
|
32
|
+
'-b', options.brokers,
|
|
33
|
+
'-t', options.topic
|
|
34
|
+
];
|
|
35
|
+
|
|
36
|
+
if (options.key) {
|
|
37
|
+
args.push('-K', ':'); // Key delimiter
|
|
38
|
+
}
|
|
39
|
+
|
|
40
|
+
if (options.partition !== undefined) {
|
|
41
|
+
args.push('-p', options.partition.toString());
|
|
42
|
+
}
|
|
43
|
+
|
|
44
|
+
if (options.compression) {
|
|
45
|
+
args.push('-z', options.compression);
|
|
46
|
+
}
|
|
47
|
+
|
|
48
|
+
if (options.acks !== undefined) {
|
|
49
|
+
const acksValue = options.acks === 'all' ? '-1' : options.acks.toString();
|
|
50
|
+
args.push('-X', `acks=${acksValue}`);
|
|
51
|
+
}
|
|
52
|
+
|
|
53
|
+
const command = args.join(' ');
|
|
54
|
+
const input = options.key ? `${options.key}:${message}` : message;
|
|
55
|
+
|
|
56
|
+
return this.execute(command, input);
|
|
57
|
+
}
|
|
58
|
+
|
|
59
|
+
/**
|
|
60
|
+
* Consume messages from Kafka topic
|
|
61
|
+
*/
|
|
62
|
+
async consume(options: KcatConsumeOptions): Promise<ConsumerMessage[]> {
|
|
63
|
+
const args: string[] = [
|
|
64
|
+
'kcat',
|
|
65
|
+
'-C', // Consumer mode
|
|
66
|
+
'-b', options.brokers,
|
|
67
|
+
'-t', options.topic,
|
|
68
|
+
'-f', '%t:%p:%o:%k:%s\\n' // Format: topic:partition:offset:key:value
|
|
69
|
+
];
|
|
70
|
+
|
|
71
|
+
if (options.offset) {
|
|
72
|
+
args.push('-o', options.offset.toString());
|
|
73
|
+
}
|
|
74
|
+
|
|
75
|
+
if (options.partition !== undefined) {
|
|
76
|
+
args.push('-p', options.partition.toString());
|
|
77
|
+
}
|
|
78
|
+
|
|
79
|
+
if (options.count) {
|
|
80
|
+
args.push('-c', options.count.toString());
|
|
81
|
+
}
|
|
82
|
+
|
|
83
|
+
if (options.groupId) {
|
|
84
|
+
args.push('-G', options.groupId);
|
|
85
|
+
}
|
|
86
|
+
|
|
87
|
+
const command = args.join(' ');
|
|
88
|
+
const result = await this.execute(command);
|
|
89
|
+
|
|
90
|
+
if (!result.success) {
|
|
91
|
+
throw new Error(`kcat consume failed: ${result.error}`);
|
|
92
|
+
}
|
|
93
|
+
|
|
94
|
+
return this.parseConsumeOutput(result.output);
|
|
95
|
+
}
|
|
96
|
+
|
|
97
|
+
/**
|
|
98
|
+
* Get cluster metadata
|
|
99
|
+
*/
|
|
100
|
+
async getMetadata(options: KcatMetadataOptions): Promise<ClusterMetadata> {
|
|
101
|
+
const args: string[] = [
|
|
102
|
+
'kcat',
|
|
103
|
+
'-L', // Metadata mode
|
|
104
|
+
'-b', options.brokers,
|
|
105
|
+
'-J' // JSON output
|
|
106
|
+
];
|
|
107
|
+
|
|
108
|
+
if (options.topic) {
|
|
109
|
+
args.push('-t', options.topic);
|
|
110
|
+
}
|
|
111
|
+
|
|
112
|
+
const command = args.join(' ');
|
|
113
|
+
const result = await this.execute(command);
|
|
114
|
+
|
|
115
|
+
if (!result.success) {
|
|
116
|
+
throw new Error(`kcat metadata failed: ${result.error}`);
|
|
117
|
+
}
|
|
118
|
+
|
|
119
|
+
return this.parseMetadataOutput(result.output);
|
|
120
|
+
}
|
|
121
|
+
|
|
122
|
+
/**
|
|
123
|
+
* Query topic offsets
|
|
124
|
+
*/
|
|
125
|
+
async queryOffsets(options: KcatQueryOptions): Promise<Record<number, { earliest: number; latest: number }>> {
|
|
126
|
+
const args: string[] = [
|
|
127
|
+
'kcat',
|
|
128
|
+
'-Q', // Query mode
|
|
129
|
+
'-b', options.brokers,
|
|
130
|
+
'-t', options.topic
|
|
131
|
+
];
|
|
132
|
+
|
|
133
|
+
if (options.partition !== undefined) {
|
|
134
|
+
args.push('-p', options.partition.toString());
|
|
135
|
+
}
|
|
136
|
+
|
|
137
|
+
const command = args.join(' ');
|
|
138
|
+
const result = await this.execute(command);
|
|
139
|
+
|
|
140
|
+
if (!result.success) {
|
|
141
|
+
throw new Error(`kcat query failed: ${result.error}`);
|
|
142
|
+
}
|
|
143
|
+
|
|
144
|
+
return this.parseQueryOutput(result.output);
|
|
145
|
+
}
|
|
146
|
+
|
|
147
|
+
/**
|
|
148
|
+
* Execute kcat command
|
|
149
|
+
*/
|
|
150
|
+
private async execute(command: string, input?: string): Promise<CLIExecutionResult> {
|
|
151
|
+
const startTime = Date.now();
|
|
152
|
+
|
|
153
|
+
try {
|
|
154
|
+
const execOptions: any = {
|
|
155
|
+
maxBuffer: 10 * 1024 * 1024, // 10MB buffer
|
|
156
|
+
timeout: 30000 // 30 second timeout
|
|
157
|
+
};
|
|
158
|
+
|
|
159
|
+
if (input) {
|
|
160
|
+
execOptions.input = input;
|
|
161
|
+
}
|
|
162
|
+
|
|
163
|
+
const { stdout, stderr } = await execAsync(command, execOptions);
|
|
164
|
+
const duration = Date.now() - startTime;
|
|
165
|
+
|
|
166
|
+
return {
|
|
167
|
+
success: true,
|
|
168
|
+
output: stdout,
|
|
169
|
+
error: stderr || undefined,
|
|
170
|
+
exitCode: 0,
|
|
171
|
+
command,
|
|
172
|
+
duration
|
|
173
|
+
};
|
|
174
|
+
} catch (error: any) {
|
|
175
|
+
const duration = Date.now() - startTime;
|
|
176
|
+
|
|
177
|
+
return {
|
|
178
|
+
success: false,
|
|
179
|
+
output: error.stdout || '',
|
|
180
|
+
error: error.stderr || error.message,
|
|
181
|
+
exitCode: error.code || 1,
|
|
182
|
+
command,
|
|
183
|
+
duration
|
|
184
|
+
};
|
|
185
|
+
}
|
|
186
|
+
}
|
|
187
|
+
|
|
188
|
+
/**
|
|
189
|
+
* Parse consume output into messages
|
|
190
|
+
*/
|
|
191
|
+
private parseConsumeOutput(output: string): ConsumerMessage[] {
|
|
192
|
+
const lines = output.trim().split('\n').filter(line => line.length > 0);
|
|
193
|
+
const messages: ConsumerMessage[] = [];
|
|
194
|
+
|
|
195
|
+
for (const line of lines) {
|
|
196
|
+
// Format: topic:partition:offset:key:value
|
|
197
|
+
const parts = line.split(':');
|
|
198
|
+
if (parts.length >= 5) {
|
|
199
|
+
messages.push({
|
|
200
|
+
topic: parts[0],
|
|
201
|
+
partition: parseInt(parts[1]),
|
|
202
|
+
offset: parseInt(parts[2]),
|
|
203
|
+
key: parts[3] || undefined,
|
|
204
|
+
value: parts.slice(4).join(':'), // Rejoin in case value contains ':'
|
|
205
|
+
timestamp: Date.now() // kcat doesn't provide timestamp in this format
|
|
206
|
+
});
|
|
207
|
+
}
|
|
208
|
+
}
|
|
209
|
+
|
|
210
|
+
return messages;
|
|
211
|
+
}
|
|
212
|
+
|
|
213
|
+
/**
|
|
214
|
+
* Parse metadata JSON output
|
|
215
|
+
*/
|
|
216
|
+
private parseMetadataOutput(output: string): ClusterMetadata {
|
|
217
|
+
const metadata = JSON.parse(output);
|
|
218
|
+
|
|
219
|
+
return {
|
|
220
|
+
clusterId: metadata.originatingBroker?.cluster || 'unknown',
|
|
221
|
+
controllerId: metadata.controllerId || -1,
|
|
222
|
+
brokers: (metadata.brokers || []).map((b: any) => ({
|
|
223
|
+
id: b.id,
|
|
224
|
+
host: b.host,
|
|
225
|
+
port: b.port,
|
|
226
|
+
rack: b.rack
|
|
227
|
+
})),
|
|
228
|
+
topics: (metadata.topics || []).map((t: any) => ({
|
|
229
|
+
name: t.topic,
|
|
230
|
+
replicationFactor: this.calculateReplicationFactor(t.partitions),
|
|
231
|
+
configs: {},
|
|
232
|
+
partitions: (t.partitions || []).map((p: any) => ({
|
|
233
|
+
id: p.partition,
|
|
234
|
+
leader: p.leader,
|
|
235
|
+
replicas: p.replicas || [],
|
|
236
|
+
isr: p.isrs || [],
|
|
237
|
+
earliestOffset: 0, // Not provided in metadata
|
|
238
|
+
latestOffset: 0 // Not provided in metadata
|
|
239
|
+
}))
|
|
240
|
+
}))
|
|
241
|
+
};
|
|
242
|
+
}
|
|
243
|
+
|
|
244
|
+
/**
|
|
245
|
+
* Calculate replication factor from partitions
|
|
246
|
+
*/
|
|
247
|
+
private calculateReplicationFactor(partitions: any[]): number {
|
|
248
|
+
if (!partitions || partitions.length === 0) return 0;
|
|
249
|
+
return partitions[0].replicas?.length || 0;
|
|
250
|
+
}
|
|
251
|
+
|
|
252
|
+
/**
|
|
253
|
+
* Parse query offset output
|
|
254
|
+
*/
|
|
255
|
+
private parseQueryOutput(output: string): Record<number, { earliest: number; latest: number }> {
|
|
256
|
+
const lines = output.trim().split('\n');
|
|
257
|
+
const offsets: Record<number, { earliest: number; latest: number }> = {};
|
|
258
|
+
|
|
259
|
+
for (const line of lines) {
|
|
260
|
+
// Format: topic partition [earliest] [latest]
|
|
261
|
+
const match = line.match(/(\d+)\s+\[(\d+)\]\s+\[(\d+)\]/);
|
|
262
|
+
if (match) {
|
|
263
|
+
const partition = parseInt(match[1]);
|
|
264
|
+
offsets[partition] = {
|
|
265
|
+
earliest: parseInt(match[2]),
|
|
266
|
+
latest: parseInt(match[3])
|
|
267
|
+
};
|
|
268
|
+
}
|
|
269
|
+
}
|
|
270
|
+
|
|
271
|
+
return offsets;
|
|
272
|
+
}
|
|
273
|
+
|
|
274
|
+
/**
|
|
275
|
+
* Check if kcat is installed
|
|
276
|
+
*/
|
|
277
|
+
static async isInstalled(): Promise<boolean> {
|
|
278
|
+
try {
|
|
279
|
+
const { stdout } = await execAsync('which kcat 2>/dev/null || which kafkacat 2>/dev/null');
|
|
280
|
+
return !!stdout.trim();
|
|
281
|
+
} catch {
|
|
282
|
+
return false;
|
|
283
|
+
}
|
|
284
|
+
}
|
|
285
|
+
|
|
286
|
+
/**
|
|
287
|
+
* Get kcat version
|
|
288
|
+
*/
|
|
289
|
+
static async getVersion(): Promise<string> {
|
|
290
|
+
try {
|
|
291
|
+
const { stdout } = await execAsync('kcat -V 2>&1 || kafkacat -V 2>&1');
|
|
292
|
+
const match = stdout.match(/kafkacat - Apache Kafka producer and consumer tool\s+(\S+)/);
|
|
293
|
+
return match ? match[1] : 'unknown';
|
|
294
|
+
} catch {
|
|
295
|
+
return 'unknown';
|
|
296
|
+
}
|
|
297
|
+
}
|
|
298
|
+
}
|
|
@@ -0,0 +1,92 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* CLI Tool Types and Interfaces
|
|
3
|
+
*
|
|
4
|
+
* Types for kcat, kcli, kaf, kafkactl wrappers
|
|
5
|
+
*/
|
|
6
|
+
|
|
7
|
+
export enum CLITool {
|
|
8
|
+
KCAT = 'kcat',
|
|
9
|
+
KCLI = 'kcli',
|
|
10
|
+
KAF = 'kaf',
|
|
11
|
+
KAFKACTL = 'kafkactl'
|
|
12
|
+
}
|
|
13
|
+
|
|
14
|
+
export interface KcatProduceOptions {
|
|
15
|
+
topic: string;
|
|
16
|
+
brokers: string;
|
|
17
|
+
key?: string;
|
|
18
|
+
partition?: number;
|
|
19
|
+
headers?: Record<string, string>;
|
|
20
|
+
compression?: 'none' | 'gzip' | 'snappy' | 'lz4' | 'zstd';
|
|
21
|
+
acks?: 0 | 1 | 'all';
|
|
22
|
+
}
|
|
23
|
+
|
|
24
|
+
export interface KcatConsumeOptions {
|
|
25
|
+
topic: string;
|
|
26
|
+
brokers: string;
|
|
27
|
+
offset?: 'beginning' | 'end' | 'stored' | number;
|
|
28
|
+
partition?: number;
|
|
29
|
+
count?: number;
|
|
30
|
+
format?: 'json' | 'avro' | 'protobuf' | 'string';
|
|
31
|
+
groupId?: string;
|
|
32
|
+
}
|
|
33
|
+
|
|
34
|
+
export interface KcatMetadataOptions {
|
|
35
|
+
brokers: string;
|
|
36
|
+
topic?: string;
|
|
37
|
+
}
|
|
38
|
+
|
|
39
|
+
export interface KcatQueryOptions {
|
|
40
|
+
brokers: string;
|
|
41
|
+
topic: string;
|
|
42
|
+
partition?: number;
|
|
43
|
+
}
|
|
44
|
+
|
|
45
|
+
export interface TopicMetadata {
|
|
46
|
+
name: string;
|
|
47
|
+
partitions: PartitionMetadata[];
|
|
48
|
+
replicationFactor: number;
|
|
49
|
+
configs: Record<string, string>;
|
|
50
|
+
}
|
|
51
|
+
|
|
52
|
+
export interface PartitionMetadata {
|
|
53
|
+
id: number;
|
|
54
|
+
leader: number;
|
|
55
|
+
replicas: number[];
|
|
56
|
+
isr: number[]; // In-Sync Replicas
|
|
57
|
+
earliestOffset: number;
|
|
58
|
+
latestOffset: number;
|
|
59
|
+
}
|
|
60
|
+
|
|
61
|
+
export interface BrokerMetadata {
|
|
62
|
+
id: number;
|
|
63
|
+
host: string;
|
|
64
|
+
port: number;
|
|
65
|
+
rack?: string;
|
|
66
|
+
}
|
|
67
|
+
|
|
68
|
+
export interface ClusterMetadata {
|
|
69
|
+
brokers: BrokerMetadata[];
|
|
70
|
+
topics: TopicMetadata[];
|
|
71
|
+
clusterId: string;
|
|
72
|
+
controllerId: number;
|
|
73
|
+
}
|
|
74
|
+
|
|
75
|
+
export interface ConsumerMessage {
|
|
76
|
+
topic: string;
|
|
77
|
+
partition: number;
|
|
78
|
+
offset: number;
|
|
79
|
+
timestamp: number;
|
|
80
|
+
key?: string;
|
|
81
|
+
value: string;
|
|
82
|
+
headers?: Record<string, string>;
|
|
83
|
+
}
|
|
84
|
+
|
|
85
|
+
export interface CLIExecutionResult {
|
|
86
|
+
success: boolean;
|
|
87
|
+
output: string;
|
|
88
|
+
error?: string;
|
|
89
|
+
exitCode: number;
|
|
90
|
+
command: string;
|
|
91
|
+
duration: number;
|
|
92
|
+
}
|
|
@@ -0,0 +1,305 @@
|
|
|
1
|
+
var ConnectorCategory = /* @__PURE__ */ ((ConnectorCategory2) => {
|
|
2
|
+
ConnectorCategory2["DATABASE"] = "database";
|
|
3
|
+
ConnectorCategory2["CLOUD_STORAGE"] = "cloud-storage";
|
|
4
|
+
ConnectorCategory2["MESSAGE_QUEUE"] = "message-queue";
|
|
5
|
+
ConnectorCategory2["DATA_WAREHOUSE"] = "data-warehouse";
|
|
6
|
+
ConnectorCategory2["SEARCH"] = "search";
|
|
7
|
+
ConnectorCategory2["MONITORING"] = "monitoring";
|
|
8
|
+
ConnectorCategory2["FILE_SYSTEM"] = "file-system";
|
|
9
|
+
return ConnectorCategory2;
|
|
10
|
+
})(ConnectorCategory || {});
|
|
11
|
+
class ConnectorCatalog {
|
|
12
|
+
/**
|
|
13
|
+
* JDBC Source Connector (Database → Kafka)
|
|
14
|
+
*
|
|
15
|
+
* Use case: Stream database tables to Kafka (CDC alternative)
|
|
16
|
+
*/
|
|
17
|
+
static jdbcSource(options) {
|
|
18
|
+
const config = {
|
|
19
|
+
name: `jdbc-source-${options.tableName}`,
|
|
20
|
+
"connector.class": "io.confluent.connect.jdbc.JdbcSourceConnector",
|
|
21
|
+
"tasks.max": "1",
|
|
22
|
+
"connection.url": options.connectionUrl,
|
|
23
|
+
"connection.user": options.user,
|
|
24
|
+
"connection.password": options.password,
|
|
25
|
+
"table.whitelist": options.tableName,
|
|
26
|
+
"mode": options.mode,
|
|
27
|
+
"topic.prefix": options.topicPrefix,
|
|
28
|
+
"poll.interval.ms": "5000"
|
|
29
|
+
};
|
|
30
|
+
if (options.mode === "incrementing" && options.incrementingColumn) {
|
|
31
|
+
config["incrementing.column.name"] = options.incrementingColumn;
|
|
32
|
+
}
|
|
33
|
+
if ((options.mode === "timestamp" || options.mode === "timestamp+incrementing") && options.timestampColumn) {
|
|
34
|
+
config["timestamp.column.name"] = options.timestampColumn;
|
|
35
|
+
}
|
|
36
|
+
return config;
|
|
37
|
+
}
|
|
38
|
+
/**
|
|
39
|
+
* JDBC Sink Connector (Kafka → Database)
|
|
40
|
+
*
|
|
41
|
+
* Use case: Write Kafka events to database tables
|
|
42
|
+
*/
|
|
43
|
+
static jdbcSink(options) {
|
|
44
|
+
return {
|
|
45
|
+
name: `jdbc-sink-${options.topics.join("-")}`,
|
|
46
|
+
"connector.class": "io.confluent.connect.jdbc.JdbcSinkConnector",
|
|
47
|
+
"tasks.max": "1",
|
|
48
|
+
"connection.url": options.connectionUrl,
|
|
49
|
+
"connection.user": options.user,
|
|
50
|
+
"connection.password": options.password,
|
|
51
|
+
"topics": options.topics.join(","),
|
|
52
|
+
"auto.create": options.autoCreate !== false ? "true" : "false",
|
|
53
|
+
"auto.evolve": options.autoEvolve !== false ? "true" : "false",
|
|
54
|
+
"insert.mode": options.insertMode || "insert",
|
|
55
|
+
"pk.mode": options.pkMode || "none",
|
|
56
|
+
"pk.fields": options.pkFields?.join(",") || ""
|
|
57
|
+
};
|
|
58
|
+
}
|
|
59
|
+
/**
|
|
60
|
+
* Debezium MySQL Source Connector (CDC)
|
|
61
|
+
*
|
|
62
|
+
* Use case: Capture all database changes in real-time
|
|
63
|
+
*/
|
|
64
|
+
static debeziumMySQL(options) {
|
|
65
|
+
return {
|
|
66
|
+
name: `debezium-mysql-${options.databaseName}`,
|
|
67
|
+
"connector.class": "io.debezium.connector.mysql.MySqlConnector",
|
|
68
|
+
"tasks.max": "1",
|
|
69
|
+
"database.hostname": options.hostname,
|
|
70
|
+
"database.port": options.port.toString(),
|
|
71
|
+
"database.user": options.user,
|
|
72
|
+
"database.password": options.password,
|
|
73
|
+
"database.server.id": options.serverId.toString(),
|
|
74
|
+
"database.server.name": options.serverName,
|
|
75
|
+
"database.include.list": options.databaseName,
|
|
76
|
+
"table.include.list": options.tableIncludeList || `${options.databaseName}.*`,
|
|
77
|
+
"database.history.kafka.bootstrap.servers": "localhost:9092",
|
|
78
|
+
"database.history.kafka.topic": `dbhistory.${options.databaseName}`
|
|
79
|
+
};
|
|
80
|
+
}
|
|
81
|
+
/**
|
|
82
|
+
* Debezium PostgreSQL Source Connector (CDC)
|
|
83
|
+
*/
|
|
84
|
+
static debeziumPostgreSQL(options) {
|
|
85
|
+
return {
|
|
86
|
+
name: `debezium-postgres-${options.databaseName}`,
|
|
87
|
+
"connector.class": "io.debezium.connector.postgresql.PostgresConnector",
|
|
88
|
+
"tasks.max": "1",
|
|
89
|
+
"database.hostname": options.hostname,
|
|
90
|
+
"database.port": options.port.toString(),
|
|
91
|
+
"database.user": options.user,
|
|
92
|
+
"database.password": options.password,
|
|
93
|
+
"database.dbname": options.databaseName,
|
|
94
|
+
"database.server.name": options.serverName,
|
|
95
|
+
"slot.name": options.slotName,
|
|
96
|
+
"publication.name": options.publicationName,
|
|
97
|
+
"plugin.name": "pgoutput"
|
|
98
|
+
};
|
|
99
|
+
}
|
|
100
|
+
/**
|
|
101
|
+
* S3 Sink Connector (Kafka → AWS S3)
|
|
102
|
+
*
|
|
103
|
+
* Use case: Archive Kafka data to S3 for analytics
|
|
104
|
+
*/
|
|
105
|
+
static s3Sink(options) {
|
|
106
|
+
const formatClass = options.format === "avro" ? "io.confluent.connect.s3.format.avro.AvroFormat" : options.format === "parquet" ? "io.confluent.connect.s3.format.parquet.ParquetFormat" : "io.confluent.connect.s3.format.json.JsonFormat";
|
|
107
|
+
return {
|
|
108
|
+
name: `s3-sink-${options.topics.join("-")}`,
|
|
109
|
+
"connector.class": "io.confluent.connect.s3.S3SinkConnector",
|
|
110
|
+
"tasks.max": "1",
|
|
111
|
+
"topics": options.topics.join(","),
|
|
112
|
+
"s3.bucket.name": options.s3BucketName,
|
|
113
|
+
"s3.region": options.s3Region,
|
|
114
|
+
"format.class": formatClass,
|
|
115
|
+
"flush.size": (options.flushSize || 1e3).toString(),
|
|
116
|
+
"rotate.interval.ms": (options.rotateIntervalMs || 36e5).toString(),
|
|
117
|
+
"partitioner.class": options.partitionerClass || "io.confluent.connect.storage.partitioner.TimeBasedPartitioner",
|
|
118
|
+
"path.format": "'year'=YYYY/'month'=MM/'day'=dd/'hour'=HH",
|
|
119
|
+
"locale": "en-US",
|
|
120
|
+
"timezone": "UTC",
|
|
121
|
+
"timestamp.extractor": "Record"
|
|
122
|
+
};
|
|
123
|
+
}
|
|
124
|
+
/**
|
|
125
|
+
* Elasticsearch Sink Connector (Kafka → Elasticsearch)
|
|
126
|
+
*
|
|
127
|
+
* Use case: Index Kafka data for full-text search
|
|
128
|
+
*/
|
|
129
|
+
static elasticsearchSink(options) {
|
|
130
|
+
return {
|
|
131
|
+
name: `elasticsearch-sink-${options.topics.join("-")}`,
|
|
132
|
+
"connector.class": "io.confluent.connect.elasticsearch.ElasticsearchSinkConnector",
|
|
133
|
+
"tasks.max": "1",
|
|
134
|
+
"topics": options.topics.join(","),
|
|
135
|
+
"connection.url": options.connectionUrl,
|
|
136
|
+
"type.name": options.typeName || "_doc",
|
|
137
|
+
"key.ignore": "true",
|
|
138
|
+
"schema.ignore": "false",
|
|
139
|
+
"batch.size": (options.batchSize || 2e3).toString(),
|
|
140
|
+
"max.buffered.records": "20000",
|
|
141
|
+
"linger.ms": "1000",
|
|
142
|
+
"flush.timeout.ms": "10000",
|
|
143
|
+
"max.in.flight.requests": "5",
|
|
144
|
+
"retry.backoff.ms": "100",
|
|
145
|
+
"max.retries": "10"
|
|
146
|
+
};
|
|
147
|
+
}
|
|
148
|
+
/**
|
|
149
|
+
* MongoDB Sink Connector (Kafka → MongoDB)
|
|
150
|
+
*
|
|
151
|
+
* Use case: Write Kafka events to MongoDB collections
|
|
152
|
+
*/
|
|
153
|
+
static mongodbSink(options) {
|
|
154
|
+
return {
|
|
155
|
+
name: `mongodb-sink-${options.topics.join("-")}`,
|
|
156
|
+
"connector.class": "com.mongodb.kafka.connect.MongoSinkConnector",
|
|
157
|
+
"tasks.max": "1",
|
|
158
|
+
"topics": options.topics.join(","),
|
|
159
|
+
"connection.uri": options.connectionUri,
|
|
160
|
+
"database": options.databaseName,
|
|
161
|
+
"collection": options.collectionName || "kafka_data",
|
|
162
|
+
"max.num.retries": "3",
|
|
163
|
+
"retries.defer.timeout": "5000"
|
|
164
|
+
};
|
|
165
|
+
}
|
|
166
|
+
/**
|
|
167
|
+
* HTTP Sink Connector (Kafka → REST API)
|
|
168
|
+
*
|
|
169
|
+
* Use case: Send Kafka events to external APIs
|
|
170
|
+
*/
|
|
171
|
+
static httpSink(options) {
|
|
172
|
+
const config = {
|
|
173
|
+
name: `http-sink-${options.topics.join("-")}`,
|
|
174
|
+
"connector.class": "io.confluent.connect.http.HttpSinkConnector",
|
|
175
|
+
"tasks.max": "1",
|
|
176
|
+
"topics": options.topics.join(","),
|
|
177
|
+
"http.api.url": options.httpApiUrl,
|
|
178
|
+
"request.method": options.httpMethod || "POST",
|
|
179
|
+
"batch.max.size": (options.batchSize || 10).toString(),
|
|
180
|
+
"retry.on.status.codes": "500-599",
|
|
181
|
+
"max.retries": "3",
|
|
182
|
+
"retry.backoff.ms": "1000"
|
|
183
|
+
};
|
|
184
|
+
if (options.headers) {
|
|
185
|
+
Object.entries(options.headers).forEach(([key, value], index) => {
|
|
186
|
+
config[`headers.${index}.name`] = key;
|
|
187
|
+
config[`headers.${index}.value`] = value;
|
|
188
|
+
});
|
|
189
|
+
}
|
|
190
|
+
return config;
|
|
191
|
+
}
|
|
192
|
+
/**
|
|
193
|
+
* HDFS Sink Connector (Kafka → Hadoop HDFS)
|
|
194
|
+
*
|
|
195
|
+
* Use case: Archive Kafka data to Hadoop for batch processing
|
|
196
|
+
*/
|
|
197
|
+
static hdfsSink(options) {
|
|
198
|
+
const formatClass = options.format === "avro" ? "io.confluent.connect.hdfs.avro.AvroFormat" : options.format === "parquet" ? "io.confluent.connect.hdfs.parquet.ParquetFormat" : "io.confluent.connect.hdfs.json.JsonFormat";
|
|
199
|
+
return {
|
|
200
|
+
name: `hdfs-sink-${options.topics.join("-")}`,
|
|
201
|
+
"connector.class": "io.confluent.connect.hdfs.HdfsSinkConnector",
|
|
202
|
+
"tasks.max": "1",
|
|
203
|
+
"topics": options.topics.join(","),
|
|
204
|
+
"hdfs.url": options.hdfsUrl,
|
|
205
|
+
"format.class": formatClass,
|
|
206
|
+
"flush.size": (options.flushSize || 1e3).toString(),
|
|
207
|
+
"rotate.interval.ms": (options.rotateIntervalMs || 36e5).toString(),
|
|
208
|
+
"partitioner.class": "io.confluent.connect.storage.partitioner.TimeBasedPartitioner",
|
|
209
|
+
"path.format": "'year'=YYYY/'month'=MM/'day'=dd/'hour'=HH",
|
|
210
|
+
"locale": "en-US",
|
|
211
|
+
"timezone": "UTC"
|
|
212
|
+
};
|
|
213
|
+
}
|
|
214
|
+
/**
|
|
215
|
+
* Snowflake Sink Connector (Kafka → Snowflake)
|
|
216
|
+
*
|
|
217
|
+
* Use case: Stream Kafka data to Snowflake data warehouse
|
|
218
|
+
*/
|
|
219
|
+
static snowflakeSink(options) {
|
|
220
|
+
return {
|
|
221
|
+
name: `snowflake-sink-${options.topics.join("-")}`,
|
|
222
|
+
"connector.class": "com.snowflake.kafka.connector.SnowflakeSinkConnector",
|
|
223
|
+
"tasks.max": "8",
|
|
224
|
+
"topics": options.topics.join(","),
|
|
225
|
+
"snowflake.url.name": options.snowflakeUrl,
|
|
226
|
+
"snowflake.user.name": options.snowflakeUser,
|
|
227
|
+
"snowflake.private.key": options.snowflakePrivateKey,
|
|
228
|
+
"snowflake.database.name": options.snowflakeDatabase,
|
|
229
|
+
"snowflake.schema.name": options.snowflakeSchema,
|
|
230
|
+
"buffer.count.records": "10000",
|
|
231
|
+
"buffer.flush.time": "60",
|
|
232
|
+
"buffer.size.bytes": "5000000"
|
|
233
|
+
};
|
|
234
|
+
}
|
|
235
|
+
/**
|
|
236
|
+
* BigQuery Sink Connector (Kafka → Google BigQuery)
|
|
237
|
+
*
|
|
238
|
+
* Use case: Stream Kafka data to BigQuery for analytics
|
|
239
|
+
*/
|
|
240
|
+
static bigQuerySink(options) {
|
|
241
|
+
return {
|
|
242
|
+
name: `bigquery-sink-${options.topics.join("-")}`,
|
|
243
|
+
"connector.class": "com.wepay.kafka.connect.bigquery.BigQuerySinkConnector",
|
|
244
|
+
"tasks.max": "1",
|
|
245
|
+
"topics": options.topics.join(","),
|
|
246
|
+
"project": options.projectId,
|
|
247
|
+
"defaultDataset": options.datasetName,
|
|
248
|
+
"autoCreateTables": options.autoCreateTables !== false ? "true" : "false",
|
|
249
|
+
"autoUpdateSchemas": "true",
|
|
250
|
+
"sanitizeTopics": "true",
|
|
251
|
+
"allowNewBigQueryFields": "true",
|
|
252
|
+
"allowBigQueryRequiredFieldRelaxation": "true"
|
|
253
|
+
};
|
|
254
|
+
}
|
|
255
|
+
}
|
|
256
|
+
class ConnectorManager {
|
|
257
|
+
/**
|
|
258
|
+
* Deploy connector via REST API
|
|
259
|
+
*/
|
|
260
|
+
static async deployConnector(connectUrl, config) {
|
|
261
|
+
const response = await fetch(`${connectUrl}/connectors`, {
|
|
262
|
+
method: "POST",
|
|
263
|
+
headers: { "Content-Type": "application/json" },
|
|
264
|
+
body: JSON.stringify({ name: config.name, config })
|
|
265
|
+
});
|
|
266
|
+
if (!response.ok) {
|
|
267
|
+
throw new Error(`Failed to deploy connector: ${await response.text()}`);
|
|
268
|
+
}
|
|
269
|
+
console.log(`\u2705 Connector deployed: ${config.name}`);
|
|
270
|
+
}
|
|
271
|
+
/**
|
|
272
|
+
* List all connectors
|
|
273
|
+
*/
|
|
274
|
+
static async listConnectors(connectUrl) {
|
|
275
|
+
const response = await fetch(`${connectUrl}/connectors`);
|
|
276
|
+
return response.json();
|
|
277
|
+
}
|
|
278
|
+
/**
|
|
279
|
+
* Get connector status
|
|
280
|
+
*/
|
|
281
|
+
static async getConnectorStatus(connectUrl, connectorName) {
|
|
282
|
+
const response = await fetch(`${connectUrl}/connectors/${connectorName}/status`);
|
|
283
|
+
return response.json();
|
|
284
|
+
}
|
|
285
|
+
/**
|
|
286
|
+
* Delete connector
|
|
287
|
+
*/
|
|
288
|
+
static async deleteConnector(connectUrl, connectorName) {
|
|
289
|
+
await fetch(`${connectUrl}/connectors/${connectorName}`, {
|
|
290
|
+
method: "DELETE"
|
|
291
|
+
});
|
|
292
|
+
console.log(`\u2705 Connector deleted: ${connectorName}`);
|
|
293
|
+
}
|
|
294
|
+
}
|
|
295
|
+
var connector_catalog_default = {
|
|
296
|
+
ConnectorCatalog,
|
|
297
|
+
ConnectorManager,
|
|
298
|
+
ConnectorCategory
|
|
299
|
+
};
|
|
300
|
+
export {
|
|
301
|
+
ConnectorCatalog,
|
|
302
|
+
ConnectorCategory,
|
|
303
|
+
ConnectorManager,
|
|
304
|
+
connector_catalog_default as default
|
|
305
|
+
};
|