omniai-anthropic 0.1.0 → 1.0.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/Gemfile +14 -0
- data/README.md +96 -16
- data/lib/omniai/anthropic/chat/completion.rb +24 -0
- data/lib/omniai/anthropic/chat/stream.rb +106 -0
- data/lib/omniai/anthropic/chat.rb +62 -0
- data/lib/omniai/anthropic/client.rb +63 -0
- data/lib/omniai/anthropic/config.rb +18 -0
- data/lib/omniai/anthropic/version.rb +1 -1
- data/lib/omniai/anthropic.rb +17 -3
- metadata +18 -13
- data/.rspec +0 -3
- data/.rubocop.yml +0 -8
- data/Rakefile +0 -12
- data/sig/omniai/anthropic.rbs +0 -6
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: f431b98530e904ca88f4860c87fcaf4d44d3960a58158ae91a20e0ff06a67e83
|
4
|
+
data.tar.gz: c8a45bf94836565c387394558641253a7665d33e95537c13d642c4f21a6f9fa9
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: c37e3b9502f43209bc5d6fe9902d51460d09f55c56f9fdcb6567ac6e7a461d7857d1cda6982afc6058d992b158c179b2aa48d285bba710c066b27cc5a00f4ef2
|
7
|
+
data.tar.gz: 6e7f50fea78c63919c101c5e6e591227b286a09806e854fff101d066a9a7ed084977f615b27dcdfe3018119ca20c0513f8ceae5668daf250d33ba115b31cc9ff
|
data/Gemfile
ADDED
data/README.md
CHANGED
@@ -1,31 +1,111 @@
|
|
1
|
-
#
|
1
|
+
# OmniAI::Anthropic
|
2
2
|
|
3
|
-
|
4
|
-
|
5
|
-
Welcome to your new gem! In this directory, you'll find the files you need to be able to package up your Ruby library into a gem. Put your Ruby code in the file `lib/omniai/anthropic`. To experiment with that code, run `bin/console` for an interactive prompt.
|
3
|
+
An Anthropic implementation of the [OmniAI](https://github.com/ksylvest/omniai) APIs.
|
6
4
|
|
7
5
|
## Installation
|
8
6
|
|
9
|
-
|
7
|
+
```sh
|
8
|
+
gem install omniai-anthropic
|
9
|
+
```
|
10
10
|
|
11
|
-
|
11
|
+
## Usage
|
12
12
|
|
13
|
-
|
13
|
+
### Client
|
14
14
|
|
15
|
-
|
15
|
+
A client is setup as follows if `ENV['ANTHROPIC_API_KEY']` exists:
|
16
16
|
|
17
|
-
|
17
|
+
```ruby
|
18
|
+
client = OmniAI::Anthropic::Client.new
|
19
|
+
```
|
18
20
|
|
19
|
-
|
21
|
+
A client may also be passed the following options:
|
22
|
+
|
23
|
+
- `api_key` (required - default is `ENV['ANTHROPIC_API_KEY']`)
|
24
|
+
- `host` (optional)
|
25
|
+
|
26
|
+
### Configuration
|
27
|
+
|
28
|
+
Global configuration is supported for the following options:
|
29
|
+
|
30
|
+
```ruby
|
31
|
+
OmniAI::Anthropic.configure do |config|
|
32
|
+
config.api_key = '...' # default: ENV['ANTHROPIC_API_KEY']
|
33
|
+
config.host = '...' # default: 'https://api.anthropic.com'
|
34
|
+
end
|
35
|
+
```
|
36
|
+
|
37
|
+
### Chat
|
38
|
+
|
39
|
+
A chat completion is generated by passing in prompts using any a variety of formats:
|
40
|
+
|
41
|
+
```ruby
|
42
|
+
completion = client.chat('Tell me a joke!')
|
43
|
+
completion.choice.message.content # 'Why did the chicken cross the road? To get to the other side.'
|
44
|
+
```
|
45
|
+
|
46
|
+
```ruby
|
47
|
+
completion = client.chat({
|
48
|
+
role: OmniAI::Chat::Role::USER,
|
49
|
+
content: 'Is it wise to jump off a bridge?'
|
50
|
+
})
|
51
|
+
completion.choice.message.content # 'No.'
|
52
|
+
```
|
53
|
+
|
54
|
+
```ruby
|
55
|
+
completion = client.chat([
|
56
|
+
{
|
57
|
+
role: OmniAI::Chat::Role::SYSTEM,
|
58
|
+
content: 'You are a helpful assistant.'
|
59
|
+
},
|
60
|
+
'What is the capital of Canada?',
|
61
|
+
])
|
62
|
+
completion.choice.message.content # 'The capital of Canada is Ottawa.'
|
63
|
+
```
|
64
|
+
|
65
|
+
#### Model
|
66
|
+
|
67
|
+
`model` takes an optional string (default is `claude-3-haiku-20240307`):
|
68
|
+
|
69
|
+
```ruby
|
70
|
+
completion = client.chat('Provide code for fibonacci', model: OmniAI::Anthropic::Chat::Model::OPUS)
|
71
|
+
completion.choice.message.content # 'def fibonacci(n)...end'
|
72
|
+
```
|
73
|
+
|
74
|
+
[Anthropic API Reference `model`](https://docs.anthropic.com/en/api/messages)
|
75
|
+
|
76
|
+
#### Temperature
|
77
|
+
|
78
|
+
`temperature` takes an optional float between `0.0` and `1.0` (defaults is `0.7`):
|
79
|
+
|
80
|
+
```ruby
|
81
|
+
completion = client.chat('Pick a number between 1 and 5', temperature: 1.0)
|
82
|
+
completion.choice.message.content # '3'
|
83
|
+
```
|
84
|
+
|
85
|
+
[Anthropic API Reference `temperature`](https://docs.anthropic.com/en/api/messages)
|
86
|
+
|
87
|
+
#### Stream
|
88
|
+
|
89
|
+
`stream` takes an optional a proc to stream responses in real-time chunks instead of waiting for a complete response:
|
20
90
|
|
21
|
-
|
91
|
+
```ruby
|
92
|
+
stream = proc do |chunk|
|
93
|
+
print(chunk.choice.delta.content) # 'Better', 'three', 'hours', ...
|
94
|
+
end
|
95
|
+
client.chat('Be poetic.', stream:)
|
96
|
+
```
|
22
97
|
|
23
|
-
|
98
|
+
[Anthropic API Reference `stream`](https://docs.anthropic.com/en/api/messages)
|
24
99
|
|
25
|
-
|
100
|
+
#### Format
|
26
101
|
|
27
|
-
|
102
|
+
`format` takes an optional symbol (`:json`) and modifies requests to send additional system text requesting JSON:
|
28
103
|
|
29
|
-
|
104
|
+
```ruby
|
105
|
+
completion = client.chat([
|
106
|
+
{ role: OmniAI::Chat::Role::USER, content: 'What is the name of the drummer for the Beatles?' }
|
107
|
+
], format: :json)
|
108
|
+
JSON.parse(completion.choice.message.content) # { "name": "Ringo" }
|
109
|
+
```
|
30
110
|
|
31
|
-
|
111
|
+
[Anthropic API Reference `control-output-format`](https://docs.anthropic.com/en/docs/control-output-format)
|
@@ -0,0 +1,24 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
module OmniAI
|
4
|
+
module Anthropic
|
5
|
+
class Chat
|
6
|
+
# A completion returned by the API.
|
7
|
+
class Completion < OmniAI::Chat::Completion
|
8
|
+
# @return [Array<OmniAI::Chat::Choice>]
|
9
|
+
def choices
|
10
|
+
@choices ||= begin
|
11
|
+
role = @data['role']
|
12
|
+
|
13
|
+
@data['content'].map do |data, index|
|
14
|
+
OmniAI::Chat::Choice.new(data: {
|
15
|
+
'index' => index,
|
16
|
+
'message' => { 'role' => role, 'content' => data['text'] },
|
17
|
+
})
|
18
|
+
end
|
19
|
+
end
|
20
|
+
end
|
21
|
+
end
|
22
|
+
end
|
23
|
+
end
|
24
|
+
end
|
@@ -0,0 +1,106 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
module OmniAI
|
4
|
+
module Anthropic
|
5
|
+
class Chat
|
6
|
+
# A stream given when streaming.
|
7
|
+
class Stream < OmniAI::Chat::Stream
|
8
|
+
module Type
|
9
|
+
PING = 'ping'
|
10
|
+
MESSAGE_START = 'message_start'
|
11
|
+
MESSAGE_STOP = 'message_stop'
|
12
|
+
MESSAGE_DELTA = 'message_delta'
|
13
|
+
CONTENT_BLOCK_START = 'content_block_start'
|
14
|
+
CONTENT_BLOCK_STOP = 'content_block_stop'
|
15
|
+
CONTENT_BLOCK_DELTA = 'content_block_delta'
|
16
|
+
end
|
17
|
+
|
18
|
+
# Process the stream into chunks by event.
|
19
|
+
class Builder
|
20
|
+
attr_reader :id, :model, :role, :content, :index
|
21
|
+
|
22
|
+
# @return [OmniAI::Chat::Chunk]
|
23
|
+
def chunk
|
24
|
+
OmniAI::Chat::Chunk.new(data: {
|
25
|
+
'id' => @id,
|
26
|
+
'model' => @model,
|
27
|
+
'choices' => [{
|
28
|
+
'index' => @index,
|
29
|
+
'delta' => { 'role' => @role, 'content' => @content },
|
30
|
+
}],
|
31
|
+
})
|
32
|
+
end
|
33
|
+
|
34
|
+
# Handler for Type::MESSAGE_START
|
35
|
+
#
|
36
|
+
# @param data [Hash]
|
37
|
+
def message_start(data)
|
38
|
+
@id = data['id']
|
39
|
+
@model = data['model']
|
40
|
+
@role = data['role']
|
41
|
+
end
|
42
|
+
|
43
|
+
# Handler for Type::MESSAGE_STOP
|
44
|
+
#
|
45
|
+
# @param data [Hash]
|
46
|
+
def message_stop(_)
|
47
|
+
@id = nil
|
48
|
+
@model = nil
|
49
|
+
@role = nil
|
50
|
+
end
|
51
|
+
|
52
|
+
# Handler for Type::CONTENT_BLOCK_START
|
53
|
+
#
|
54
|
+
# @param data [Hash]
|
55
|
+
def content_block_start(data)
|
56
|
+
@index = data['index']
|
57
|
+
end
|
58
|
+
|
59
|
+
# Handler for Type::CONTENT_BLOCK_STOP
|
60
|
+
#
|
61
|
+
# @param data [Hash]
|
62
|
+
def content_block_stop(_)
|
63
|
+
@index = nil
|
64
|
+
end
|
65
|
+
|
66
|
+
# Handler for Type::CONTENT_BLOCK_DELTA
|
67
|
+
#
|
68
|
+
# @param data [Hash]
|
69
|
+
def content_block_delta(data)
|
70
|
+
return unless data['delta']['type'].eql?('text_delta')
|
71
|
+
|
72
|
+
@content = data['delta']['text']
|
73
|
+
end
|
74
|
+
end
|
75
|
+
|
76
|
+
# @yield [OmniAI::Chat::Chunk]
|
77
|
+
def stream!(&block)
|
78
|
+
builder = Builder.new
|
79
|
+
|
80
|
+
@response.body.each do |chunk|
|
81
|
+
@parser.feed(chunk) do |type, data|
|
82
|
+
process(type:, data: JSON.parse(data), builder:, &block)
|
83
|
+
end
|
84
|
+
end
|
85
|
+
end
|
86
|
+
|
87
|
+
private
|
88
|
+
|
89
|
+
# @param type [String]
|
90
|
+
# @param data [Hash]
|
91
|
+
# @param builder [Builder]
|
92
|
+
def process(type:, data:, builder:, &)
|
93
|
+
case type
|
94
|
+
when Type::MESSAGE_START then builder.message_start(data)
|
95
|
+
when Type::CONTENT_BLOCK_START then builder.content_block_start(data)
|
96
|
+
when Type::CONTENT_BLOCK_STOP then builder.content_block_stop(data)
|
97
|
+
when Type::MESSAGE_STOP then builder.message_stop(data)
|
98
|
+
when Type::CONTENT_BLOCK_DELTA
|
99
|
+
builder.content_block_delta(data)
|
100
|
+
yield(builder.chunk)
|
101
|
+
end
|
102
|
+
end
|
103
|
+
end
|
104
|
+
end
|
105
|
+
end
|
106
|
+
end
|
@@ -0,0 +1,62 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
module OmniAI
|
4
|
+
module Anthropic
|
5
|
+
# A Anthropic chat implementation.
|
6
|
+
#
|
7
|
+
# Usage:
|
8
|
+
#
|
9
|
+
# chat = OmniAI::Anthropic::Chat.new(client: client)
|
10
|
+
# chat.completion('Tell me a joke.')
|
11
|
+
# chat.completion(['Tell me a joke.'])
|
12
|
+
# chat.completion({ role: 'user', content: 'Tell me a joke.' })
|
13
|
+
# chat.completion([{ role: 'system', content: 'Tell me a joke.' }])
|
14
|
+
class Chat < OmniAI::Chat
|
15
|
+
module Model
|
16
|
+
HAIKU = 'claude-3-haiku-20240307'
|
17
|
+
SONNET = 'claude-3-sonnet-20240229'
|
18
|
+
OPUS = 'claude-3-opus-20240229'
|
19
|
+
end
|
20
|
+
|
21
|
+
protected
|
22
|
+
|
23
|
+
# @param response [HTTP::Response]
|
24
|
+
# @return [OmniAI::Anthropic::Chat::Stream]
|
25
|
+
def stream!(response:)
|
26
|
+
raise Error, "#{self.class.name}#stream! unstreamable" unless @stream
|
27
|
+
|
28
|
+
Stream.new(response:).stream! { |chunk| @stream.call(chunk) }
|
29
|
+
end
|
30
|
+
|
31
|
+
# @param response [HTTP::Response]
|
32
|
+
# @param response [OmniAI::Anthropic::Chat::Completion]
|
33
|
+
def complete!(response:)
|
34
|
+
Completion.new(data: response.parse)
|
35
|
+
end
|
36
|
+
|
37
|
+
# @return [Hash]
|
38
|
+
def payload
|
39
|
+
OmniAI::Anthropic.config.chat_options.merge({
|
40
|
+
model: @model,
|
41
|
+
messages: messages.filter { |message| !message[:role].eql?(OmniAI::Chat::Role::SYSTEM) },
|
42
|
+
system:,
|
43
|
+
stream: @stream.nil? ? nil : !@stream.nil?,
|
44
|
+
temperature: @temperature,
|
45
|
+
}).compact
|
46
|
+
end
|
47
|
+
|
48
|
+
# @return [String, nil]
|
49
|
+
def system
|
50
|
+
messages = self.messages.filter { |message| message[:role].eql?(OmniAI::Chat::Role::SYSTEM) }
|
51
|
+
messages << { role: OmniAI::Chat::Role::SYSTEM, content: OmniAI::Chat::JSON_PROMPT } if @format.eql?(:json)
|
52
|
+
|
53
|
+
messages.map { |message| message[:content] }.join("\n\n") if messages.any?
|
54
|
+
end
|
55
|
+
|
56
|
+
# @return [String]
|
57
|
+
def path
|
58
|
+
"/#{OmniAI::Anthropic::Client::VERSION}/messages"
|
59
|
+
end
|
60
|
+
end
|
61
|
+
end
|
62
|
+
end
|
@@ -0,0 +1,63 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
module OmniAI
|
4
|
+
module Anthropic
|
5
|
+
# An Anthropic client implementation. Usage:
|
6
|
+
#
|
7
|
+
# w/ `api_key``:
|
8
|
+
# client = OmniAI::Anthropic::Client.new(api_key: '...')
|
9
|
+
#
|
10
|
+
# w/ ENV['ANTHROPIC_API_KEY']:
|
11
|
+
#
|
12
|
+
# ENV['ANTHROPIC_API_KEY'] = '...'
|
13
|
+
# client = OmniAI::Anthropic::Client.new
|
14
|
+
#
|
15
|
+
# w/ config:
|
16
|
+
#
|
17
|
+
# OmniAI::Anthropic.configure do |config|
|
18
|
+
# config.api_key = '...'
|
19
|
+
# end
|
20
|
+
#
|
21
|
+
# client = OmniAI::Anthropic::Client.new
|
22
|
+
class Client < OmniAI::Client
|
23
|
+
VERSION = 'v1'
|
24
|
+
|
25
|
+
# @param api_key [String] optional - defaults to `OmniAI::Anthropic.config.api_key`
|
26
|
+
# @param host [String] optional - defaults to `OmniAI::Anthropic.config.host`
|
27
|
+
def initialize(
|
28
|
+
api_key: OmniAI::Anthropic.config.api_key,
|
29
|
+
version: OmniAI::Anthropic.config.version,
|
30
|
+
logger: OmniAI::Anthropic.config.logger,
|
31
|
+
host: OmniAI::Anthropic.config.host
|
32
|
+
)
|
33
|
+
raise(ArgumentError, %(ENV['ANTHROPIC_API_KEY'] must be defined or `api_key` must be passed)) if api_key.nil?
|
34
|
+
|
35
|
+
super(api_key:, logger:)
|
36
|
+
|
37
|
+
@host = host
|
38
|
+
@version = version
|
39
|
+
end
|
40
|
+
|
41
|
+
# @return [HTTP::Client]
|
42
|
+
def connection
|
43
|
+
HTTP
|
44
|
+
.headers('x-api-key': @api_key)
|
45
|
+
.headers('anthropic-version': @version)
|
46
|
+
.persistent('https://api.anthropic.com')
|
47
|
+
end
|
48
|
+
|
49
|
+
# @raise [OmniAI::Error]
|
50
|
+
#
|
51
|
+
# @param messages [String, Array, Hash]
|
52
|
+
# @param model [String] optional
|
53
|
+
# @param format [Symbol] optional :text or :json
|
54
|
+
# @param temperature [Float, nil] optional
|
55
|
+
# @param stream [Proc, nil] optional
|
56
|
+
#
|
57
|
+
# @return [OmniAI::Chat::Completion]
|
58
|
+
def chat(messages, model: Chat::Model::HAIKU, temperature: nil, format: nil, stream: nil)
|
59
|
+
Chat.process!(messages, model:, temperature:, format:, stream:, client: self)
|
60
|
+
end
|
61
|
+
end
|
62
|
+
end
|
63
|
+
end
|
@@ -0,0 +1,18 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
module OmniAI
|
4
|
+
module Anthropic
|
5
|
+
# Config for the Anthropic `api_key` / `host` / `logger` / `version` / `chat_options`.
|
6
|
+
class Config < OmniAI::Config
|
7
|
+
attr_accessor :version, :chat_options
|
8
|
+
|
9
|
+
def initialize
|
10
|
+
super
|
11
|
+
@api_key = ENV.fetch('ANTHROPIC_API_KEY', nil)
|
12
|
+
@host = ENV.fetch('ANTHROPIC_HOST', 'https://api.anthropic.com')
|
13
|
+
@version = ENV.fetch('ANTHROPIC_VERSION', '2023-06-01')
|
14
|
+
@chat_options = { max_tokens: 4096 }
|
15
|
+
end
|
16
|
+
end
|
17
|
+
end
|
18
|
+
end
|
data/lib/omniai/anthropic.rb
CHANGED
@@ -1,10 +1,24 @@
|
|
1
1
|
# frozen_string_literal: true
|
2
2
|
|
3
|
-
|
3
|
+
require 'event_stream_parser'
|
4
|
+
require 'omniai'
|
5
|
+
require 'zeitwerk'
|
6
|
+
|
7
|
+
loader = Zeitwerk::Loader.for_gem
|
8
|
+
loader.push_dir(__dir__, namespace: OmniAI)
|
9
|
+
loader.setup
|
4
10
|
|
5
11
|
module OmniAI
|
12
|
+
# A namespace for everything Anthropic.
|
6
13
|
module Anthropic
|
7
|
-
|
8
|
-
|
14
|
+
# @return [OmniAI::Anthropic::Config]
|
15
|
+
def self.config
|
16
|
+
@config ||= Config.new
|
17
|
+
end
|
18
|
+
|
19
|
+
# @yield [OmniAI::Anthropic::Config]
|
20
|
+
def self.configure
|
21
|
+
yield config
|
22
|
+
end
|
9
23
|
end
|
10
24
|
end
|
metadata
CHANGED
@@ -1,17 +1,17 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: omniai-anthropic
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 0.1
|
4
|
+
version: 1.0.1
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- Kevin Sylvestre
|
8
8
|
autorequire:
|
9
9
|
bindir: exe
|
10
10
|
cert_chain: []
|
11
|
-
date: 2024-06-
|
11
|
+
date: 2024-06-14 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
|
-
name:
|
14
|
+
name: event_stream_parser
|
15
15
|
requirement: !ruby/object:Gem::Requirement
|
16
16
|
requirements:
|
17
17
|
- - ">="
|
@@ -25,13 +25,13 @@ dependencies:
|
|
25
25
|
- !ruby/object:Gem::Version
|
26
26
|
version: '0'
|
27
27
|
- !ruby/object:Gem::Dependency
|
28
|
-
name:
|
28
|
+
name: omniai
|
29
29
|
requirement: !ruby/object:Gem::Requirement
|
30
30
|
requirements:
|
31
31
|
- - ">="
|
32
32
|
- !ruby/object:Gem::Version
|
33
33
|
version: '0'
|
34
|
-
type: :
|
34
|
+
type: :runtime
|
35
35
|
prerelease: false
|
36
36
|
version_requirements: !ruby/object:Gem::Requirement
|
37
37
|
requirements:
|
@@ -39,13 +39,13 @@ dependencies:
|
|
39
39
|
- !ruby/object:Gem::Version
|
40
40
|
version: '0'
|
41
41
|
- !ruby/object:Gem::Dependency
|
42
|
-
name:
|
42
|
+
name: zeitwerk
|
43
43
|
requirement: !ruby/object:Gem::Requirement
|
44
44
|
requirements:
|
45
45
|
- - ">="
|
46
46
|
- !ruby/object:Gem::Version
|
47
47
|
version: '0'
|
48
|
-
type: :
|
48
|
+
type: :runtime
|
49
49
|
prerelease: false
|
50
50
|
version_requirements: !ruby/object:Gem::Requirement
|
51
51
|
requirements:
|
@@ -59,16 +59,21 @@ executables: []
|
|
59
59
|
extensions: []
|
60
60
|
extra_rdoc_files: []
|
61
61
|
files:
|
62
|
-
-
|
63
|
-
- ".rubocop.yml"
|
62
|
+
- Gemfile
|
64
63
|
- README.md
|
65
|
-
- Rakefile
|
66
64
|
- lib/omniai/anthropic.rb
|
65
|
+
- lib/omniai/anthropic/chat.rb
|
66
|
+
- lib/omniai/anthropic/chat/completion.rb
|
67
|
+
- lib/omniai/anthropic/chat/stream.rb
|
68
|
+
- lib/omniai/anthropic/client.rb
|
69
|
+
- lib/omniai/anthropic/config.rb
|
67
70
|
- lib/omniai/anthropic/version.rb
|
68
|
-
- sig/omniai/anthropic.rbs
|
69
71
|
homepage: https://github.com/ksylvest/omniai-anthropic
|
70
72
|
licenses: []
|
71
|
-
metadata:
|
73
|
+
metadata:
|
74
|
+
homepage_uri: https://github.com/ksylvest/omniai-anthropic
|
75
|
+
changelog_uri: https://github.com/ksylvest/omniai-anthropic/releases
|
76
|
+
rubygems_mfa_required: 'true'
|
72
77
|
post_install_message:
|
73
78
|
rdoc_options: []
|
74
79
|
require_paths:
|
@@ -77,7 +82,7 @@ required_ruby_version: !ruby/object:Gem::Requirement
|
|
77
82
|
requirements:
|
78
83
|
- - ">="
|
79
84
|
- !ruby/object:Gem::Version
|
80
|
-
version: 3.
|
85
|
+
version: 3.3.0
|
81
86
|
required_rubygems_version: !ruby/object:Gem::Requirement
|
82
87
|
requirements:
|
83
88
|
- - ">="
|
data/.rspec
DELETED
data/.rubocop.yml
DELETED
data/Rakefile
DELETED