influxdb-client 1.1.0.pre.203 → 1.1.0.pre.323
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/.codecov.yml +3 -0
- data/.gitignore +2 -0
- data/.rubocop.yml +4 -0
- data/CHANGELOG.md +3 -0
- data/README.md +29 -0
- data/bin/generate-sources.sh +30 -0
- data/bin/pom.xml +34 -0
- data/bin/swagger.yml +9867 -0
- data/lib/influxdb2/client.rb +3 -0
- data/lib/influxdb2/client/client.rb +7 -0
- data/lib/influxdb2/client/default_api.rb +68 -0
- data/lib/influxdb2/client/flux_csv_parser.rb +210 -0
- data/lib/influxdb2/client/flux_table.rb +99 -0
- data/lib/influxdb2/client/models/dialect.rb +317 -0
- data/lib/influxdb2/client/models/query.rb +284 -0
- data/lib/influxdb2/client/query_api.rb +79 -0
- data/lib/influxdb2/client/write_api.rb +2 -39
- data/test/influxdb/flux_csv_parser_test.rb +328 -0
- data/test/influxdb/query_api_integration_test.rb +58 -0
- data/test/influxdb/query_api_test.rb +75 -0
- data/test/test_helper.rb +4 -1
- metadata +18 -2
@@ -0,0 +1,284 @@
|
|
1
|
+
=begin
|
2
|
+
#Influx API Service
|
3
|
+
|
4
|
+
#No description provided (generated by Openapi Generator https://github.com/openapitools/openapi-generator)
|
5
|
+
|
6
|
+
OpenAPI spec version: 0.1.0
|
7
|
+
|
8
|
+
Generated by: https://openapi-generator.tech
|
9
|
+
OpenAPI Generator version: 3.3.4
|
10
|
+
|
11
|
+
=end
|
12
|
+
|
13
|
+
require 'date'
|
14
|
+
|
15
|
+
module InfluxDB2
|
16
|
+
# Query influx with specific return formatting.
|
17
|
+
class Query
|
18
|
+
attr_accessor :extern
|
19
|
+
|
20
|
+
# Query script to execute.
|
21
|
+
attr_accessor :query
|
22
|
+
|
23
|
+
# The type of query.
|
24
|
+
attr_accessor :type
|
25
|
+
|
26
|
+
# Required for `influxql` type queries.
|
27
|
+
attr_accessor :db
|
28
|
+
|
29
|
+
# Required for `influxql` type queries.
|
30
|
+
attr_accessor :rp
|
31
|
+
|
32
|
+
# Required for `influxql` type queries.
|
33
|
+
attr_accessor :cluster
|
34
|
+
|
35
|
+
attr_accessor :dialect
|
36
|
+
|
37
|
+
class EnumAttributeValidator
|
38
|
+
attr_reader :datatype
|
39
|
+
attr_reader :allowable_values
|
40
|
+
|
41
|
+
def initialize(datatype, allowable_values)
|
42
|
+
@allowable_values = allowable_values.map do |value|
|
43
|
+
case datatype.to_s
|
44
|
+
when /Integer/i
|
45
|
+
value.to_i
|
46
|
+
when /Float/i
|
47
|
+
value.to_f
|
48
|
+
else
|
49
|
+
value
|
50
|
+
end
|
51
|
+
end
|
52
|
+
end
|
53
|
+
|
54
|
+
def valid?(value)
|
55
|
+
!value || allowable_values.include?(value)
|
56
|
+
end
|
57
|
+
end
|
58
|
+
|
59
|
+
# Attribute mapping from ruby-style variable name to JSON key.
|
60
|
+
def self.attribute_map
|
61
|
+
{
|
62
|
+
:'extern' => :'extern',
|
63
|
+
:'query' => :'query',
|
64
|
+
:'type' => :'type',
|
65
|
+
:'db' => :'db',
|
66
|
+
:'rp' => :'rp',
|
67
|
+
:'cluster' => :'cluster',
|
68
|
+
:'dialect' => :'dialect'
|
69
|
+
}
|
70
|
+
end
|
71
|
+
|
72
|
+
# Attribute type mapping.
|
73
|
+
def self.openapi_types
|
74
|
+
{
|
75
|
+
:'extern' => :'File',
|
76
|
+
:'query' => :'String',
|
77
|
+
:'type' => :'String',
|
78
|
+
:'db' => :'String',
|
79
|
+
:'rp' => :'String',
|
80
|
+
:'cluster' => :'String',
|
81
|
+
:'dialect' => :'Dialect'
|
82
|
+
}
|
83
|
+
end
|
84
|
+
|
85
|
+
# Initializes the object
|
86
|
+
# @param [Hash] attributes Model attributes in the form of hash
|
87
|
+
def initialize(attributes = {})
|
88
|
+
return unless attributes.is_a?(Hash)
|
89
|
+
|
90
|
+
# convert string to symbol for hash key
|
91
|
+
attributes = attributes.each_with_object({}) { |(k, v), h| h[k.to_sym] = v }
|
92
|
+
|
93
|
+
if attributes.has_key?(:'extern')
|
94
|
+
self.extern = attributes[:'extern']
|
95
|
+
end
|
96
|
+
|
97
|
+
if attributes.has_key?(:'query')
|
98
|
+
self.query = attributes[:'query']
|
99
|
+
end
|
100
|
+
|
101
|
+
if attributes.has_key?(:'type')
|
102
|
+
self.type = attributes[:'type']
|
103
|
+
else
|
104
|
+
self.type = 'flux'
|
105
|
+
end
|
106
|
+
|
107
|
+
if attributes.has_key?(:'db')
|
108
|
+
self.db = attributes[:'db']
|
109
|
+
end
|
110
|
+
|
111
|
+
if attributes.has_key?(:'rp')
|
112
|
+
self.rp = attributes[:'rp']
|
113
|
+
end
|
114
|
+
|
115
|
+
if attributes.has_key?(:'cluster')
|
116
|
+
self.cluster = attributes[:'cluster']
|
117
|
+
end
|
118
|
+
|
119
|
+
if attributes.has_key?(:'dialect')
|
120
|
+
self.dialect = attributes[:'dialect']
|
121
|
+
end
|
122
|
+
end
|
123
|
+
|
124
|
+
# Show invalid properties with the reasons. Usually used together with valid?
|
125
|
+
# @return Array for valid properties with the reasons
|
126
|
+
def list_invalid_properties
|
127
|
+
invalid_properties = Array.new
|
128
|
+
if @query.nil?
|
129
|
+
invalid_properties.push('invalid value for "query", query cannot be nil.')
|
130
|
+
end
|
131
|
+
|
132
|
+
invalid_properties
|
133
|
+
end
|
134
|
+
|
135
|
+
# Check to see if the all the properties in the model are valid
|
136
|
+
# @return true if the model is valid
|
137
|
+
def valid?
|
138
|
+
return false if @query.nil?
|
139
|
+
type_validator = EnumAttributeValidator.new('String', ['flux', 'influxql'])
|
140
|
+
return false unless type_validator.valid?(@type)
|
141
|
+
true
|
142
|
+
end
|
143
|
+
|
144
|
+
# Custom attribute writer method checking allowed values (enum).
|
145
|
+
# @param [Object] type Object to be assigned
|
146
|
+
def type=(type)
|
147
|
+
validator = EnumAttributeValidator.new('String', ['flux', 'influxql'])
|
148
|
+
unless validator.valid?(type)
|
149
|
+
fail ArgumentError, 'invalid value for "type", must be one of #{validator.allowable_values}.'
|
150
|
+
end
|
151
|
+
@type = type
|
152
|
+
end
|
153
|
+
|
154
|
+
# Checks equality by comparing each attribute.
|
155
|
+
# @param [Object] Object to be compared
|
156
|
+
def ==(o)
|
157
|
+
return true if self.equal?(o)
|
158
|
+
self.class == o.class &&
|
159
|
+
extern == o.extern &&
|
160
|
+
query == o.query &&
|
161
|
+
type == o.type &&
|
162
|
+
db == o.db &&
|
163
|
+
rp == o.rp &&
|
164
|
+
cluster == o.cluster &&
|
165
|
+
dialect == o.dialect
|
166
|
+
end
|
167
|
+
|
168
|
+
# @see the `==` method
|
169
|
+
# @param [Object] Object to be compared
|
170
|
+
def eql?(o)
|
171
|
+
self == o
|
172
|
+
end
|
173
|
+
|
174
|
+
# Calculates hash code according to all attributes.
|
175
|
+
# @return [Fixnum] Hash code
|
176
|
+
def hash
|
177
|
+
[extern, query, type, db, rp, cluster, dialect].hash
|
178
|
+
end
|
179
|
+
|
180
|
+
# Builds the object from hash
|
181
|
+
# @param [Hash] attributes Model attributes in the form of hash
|
182
|
+
# @return [Object] Returns the model itself
|
183
|
+
def build_from_hash(attributes)
|
184
|
+
return nil unless attributes.is_a?(Hash)
|
185
|
+
self.class.openapi_types.each_pair do |key, type|
|
186
|
+
if type =~ /\AArray<(.*)>/i
|
187
|
+
# check to ensure the input is an array given that the the attribute
|
188
|
+
# is documented as an array but the input is not
|
189
|
+
if attributes[self.class.attribute_map[key]].is_a?(Array)
|
190
|
+
self.send("#{key}=", attributes[self.class.attribute_map[key]].map { |v| _deserialize($1, v) })
|
191
|
+
end
|
192
|
+
elsif !attributes[self.class.attribute_map[key]].nil?
|
193
|
+
self.send("#{key}=", _deserialize(type, attributes[self.class.attribute_map[key]]))
|
194
|
+
end # or else data not found in attributes(hash), not an issue as the data can be optional
|
195
|
+
end
|
196
|
+
|
197
|
+
self
|
198
|
+
end
|
199
|
+
|
200
|
+
# Deserializes the data based on type
|
201
|
+
# @param string type Data type
|
202
|
+
# @param string value Value to be deserialized
|
203
|
+
# @return [Object] Deserialized data
|
204
|
+
def _deserialize(type, value)
|
205
|
+
case type.to_sym
|
206
|
+
when :DateTime
|
207
|
+
DateTime.parse(value)
|
208
|
+
when :Date
|
209
|
+
Date.parse(value)
|
210
|
+
when :String
|
211
|
+
value.to_s
|
212
|
+
when :Integer
|
213
|
+
value.to_i
|
214
|
+
when :Float
|
215
|
+
value.to_f
|
216
|
+
when :BOOLEAN
|
217
|
+
if value.to_s =~ /\A(true|t|yes|y|1)\z/i
|
218
|
+
true
|
219
|
+
else
|
220
|
+
false
|
221
|
+
end
|
222
|
+
when :Object
|
223
|
+
# generic object (usually a Hash), return directly
|
224
|
+
value
|
225
|
+
when /\AArray<(?<inner_type>.+)>\z/
|
226
|
+
inner_type = Regexp.last_match[:inner_type]
|
227
|
+
value.map { |v| _deserialize(inner_type, v) }
|
228
|
+
when /\AHash<(?<k_type>.+?), (?<v_type>.+)>\z/
|
229
|
+
k_type = Regexp.last_match[:k_type]
|
230
|
+
v_type = Regexp.last_match[:v_type]
|
231
|
+
{}.tap do |hash|
|
232
|
+
value.each do |k, v|
|
233
|
+
hash[_deserialize(k_type, k)] = _deserialize(v_type, v)
|
234
|
+
end
|
235
|
+
end
|
236
|
+
else # model
|
237
|
+
temp_model = InfluxDB2.const_get(type).new
|
238
|
+
temp_model.build_from_hash(value)
|
239
|
+
end
|
240
|
+
end
|
241
|
+
|
242
|
+
# Returns the string representation of the object
|
243
|
+
# @return [String] String presentation of the object
|
244
|
+
def to_s
|
245
|
+
to_hash.to_s
|
246
|
+
end
|
247
|
+
|
248
|
+
# to_body is an alias to to_hash (backward compatibility)
|
249
|
+
# @return [Hash] Returns the object in the form of hash
|
250
|
+
def to_body
|
251
|
+
to_hash
|
252
|
+
end
|
253
|
+
|
254
|
+
# Returns the object in the form of hash
|
255
|
+
# @return [Hash] Returns the object in the form of hash
|
256
|
+
def to_hash
|
257
|
+
hash = {}
|
258
|
+
self.class.attribute_map.each_pair do |attr, param|
|
259
|
+
value = self.send(attr)
|
260
|
+
next if value.nil?
|
261
|
+
hash[param] = _to_hash(value)
|
262
|
+
end
|
263
|
+
hash
|
264
|
+
end
|
265
|
+
|
266
|
+
# Outputs non-array value in the form of hash
|
267
|
+
# For object, use to_hash. Otherwise, just return the value
|
268
|
+
# @param [Object] value Any valid value
|
269
|
+
# @return [Hash] Returns the value in the form of hash
|
270
|
+
def _to_hash(value)
|
271
|
+
if value.is_a?(Array)
|
272
|
+
value.compact.map { |v| _to_hash(v) }
|
273
|
+
elsif value.is_a?(Hash)
|
274
|
+
{}.tap do |hash|
|
275
|
+
value.each { |k, v| hash[k] = _to_hash(v) }
|
276
|
+
end
|
277
|
+
elsif value.respond_to? :to_hash
|
278
|
+
value.to_hash
|
279
|
+
else
|
280
|
+
value
|
281
|
+
end
|
282
|
+
end
|
283
|
+
end
|
284
|
+
end
|
@@ -0,0 +1,79 @@
|
|
1
|
+
# The MIT License
|
2
|
+
#
|
3
|
+
# Permission is hereby granted, free of charge, to any person obtaining a copy
|
4
|
+
# of this software and associated documentation files (the "Software"), to deal
|
5
|
+
# in the Software without restriction, including without limitation the rights
|
6
|
+
# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
7
|
+
# copies of the Software, and to permit persons to whom the Software is
|
8
|
+
# furnished to do so, subject to the following conditions:
|
9
|
+
#
|
10
|
+
# The above copyright notice and this permission notice shall be included in
|
11
|
+
# all copies or substantial portions of the Software.
|
12
|
+
#
|
13
|
+
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
14
|
+
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
15
|
+
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
16
|
+
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
17
|
+
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
18
|
+
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
|
19
|
+
# THE SOFTWARE.
|
20
|
+
require_relative 'models/dialect'
|
21
|
+
require_relative 'models/query'
|
22
|
+
require_relative 'flux_csv_parser'
|
23
|
+
require 'json'
|
24
|
+
|
25
|
+
module InfluxDB2
|
26
|
+
# The client of the InfluxDB 2.0 that implement Query HTTP API endpoint.
|
27
|
+
#
|
28
|
+
class QueryApi < DefaultApi
|
29
|
+
DEFAULT_DIALECT = InfluxDB2::Dialect.new(header: true, delimiter: ',', comment_prefix: '#',
|
30
|
+
annotations: %w[datatype group default])
|
31
|
+
|
32
|
+
# @param [Hash] options The options to be used by the client.
|
33
|
+
def initialize(options:)
|
34
|
+
super(options: options)
|
35
|
+
end
|
36
|
+
|
37
|
+
# @param [Object] query the flux query to execute. The data could be represent by [String], [Query]
|
38
|
+
# @param [String] org specifies the source organization
|
39
|
+
# @return [String] result of query
|
40
|
+
def query_raw(query: nil, org: nil, dialect: DEFAULT_DIALECT)
|
41
|
+
org_param = org || @options[:org]
|
42
|
+
_check('org', org_param)
|
43
|
+
|
44
|
+
payload = _generate_payload(query, dialect)
|
45
|
+
return nil if payload.nil?
|
46
|
+
|
47
|
+
uri = URI.parse(File.join(@options[:url], '/api/v2/query'))
|
48
|
+
uri.query = URI.encode_www_form(org: org_param)
|
49
|
+
|
50
|
+
_post(payload.to_body.to_json, uri).read_body
|
51
|
+
end
|
52
|
+
|
53
|
+
# @param [Object] query the flux query to execute. The data could be represent by [String], [Query]
|
54
|
+
# @param [String] org specifies the source organization
|
55
|
+
# @return [Array] list of FluxTables which are matched the query
|
56
|
+
def query(query: nil, org: nil, dialect: DEFAULT_DIALECT)
|
57
|
+
response = query_raw(query: query, org: org, dialect: dialect)
|
58
|
+
parser = InfluxDB2::FluxCsvParser.new
|
59
|
+
|
60
|
+
parser.parse(response)
|
61
|
+
end
|
62
|
+
|
63
|
+
private
|
64
|
+
|
65
|
+
def _generate_payload(query, dialect)
|
66
|
+
if query.nil?
|
67
|
+
nil
|
68
|
+
elsif query.is_a?(Query)
|
69
|
+
query
|
70
|
+
elsif query.is_a?(String)
|
71
|
+
if query.empty?
|
72
|
+
nil
|
73
|
+
else
|
74
|
+
Query.new(query: query, dialect: dialect, type: nil)
|
75
|
+
end
|
76
|
+
end
|
77
|
+
end
|
78
|
+
end
|
79
|
+
end
|
@@ -37,14 +37,10 @@ module InfluxDB2
|
|
37
37
|
|
38
38
|
# Write time series data into InfluxDB.
|
39
39
|
#
|
40
|
-
class WriteApi
|
41
|
-
DEFAULT_TIMEOUT = 10
|
42
|
-
DEFAULT_REDIRECT_COUNT = 10
|
43
|
-
|
40
|
+
class WriteApi < DefaultApi
|
44
41
|
# @param [Hash] options The options to be used by the client.
|
45
42
|
def initialize(options:)
|
46
|
-
|
47
|
-
@max_redirect_count = @options[:max_redirect_count] || DEFAULT_REDIRECT_COUNT
|
43
|
+
super(options: options)
|
48
44
|
end
|
49
45
|
|
50
46
|
# Write data into specified Bucket.
|
@@ -98,39 +94,6 @@ module InfluxDB2
|
|
98
94
|
|
99
95
|
private
|
100
96
|
|
101
|
-
def _post(payload, uri, limit = @max_redirect_count)
|
102
|
-
raise InfluxError.from_message("Too many HTTP redirects. Exceeded limit: #{@max_redirect_count}") if limit.zero?
|
103
|
-
|
104
|
-
http = Net::HTTP.new(uri.host, uri.port)
|
105
|
-
http.open_timeout = @options[:open_timeout] || DEFAULT_TIMEOUT
|
106
|
-
http.write_timeout = @options[:write_timeout] || DEFAULT_TIMEOUT if Net::HTTP.method_defined? :write_timeout
|
107
|
-
http.read_timeout = @options[:read_timeout] || DEFAULT_TIMEOUT
|
108
|
-
http.use_ssl = @options[:use_ssl].nil? ? true : @options[:use_ssl]
|
109
|
-
|
110
|
-
request = Net::HTTP::Post.new(uri.request_uri)
|
111
|
-
request['Authorization'] = "Token #{@options[:token]}"
|
112
|
-
request.body = payload
|
113
|
-
|
114
|
-
begin
|
115
|
-
response = http.request(request)
|
116
|
-
case response
|
117
|
-
when Net::HTTPSuccess then
|
118
|
-
response
|
119
|
-
when Net::HTTPRedirection then
|
120
|
-
location = response['location']
|
121
|
-
_post(payload, URI.parse(location), limit - 1)
|
122
|
-
else
|
123
|
-
raise InfluxError.from_response(response)
|
124
|
-
end
|
125
|
-
ensure
|
126
|
-
http.finish if http.started?
|
127
|
-
end
|
128
|
-
end
|
129
|
-
|
130
|
-
def _check(key, value)
|
131
|
-
raise ArgumentError, "The '#{key}' should be defined as argument or default option: #{@options}" if value.nil?
|
132
|
-
end
|
133
|
-
|
134
97
|
def _generate_payload(data)
|
135
98
|
if data.nil?
|
136
99
|
nil
|
@@ -0,0 +1,328 @@
|
|
1
|
+
# The MIT License
|
2
|
+
#
|
3
|
+
# Permission is hereby granted, free of charge, to any person obtaining a copy
|
4
|
+
# of this software and associated documentation files (the "Software"), to deal
|
5
|
+
# in the Software without restriction, including without limitation the rights
|
6
|
+
# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
7
|
+
# copies of the Software, and to permit persons to whom the Software is
|
8
|
+
# furnished to do so, subject to the following conditions:
|
9
|
+
#
|
10
|
+
# The above copyright notice and this permission notice shall be included in
|
11
|
+
# all copies or substantial portions of the Software.
|
12
|
+
#
|
13
|
+
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
14
|
+
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
15
|
+
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
16
|
+
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
17
|
+
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
18
|
+
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
|
19
|
+
# THE SOFTWARE.
|
20
|
+
|
21
|
+
require 'test_helper'
|
22
|
+
|
23
|
+
class FluxCsvParserTest < MiniTest::Test
|
24
|
+
def setup
|
25
|
+
@parser = InfluxDB2::FluxCsvParser.new
|
26
|
+
end
|
27
|
+
|
28
|
+
def test_multiple_values
|
29
|
+
data = "#datatype,string,long,dateTime:RFC3339,dateTime:RFC3339,string,string,string,string,long,long,string\n" \
|
30
|
+
"#group,false,false,true,true,true,true,true,true,false,false,false\n" \
|
31
|
+
"#default,_result,,,,,,,,,,\n" \
|
32
|
+
",result,table,_start,_stop,_field,_measurement,host,region,_value2,value1,value_str\n" \
|
33
|
+
",,0,1677-09-21T00:12:43.145224192Z,2018-07-16T11:21:02.547596934Z,free,mem,A,west,121,11,test\n" \
|
34
|
+
",,1,1677-09-21T00:12:43.145224192Z,2018-07-16T11:21:02.547596934Z,free,mem,B,west,484,22,test\n" \
|
35
|
+
",,2,1677-09-21T00:12:43.145224192Z,2018-07-16T11:21:02.547596934Z,usage_system,cpu,A,west,1444,38,test\n" \
|
36
|
+
',,3,1677-09-21T00:12:43.145224192Z,2018-07-16T11:21:02.547596934Z,user_usage,cpu,A,west,2401,49,test'
|
37
|
+
|
38
|
+
tables = @parser.parse(data)
|
39
|
+
|
40
|
+
column_headers = tables[0].columns
|
41
|
+
assert_equal 11, column_headers.size
|
42
|
+
|
43
|
+
values = [false, false, true, true, true, true, true, true, false, false, false]
|
44
|
+
_assert_columns(column_headers, values: values)
|
45
|
+
assert_equal 4, tables.size
|
46
|
+
|
47
|
+
_assert_multiple_record(tables)
|
48
|
+
end
|
49
|
+
|
50
|
+
def test_parse_shortcut
|
51
|
+
data = '#datatype,string,long,dateTime:RFC3339,dateTime:RFC3339,' \
|
52
|
+
"dateTime:RFC3339,long,string,string,string,boolean\n" \
|
53
|
+
"#group,false,false,false,false,false,false,false,false,false,true\n" \
|
54
|
+
"#default,_result,,,,,,,,,true\n" \
|
55
|
+
",result,table,_start,_stop,_time,_value,_field,_measurement,host,value\n" \
|
56
|
+
",,0,1970-01-01T00:00:10Z,1970-01-01T00:00:20Z,1970-01-01T00:00:10Z,10,free,mem,A,true\n"
|
57
|
+
|
58
|
+
tables = @parser.parse(data)
|
59
|
+
|
60
|
+
assert_equal 1, tables.size
|
61
|
+
assert_equal 1, tables[0].records.size
|
62
|
+
|
63
|
+
record = tables[0].records[0]
|
64
|
+
|
65
|
+
assert_equal _parse_time('1970-01-01T00:00:10Z'), record.start
|
66
|
+
assert_equal _parse_time('1970-01-01T00:00:20Z'), record.stop
|
67
|
+
assert_equal _parse_time('1970-01-01T00:00:10Z'), record.time
|
68
|
+
assert_equal 10, record.value
|
69
|
+
assert_equal 'free', record.field
|
70
|
+
assert_equal 'mem', record.measurement
|
71
|
+
end
|
72
|
+
|
73
|
+
def test_mapping_boolean
|
74
|
+
data = '#datatype,string,long,dateTime:RFC3339,dateTime:RFC3339,' \
|
75
|
+
"dateTime:RFC3339,long,string,string,string,boolean\n" \
|
76
|
+
"#group,false,false,false,false,false,false,false,false,false,true\n" \
|
77
|
+
"#default,_result,,,,,,,,,true\n" \
|
78
|
+
",result,table,_start,_stop,_time,_value,_field,_measurement,host,value\n" \
|
79
|
+
",,0,1970-01-01T00:00:10Z,1970-01-01T00:00:20Z,1970-01-01T00:00:10Z,10,free,mem,A,true\n" \
|
80
|
+
",,0,1970-01-01T00:00:10Z,1970-01-01T00:00:20Z,1970-01-01T00:00:10Z,10,free,mem,A,false\n" \
|
81
|
+
",,0,1970-01-01T00:00:10Z,1970-01-01T00:00:20Z,1970-01-01T00:00:10Z,10,free,mem,A,x\n" \
|
82
|
+
",,0,1970-01-01T00:00:10Z,1970-01-01T00:00:20Z,1970-01-01T00:00:10Z,10,free,mem,A,\n"
|
83
|
+
|
84
|
+
tables = @parser.parse(data)
|
85
|
+
records = tables[0].records
|
86
|
+
|
87
|
+
assert_equal true, records[0].values['value']
|
88
|
+
assert_equal false, records[1].values['value']
|
89
|
+
assert_equal false, records[2].values['value']
|
90
|
+
assert_equal true, records[3].values['value']
|
91
|
+
end
|
92
|
+
|
93
|
+
def test_mapping_unsigned_long
|
94
|
+
data = '#datatype,string,long,dateTime:RFC3339,dateTime:RFC3339,' \
|
95
|
+
"dateTime:RFC3339,long,string,string,string,unsignedLong\n" \
|
96
|
+
"#group,false,false,false,false,false,false,false,false,false,true\n" \
|
97
|
+
"#default,_result,,,,,,,,,\n" \
|
98
|
+
",result,table,_start,_stop,_time,_value,_field,_measurement,host,value\n" \
|
99
|
+
",,0,1970-01-01T00:00:10Z,1970-01-01T00:00:20Z,1970-01-01T00:00:10Z,10,free,mem,A,17916881237904312345\n" \
|
100
|
+
",,0,1970-01-01T00:00:10Z,1970-01-01T00:00:20Z,1970-01-01T00:00:10Z,10,free,mem,A,\n"
|
101
|
+
|
102
|
+
expected = 17_916_881_237_904_312_345
|
103
|
+
|
104
|
+
tables = @parser.parse(data)
|
105
|
+
records = tables[0].records
|
106
|
+
|
107
|
+
assert_equal expected, records[0].values['value']
|
108
|
+
assert_nil records[1].values['value']
|
109
|
+
end
|
110
|
+
|
111
|
+
def test_mapping_double
|
112
|
+
data = '#datatype,string,long,dateTime:RFC3339,dateTime:RFC3339,' \
|
113
|
+
"dateTime:RFC3339,long,string,string,string,double\n" \
|
114
|
+
"#group,false,false,false,false,false,false,false,false,false,true\n" \
|
115
|
+
"#default,_result,,,,,,,,,\n" \
|
116
|
+
",result,table,_start,_stop,_time,_value,_field,_measurement,host,value\n" \
|
117
|
+
",,0,1970-01-01T00:00:10Z,1970-01-01T00:00:20Z,1970-01-01T00:00:10Z,10,free,mem,A,12.25\n" \
|
118
|
+
",,0,1970-01-01T00:00:10Z,1970-01-01T00:00:20Z,1970-01-01T00:00:10Z,10,free,mem,A,\n" \
|
119
|
+
|
120
|
+
tables = @parser.parse(data)
|
121
|
+
records = tables[0].records
|
122
|
+
|
123
|
+
assert_equal 12.25, records[0].values['value']
|
124
|
+
assert_nil records[1].values['value']
|
125
|
+
end
|
126
|
+
|
127
|
+
def test_mapping_base64_binary
|
128
|
+
binary_data = 'test value'
|
129
|
+
encoded_data = Base64.encode64(binary_data)
|
130
|
+
|
131
|
+
data = '#datatype,string,long,dateTime:RFC3339,dateTime:RFC3339,' \
|
132
|
+
"dateTime:RFC3339,long,string,string,string,base64Binary\n" \
|
133
|
+
"#group,false,false,false,false,false,false,false,false,false,true\n" \
|
134
|
+
"#default,_result,,,,,,,,,\n" \
|
135
|
+
",result,table,_start,_stop,_time,_value,_field,_measurement,host,value\n" \
|
136
|
+
',,0,1970-01-01T00:00:10Z,1970-01-01T00:00:20Z,1970-01-01T00:00:10Z,10,free,mem,A,' + encoded_data + "\n" \
|
137
|
+
",,0,1970-01-01T00:00:10Z,1970-01-01T00:00:20Z,1970-01-01T00:00:10Z,10,free,mem,A,\n"
|
138
|
+
|
139
|
+
tables = @parser.parse(data)
|
140
|
+
records = tables[0].records
|
141
|
+
|
142
|
+
value = records[0].values['value']
|
143
|
+
|
144
|
+
assert !value.nil?
|
145
|
+
assert_equal binary_data, value
|
146
|
+
|
147
|
+
assert_nil records[1].values['value']
|
148
|
+
end
|
149
|
+
|
150
|
+
def test_mapping_rfc3339
|
151
|
+
data = '#datatype,string,long,dateTime:RFC3339,dateTime:RFC3339,' \
|
152
|
+
"dateTime:RFC3339,long,string,string,string,dateTime:RFC3339\n" \
|
153
|
+
"#group,false,false,false,false,false,false,false,false,false,true\n" \
|
154
|
+
"#default,_result,,,,,,,,,\n" \
|
155
|
+
",result,table,_start,_stop,_time,_value,_field,_measurement,host,value\n" \
|
156
|
+
",,0,1970-01-01T00:00:10Z,1970-01-01T00:00:20Z,1970-01-01T00:00:10Z,10,free,mem,A,1970-01-01T00:00:10Z\n" \
|
157
|
+
",,0,1970-01-01T00:00:10Z,1970-01-01T00:00:20Z,1970-01-01T00:00:10Z,10,free,mem,A,\n"
|
158
|
+
|
159
|
+
tables = @parser.parse(data)
|
160
|
+
records = tables[0].records
|
161
|
+
|
162
|
+
assert_equal Time.parse('1970-01-01T00:00:10Z').to_datetime.rfc3339, records[0].values['value']
|
163
|
+
assert_nil records[1].values['value']
|
164
|
+
end
|
165
|
+
|
166
|
+
def test_mapping_duration
|
167
|
+
data = '#datatype,string,long,dateTime:RFC3339,dateTime:RFC3339' \
|
168
|
+
",dateTime:RFC3339,long,string,string,string,duration\n" \
|
169
|
+
"#group,false,false,false,false,false,false,false,false,false,true\n" \
|
170
|
+
"#default,_result,,,,,,,,,\n" \
|
171
|
+
",result,table,_start,_stop,_time,_value,_field,_measurement,host,value\n" \
|
172
|
+
",,0,1970-01-01T00:00:10Z,1970-01-01T00:00:20Z,1970-01-01T00:00:10Z,10,free,mem,A,125\n" \
|
173
|
+
",,0,1970-01-01T00:00:10Z,1970-01-01T00:00:20Z,1970-01-01T00:00:10Z,10,free,mem,A,\n"
|
174
|
+
|
175
|
+
tables = @parser.parse(data)
|
176
|
+
records = tables[0].records
|
177
|
+
|
178
|
+
assert_equal 125, records[0].values['value']
|
179
|
+
assert_nil records[1].values['value']
|
180
|
+
end
|
181
|
+
|
182
|
+
def test_group_key
|
183
|
+
data = '#datatype,string,long,dateTime:RFC3339,dateTime:RFC3339,' \
|
184
|
+
"dateTime:RFC3339,long,string,string,string,duration\n" \
|
185
|
+
"#group,false,false,false,false,true,false,false,false,false,true\n" \
|
186
|
+
"#default,_result,,,,,,,,,\n" \
|
187
|
+
",result,table,_start,_stop,_time,_value,_field,_measurement,host,value\n" \
|
188
|
+
",,0,1970-01-01T00:00:10Z,1970-01-01T00:00:20Z,1970-01-01T00:00:10Z,10,free,mem,A,125\n" \
|
189
|
+
",,0,1970-01-01T00:00:10Z,1970-01-01T00:00:20Z,1970-01-01T00:00:10Z,10,free,mem,A,\n" \
|
190
|
+
|
191
|
+
tables = @parser.parse(data)
|
192
|
+
|
193
|
+
assert_equal 10, tables[0].columns.size
|
194
|
+
assert_equal 2, tables[0].group_key.size
|
195
|
+
end
|
196
|
+
|
197
|
+
def test_unknown_type_as_string
|
198
|
+
data = '#datatype,string,long,dateTime:RFC3339,dateTime:RFC3339,' \
|
199
|
+
"dateTime:RFC3339,long,string,string,string,unknown\n" \
|
200
|
+
"#group,false,false,false,false,false,false,false,false,false,true\n" \
|
201
|
+
"#default,_result,,,,,,,,,\n" \
|
202
|
+
",result,table,_start,_stop,_time,_value,_field,_measurement,host,value\n" \
|
203
|
+
",,0,1970-01-01T00:00:10Z,1970-01-01T00:00:20Z,1970-01-01T00:00:10Z,10,free,mem,A,12.25\n" \
|
204
|
+
",,0,1970-01-01T00:00:10Z,1970-01-01T00:00:20Z,1970-01-01T00:00:10Z,10,free,mem,A,\n"
|
205
|
+
|
206
|
+
tables = @parser.parse(data)
|
207
|
+
records = tables[0].records
|
208
|
+
|
209
|
+
assert_equal '12.25', records[0].values['value']
|
210
|
+
assert_nil records[1].values['value']
|
211
|
+
end
|
212
|
+
|
213
|
+
private
|
214
|
+
|
215
|
+
def _parse_time(time)
|
216
|
+
Time.parse(time).to_datetime.rfc3339
|
217
|
+
end
|
218
|
+
|
219
|
+
def _assert_record(flux_record, values: nil, size: 0, value: nil)
|
220
|
+
values.keys.each do |key|
|
221
|
+
assert_equal values[key], flux_record.values[key]
|
222
|
+
end
|
223
|
+
|
224
|
+
if value.nil?
|
225
|
+
assert_nil value
|
226
|
+
else
|
227
|
+
assert_equal value, flux_record.value
|
228
|
+
end
|
229
|
+
|
230
|
+
assert_equal size, flux_record.values.size
|
231
|
+
end
|
232
|
+
|
233
|
+
def _assert_columns(column_headers, values: nil)
|
234
|
+
i = 0
|
235
|
+
values.each do |value|
|
236
|
+
assert_equal value, column_headers[i].group
|
237
|
+
i += 1
|
238
|
+
end
|
239
|
+
end
|
240
|
+
|
241
|
+
def _assert_multiple_record(tables)
|
242
|
+
# Record 1
|
243
|
+
table_records = tables[0].records
|
244
|
+
assert_equal 1, table_records.size
|
245
|
+
|
246
|
+
values = { 'table' => 0, 'host' => 'A', 'region' => 'west', 'value1' => 11, '_value2' => 121,
|
247
|
+
'value_str' => 'test' }
|
248
|
+
|
249
|
+
_assert_record(table_records[0], values: values, size: 11)
|
250
|
+
|
251
|
+
# Record 2
|
252
|
+
table_records = tables[1].records
|
253
|
+
assert_equal 1, table_records.size
|
254
|
+
|
255
|
+
values = { 'table' => 1, 'host' => 'B', 'region' => 'west', 'value1' => 22, '_value2' => 484,
|
256
|
+
'value_str' => 'test' }
|
257
|
+
|
258
|
+
_assert_record(table_records[0], values: values, size: 11)
|
259
|
+
|
260
|
+
# Record 3
|
261
|
+
table_records = tables[2].records
|
262
|
+
assert_equal 1, table_records.size
|
263
|
+
|
264
|
+
values = { 'table' => 2, 'host' => 'A', 'region' => 'west', 'value1' => 38, '_value2' => 1444,
|
265
|
+
'value_str' => 'test' }
|
266
|
+
|
267
|
+
_assert_record(table_records[0], values: values, size: 11)
|
268
|
+
|
269
|
+
# Record 4
|
270
|
+
table_records = tables[3].records
|
271
|
+
assert_equal 1, table_records.size
|
272
|
+
|
273
|
+
values = { 'table' => 3, 'host' => 'A', 'region' => 'west', 'value1' => 49, '_value2' => 2401,
|
274
|
+
'value_str' => 'test' }
|
275
|
+
|
276
|
+
_assert_record(table_records[0], values: values, size: 11)
|
277
|
+
end
|
278
|
+
end
|
279
|
+
|
280
|
+
class FluxCsvParserErrorTest < MiniTest::Test
|
281
|
+
def setup
|
282
|
+
@parser = InfluxDB2::FluxCsvParser.new
|
283
|
+
end
|
284
|
+
|
285
|
+
def test_error
|
286
|
+
data = "#datatype,string,string\n" \
|
287
|
+
"#group,true,true\n" \
|
288
|
+
"#default,,\n" \
|
289
|
+
",error,reference\n" \
|
290
|
+
',failed to create physical plan: invalid time bounds from procedure from: bounds contain zero time,897'
|
291
|
+
|
292
|
+
error = assert_raises InfluxDB2::FluxQueryError do
|
293
|
+
@parser.parse(data)
|
294
|
+
end
|
295
|
+
|
296
|
+
assert_equal 'failed to create physical plan: invalid time bounds from procedure from: bounds contain zero time',
|
297
|
+
error.message
|
298
|
+
assert_equal 897, error.reference
|
299
|
+
end
|
300
|
+
|
301
|
+
def test_error_without_reference
|
302
|
+
data = "#datatype,string,string\n" \
|
303
|
+
"#group,true,true\n" \
|
304
|
+
"#default,,\n" \
|
305
|
+
",error,reference\n" \
|
306
|
+
',failed to create physical plan: invalid time bounds from procedure from: bounds contain zero time,'
|
307
|
+
|
308
|
+
error = assert_raises InfluxDB2::FluxQueryError do
|
309
|
+
@parser.parse(data)
|
310
|
+
end
|
311
|
+
|
312
|
+
assert_equal 'failed to create physical plan: invalid time bounds from procedure from: bounds contain zero time',
|
313
|
+
error.message
|
314
|
+
assert_equal 0, error.reference
|
315
|
+
end
|
316
|
+
|
317
|
+
def test_without_table_definition
|
318
|
+
data = ",result,table,_start,_stop,_time,_value,_field,_measurement,host,value\n" \
|
319
|
+
",,0,1970-01-01T00:00:10Z,1970-01-01T00:00:20Z,1970-01-01T00:00:10Z,10,free,mem,A,12.25\n" \
|
320
|
+
",,0,1970-01-01T00:00:10Z,1970-01-01T00:00:20Z,1970-01-01T00:00:10Z,10,free,mem,A,\n"
|
321
|
+
|
322
|
+
error = assert_raises InfluxDB2::FluxCsvParserError do
|
323
|
+
@parser.parse(data)
|
324
|
+
end
|
325
|
+
|
326
|
+
assert_equal 'Unable to parse CSV response. FluxTable definition was not found.', error.message
|
327
|
+
end
|
328
|
+
end
|