hocon 1.2.5 → 1.2.6

Sign up to get free protection for your applications and to get access to all the features.
Files changed (50) hide show
  1. checksums.yaml +4 -4
  2. data/CHANGELOG.md +6 -1
  3. data/lib/hocon/version.rb +1 -1
  4. metadata +1 -47
  5. data/spec/fixtures/hocon/by_extension/cat.conf +0 -4
  6. data/spec/fixtures/hocon/by_extension/cat.test +0 -4
  7. data/spec/fixtures/hocon/by_extension/cat.test-json +0 -3
  8. data/spec/fixtures/hocon/with_substitution/subst.conf +0 -2
  9. data/spec/fixtures/parse_render/example1/input.conf +0 -21
  10. data/spec/fixtures/parse_render/example1/output.conf +0 -26
  11. data/spec/fixtures/parse_render/example1/output_nocomments.conf +0 -17
  12. data/spec/fixtures/parse_render/example2/input.conf +0 -10
  13. data/spec/fixtures/parse_render/example2/output.conf +0 -17
  14. data/spec/fixtures/parse_render/example2/output_nocomments.conf +0 -17
  15. data/spec/fixtures/parse_render/example3/input.conf +0 -2
  16. data/spec/fixtures/parse_render/example3/output.conf +0 -2
  17. data/spec/fixtures/parse_render/example4/input.json +0 -6
  18. data/spec/fixtures/parse_render/example4/output.conf +0 -6
  19. data/spec/fixtures/test_utils/resources/bom.conf +0 -2
  20. data/spec/fixtures/test_utils/resources/cycle.conf +0 -1
  21. data/spec/fixtures/test_utils/resources/file-include.conf +0 -5
  22. data/spec/fixtures/test_utils/resources/include-from-list.conf +0 -4
  23. data/spec/fixtures/test_utils/resources/subdir/bar.conf +0 -1
  24. data/spec/fixtures/test_utils/resources/subdir/baz.conf +0 -1
  25. data/spec/fixtures/test_utils/resources/subdir/foo.conf +0 -5
  26. data/spec/fixtures/test_utils/resources/test01.conf +0 -80
  27. data/spec/fixtures/test_utils/resources/test01.json +0 -4
  28. data/spec/fixtures/test_utils/resources/test03.conf +0 -36
  29. data/spec/fixtures/test_utils/resources/utf16.conf +0 -0
  30. data/spec/fixtures/test_utils/resources/utf8.conf +0 -2
  31. data/spec/fixtures/test_utils/resources//341/232/240/341/233/207/341/232/273.conf +0 -2
  32. data/spec/spec_helper.rb +0 -43
  33. data/spec/test_utils.rb +0 -758
  34. data/spec/unit/cli/cli_spec.rb +0 -157
  35. data/spec/unit/hocon/README.md +0 -7
  36. data/spec/unit/hocon/hocon_spec.rb +0 -114
  37. data/spec/unit/typesafe/config/README.md +0 -4
  38. data/spec/unit/typesafe/config/concatenation_spec.rb +0 -417
  39. data/spec/unit/typesafe/config/conf_parser_spec.rb +0 -831
  40. data/spec/unit/typesafe/config/config_document_parser_spec.rb +0 -494
  41. data/spec/unit/typesafe/config/config_document_spec.rb +0 -576
  42. data/spec/unit/typesafe/config/config_factory_spec.rb +0 -120
  43. data/spec/unit/typesafe/config/config_node_spec.rb +0 -552
  44. data/spec/unit/typesafe/config/config_value_factory_spec.rb +0 -85
  45. data/spec/unit/typesafe/config/config_value_spec.rb +0 -959
  46. data/spec/unit/typesafe/config/path_spec.rb +0 -261
  47. data/spec/unit/typesafe/config/public_api_spec.rb +0 -520
  48. data/spec/unit/typesafe/config/simple_config_spec.rb +0 -112
  49. data/spec/unit/typesafe/config/token_spec.rb +0 -188
  50. data/spec/unit/typesafe/config/tokenizer_spec.rb +0 -801
@@ -1,112 +0,0 @@
1
- # encoding: utf-8
2
-
3
- require 'spec_helper'
4
- require 'hocon/config_factory'
5
- require 'hocon/config_render_options'
6
- require 'hocon/config_value_factory'
7
-
8
- describe Hocon::Impl::SimpleConfig do
9
- let(:render_options) { Hocon::ConfigRenderOptions.defaults }
10
-
11
- before do
12
- render_options.origin_comments = false
13
- render_options.json = false
14
- end
15
-
16
- shared_examples_for "config_value_retrieval_single_value" do
17
- let(:input_file) { "#{FIXTURE_DIR}/parse_render/#{example[:name]}/input.conf" }
18
- it "should allow you to get a value for a specific configuration setting" do
19
- expect(conf.get_value(setting).transform_to_string).to eq(expected_setting)
20
- end
21
- end
22
-
23
- shared_examples_for "config_value_retrieval_config_list" do
24
- let(:input_file) { "#{FIXTURE_DIR}/parse_render/#{example[:name]}/input.conf" }
25
- it "should allow you to get a value for a setting whose value is a data structure" do
26
- expect(conf.get_value(setting).
27
- render_value_to_sb(StringIO.new, 2, nil,
28
- Hocon::ConfigRenderOptions.new(false, false, false, false)).
29
- string).to eq(expected_setting)
30
- end
31
- end
32
-
33
- shared_examples_for "has_path_check" do
34
- let(:input_file) { "#{FIXTURE_DIR}/parse_render/#{example[:name]}/input.conf" }
35
- it "should return true if a path exists" do
36
- expect(conf.has_path?(setting)).to eql(true)
37
- end
38
-
39
- it "should return false if a path does not exist" do
40
- expect(conf.has_path?(false_setting)).to eq(false)
41
- end
42
- end
43
-
44
- shared_examples_for "add_value_to_config" do
45
- let(:input_file) { "#{FIXTURE_DIR}/parse_render/#{example[:name]}/input.conf" }
46
- it "should add desired setting with desired value" do
47
- modified_conf = conf.with_value(setting_to_add, value_to_add)
48
- expect(modified_conf.get_value(setting_to_add)).to eq(value_to_add)
49
- end
50
- end
51
-
52
- shared_examples_for "add_data_structures_to_config" do
53
- let(:input_file) { "#{FIXTURE_DIR}/parse_render/#{example[:name]}/input.conf" }
54
- it "should add a nested map to a config" do
55
- map = Hocon::ConfigValueFactory.from_any_ref({"a" => "b", "c" => {"d" => "e"}}, nil)
56
- modified_conf = conf.with_value(setting_to_add, map)
57
- expect(modified_conf.get_value(setting_to_add)).to eq(map)
58
- end
59
-
60
- it "should add an array to a config" do
61
- array = Hocon::ConfigValueFactory.from_any_ref([1,2,3,4,5], nil)
62
- modified_conf = conf.with_value(setting_to_add, array)
63
- expect(modified_conf.get_value(setting_to_add)).to eq(array)
64
- end
65
- end
66
-
67
- shared_examples_for "remove_value_from_config" do
68
- let(:input_file) { "#{FIXTURE_DIR}/parse_render/#{example[:name]}/input.conf" }
69
- it "should remove desired setting" do
70
- modified_conf = conf.without_path(setting_to_remove)
71
- expect(modified_conf.has_path?(setting_to_remove)).to be false
72
- end
73
- end
74
-
75
- context "example1" do
76
- let(:example) { EXAMPLE1 }
77
- let(:setting) { "foo.bar.yahoo" }
78
- let(:expected_setting) { "yippee" }
79
- let(:false_setting) { "non-existent" }
80
- let(:setting_to_add) { "foo.bar.test" }
81
- let(:value_to_add) { Hocon::Impl::ConfigString.new(nil, "This is a test string") }
82
- let(:setting_to_remove) { "foo.bar" }
83
-
84
- context "parsing a .conf file" do
85
- let(:conf) { Hocon::ConfigFactory.parse_file(input_file) }
86
- include_examples "config_value_retrieval_single_value"
87
- include_examples "has_path_check"
88
- include_examples "add_value_to_config"
89
- include_examples "add_data_structures_to_config"
90
- include_examples "remove_value_from_config"
91
- end
92
- end
93
-
94
- context "example2" do
95
- let(:example) { EXAMPLE2 }
96
- let(:setting) { "jruby-puppet.jruby-pools" }
97
- let(:expected_setting) { "[{environment=production}]" }
98
- let(:false_setting) { "jruby-puppet-false" }
99
- let(:setting_to_add) { "top" }
100
- let(:value_to_add) { Hocon::Impl::ConfigInt.new(nil, 12345, "12345") }
101
- let(:setting_to_remove) { "jruby-puppet.master-conf-dir" }
102
-
103
- context "parsing a .conf file" do
104
- let(:conf) { Hocon::ConfigFactory.parse_file(input_file) }
105
- include_examples "config_value_retrieval_config_list"
106
- include_examples "has_path_check"
107
- include_examples "add_value_to_config"
108
- include_examples "add_data_structures_to_config"
109
- include_examples "remove_value_from_config"
110
- end
111
- end
112
- end
@@ -1,188 +0,0 @@
1
- # encoding: utf-8
2
-
3
- require 'spec_helper'
4
- require 'hocon'
5
- require 'test_utils'
6
- require 'pp'
7
-
8
-
9
- describe Hocon::Impl::Token do
10
- Tokens = Hocon::Impl::Tokens
11
-
12
- ####################
13
- # Equality
14
- ####################
15
- context "check token equality" do
16
- context "syntax tokens" do
17
- let(:first_object) { Tokens::START }
18
- let(:second_object) { Tokens::START }
19
-
20
- include_examples "object_equality"
21
- end
22
-
23
- context "integer tokens" do
24
- let(:first_object) { TestUtils.token_int(42) }
25
- let(:second_object) { TestUtils.token_int(42) }
26
-
27
- include_examples "object_equality"
28
- end
29
-
30
- context "truth tokens" do
31
- let(:first_object) { TestUtils.token_true }
32
- let(:second_object) { TestUtils.token_true }
33
-
34
- include_examples "object_equality"
35
- end
36
-
37
- context "int and double of the same value" do
38
- let(:first_object) { TestUtils.token_int(10) }
39
- let(:second_object) { TestUtils.token_double(10.0) }
40
-
41
- include_examples "object_equality"
42
- end
43
-
44
- context "double tokens" do
45
- let(:first_object) { TestUtils.token_int(3.14) }
46
- let(:second_object) { TestUtils.token_int(3.14) }
47
-
48
- include_examples "object_equality"
49
- end
50
-
51
- context "quoted string tokens" do
52
- let(:first_object) { TestUtils.token_string("foo") }
53
- let(:second_object) { TestUtils.token_string("foo") }
54
-
55
- include_examples "object_equality"
56
- end
57
-
58
- context "unquoted string tokens" do
59
- let(:first_object) { TestUtils.token_unquoted("foo") }
60
- let(:second_object) { TestUtils.token_unquoted("foo") }
61
-
62
- include_examples "object_equality"
63
- end
64
-
65
- context "key substitution tokens" do
66
- let(:first_object) { TestUtils.token_key_substitution("foo") }
67
- let(:second_object) { TestUtils.token_key_substitution("foo") }
68
-
69
- include_examples "object_equality"
70
- end
71
-
72
- context "null tokens" do
73
- let(:first_object) { TestUtils.token_null }
74
- let(:second_object) { TestUtils.token_null }
75
-
76
- include_examples "object_equality"
77
- end
78
-
79
- context "newline tokens" do
80
- let(:first_object) { TestUtils.token_line(10) }
81
- let(:second_object) { TestUtils.token_line(10) }
82
-
83
- include_examples "object_equality"
84
- end
85
- end
86
-
87
-
88
- ####################
89
- # Inequality
90
- ####################
91
- context "check token inequality" do
92
- context "syntax tokens" do
93
- let(:first_object) { Tokens::START }
94
- let(:second_object) { Tokens::OPEN_CURLY }
95
-
96
- include_examples "object_inequality"
97
- end
98
-
99
- context "integer tokens" do
100
- let(:first_object) { TestUtils.token_int(42) }
101
- let(:second_object) { TestUtils.token_int(43) }
102
-
103
- include_examples "object_inequality"
104
- end
105
-
106
- context "double tokens" do
107
- let(:first_object) { TestUtils.token_int(3.14) }
108
- let(:second_object) { TestUtils.token_int(4.14) }
109
-
110
- include_examples "object_inequality"
111
- end
112
-
113
- context "truth tokens" do
114
- let(:first_object) { TestUtils.token_true }
115
- let(:second_object) { TestUtils.token_false }
116
-
117
- include_examples "object_inequality"
118
- end
119
-
120
- context "quoted string tokens" do
121
- let(:first_object) { TestUtils.token_string("foo") }
122
- let(:second_object) { TestUtils.token_string("bar") }
123
-
124
- include_examples "object_inequality"
125
- end
126
-
127
- context "unquoted string tokens" do
128
- let(:first_object) { TestUtils.token_unquoted("foo") }
129
- let(:second_object) { TestUtils.token_unquoted("bar") }
130
-
131
- include_examples "object_inequality"
132
- end
133
-
134
- context "key substitution tokens" do
135
- let(:first_object) { TestUtils.token_key_substitution("foo") }
136
- let(:second_object) { TestUtils.token_key_substitution("bar") }
137
-
138
- include_examples "object_inequality"
139
- end
140
-
141
- context "newline tokens" do
142
- let(:first_object) { TestUtils.token_line(10) }
143
- let(:second_object) { TestUtils.token_line(11) }
144
-
145
- include_examples "object_inequality"
146
- end
147
-
148
- context "true and int tokens" do
149
- let(:first_object) { TestUtils.token_true }
150
- let(:second_object) { TestUtils.token_int(1) }
151
-
152
- include_examples "object_inequality"
153
- end
154
-
155
- context "string 'true' and true tokens" do
156
- let(:first_object) { TestUtils.token_true }
157
- let(:second_object) { TestUtils.token_string("true") }
158
-
159
- include_examples "object_inequality"
160
- end
161
-
162
- context "int and double of slightly different values" do
163
- let(:first_object) { TestUtils.token_int(10) }
164
- let(:second_object) { TestUtils.token_double(10.000001) }
165
-
166
- include_examples "object_inequality"
167
- end
168
- end
169
-
170
- context "Check that to_s doesn't throw exception" do
171
- it "shouldn't throw an exception" do
172
- # just be sure to_s doesn't throw an exception. It's for debugging
173
- # so its exact output doesn't matter a lot
174
- TestUtils.token_true.to_s
175
- TestUtils.token_false.to_s
176
- TestUtils.token_int(42).to_s
177
- TestUtils.token_double(3.14).to_s
178
- TestUtils.token_null.to_s
179
- TestUtils.token_unquoted("foo").to_s
180
- TestUtils.token_string("bar").to_s
181
- TestUtils.token_key_substitution("a").to_s
182
- TestUtils.token_line(10).to_s
183
- Tokens::START.to_s
184
- Tokens::EOF.to_s
185
- Tokens::COLON.to_s
186
- end
187
- end
188
- end
@@ -1,801 +0,0 @@
1
- # encoding: utf-8
2
-
3
- require 'spec_helper'
4
- require 'hocon'
5
- require 'test_utils'
6
- require 'pp'
7
-
8
-
9
- describe Hocon::Impl::Tokenizer do
10
- Tokens = Hocon::Impl::Tokens
11
-
12
- shared_examples_for "token_matching" do
13
- it "should match the tokenized string to the list of expected tokens" do
14
- tokenized_from_string = TestUtils.tokenize_as_list(test_string)
15
- tokenized_as_string = TestUtils.tokenize_as_string(test_string)
16
-
17
- # Add START and EOF tokens
18
- wrapped_tokens = TestUtils.wrap_tokens(expected_tokens)
19
-
20
- # Compare the two lists of tokens
21
- expect(tokenized_from_string).to eq(wrapped_tokens)
22
- expect(tokenized_as_string).to eq(test_string)
23
- end
24
- end
25
-
26
- shared_examples_for "strings_with_problems" do
27
- it "should find a problem when tokenizing" do
28
- token_list = TestUtils.tokenize_as_list(test_string)
29
- expect(token_list.map { |token| Tokens.problem?(token) }).to include(true)
30
- end
31
- end
32
-
33
- ####################
34
- # Whitespace
35
- ####################
36
- context "tokenizing whitespace" do
37
- context "tokenize empty string" do
38
- let(:test_string) { "" }
39
- let(:expected_tokens) { [] }
40
-
41
- include_examples "token_matching"
42
- end
43
-
44
- context "tokenize newlines" do
45
- let(:test_string) { "\n\n" }
46
- let(:expected_tokens) { [TestUtils.token_line(1),
47
- TestUtils.token_line(2)] }
48
-
49
- include_examples "token_matching"
50
- end
51
-
52
- context "tokenize unquoted text should keep spaces" do
53
- let(:test_string) { " foo \n" }
54
- let(:expected_tokens) { [TestUtils.token_whitespace(" "),
55
- TestUtils.token_unquoted("foo"),
56
- TestUtils.token_whitespace(" "),
57
- TestUtils.token_line(1)] }
58
-
59
- include_examples "token_matching"
60
- end
61
-
62
- context "tokenize unquoted text with internal spaces should keep spaces" do
63
- let(:test_string) { " foo bar baz \n" }
64
- let(:expected_tokens) { [TestUtils.token_whitespace(" "),
65
- TestUtils.token_unquoted("foo"),
66
- TestUtils.token_unquoted(" "),
67
- TestUtils.token_unquoted("bar"),
68
- TestUtils.token_unquoted(" "),
69
- TestUtils.token_unquoted("baz"),
70
- TestUtils.token_whitespace(" "),
71
- TestUtils.token_line(1)] }
72
-
73
- include_examples "token_matching"
74
- end
75
- end
76
-
77
- ####################
78
- # Booleans and Null
79
- ####################
80
- context "tokenizing booleans and null" do
81
- context "tokenize true and unquoted text" do
82
- let(:test_string) { "truefoo" }
83
- let(:expected_tokens) { [TestUtils.token_true,
84
- TestUtils.token_unquoted("foo")] }
85
-
86
- include_examples "token_matching"
87
- end
88
-
89
- context "tokenize false and unquoted text" do
90
- let(:test_string) { "falsefoo" }
91
- let(:expected_tokens) { [TestUtils.token_false,
92
- TestUtils.token_unquoted("foo")] }
93
-
94
- include_examples "token_matching"
95
- end
96
-
97
- context "tokenize null and unquoted text" do
98
- let(:test_string) { "nullfoo" }
99
- let(:expected_tokens) { [TestUtils.token_null,
100
- TestUtils.token_unquoted("foo")] }
101
-
102
- include_examples "token_matching"
103
- end
104
-
105
- context "tokenize unquoted text containing true" do
106
- let(:test_string) { "footrue" }
107
- let(:expected_tokens) { [TestUtils.token_unquoted("footrue")] }
108
-
109
- include_examples "token_matching"
110
- end
111
-
112
- context "tokenize unquoted text containing space and true" do
113
- let(:test_string) { "foo true" }
114
- let(:expected_tokens) { [TestUtils.token_unquoted("foo"),
115
- TestUtils.token_unquoted(" "),
116
- TestUtils.token_true] }
117
-
118
- include_examples "token_matching"
119
- end
120
-
121
- context "tokenize true and space and unquoted text" do
122
- let(:test_string) { "true foo" }
123
- let(:expected_tokens) { [TestUtils.token_true,
124
- TestUtils.token_unquoted(" "),
125
- TestUtils.token_unquoted("foo")] }
126
-
127
- include_examples "token_matching"
128
- end
129
- end
130
-
131
- ####################
132
- # Slashes
133
- ####################
134
- context "tokenizing slashes" do
135
- context "tokenize unquoted text containing slash" do
136
- let(:test_string) { "a/b/c/" }
137
- let(:expected_tokens) { [TestUtils.token_unquoted("a/b/c/")] }
138
-
139
- include_examples "token_matching"
140
- end
141
-
142
- context "tokenize slash" do
143
- let(:test_string) { "/" }
144
- let(:expected_tokens) { [TestUtils.token_unquoted("/")] }
145
-
146
- include_examples "token_matching"
147
- end
148
-
149
- context "tokenize slash space slash" do
150
- let(:test_string) { "/ /" }
151
- let(:expected_tokens) { [TestUtils.token_unquoted("/"),
152
- TestUtils.token_unquoted(" "),
153
- TestUtils.token_unquoted("/")] }
154
-
155
- include_examples "token_matching"
156
- end
157
-
158
- ####################
159
- # Quotes
160
- ####################
161
- context "tokenize mixed unquoted and quoted" do
162
- let(:test_string) { " foo\"bar\"baz \n" }
163
- let(:expected_tokens) { [TestUtils.token_whitespace(" "),
164
- TestUtils.token_unquoted("foo"),
165
- TestUtils.token_string("bar"),
166
- TestUtils.token_unquoted("baz"),
167
- TestUtils.token_whitespace(" "),
168
- TestUtils.token_line(1)] }
169
-
170
- include_examples "token_matching"
171
- end
172
-
173
- context "tokenize empty triple quoted string" do
174
- let(:test_string) { '""""""' }
175
- let(:expected_tokens) { [TestUtils.token_string("")] }
176
-
177
- include_examples "token_matching"
178
- end
179
-
180
- context "tokenize trivial triple quoted string" do
181
- let(:test_string) { '"""bar"""' }
182
- let(:expected_tokens) { [TestUtils.token_string("bar")] }
183
-
184
- include_examples "token_matching"
185
- end
186
-
187
- context "tokenize no escapes in triple quoted string" do
188
- let(:test_string) { '"""\n"""' }
189
- let(:expected_tokens) { [TestUtils.token_string('\n')] }
190
-
191
- include_examples "token_matching"
192
- end
193
-
194
- context "tokenize trailing quotes in triple quoted string" do
195
- let(:test_string) { '"""""""""' }
196
- let(:expected_tokens) { [TestUtils.token_string('"""')] }
197
-
198
- include_examples "token_matching"
199
- end
200
-
201
- context "tokenize new line in triple quoted string" do
202
- let(:test_string) { '"""foo\nbar"""' }
203
- let(:expected_tokens) { [TestUtils.token_string('foo\nbar')] }
204
-
205
- include_examples "token_matching"
206
- end
207
- end
208
-
209
- ####################
210
- # Find problems when tokenizing
211
- ####################
212
- context "finding problems when tokenizing" do
213
- context "nothing after backslash" do
214
- let(:test_string) { ' "\" ' }
215
- include_examples "strings_with_problems"
216
- end
217
-
218
- context "there is no \q escape sequence" do
219
- let(:test_string) { ' "\q" ' }
220
- include_examples "strings_with_problems"
221
- end
222
-
223
- context "unicode byte sequence missing a byte" do
224
- let(:test_string) { '"\u123"' }
225
- include_examples "strings_with_problems"
226
- end
227
-
228
- context "unicode byte sequence missing two bytes" do
229
- let(:test_string) { '"\u12"' }
230
- include_examples "strings_with_problems"
231
- end
232
-
233
- context "unicode byte sequence missing three bytes" do
234
- let(:test_string) { '"\u1"' }
235
- include_examples "strings_with_problems"
236
- end
237
-
238
- context "unicode byte missing" do
239
- let(:test_string) { '"\u"' }
240
- include_examples "strings_with_problems"
241
- end
242
-
243
- context "just a single quote" do
244
- let(:test_string) { '"' }
245
- include_examples "strings_with_problems"
246
- end
247
-
248
- context "no end quote" do
249
- let(:test_string) { ' "abcdefg' }
250
- include_examples "strings_with_problems"
251
- end
252
-
253
- context "file ends with a backslash" do
254
- let(:test_string) { '\"\\' }
255
- include_examples "strings_with_problems"
256
- end
257
-
258
- context "file ends with a $" do
259
- let(:test_string) { "$" }
260
- include_examples "strings_with_problems"
261
- end
262
-
263
- context "file ends with a ${" do
264
- let(:test_string) { "${" }
265
- include_examples "strings_with_problems"
266
- end
267
- end
268
-
269
- ####################
270
- # Numbers
271
- ####################
272
- context "tokenizing numbers" do
273
- context "parse positive float" do
274
- let(:test_string) { "1.2" }
275
- let(:expected_tokens) { [TestUtils.token_double(1.2)] }
276
- include_examples "token_matching"
277
- end
278
-
279
- context "parse negative float" do
280
- let(:test_string) { "-1.2" }
281
- let(:expected_tokens) { [TestUtils.token_double(-1.2)] }
282
- include_examples "token_matching"
283
- end
284
-
285
- context "parse exponent notation" do
286
- let(:test_string) { "1e6" }
287
- let(:expected_tokens) { [TestUtils.token_double(1e6)] }
288
- include_examples "token_matching"
289
- end
290
-
291
- context "parse negative exponent" do
292
- let(:test_string) { "1e-6" }
293
- let(:expected_tokens) { [TestUtils.token_double(1e-6)] }
294
- include_examples "token_matching"
295
- end
296
-
297
- context "parse exponent with capital E" do
298
- let(:test_string) { "1E-6" }
299
- let(:expected_tokens) { [TestUtils.token_double(1e-6)] }
300
- include_examples "token_matching"
301
- end
302
-
303
- context "parse negative int" do
304
- let(:test_string) { "-1" }
305
- let(:expected_tokens) { [TestUtils.token_int(-1)] }
306
- include_examples "token_matching"
307
- end
308
- end
309
-
310
- ####################
311
- # Comments
312
- ####################
313
- context "tokenizing comments" do
314
- context "tokenize two slashes as comment" do
315
- let(:test_string) { "//" }
316
- let(:expected_tokens) { [TestUtils.token_comment_double_slash("")] }
317
-
318
- include_examples "token_matching"
319
- end
320
-
321
- context "tokenize two slashes in string as string" do
322
- let(:test_string) { '"//bar"' }
323
- let(:expected_tokens) { [TestUtils.token_string("//bar")] }
324
-
325
- include_examples "token_matching"
326
- end
327
-
328
- context "tokenize hash in string as string" do
329
- let(:test_string) { '"#bar"' }
330
- let(:expected_tokens) { [TestUtils.token_string("#bar")] }
331
-
332
- include_examples "token_matching"
333
- end
334
-
335
- context "tokenize slash comment after unquoted text" do
336
- let(:test_string) { "bar//comment" }
337
- let(:expected_tokens) { [TestUtils.token_unquoted("bar"),
338
- TestUtils.token_comment_double_slash("comment")] }
339
-
340
- include_examples "token_matching"
341
- end
342
-
343
- context "tokenize hash comment after unquoted text" do
344
- let(:test_string) { "bar#comment" }
345
- let(:expected_tokens) { [TestUtils.token_unquoted("bar"),
346
- TestUtils.token_comment_hash("comment")] }
347
-
348
- include_examples "token_matching"
349
- end
350
-
351
- context "tokenize slash comment after int" do
352
- let(:test_string) { "10//comment" }
353
- let(:expected_tokens) { [TestUtils.token_int(10),
354
- TestUtils.token_comment_double_slash("comment")] }
355
-
356
- include_examples "token_matching"
357
- end
358
-
359
- context "tokenize hash comment after int" do
360
- let(:test_string) { "10#comment" }
361
- let(:expected_tokens) { [TestUtils.token_int(10),
362
- TestUtils.token_comment_hash("comment")] }
363
-
364
- include_examples "token_matching"
365
- end
366
-
367
- context "tokenize hash comment after int" do
368
- let(:test_string) { "10#comment" }
369
- let(:expected_tokens) { [TestUtils.token_int(10),
370
- TestUtils.token_comment_hash("comment")] }
371
-
372
- include_examples "token_matching"
373
- end
374
-
375
- context "tokenize slash comment after float" do
376
- let(:test_string) { "3.14//comment" }
377
- let(:expected_tokens) { [TestUtils.token_double(3.14),
378
- TestUtils.token_comment_double_slash("comment")] }
379
-
380
- include_examples "token_matching"
381
- end
382
-
383
- context "tokenize hash comment after float" do
384
- let(:test_string) { "3.14#comment" }
385
- let(:expected_tokens) { [TestUtils.token_double(3.14),
386
- TestUtils.token_comment_hash("comment")] }
387
-
388
- include_examples "token_matching"
389
- end
390
-
391
- context "tokenize slash comment with newline" do
392
- let(:test_string) { "10//comment\n12" }
393
- let(:expected_tokens) { [TestUtils.token_int(10),
394
- TestUtils.token_comment_double_slash("comment"),
395
- TestUtils.token_line(1),
396
- TestUtils.token_int(12)] }
397
-
398
- include_examples "token_matching"
399
- end
400
-
401
- context "tokenize hash comment with newline" do
402
- let(:test_string) { "10#comment\n12" }
403
- let(:expected_tokens) { [TestUtils.token_int(10),
404
- TestUtils.token_comment_hash("comment"),
405
- TestUtils.token_line(1),
406
- TestUtils.token_int(12)] }
407
-
408
- include_examples "token_matching"
409
- end
410
-
411
- context "tokenize slash comments on two consecutive lines" do
412
- let(:test_string) { "//comment\n//comment2" }
413
- let(:expected_tokens) { [TestUtils.token_comment_double_slash("comment"),
414
- TestUtils.token_line(1),
415
- TestUtils.token_comment_double_slash("comment2")] }
416
-
417
- include_examples "token_matching"
418
- end
419
-
420
- context "tokenize hash comments on two consecutive lines" do
421
- let(:test_string) { "#comment\n#comment2" }
422
- let(:expected_tokens) { [TestUtils.token_comment_hash("comment"),
423
- TestUtils.token_line(1),
424
- TestUtils.token_comment_hash("comment2")] }
425
- include_examples "token_matching"
426
- end
427
-
428
- context "tokenize slash comments on multiple lines with whitespace" do
429
- let(:test_string) { " //comment\r\n //comment2 \n//comment3 \n\n//comment4" }
430
- let(:expected_tokens) { [TestUtils.token_whitespace(" "),
431
- TestUtils.token_comment_double_slash("comment\r"),
432
- TestUtils.token_line(1),
433
- TestUtils.token_whitespace(" "),
434
- TestUtils.token_comment_double_slash("comment2 "),
435
- TestUtils.token_line(2),
436
- TestUtils.token_comment_double_slash("comment3 "),
437
- TestUtils.token_line(3),
438
- TestUtils.token_line(4),
439
- TestUtils.token_comment_double_slash("comment4")] }
440
-
441
- include_examples "token_matching"
442
- end
443
-
444
- context "tokenize hash comments on multiple lines with whitespace" do
445
- let(:test_string) { " #comment\r\n #comment2 \n#comment3 \n\n#comment4" }
446
- let(:expected_tokens) { [TestUtils.token_whitespace(" "),
447
- TestUtils.token_comment_hash("comment\r"),
448
- TestUtils.token_line(1),
449
- TestUtils.token_whitespace(" "),
450
- TestUtils.token_comment_hash("comment2 "),
451
- TestUtils.token_line(2),
452
- TestUtils.token_comment_hash("comment3 "),
453
- TestUtils.token_line(3),
454
- TestUtils.token_line(4),
455
- TestUtils.token_comment_hash("comment4")] }
456
-
457
- include_examples "token_matching"
458
- end
459
- end
460
-
461
- ####################
462
- # Brackets, braces
463
- ####################
464
- context "tokenizing brackets and braces" do
465
- context "tokenize open curly braces" do
466
- let(:test_string) { "{{" }
467
- let(:expected_tokens) { [Tokens::OPEN_CURLY, Tokens::OPEN_CURLY] }
468
-
469
- include_examples "token_matching"
470
- end
471
-
472
- context "tokenize close curly braces" do
473
- let(:test_string) { "}}" }
474
- let(:expected_tokens) { [Tokens::CLOSE_CURLY, Tokens::CLOSE_CURLY] }
475
-
476
- include_examples "token_matching"
477
- end
478
-
479
- context "tokenize open and close curly braces" do
480
- let(:test_string) { "{}" }
481
- let(:expected_tokens) { [Tokens::OPEN_CURLY, Tokens::CLOSE_CURLY] }
482
-
483
- include_examples "token_matching"
484
- end
485
-
486
- context "tokenize open and close curly braces" do
487
- let(:test_string) { "{}" }
488
- let(:expected_tokens) { [Tokens::OPEN_CURLY, Tokens::CLOSE_CURLY] }
489
-
490
- include_examples "token_matching"
491
- end
492
-
493
- context "tokenize open square brackets" do
494
- let(:test_string) { "[[" }
495
- let(:expected_tokens) { [Tokens::OPEN_SQUARE, Tokens::OPEN_SQUARE] }
496
-
497
- include_examples "token_matching"
498
- end
499
-
500
- context "tokenize close square brackets" do
501
- let(:test_string) { "]]" }
502
- let(:expected_tokens) { [Tokens::CLOSE_SQUARE, Tokens::CLOSE_SQUARE] }
503
-
504
- include_examples "token_matching"
505
- end
506
-
507
- context "tokenize open and close square brackets" do
508
- let(:test_string) { "[]" }
509
- let(:expected_tokens) { [Tokens::OPEN_SQUARE, Tokens::CLOSE_SQUARE] }
510
-
511
- include_examples "token_matching"
512
- end
513
- end
514
-
515
- ####################
516
- # comma, colon, equals, plus equals
517
- ####################
518
- context "tokenizing comma, colon, equals, and plus equals" do
519
- context "tokenize comma" do
520
- let(:test_string) { "," }
521
- let(:expected_tokens) { [Tokens::COMMA] }
522
-
523
- include_examples "token_matching"
524
- end
525
-
526
- context "tokenize colon" do
527
- let(:test_string) { ":" }
528
- let(:expected_tokens) { [Tokens::COLON] }
529
-
530
- include_examples "token_matching"
531
- end
532
-
533
- context "tokenize equals" do
534
- let(:test_string) { "=" }
535
- let(:expected_tokens) { [Tokens::EQUALS] }
536
-
537
- include_examples "token_matching"
538
- end
539
-
540
- context "tokenize plus equals" do
541
- let(:test_string) { "+=" }
542
- let(:expected_tokens) { [Tokens::PLUS_EQUALS] }
543
-
544
- include_examples "token_matching"
545
- end
546
-
547
- context "tokenize comma, colon, plus equals, and equals together" do
548
- let(:test_string) { "=:,+=" }
549
- let(:expected_tokens) { [Tokens::EQUALS,
550
- Tokens::COLON,
551
- Tokens::COMMA,
552
- Tokens::PLUS_EQUALS] }
553
-
554
- include_examples "token_matching"
555
- end
556
- end
557
-
558
- ####################
559
- # Substitutions
560
- ####################
561
- context "tokenizing substitutions" do
562
- context "tokenize substitution" do
563
- let(:test_string) { "${a.b}" }
564
- let(:expected_tokens) { [TestUtils.token_substitution(TestUtils.token_unquoted("a.b"))] }
565
-
566
- include_examples "token_matching"
567
- end
568
-
569
- context "tokenize optional substitution" do
570
- let(:test_string) { "${?x.y}" }
571
- let(:expected_tokens) { [TestUtils.token_optional_substitution(TestUtils.token_unquoted("x.y"))] }
572
-
573
- include_examples "token_matching"
574
- end
575
-
576
- context "tokenize key substitution" do
577
- let(:test_string) { '${"c.d"}' }
578
- let(:expected_tokens) { [TestUtils.token_key_substitution("c.d")] }
579
-
580
- include_examples "token_matching"
581
- end
582
- end
583
-
584
- ####################
585
- # Unicode and escape characters
586
- ####################
587
- context "tokenizing unicode and escape characters" do
588
- context "tokenize unicode infinity symbol" do
589
- let(:test_string) { '"\u221E"' }
590
- let(:expected_tokens) { [TestUtils.token_string("\u{221E}")] }
591
-
592
- include_examples "token_matching"
593
- end
594
-
595
- context "tokenize null byte" do
596
- let(:test_string) { ' "\u0000" ' }
597
- let(:expected_tokens) { [TestUtils.token_whitespace(" "),
598
- TestUtils.token_string("\u0000"),
599
- TestUtils.token_whitespace(" ")] }
600
-
601
- include_examples "token_matching"
602
- end
603
-
604
- context "tokenize various espace codes" do
605
- let(:test_string) { ' "\"\\\/\b\f\n\r\t" ' }
606
- let(:expected_tokens) { [TestUtils.token_whitespace(" "),
607
- TestUtils.token_string("\"\\/\b\f\n\r\t"),
608
- TestUtils.token_whitespace(" ")] }
609
-
610
- include_examples "token_matching"
611
- end
612
-
613
- context "tokenize unicode F" do
614
- let(:test_string) { ' "\u0046" ' }
615
- let(:expected_tokens) { [TestUtils.token_whitespace(" "),
616
- TestUtils.token_string("F"),
617
- TestUtils.token_whitespace(" ")] }
618
-
619
- include_examples "token_matching"
620
- end
621
-
622
- context "tokenize two unicode Fs" do
623
- let(:test_string) { ' "\u0046\u0046" ' }
624
- let(:expected_tokens) { [TestUtils.token_whitespace(" "),
625
- TestUtils.token_string("FF"),
626
- TestUtils.token_whitespace(" ")] }
627
-
628
- include_examples "token_matching"
629
- end
630
- end
631
-
632
- ####################
633
- # Reserved Characters
634
- ####################
635
- context "Finding problems with using reserved characters" do
636
- context "problem with reserved character +" do
637
- let(:test_string) { "+" }
638
- include_examples "strings_with_problems"
639
- end
640
-
641
- context "problem with reserved character `" do
642
- let(:test_string) { "`" }
643
- include_examples "strings_with_problems"
644
- end
645
-
646
- context "problem with reserved character ^" do
647
- let(:test_string) { "^" }
648
- include_examples "strings_with_problems"
649
- end
650
-
651
- context "problem with reserved character ?" do
652
- let(:test_string) { "?" }
653
- include_examples "strings_with_problems"
654
- end
655
-
656
- context "problem with reserved character !" do
657
- let(:test_string) { "!" }
658
- include_examples "strings_with_problems"
659
- end
660
-
661
- context "problem with reserved character @" do
662
- let(:test_string) { "@" }
663
- include_examples "strings_with_problems"
664
- end
665
-
666
- context "problem with reserved character *" do
667
- let(:test_string) { "*" }
668
- include_examples "strings_with_problems"
669
- end
670
-
671
- context "problem with reserved character &" do
672
- let(:test_string) { "&" }
673
- include_examples "strings_with_problems"
674
- end
675
-
676
- context "problem with reserved character \\" do
677
- let(:test_string) { "\\" }
678
- include_examples "strings_with_problems"
679
- end
680
- end
681
-
682
- ####################
683
- # Combine all types
684
- ####################
685
- context "Tokenizing all types together" do
686
- context "tokenize all types no spaces" do
687
- let(:test_string) { ',:=}{][+="foo""""bar"""true3.14false42null${a.b}${?x.y}${"c.d"}' + "\n" }
688
- let(:expected_tokens) { [Tokens::COMMA,
689
- Tokens::COLON,
690
- Tokens::EQUALS,
691
- Tokens::CLOSE_CURLY,
692
- Tokens::OPEN_CURLY,
693
- Tokens::CLOSE_SQUARE,
694
- Tokens::OPEN_SQUARE,
695
- Tokens::PLUS_EQUALS,
696
- TestUtils.token_string("foo"),
697
- TestUtils.token_string("bar"),
698
- TestUtils.token_true,
699
- TestUtils.token_double(3.14),
700
- TestUtils.token_false,
701
- TestUtils.token_int(42),
702
- TestUtils.token_null,
703
- TestUtils.token_substitution(TestUtils.token_unquoted("a.b")),
704
- TestUtils.token_optional_substitution(TestUtils.token_unquoted("x.y")),
705
- TestUtils.token_key_substitution("c.d"),
706
- TestUtils.token_line(1)] }
707
-
708
- include_examples "token_matching"
709
- end
710
-
711
- context "tokenize all types single spaces" do
712
- let(:test_string) { ' , : = } { ] [ += "foo" """bar""" 42 true 3.14 false null ${a.b} ${?x.y} ${"c.d"} ' + "\n " }
713
- let(:expected_tokens) { [TestUtils.token_whitespace(" "),
714
- Tokens::COMMA,
715
- TestUtils.token_whitespace(" "),
716
- Tokens::COLON,
717
- TestUtils.token_whitespace(" "),
718
- Tokens::EQUALS,
719
- TestUtils.token_whitespace(" "),
720
- Tokens::CLOSE_CURLY,
721
- TestUtils.token_whitespace(" "),
722
- Tokens::OPEN_CURLY,
723
- TestUtils.token_whitespace(" "),
724
- Tokens::CLOSE_SQUARE,
725
- TestUtils.token_whitespace(" "),
726
- Tokens::OPEN_SQUARE,
727
- TestUtils.token_whitespace(" "),
728
- Tokens::PLUS_EQUALS,
729
- TestUtils.token_whitespace(" "),
730
- TestUtils.token_string("foo"),
731
- TestUtils.token_unquoted(" "),
732
- TestUtils.token_string("bar"),
733
- TestUtils.token_unquoted(" "),
734
- TestUtils.token_int(42),
735
- TestUtils.token_unquoted(" "),
736
- TestUtils.token_true,
737
- TestUtils.token_unquoted(" "),
738
- TestUtils.token_double(3.14),
739
- TestUtils.token_unquoted(" "),
740
- TestUtils.token_false,
741
- TestUtils.token_unquoted(" "),
742
- TestUtils.token_null,
743
- TestUtils.token_unquoted(" "),
744
- TestUtils.token_substitution(TestUtils.token_unquoted("a.b")),
745
- TestUtils.token_unquoted(" "),
746
- TestUtils.token_optional_substitution(TestUtils.token_unquoted("x.y")),
747
- TestUtils.token_unquoted(" "),
748
- TestUtils.token_key_substitution("c.d"),
749
- TestUtils.token_whitespace(" "),
750
- TestUtils.token_line(1),
751
- TestUtils.token_whitespace(" ")] }
752
-
753
- include_examples "token_matching"
754
- end
755
-
756
- context "tokenize all types multiple spaces" do
757
- let(:test_string) { ' , : = } { ] [ += "foo" """bar""" 42 true 3.14 false null ${a.b} ${?x.y} ${"c.d"} ' + "\n " }
758
- let(:expected_tokens) { [TestUtils.token_whitespace(" "),
759
- Tokens::COMMA,
760
- TestUtils.token_whitespace(" "),
761
- Tokens::COLON,
762
- TestUtils.token_whitespace(" "),
763
- Tokens::EQUALS,
764
- TestUtils.token_whitespace(" "),
765
- Tokens::CLOSE_CURLY,
766
- TestUtils.token_whitespace(" "),
767
- Tokens::OPEN_CURLY,
768
- TestUtils.token_whitespace(" "),
769
- Tokens::CLOSE_SQUARE,
770
- TestUtils.token_whitespace(" "),
771
- Tokens::OPEN_SQUARE,
772
- TestUtils.token_whitespace(" "),
773
- Tokens::PLUS_EQUALS,
774
- TestUtils.token_whitespace(" "),
775
- TestUtils.token_string("foo"),
776
- TestUtils.token_unquoted(" "),
777
- TestUtils.token_string("bar"),
778
- TestUtils.token_unquoted(" "),
779
- TestUtils.token_int(42),
780
- TestUtils.token_unquoted(" "),
781
- TestUtils.token_true,
782
- TestUtils.token_unquoted(" "),
783
- TestUtils.token_double(3.14),
784
- TestUtils.token_unquoted(" "),
785
- TestUtils.token_false,
786
- TestUtils.token_unquoted(" "),
787
- TestUtils.token_null,
788
- TestUtils.token_unquoted(" "),
789
- TestUtils.token_substitution(TestUtils.token_unquoted("a.b")),
790
- TestUtils.token_unquoted(" "),
791
- TestUtils.token_optional_substitution(TestUtils.token_unquoted("x.y")),
792
- TestUtils.token_unquoted(" "),
793
- TestUtils.token_key_substitution("c.d"),
794
- TestUtils.token_whitespace(" "),
795
- TestUtils.token_line(1),
796
- TestUtils.token_whitespace(" ")] }
797
-
798
- include_examples "token_matching"
799
- end
800
- end
801
- end