gherkin 1.0.2-java
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- data/.gitignore +7 -0
- data/.mailmap +2 -0
- data/History.txt +9 -0
- data/LICENSE +20 -0
- data/README.rdoc +38 -0
- data/Rakefile +48 -0
- data/VERSION.yml +4 -0
- data/bin/gherkin +5 -0
- data/cucumber.yml +3 -0
- data/features/feature_parser.feature +206 -0
- data/features/native_lexer.feature +19 -0
- data/features/parser_with_native_lexer.feature +205 -0
- data/features/pretty_printer.feature +14 -0
- data/features/step_definitions/gherkin_steps.rb +34 -0
- data/features/step_definitions/pretty_printer_steps.rb +56 -0
- data/features/steps_parser.feature +46 -0
- data/features/support/env.rb +33 -0
- data/gherkin.gemspec +155 -0
- data/java/.gitignore +2 -0
- data/java/Gherkin.iml +24 -0
- data/java/build.xml +13 -0
- data/java/src/gherkin/FixJava.java +34 -0
- data/java/src/gherkin/Lexer.java +5 -0
- data/java/src/gherkin/LexingError.java +7 -0
- data/java/src/gherkin/Listener.java +27 -0
- data/java/src/gherkin/ParseError.java +22 -0
- data/java/src/gherkin/Parser.java +185 -0
- data/java/src/gherkin/lexer/.gitignore +1 -0
- data/java/src/gherkin/parser/StateMachineReader.java +62 -0
- data/lib/.gitignore +4 -0
- data/lib/gherkin.rb +2 -0
- data/lib/gherkin/c_lexer.rb +10 -0
- data/lib/gherkin/cli/main.rb +34 -0
- data/lib/gherkin/core_ext/array.rb +5 -0
- data/lib/gherkin/i18n.rb +87 -0
- data/lib/gherkin/i18n.yml +535 -0
- data/lib/gherkin/i18n_lexer.rb +29 -0
- data/lib/gherkin/java_lexer.rb +10 -0
- data/lib/gherkin/lexer.rb +44 -0
- data/lib/gherkin/parser.rb +19 -0
- data/lib/gherkin/parser/meta.txt +4 -0
- data/lib/gherkin/parser/root.txt +9 -0
- data/lib/gherkin/parser/steps.txt +3 -0
- data/lib/gherkin/rb_lexer.rb +10 -0
- data/lib/gherkin/rb_lexer/.gitignore +1 -0
- data/lib/gherkin/rb_lexer/README.rdoc +8 -0
- data/lib/gherkin/rb_parser.rb +117 -0
- data/lib/gherkin/tools.rb +8 -0
- data/lib/gherkin/tools/files.rb +30 -0
- data/lib/gherkin/tools/pretty_listener.rb +84 -0
- data/lib/gherkin/tools/reformat.rb +19 -0
- data/lib/gherkin/tools/stats.rb +21 -0
- data/lib/gherkin/tools/stats_listener.rb +50 -0
- data/nativegems.sh +5 -0
- data/ragel/i18n/.gitignore +1 -0
- data/ragel/lexer.c.rl.erb +403 -0
- data/ragel/lexer.java.rl.erb +200 -0
- data/ragel/lexer.rb.rl.erb +171 -0
- data/ragel/lexer_common.rl.erb +46 -0
- data/spec/gherkin/c_lexer_spec.rb +21 -0
- data/spec/gherkin/fixtures/1.feature +8 -0
- data/spec/gherkin/fixtures/complex.feature +43 -0
- data/spec/gherkin/fixtures/i18n_fr.feature +13 -0
- data/spec/gherkin/fixtures/i18n_no.feature +6 -0
- data/spec/gherkin/fixtures/i18n_zh-CN.feature +8 -0
- data/spec/gherkin/fixtures/simple.feature +3 -0
- data/spec/gherkin/fixtures/simple_with_comments.feature +7 -0
- data/spec/gherkin/fixtures/simple_with_tags.feature +11 -0
- data/spec/gherkin/i18n_lexer_spec.rb +22 -0
- data/spec/gherkin/i18n_spec.rb +57 -0
- data/spec/gherkin/java_lexer_spec.rb +20 -0
- data/spec/gherkin/parser_spec.rb +28 -0
- data/spec/gherkin/rb_lexer_spec.rb +18 -0
- data/spec/gherkin/sexp_recorder.rb +29 -0
- data/spec/gherkin/shared/lexer_spec.rb +433 -0
- data/spec/gherkin/shared/py_string_spec.rb +124 -0
- data/spec/gherkin/shared/table_spec.rb +97 -0
- data/spec/gherkin/shared/tags_spec.rb +50 -0
- data/spec/spec_helper.rb +53 -0
- data/tasks/bench.rake +186 -0
- data/tasks/bench/feature_builder.rb +49 -0
- data/tasks/bench/generated/.gitignore +1 -0
- data/tasks/bench/null_listener.rb +4 -0
- data/tasks/compile.rake +70 -0
- data/tasks/cucumber.rake +20 -0
- data/tasks/ragel_task.rb +70 -0
- data/tasks/rdoc.rake +12 -0
- data/tasks/rspec.rake +15 -0
- metadata +196 -0
data/nativegems.sh
ADDED
@@ -0,0 +1 @@
|
|
1
|
+
*.rl
|
@@ -0,0 +1,403 @@
|
|
1
|
+
#include <assert.h>
|
2
|
+
#include <ruby.h>
|
3
|
+
|
4
|
+
#if defined(_WIN32)
|
5
|
+
#include <stddef.h>
|
6
|
+
#endif
|
7
|
+
|
8
|
+
#ifdef HAVE_RUBY_RE_H
|
9
|
+
#include <ruby/re.h>
|
10
|
+
#else
|
11
|
+
#include <re.h>
|
12
|
+
#endif
|
13
|
+
|
14
|
+
#ifdef HAVE_RUBY_ENCODING_H
|
15
|
+
#include <ruby/encoding.h>
|
16
|
+
#define ENCODED_STR_NEW(ptr, len) \
|
17
|
+
rb_enc_str_new(ptr, len, rb_utf8_encoding());
|
18
|
+
#else
|
19
|
+
#define ENCODED_STR_NEW(ptr, len) \
|
20
|
+
rb_str_new(ptr, len);
|
21
|
+
#endif
|
22
|
+
|
23
|
+
#ifndef RSTRING_PTR
|
24
|
+
#define RSTRING_PTR(s) (RSTRING(s)->ptr)
|
25
|
+
#endif
|
26
|
+
|
27
|
+
#ifndef RSTRING_LEN
|
28
|
+
#define RSTRING_LEN(s) (RSTRING(s)->len)
|
29
|
+
#endif
|
30
|
+
|
31
|
+
#define DATA_GET(FROM, TYPE, NAME) \
|
32
|
+
Data_Get_Struct(FROM, TYPE, NAME); \
|
33
|
+
if (NAME == NULL) { \
|
34
|
+
rb_raise(rb_eArgError, "NULL found for " # NAME " when it shouldn't be."); \
|
35
|
+
}
|
36
|
+
|
37
|
+
typedef struct lexer_state {
|
38
|
+
int content_len;
|
39
|
+
int line_number;
|
40
|
+
int current_line;
|
41
|
+
int start_col;
|
42
|
+
size_t mark;
|
43
|
+
size_t keyword_start;
|
44
|
+
size_t keyword_end;
|
45
|
+
size_t next_keyword_start;
|
46
|
+
size_t content_start;
|
47
|
+
size_t content_end;
|
48
|
+
size_t field_len;
|
49
|
+
size_t query_start;
|
50
|
+
size_t last_newline;
|
51
|
+
size_t final_newline;
|
52
|
+
} lexer_state;
|
53
|
+
|
54
|
+
static VALUE mGherkin;
|
55
|
+
static VALUE mLexer;
|
56
|
+
static VALUE mCLexer;
|
57
|
+
static VALUE cI18nLexer;
|
58
|
+
static VALUE rb_eGherkinLexerError;
|
59
|
+
|
60
|
+
#define LEN(AT, P) (P - data - lexer->AT)
|
61
|
+
#define MARK(M, P) (lexer->M = (P) - data)
|
62
|
+
#define PTR_TO(P) (data + lexer->P)
|
63
|
+
|
64
|
+
#define STORE_KW_END_CON(EVENT) \
|
65
|
+
store_kw_con(listener, # EVENT, \
|
66
|
+
PTR_TO(keyword_start), LEN(keyword_start, PTR_TO(keyword_end - 1)), \
|
67
|
+
PTR_TO(content_start), LEN(content_start, PTR_TO(content_end)), \
|
68
|
+
lexer->current_line); \
|
69
|
+
if (lexer->content_end != 0) { \
|
70
|
+
p = PTR_TO(content_end - 1); \
|
71
|
+
} \
|
72
|
+
lexer->content_end = 0;
|
73
|
+
|
74
|
+
#define STORE_ATTR(ATTR) \
|
75
|
+
store_attr(listener, # ATTR, \
|
76
|
+
PTR_TO(content_start), LEN(content_start, p), \
|
77
|
+
lexer->line_number);
|
78
|
+
|
79
|
+
%%{
|
80
|
+
machine lexer;
|
81
|
+
|
82
|
+
action begin_content {
|
83
|
+
MARK(content_start, p);
|
84
|
+
lexer->current_line = lexer->line_number;
|
85
|
+
}
|
86
|
+
|
87
|
+
action begin_pystring_content {
|
88
|
+
MARK(content_start, p);
|
89
|
+
}
|
90
|
+
|
91
|
+
action start_pystring {
|
92
|
+
lexer->current_line = lexer->line_number;
|
93
|
+
lexer->start_col = p - data - lexer->last_newline;
|
94
|
+
}
|
95
|
+
|
96
|
+
action store_pystring_content {
|
97
|
+
int len = LEN(content_start, PTR_TO(final_newline));
|
98
|
+
|
99
|
+
if (len < 0) len = 0;
|
100
|
+
|
101
|
+
store_pystring_content(listener, lexer->start_col, PTR_TO(content_start), len, lexer->current_line);
|
102
|
+
}
|
103
|
+
|
104
|
+
action store_feature_content {
|
105
|
+
STORE_KW_END_CON(feature)
|
106
|
+
}
|
107
|
+
|
108
|
+
action store_background_content {
|
109
|
+
STORE_KW_END_CON(background)
|
110
|
+
}
|
111
|
+
|
112
|
+
action store_scenario_content {
|
113
|
+
STORE_KW_END_CON(scenario)
|
114
|
+
}
|
115
|
+
|
116
|
+
action store_scenario_outline_content {
|
117
|
+
STORE_KW_END_CON(scenario_outline)
|
118
|
+
}
|
119
|
+
|
120
|
+
action store_examples_content {
|
121
|
+
STORE_KW_END_CON(examples)
|
122
|
+
}
|
123
|
+
|
124
|
+
action store_step_content {
|
125
|
+
store_kw_con(listener, "step",
|
126
|
+
PTR_TO(keyword_start), LEN(keyword_start, PTR_TO(keyword_end)),
|
127
|
+
PTR_TO(content_start), LEN(content_start, p),
|
128
|
+
lexer->current_line);
|
129
|
+
}
|
130
|
+
|
131
|
+
action store_comment_content {
|
132
|
+
STORE_ATTR(comment)
|
133
|
+
lexer->mark = 0;
|
134
|
+
}
|
135
|
+
|
136
|
+
action store_tag_content {
|
137
|
+
STORE_ATTR(tag)
|
138
|
+
lexer->mark = 0;
|
139
|
+
}
|
140
|
+
|
141
|
+
action inc_line_number {
|
142
|
+
lexer->line_number += 1;
|
143
|
+
MARK(final_newline, p);
|
144
|
+
}
|
145
|
+
|
146
|
+
action last_newline {
|
147
|
+
MARK(last_newline, p + 1);
|
148
|
+
}
|
149
|
+
|
150
|
+
action start_keyword {
|
151
|
+
if (lexer->mark == 0) {
|
152
|
+
MARK(mark, p);
|
153
|
+
}
|
154
|
+
}
|
155
|
+
|
156
|
+
action end_keyword {
|
157
|
+
MARK(keyword_end, p);
|
158
|
+
MARK(keyword_start, PTR_TO(mark));
|
159
|
+
MARK(content_start, p + 1);
|
160
|
+
lexer->mark = 0;
|
161
|
+
}
|
162
|
+
|
163
|
+
action next_keyword_start {
|
164
|
+
MARK(content_end, p);
|
165
|
+
}
|
166
|
+
|
167
|
+
action start_table {
|
168
|
+
p = p - 1;
|
169
|
+
lexer->current_line = lexer->line_number;
|
170
|
+
rb_ary_clear(rows);
|
171
|
+
rb_ary_clear(current_row);
|
172
|
+
}
|
173
|
+
|
174
|
+
action begin_cell_content {
|
175
|
+
MARK(content_start, p);
|
176
|
+
}
|
177
|
+
|
178
|
+
action store_cell_content {
|
179
|
+
VALUE con = Qnil;
|
180
|
+
con = ENCODED_STR_NEW(PTR_TO(content_start), LEN(content_start, p));
|
181
|
+
rb_funcall(con, rb_intern("strip!"), 0);
|
182
|
+
|
183
|
+
rb_ary_push(current_row, con);
|
184
|
+
}
|
185
|
+
|
186
|
+
action start_row {
|
187
|
+
current_row = rb_ary_new();
|
188
|
+
}
|
189
|
+
|
190
|
+
action store_row {
|
191
|
+
rb_ary_push(rows, current_row);
|
192
|
+
}
|
193
|
+
|
194
|
+
action store_table {
|
195
|
+
rb_funcall(listener, rb_intern("table"), 2, rows, INT2FIX(lexer->current_line));
|
196
|
+
}
|
197
|
+
|
198
|
+
action end_feature {
|
199
|
+
if (cs < lexer_first_final) {
|
200
|
+
if (raise_lexer_error != NULL) {
|
201
|
+
int count = 0;
|
202
|
+
int newstr_count = 0;
|
203
|
+
size_t len;
|
204
|
+
const char *buff;
|
205
|
+
if (lexer->last_newline != 0) {
|
206
|
+
len = LEN(last_newline, eof);
|
207
|
+
buff = PTR_TO(last_newline);
|
208
|
+
} else {
|
209
|
+
len = strlen(data);
|
210
|
+
buff = data;
|
211
|
+
}
|
212
|
+
|
213
|
+
char newstr[len];
|
214
|
+
|
215
|
+
for (count = 0; count < len; count++) {
|
216
|
+
if(buff[count] == 10) {
|
217
|
+
newstr[newstr_count] = '\0'; // terminate new string at first newline found
|
218
|
+
break;
|
219
|
+
} else {
|
220
|
+
if (buff[count] == '%') {
|
221
|
+
newstr[newstr_count++] = buff[count];
|
222
|
+
newstr[newstr_count] = buff[count];
|
223
|
+
} else {
|
224
|
+
newstr[newstr_count] = buff[count];
|
225
|
+
}
|
226
|
+
}
|
227
|
+
newstr_count++;
|
228
|
+
}
|
229
|
+
|
230
|
+
int line = lexer->line_number;
|
231
|
+
lexer_init(lexer); // Re-initialize so we can scan again with the same lexer
|
232
|
+
raise_lexer_error(listener, newstr, line);
|
233
|
+
}
|
234
|
+
}
|
235
|
+
}
|
236
|
+
|
237
|
+
include lexer_common "lexer_common.<%= @i18n.sanitized_key %>.rl";
|
238
|
+
|
239
|
+
}%%
|
240
|
+
|
241
|
+
/** Data **/
|
242
|
+
%% write data;
|
243
|
+
|
244
|
+
static VALUE
|
245
|
+
strip_i(VALUE str, VALUE ary)
|
246
|
+
{
|
247
|
+
rb_funcall(str, rb_intern("strip!"), 0);
|
248
|
+
rb_ary_push(ary, str);
|
249
|
+
|
250
|
+
return Qnil;
|
251
|
+
}
|
252
|
+
|
253
|
+
static VALUE
|
254
|
+
multiline_strip(VALUE text)
|
255
|
+
{
|
256
|
+
VALUE map = rb_ary_new();
|
257
|
+
VALUE split = rb_str_split(text, "\n");
|
258
|
+
|
259
|
+
rb_iterate(rb_each, split, strip_i, map);
|
260
|
+
|
261
|
+
return rb_ary_join(split, rb_str_new2("\n"));
|
262
|
+
}
|
263
|
+
|
264
|
+
static void
|
265
|
+
store_kw_con(VALUE listener, const char * event_name,
|
266
|
+
const char * keyword_at, size_t keyword_length,
|
267
|
+
const char * at, size_t length,
|
268
|
+
int current_line)
|
269
|
+
{
|
270
|
+
VALUE con = Qnil, kw = Qnil;
|
271
|
+
kw = ENCODED_STR_NEW(keyword_at, keyword_length);
|
272
|
+
con = ENCODED_STR_NEW(at, length);
|
273
|
+
con = multiline_strip(con);
|
274
|
+
rb_funcall(con, rb_intern("strip!"), 0);
|
275
|
+
rb_funcall(kw, rb_intern("strip!"), 0);
|
276
|
+
rb_funcall(listener, rb_intern(event_name), 3, kw, con, INT2FIX(current_line));
|
277
|
+
}
|
278
|
+
|
279
|
+
static void
|
280
|
+
store_attr(VALUE listener, const char * attr_type,
|
281
|
+
const char * at, size_t length,
|
282
|
+
int line)
|
283
|
+
{
|
284
|
+
VALUE val = ENCODED_STR_NEW(at, length);
|
285
|
+
rb_funcall(listener, rb_intern(attr_type), 2, val, INT2FIX(line));
|
286
|
+
}
|
287
|
+
|
288
|
+
static void
|
289
|
+
store_pystring_content(VALUE listener,
|
290
|
+
int start_col,
|
291
|
+
const char *at, size_t length,
|
292
|
+
int current_line)
|
293
|
+
{
|
294
|
+
VALUE con = ENCODED_STR_NEW(at, length);
|
295
|
+
// Gherkin will crash gracefully if the string representation of start_col pushes the pattern past 32 characters
|
296
|
+
char pat[32];
|
297
|
+
snprintf(pat, 32, "^ {0,%d}", start_col);
|
298
|
+
VALUE re = rb_reg_regcomp(rb_str_new2(pat));
|
299
|
+
rb_funcall(con, rb_intern("gsub!"), 2, re, rb_str_new2(""));
|
300
|
+
rb_funcall(listener, rb_intern("py_string"), 2, con, INT2FIX(current_line));
|
301
|
+
}
|
302
|
+
|
303
|
+
static void
|
304
|
+
raise_lexer_error(VALUE listener, const char * at, int line)
|
305
|
+
{
|
306
|
+
rb_raise(rb_eGherkinLexerError, "Lexing error on line %d: '%s'.", line, at);
|
307
|
+
}
|
308
|
+
|
309
|
+
static void lexer_init(lexer_state *lexer) {
|
310
|
+
lexer->content_start = 0;
|
311
|
+
lexer->content_end = 0;
|
312
|
+
lexer->content_len = 0;
|
313
|
+
lexer->mark = 0;
|
314
|
+
lexer->field_len = 0;
|
315
|
+
lexer->keyword_start = 0;
|
316
|
+
lexer->keyword_end = 0;
|
317
|
+
lexer->next_keyword_start = 0;
|
318
|
+
lexer->line_number = 1;
|
319
|
+
lexer->last_newline = 0;
|
320
|
+
lexer->final_newline = 0;
|
321
|
+
lexer->start_col = 0;
|
322
|
+
}
|
323
|
+
|
324
|
+
static VALUE CLexer_alloc(VALUE klass)
|
325
|
+
{
|
326
|
+
VALUE obj;
|
327
|
+
lexer_state *lxr = ALLOC(lexer_state);
|
328
|
+
lexer_init(lxr);
|
329
|
+
|
330
|
+
obj = Data_Wrap_Struct(klass, NULL, -1, lxr);
|
331
|
+
|
332
|
+
return obj;
|
333
|
+
}
|
334
|
+
|
335
|
+
static VALUE CLexer_init(VALUE self, VALUE listener)
|
336
|
+
{
|
337
|
+
rb_iv_set(self, "@listener", listener);
|
338
|
+
|
339
|
+
lexer_state *lxr = NULL;
|
340
|
+
DATA_GET(self, lexer_state, lxr);
|
341
|
+
lexer_init(lxr);
|
342
|
+
|
343
|
+
return self;
|
344
|
+
}
|
345
|
+
|
346
|
+
static VALUE CLexer_scan(VALUE self, VALUE input)
|
347
|
+
{
|
348
|
+
lexer_state *lexer = NULL;
|
349
|
+
DATA_GET(self, lexer_state, lexer);
|
350
|
+
|
351
|
+
VALUE input_copy = rb_str_dup(input);
|
352
|
+
rb_str_append(input_copy, rb_str_new2("\n%_FEATURE_END_%"));
|
353
|
+
char *data = RSTRING_PTR(input_copy);
|
354
|
+
long len = RSTRING_LEN(input_copy);
|
355
|
+
|
356
|
+
if (len == 0) {
|
357
|
+
rb_raise(rb_eGherkinLexerError, "No content to lex.");
|
358
|
+
} else {
|
359
|
+
const char *p, *pe, *eof;
|
360
|
+
int cs = 0;
|
361
|
+
|
362
|
+
VALUE listener = rb_iv_get(self, "@listener");
|
363
|
+
VALUE rows = rb_ary_new();
|
364
|
+
VALUE current_row = rb_ary_new();
|
365
|
+
|
366
|
+
p = data;
|
367
|
+
pe = data + len;
|
368
|
+
eof = pe;
|
369
|
+
|
370
|
+
assert(*pe == '\0' && "pointer does not end on NULL");
|
371
|
+
assert(pe - p == len && "pointers aren't same distance");
|
372
|
+
|
373
|
+
%% write init;
|
374
|
+
%% write exec;
|
375
|
+
|
376
|
+
assert(p <= pe && "data overflow after parsing execute");
|
377
|
+
assert(lexer->content_start <= len && "content starts after data end");
|
378
|
+
assert(lexer->mark < len && "mark is after data end");
|
379
|
+
assert(lexer->field_len <= len && "field has length longer than the whole data");
|
380
|
+
|
381
|
+
// Reset lexer by re-initializing the whole thing
|
382
|
+
lexer_init(lexer);
|
383
|
+
|
384
|
+
if (cs == lexer_error) {
|
385
|
+
rb_raise(rb_eGherkinLexerError, "Invalid format, lexing fails.");
|
386
|
+
} else {
|
387
|
+
return Qtrue;
|
388
|
+
}
|
389
|
+
}
|
390
|
+
}
|
391
|
+
|
392
|
+
void Init_gherkin_lexer_<%= @i18n.sanitized_key %>()
|
393
|
+
{
|
394
|
+
mGherkin = rb_define_module("Gherkin");
|
395
|
+
mLexer = rb_const_get(mGherkin, rb_intern("Lexer"));
|
396
|
+
rb_eGherkinLexerError = rb_const_get(mLexer, rb_intern("LexingError"));
|
397
|
+
|
398
|
+
mCLexer = rb_define_module_under(mGherkin, "CLexer");
|
399
|
+
cI18nLexer = rb_define_class_under(mCLexer, "<%= @i18n.sanitized_key.capitalize %>", rb_cObject);
|
400
|
+
rb_define_alloc_func(cI18nLexer, CLexer_alloc);
|
401
|
+
rb_define_method(cI18nLexer, "initialize", CLexer_init, 1);
|
402
|
+
rb_define_method(cI18nLexer, "scan", CLexer_scan, 1);
|
403
|
+
}
|
@@ -0,0 +1,200 @@
|
|
1
|
+
package gherkin.lexer;
|
2
|
+
|
3
|
+
import java.util.List;
|
4
|
+
import java.util.ArrayList;
|
5
|
+
import java.util.regex.Pattern;
|
6
|
+
import gherkin.Lexer;
|
7
|
+
import gherkin.Listener;
|
8
|
+
import gherkin.LexingError;
|
9
|
+
|
10
|
+
public class <%= @i18n.sanitized_key.capitalize %> implements Lexer {
|
11
|
+
%%{
|
12
|
+
machine lexer;
|
13
|
+
alphtype byte;
|
14
|
+
|
15
|
+
action begin_content {
|
16
|
+
contentStart = p;
|
17
|
+
currentLine = lineNumber;
|
18
|
+
}
|
19
|
+
|
20
|
+
action start_pystring {
|
21
|
+
currentLine = lineNumber;
|
22
|
+
startCol = p - lastNewline;
|
23
|
+
}
|
24
|
+
|
25
|
+
action begin_pystring_content {
|
26
|
+
contentStart = p;
|
27
|
+
}
|
28
|
+
|
29
|
+
action store_pystring_content {
|
30
|
+
String con = unindent(startCol, substring(data, contentStart, nextKeywordStart-1).replaceFirst("(\\r?\\n)?( )*\\Z", ""));
|
31
|
+
listener.py_string(con, currentLine);
|
32
|
+
}
|
33
|
+
|
34
|
+
action store_feature_content {
|
35
|
+
String con = multilineStrip(keywordContent(data, p, eof, nextKeywordStart, contentStart).trim());
|
36
|
+
listener.feature(keyword, con, currentLine);
|
37
|
+
if(nextKeywordStart != -1) p = nextKeywordStart - 1;
|
38
|
+
nextKeywordStart = -1;
|
39
|
+
}
|
40
|
+
|
41
|
+
action store_background_content {
|
42
|
+
String con = multilineStrip(keywordContent(data, p, eof, nextKeywordStart, contentStart));
|
43
|
+
listener.background(keyword, con, currentLine);
|
44
|
+
if(nextKeywordStart != -1) p = nextKeywordStart - 1;
|
45
|
+
nextKeywordStart = -1;
|
46
|
+
}
|
47
|
+
|
48
|
+
action store_scenario_content {
|
49
|
+
String con = multilineStrip(keywordContent(data, p, eof, nextKeywordStart, contentStart));
|
50
|
+
listener.scenario(keyword, con, currentLine);
|
51
|
+
if(nextKeywordStart != -1) p = nextKeywordStart - 1;
|
52
|
+
nextKeywordStart = -1;
|
53
|
+
}
|
54
|
+
|
55
|
+
action store_scenario_outline_content {
|
56
|
+
String con = multilineStrip(keywordContent(data, p, eof, nextKeywordStart, contentStart));
|
57
|
+
listener.scenario_outline(keyword, con, currentLine);
|
58
|
+
if(nextKeywordStart != -1) p = nextKeywordStart - 1;
|
59
|
+
nextKeywordStart = -1;
|
60
|
+
}
|
61
|
+
|
62
|
+
action store_examples_content {
|
63
|
+
String con = multilineStrip(keywordContent(data, p, eof, nextKeywordStart, contentStart));
|
64
|
+
listener.examples(keyword, con, currentLine);
|
65
|
+
if(nextKeywordStart != -1) p = nextKeywordStart - 1;
|
66
|
+
nextKeywordStart = -1;
|
67
|
+
}
|
68
|
+
|
69
|
+
action store_step_content {
|
70
|
+
listener.step(keyword, substring(data, contentStart, p).trim(), currentLine);
|
71
|
+
}
|
72
|
+
|
73
|
+
action store_comment_content {
|
74
|
+
listener.comment(substring(data, contentStart, p).trim(), lineNumber);
|
75
|
+
keywordStart = -1;
|
76
|
+
}
|
77
|
+
|
78
|
+
action store_tag_content {
|
79
|
+
listener.tag(substring(data, contentStart, p).trim(), currentLine);
|
80
|
+
keywordStart = -1;
|
81
|
+
}
|
82
|
+
|
83
|
+
action inc_line_number {
|
84
|
+
lineNumber++;
|
85
|
+
}
|
86
|
+
|
87
|
+
action last_newline {
|
88
|
+
lastNewline = p + 1;
|
89
|
+
}
|
90
|
+
|
91
|
+
action start_keyword {
|
92
|
+
if(keywordStart == -1) keywordStart = p;
|
93
|
+
}
|
94
|
+
|
95
|
+
action end_keyword {
|
96
|
+
keyword = substring(data, keywordStart, p).replaceFirst(":$","").trim();
|
97
|
+
keywordStart = -1;
|
98
|
+
}
|
99
|
+
|
100
|
+
action next_keyword_start {
|
101
|
+
nextKeywordStart = p;
|
102
|
+
}
|
103
|
+
|
104
|
+
action start_table {
|
105
|
+
p = p - 1;
|
106
|
+
rows = new ArrayList<List<String>>();
|
107
|
+
currentLine = lineNumber;
|
108
|
+
}
|
109
|
+
|
110
|
+
action start_row {
|
111
|
+
currentRow = new ArrayList<String>();
|
112
|
+
}
|
113
|
+
|
114
|
+
action begin_cell_content {
|
115
|
+
contentStart = p;
|
116
|
+
}
|
117
|
+
|
118
|
+
action store_cell_content {
|
119
|
+
currentRow.add(substring(data, contentStart, p).trim());
|
120
|
+
}
|
121
|
+
|
122
|
+
action store_row {
|
123
|
+
rows.add(currentRow);
|
124
|
+
}
|
125
|
+
|
126
|
+
action store_table {
|
127
|
+
if(!rows.isEmpty()) {
|
128
|
+
listener.table(rows, currentLine);
|
129
|
+
}
|
130
|
+
}
|
131
|
+
|
132
|
+
action end_feature {
|
133
|
+
if(cs < lexer_first_final) {
|
134
|
+
String content = currentLineContent(data, lastNewline);
|
135
|
+
throw new LexingError("Lexing error on line " + lineNumber);
|
136
|
+
}
|
137
|
+
}
|
138
|
+
|
139
|
+
include lexer_common "lexer_common.<%= @i18n.sanitized_key %>.rl";
|
140
|
+
}%%
|
141
|
+
|
142
|
+
private final Listener listener;
|
143
|
+
|
144
|
+
public <%= @i18n.sanitized_key.capitalize %>(Listener listener) {
|
145
|
+
this.listener = listener;
|
146
|
+
}
|
147
|
+
|
148
|
+
%% write data noerror;
|
149
|
+
|
150
|
+
public void scan(CharSequence inputSequence) {
|
151
|
+
String input = inputSequence.toString() + "\n%_FEATURE_END_%";
|
152
|
+
byte[] data = input.getBytes();
|
153
|
+
int cs, p = 0, pe = data.length;
|
154
|
+
int eof = pe;
|
155
|
+
|
156
|
+
int lineNumber = 1;
|
157
|
+
int lastNewline = 0;
|
158
|
+
|
159
|
+
int contentStart = -1;
|
160
|
+
int currentLine = -1;
|
161
|
+
int startCol = -1;
|
162
|
+
int nextKeywordStart = -1;
|
163
|
+
int keywordStart = -1;
|
164
|
+
String keyword = null;
|
165
|
+
List<List<String>> rows = null;
|
166
|
+
List<String> currentRow = null;
|
167
|
+
|
168
|
+
%% write init;
|
169
|
+
%% write exec;
|
170
|
+
}
|
171
|
+
|
172
|
+
private String keywordContent(byte[] data, int p, int eof, int nextKeywordStart, int contentStart) {
|
173
|
+
int endPoint = (nextKeywordStart == -1 || (p == eof)) ? p : nextKeywordStart;
|
174
|
+
return substring(data, contentStart, endPoint);
|
175
|
+
}
|
176
|
+
|
177
|
+
private String multilineStrip(String text) {
|
178
|
+
StringBuffer result = new StringBuffer();
|
179
|
+
for(String s : text.split("\n")) {
|
180
|
+
result.append(s.trim()).append("\n");
|
181
|
+
}
|
182
|
+
return result.toString().trim();
|
183
|
+
}
|
184
|
+
|
185
|
+
private String unindent(int startCol, String text) {
|
186
|
+
return Pattern.compile("^ {0," + startCol + "}", Pattern.MULTILINE).matcher(text).replaceAll("");
|
187
|
+
}
|
188
|
+
|
189
|
+
private String currentLineContent(byte[] data, int lastNewline) {
|
190
|
+
return substring(data, lastNewline, data.length).trim();
|
191
|
+
}
|
192
|
+
|
193
|
+
private String substring(byte[] data, int start, int end) {
|
194
|
+
try {
|
195
|
+
return new String(data, start, end-start, "utf-8");
|
196
|
+
} catch(java.io.UnsupportedEncodingException e) {
|
197
|
+
throw new RuntimeException("Internal error", e);
|
198
|
+
}
|
199
|
+
}
|
200
|
+
}
|