gherkin 0.0.4-i386-mingw32

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (79) hide show
  1. data/.gitignore +7 -0
  2. data/LICENSE +20 -0
  3. data/README.rdoc +66 -0
  4. data/Rakefile +49 -0
  5. data/VERSION.yml +4 -0
  6. data/bin/gherkin +10 -0
  7. data/cucumber.yml +3 -0
  8. data/features/feature_parser.feature +206 -0
  9. data/features/native_lexer.feature +19 -0
  10. data/features/parser_with_native_lexer.feature +205 -0
  11. data/features/pretty_printer.feature +14 -0
  12. data/features/step_definitions/gherkin_steps.rb +34 -0
  13. data/features/step_definitions/pretty_printer_steps.rb +56 -0
  14. data/features/steps_parser.feature +46 -0
  15. data/features/support/env.rb +33 -0
  16. data/gherkin.gemspec +180 -0
  17. data/java/.gitignore +2 -0
  18. data/java/Gherkin.iml +24 -0
  19. data/java/build.xml +13 -0
  20. data/java/src/gherkin/FixJava.java +34 -0
  21. data/java/src/gherkin/Lexer.java +5 -0
  22. data/java/src/gherkin/LexingError.java +7 -0
  23. data/java/src/gherkin/Listener.java +27 -0
  24. data/java/src/gherkin/ParseError.java +22 -0
  25. data/java/src/gherkin/Parser.java +185 -0
  26. data/java/src/gherkin/lexer/.gitignore +1 -0
  27. data/java/src/gherkin/parser/StateMachineReader.java +62 -0
  28. data/lib/.gitignore +4 -0
  29. data/lib/gherkin.rb +2 -0
  30. data/lib/gherkin/c_lexer.rb +10 -0
  31. data/lib/gherkin/core_ext/array.rb +5 -0
  32. data/lib/gherkin/i18n.yml +535 -0
  33. data/lib/gherkin/i18n_lexer.rb +29 -0
  34. data/lib/gherkin/java_lexer.rb +10 -0
  35. data/lib/gherkin/lexer.rb +43 -0
  36. data/lib/gherkin/parser.rb +19 -0
  37. data/lib/gherkin/parser/meta.txt +4 -0
  38. data/lib/gherkin/parser/root.txt +9 -0
  39. data/lib/gherkin/parser/steps.txt +3 -0
  40. data/lib/gherkin/rb_lexer.rb +10 -0
  41. data/lib/gherkin/rb_lexer/.gitignore +1 -0
  42. data/lib/gherkin/rb_lexer/README.rdoc +8 -0
  43. data/lib/gherkin/rb_parser.rb +117 -0
  44. data/lib/gherkin/tools/pretty_printer.rb +83 -0
  45. data/nativegems.sh +5 -0
  46. data/ragel/i18n/.gitignore +1 -0
  47. data/ragel/lexer.c.rl.erb +401 -0
  48. data/ragel/lexer.java.rl.erb +200 -0
  49. data/ragel/lexer.rb.rl.erb +171 -0
  50. data/ragel/lexer_common.rl.erb +46 -0
  51. data/spec/gherkin/c_lexer_spec.rb +21 -0
  52. data/spec/gherkin/fixtures/1.feature +8 -0
  53. data/spec/gherkin/fixtures/complex.feature +43 -0
  54. data/spec/gherkin/fixtures/i18n_fr.feature +13 -0
  55. data/spec/gherkin/fixtures/i18n_no.feature +6 -0
  56. data/spec/gherkin/fixtures/i18n_zh-CN.feature +8 -0
  57. data/spec/gherkin/fixtures/simple.feature +3 -0
  58. data/spec/gherkin/fixtures/simple_with_comments.feature +7 -0
  59. data/spec/gherkin/fixtures/simple_with_tags.feature +11 -0
  60. data/spec/gherkin/i18n_spec.rb +57 -0
  61. data/spec/gherkin/java_lexer_spec.rb +20 -0
  62. data/spec/gherkin/parser_spec.rb +28 -0
  63. data/spec/gherkin/rb_lexer_spec.rb +18 -0
  64. data/spec/gherkin/sexp_recorder.rb +29 -0
  65. data/spec/gherkin/shared/lexer_spec.rb +433 -0
  66. data/spec/gherkin/shared/py_string_spec.rb +124 -0
  67. data/spec/gherkin/shared/table_spec.rb +97 -0
  68. data/spec/gherkin/shared/tags_spec.rb +50 -0
  69. data/spec/spec_helper.rb +53 -0
  70. data/tasks/bench.rake +193 -0
  71. data/tasks/bench/feature_builder.rb +49 -0
  72. data/tasks/bench/generated/.gitignore +1 -0
  73. data/tasks/bench/null_listener.rb +4 -0
  74. data/tasks/compile.rake +70 -0
  75. data/tasks/cucumber.rake +20 -0
  76. data/tasks/ragel_task.rb +83 -0
  77. data/tasks/rdoc.rake +12 -0
  78. data/tasks/rspec.rake +15 -0
  79. metadata +214 -0
@@ -0,0 +1,200 @@
1
+ package gherkin.lexer;
2
+
3
+ import java.util.List;
4
+ import java.util.ArrayList;
5
+ import java.util.regex.Pattern;
6
+ import gherkin.Lexer;
7
+ import gherkin.Listener;
8
+ import gherkin.LexingError;
9
+
10
+ public class <%= @i18n.capitalize %> implements Lexer {
11
+ %%{
12
+ machine lexer;
13
+ alphtype byte;
14
+
15
+ action begin_content {
16
+ contentStart = p;
17
+ currentLine = lineNumber;
18
+ }
19
+
20
+ action start_pystring {
21
+ currentLine = lineNumber;
22
+ startCol = p - lastNewline;
23
+ }
24
+
25
+ action begin_pystring_content {
26
+ contentStart = p;
27
+ }
28
+
29
+ action store_pystring_content {
30
+ String con = unindent(startCol, substring(data, contentStart, nextKeywordStart-1).replaceFirst("(\\r?\\n)?( )*\\Z", ""));
31
+ listener.py_string(con, currentLine);
32
+ }
33
+
34
+ action store_feature_content {
35
+ String con = multilineStrip(keywordContent(data, p, eof, nextKeywordStart, contentStart).trim());
36
+ listener.feature(keyword, con, currentLine);
37
+ if(nextKeywordStart != -1) p = nextKeywordStart - 1;
38
+ nextKeywordStart = -1;
39
+ }
40
+
41
+ action store_background_content {
42
+ String con = multilineStrip(keywordContent(data, p, eof, nextKeywordStart, contentStart));
43
+ listener.background(keyword, con, currentLine);
44
+ if(nextKeywordStart != -1) p = nextKeywordStart - 1;
45
+ nextKeywordStart = -1;
46
+ }
47
+
48
+ action store_scenario_content {
49
+ String con = multilineStrip(keywordContent(data, p, eof, nextKeywordStart, contentStart));
50
+ listener.scenario(keyword, con, currentLine);
51
+ if(nextKeywordStart != -1) p = nextKeywordStart - 1;
52
+ nextKeywordStart = -1;
53
+ }
54
+
55
+ action store_scenario_outline_content {
56
+ String con = multilineStrip(keywordContent(data, p, eof, nextKeywordStart, contentStart));
57
+ listener.scenario_outline(keyword, con, currentLine);
58
+ if(nextKeywordStart != -1) p = nextKeywordStart - 1;
59
+ nextKeywordStart = -1;
60
+ }
61
+
62
+ action store_examples_content {
63
+ String con = multilineStrip(keywordContent(data, p, eof, nextKeywordStart, contentStart));
64
+ listener.examples(keyword, con, currentLine);
65
+ if(nextKeywordStart != -1) p = nextKeywordStart - 1;
66
+ nextKeywordStart = -1;
67
+ }
68
+
69
+ action store_step_content {
70
+ listener.step(keyword, substring(data, contentStart, p).trim(), currentLine);
71
+ }
72
+
73
+ action store_comment_content {
74
+ listener.comment(substring(data, contentStart, p).trim(), lineNumber);
75
+ keywordStart = -1;
76
+ }
77
+
78
+ action store_tag_content {
79
+ listener.tag(substring(data, contentStart, p).trim(), currentLine);
80
+ keywordStart = -1;
81
+ }
82
+
83
+ action inc_line_number {
84
+ lineNumber++;
85
+ }
86
+
87
+ action last_newline {
88
+ lastNewline = p + 1;
89
+ }
90
+
91
+ action start_keyword {
92
+ if(keywordStart == -1) keywordStart = p;
93
+ }
94
+
95
+ action end_keyword {
96
+ keyword = substring(data, keywordStart, p).replaceFirst(":$","").trim();
97
+ keywordStart = -1;
98
+ }
99
+
100
+ action next_keyword_start {
101
+ nextKeywordStart = p;
102
+ }
103
+
104
+ action start_table {
105
+ p = p - 1;
106
+ rows = new ArrayList<List<String>>();
107
+ currentLine = lineNumber;
108
+ }
109
+
110
+ action start_row {
111
+ currentRow = new ArrayList<String>();
112
+ }
113
+
114
+ action begin_cell_content {
115
+ contentStart = p;
116
+ }
117
+
118
+ action store_cell_content {
119
+ currentRow.add(substring(data, contentStart, p).trim());
120
+ }
121
+
122
+ action store_row {
123
+ rows.add(currentRow);
124
+ }
125
+
126
+ action store_table {
127
+ if(!rows.isEmpty()) {
128
+ listener.table(rows, currentLine);
129
+ }
130
+ }
131
+
132
+ action end_feature {
133
+ if(cs < lexer_first_final) {
134
+ String content = currentLineContent(data, lastNewline);
135
+ throw new LexingError("Lexing error on line " + lineNumber);
136
+ }
137
+ }
138
+
139
+ include lexer_common "lexer_common.<%= @i18n %>.rl";
140
+ }%%
141
+
142
+ private final Listener listener;
143
+
144
+ public <%= @i18n.capitalize %>(Listener listener) {
145
+ this.listener = listener;
146
+ }
147
+
148
+ %% write data noerror;
149
+
150
+ public void scan(CharSequence inputSequence) {
151
+ String input = inputSequence.toString() + "\n%_FEATURE_END_%";
152
+ byte[] data = input.getBytes();
153
+ int cs, p = 0, pe = data.length;
154
+ int eof = pe;
155
+
156
+ int lineNumber = 1;
157
+ int lastNewline = 0;
158
+
159
+ int contentStart = -1;
160
+ int currentLine = -1;
161
+ int startCol = -1;
162
+ int nextKeywordStart = -1;
163
+ int keywordStart = -1;
164
+ String keyword = null;
165
+ List<List<String>> rows = null;
166
+ List<String> currentRow = null;
167
+
168
+ %% write init;
169
+ %% write exec;
170
+ }
171
+
172
+ private String keywordContent(byte[] data, int p, int eof, int nextKeywordStart, int contentStart) {
173
+ int endPoint = (nextKeywordStart == -1 || (p == eof)) ? p : nextKeywordStart;
174
+ return substring(data, contentStart, endPoint);
175
+ }
176
+
177
+ private String multilineStrip(String text) {
178
+ StringBuffer result = new StringBuffer();
179
+ for(String s : text.split("\n")) {
180
+ result.append(s.trim()).append("\n");
181
+ }
182
+ return result.toString().trim();
183
+ }
184
+
185
+ private String unindent(int startCol, String text) {
186
+ return Pattern.compile("^ {0," + startCol + "}", Pattern.MULTILINE).matcher(text).replaceAll("");
187
+ }
188
+
189
+ private String currentLineContent(byte[] data, int lastNewline) {
190
+ return substring(data, lastNewline, data.length).trim();
191
+ }
192
+
193
+ private String substring(byte[] data, int start, int end) {
194
+ try {
195
+ return new String(data, start, end-start, "utf-8");
196
+ } catch(java.io.UnsupportedEncodingException e) {
197
+ throw new RuntimeException("Internal error", e);
198
+ }
199
+ }
200
+ }
@@ -0,0 +1,171 @@
1
+ require 'gherkin/core_ext/array'
2
+
3
+ module Gherkin
4
+ module RbLexer
5
+ class <%= @i18n.capitalize %> #:nodoc:
6
+ %%{
7
+ machine lexer;
8
+
9
+ action begin_content {
10
+ @content_start = p
11
+ @current_line = @line_number
12
+ }
13
+
14
+ action start_pystring {
15
+ @current_line = @line_number
16
+ @start_col = p - @last_newline
17
+ }
18
+
19
+ action begin_pystring_content {
20
+ @content_start = p
21
+ }
22
+
23
+ action store_pystring_content {
24
+ con = unindent(@start_col, data[@content_start...@next_keyword_start-1].utf8_pack("c*").sub(/(\r?\n)?( )*\Z/, ''))
25
+ @listener.py_string(con, @current_line)
26
+ }
27
+
28
+ action store_feature_content {
29
+ store_keyword_content(:feature, data, p, eof) { |con| multiline_strip(con) }
30
+ p = @next_keyword_start - 1 if @next_keyword_start
31
+ @next_keyword_start = nil
32
+ }
33
+
34
+ action store_background_content {
35
+ store_keyword_content(:background, data, p, eof) { |con| multiline_strip(con) }
36
+ p = @next_keyword_start - 1 if @next_keyword_start
37
+ @next_keyword_start = nil
38
+ }
39
+
40
+ action store_scenario_content {
41
+ store_keyword_content(:scenario, data, p, eof) { |con| multiline_strip(con) }
42
+ p = @next_keyword_start - 1 if @next_keyword_start
43
+ @next_keyword_start = nil
44
+ }
45
+
46
+ action store_scenario_outline_content {
47
+ store_keyword_content(:scenario_outline, data, p, eof) { |con| multiline_strip(con) }
48
+ p = @next_keyword_start - 1 if @next_keyword_start
49
+ @next_keyword_start = nil
50
+ }
51
+
52
+ action store_examples_content {
53
+ store_keyword_content(:examples, data, p, eof) { |con| multiline_strip(con) }
54
+ p = @next_keyword_start - 1 if @next_keyword_start
55
+ @next_keyword_start = nil
56
+ }
57
+
58
+ action store_step_content {
59
+ con = data[@content_start...p].utf8_pack("c*").strip
60
+ @listener.step(@keyword, con, @current_line)
61
+ }
62
+
63
+ action store_comment_content {
64
+ con = data[@content_start...p].utf8_pack("c*").strip
65
+ @listener.comment(con, @line_number)
66
+ @keyword_start = nil
67
+ }
68
+
69
+ action store_tag_content {
70
+ con = data[@content_start...p].utf8_pack("c*").strip
71
+ @listener.tag(con, @current_line)
72
+ @keyword_start = nil
73
+ }
74
+
75
+ action inc_line_number {
76
+ @line_number += 1
77
+ }
78
+
79
+ action last_newline {
80
+ @last_newline = p + 1
81
+ }
82
+
83
+ action start_keyword {
84
+ @keyword_start ||= p
85
+ }
86
+
87
+ action end_keyword {
88
+ @keyword = data[@keyword_start...p].utf8_pack("c*").sub(/:$/,'').strip
89
+ @keyword_start = nil
90
+ }
91
+
92
+ action next_keyword_start {
93
+ @next_keyword_start = p
94
+ }
95
+
96
+ action start_table {
97
+ p = p - 1
98
+ @rows = []
99
+ @current_line = @line_number
100
+ }
101
+
102
+ action start_row {
103
+ current_row = []
104
+ }
105
+
106
+ action begin_cell_content {
107
+ @content_start = p
108
+ }
109
+
110
+ action store_cell_content {
111
+ con = data[@content_start...p].utf8_pack("c*").strip
112
+ current_row << con
113
+ }
114
+
115
+ action store_row {
116
+ @rows << current_row
117
+ }
118
+
119
+ action store_table {
120
+ if @rows.size != 0
121
+ @listener.table(@rows, @current_line)
122
+ end
123
+ }
124
+
125
+ action end_feature {
126
+ if cs < lexer_first_final
127
+ content = current_line_content(data, p)
128
+ raise Lexer::LexingError.new("Lexing error on line %d: '%s'." % [@line_number, content])
129
+ end
130
+ }
131
+
132
+ include lexer_common "lexer_common.<%= @i18n %>.rl";
133
+ }%%
134
+
135
+ def initialize(listener)
136
+ @listener = listener
137
+ %% write data;
138
+ end
139
+
140
+ def scan(data)
141
+ data = (data + "\n%_FEATURE_END_%").unpack("c*") # Explicit EOF simplifies things considerably
142
+ eof = pe = data.length
143
+
144
+ @line_number = 1
145
+ @last_newline = 0
146
+
147
+ %% write init;
148
+ %% write exec;
149
+ end
150
+
151
+ def multiline_strip(text)
152
+ text.split("\n").map{|s| s.strip}.join("\n").strip
153
+ end
154
+
155
+ def unindent(startcol, text)
156
+ text.gsub(/^ {0,#{startcol}}/, "")
157
+ end
158
+
159
+ def store_keyword_content(event, data, p, eof)
160
+ end_point = (!@next_keyword_start or (p == eof)) ? p : @next_keyword_start
161
+ con = yield data[@content_start...end_point].utf8_pack("c*")
162
+ @listener.send(event, @keyword, con, @current_line)
163
+ end
164
+
165
+ def current_line_content(data, p)
166
+ rest = data[@last_newline..-1]
167
+ rest[0..rest.index(10)||-1].utf8_pack("c*").strip
168
+ end
169
+ end
170
+ end
171
+ end
@@ -0,0 +1,46 @@
1
+ %%{
2
+ machine lexer_common;
3
+
4
+ # Language specific
5
+ I18N_Feature = <%= keywords['feature'] %> >start_keyword %end_keyword;
6
+ I18N_Background = <%= keywords['background'] %> >start_keyword %end_keyword;
7
+ I18N_ScenarioOutline = <%= keywords['scenario_outline'] %> >start_keyword %end_keyword;
8
+ I18N_Scenario = <%= keywords['scenario'] %> >start_keyword %end_keyword;
9
+ I18N_Step = (<%= keywords['given'] %> | <%= keywords['when'] %> | <%= keywords['and'] %> | <%= keywords['then'] %> | <%= keywords['but'] %>) >start_keyword %end_keyword;
10
+ I18N_Examples = <%= keywords['examples'] %> >start_keyword %end_keyword;
11
+
12
+ EOF = '%_FEATURE_END_%'; # Explicit EOF added before scanning begins
13
+ EOL = ('\r'? '\n') @inc_line_number @last_newline;
14
+
15
+ FeatureHeadingEnd = EOL+ space* (I18N_Background | I18N_Scenario | I18N_ScenarioOutline | '@' | '#' | EOF) >next_keyword_start;
16
+ ScenarioHeadingEnd = EOL+ space* ( I18N_Scenario | I18N_ScenarioOutline | I18N_Step | '@' | '#' | EOF ) >next_keyword_start;
17
+ BackgroundHeadingEnd = EOL+ space* ( I18N_Scenario | I18N_ScenarioOutline | I18N_Step | '@' | '#'| EOF ) >next_keyword_start;
18
+ ScenarioOutlineHeadingEnd = EOL+ space* ( I18N_Scenario | I18N_Step | '@' | '#' | EOF ) >next_keyword_start;
19
+ ExamplesHeadingEnd = EOL+ space* '|' >next_keyword_start;
20
+
21
+ FeatureHeading = space* I18N_Feature %begin_content ^FeatureHeadingEnd* :>> FeatureHeadingEnd @store_feature_content;
22
+ BackgroundHeading = space* I18N_Background %begin_content ^BackgroundHeadingEnd* :>> BackgroundHeadingEnd @store_background_content;
23
+ ScenarioHeading = space* I18N_Scenario %begin_content ^ScenarioHeadingEnd* :>> ScenarioHeadingEnd @store_scenario_content;
24
+ ScenarioOutlineHeading = space* I18N_ScenarioOutline %begin_content ^ScenarioOutlineHeadingEnd* :>> ScenarioOutlineHeadingEnd @store_scenario_outline_content;
25
+ ExamplesHeading = space* I18N_Examples %begin_content ^ExamplesHeadingEnd* :>> ExamplesHeadingEnd @store_examples_content;
26
+
27
+ Step = space* I18N_Step %begin_content ^EOL+ %store_step_content EOL+;
28
+ Comment = space* '#' >begin_content ^EOL* %store_comment_content EOL+;
29
+
30
+ Tag = ( '@' [^@\r\n\t ]+ >begin_content ) %store_tag_content;
31
+ Tags = space* (Tag space*)+ EOL+;
32
+
33
+ StartTable = space* '|' >start_table;
34
+ EndTable = EOL space* ^('|') >next_keyword_start;
35
+ Cell = '|' (any - '|')* >begin_cell_content %store_cell_content;
36
+ Row = space* Cell* >start_row '|' :>> (space* EOL+ space*) %store_row;
37
+ Table = StartTable :>> Row+ %store_table <: EndTable?;
38
+
39
+ StartPyString = '"""' >start_pystring space* :>> EOL;
40
+ EndPyString = (space* '"""') >next_keyword_start;
41
+ PyString = space* StartPyString %begin_pystring_content (^EOL | EOL)* :>> EndPyString %store_pystring_content space* EOL+;
42
+
43
+ Tokens = (space | EOL)* (Tags | Comment | FeatureHeading | BackgroundHeading | ScenarioHeading | ScenarioOutlineHeading | ExamplesHeading | Step | Table | PyString)* (space | EOL)* EOF;
44
+
45
+ main := Tokens %end_feature @!end_feature;
46
+ }%%
@@ -0,0 +1,21 @@
1
+ #encoding: utf-8
2
+ unless defined?(JRUBY_VERSION)
3
+ require File.expand_path(File.dirname(__FILE__) + '/../spec_helper')
4
+ require 'gherkin/c_lexer'
5
+
6
+ module Gherkin
7
+ module Lexer
8
+ describe "C Lexer" do
9
+ before do
10
+ @listener = Gherkin::SexpRecorder.new
11
+ @lexer = Gherkin::CLexer['en'].new(@listener)
12
+ end
13
+
14
+ it_should_behave_like "a Gherkin lexer"
15
+ it_should_behave_like "a Gherkin lexer lexing tags"
16
+ it_should_behave_like "a Gherkin lexer lexing py_strings"
17
+ it_should_behave_like "a Gherkin lexer lexing tables"
18
+ end
19
+ end
20
+ end
21
+ end
@@ -0,0 +1,8 @@
1
+ Feature: Logging in
2
+ So that I can be myself
3
+ # Comment
4
+ Scenario: Anonymous user can get a login form.
5
+ Scenery here
6
+
7
+ @tag
8
+ Scenario: Another one