pygments.rb 0.6.3 → 1.0.0

Sign up to get free protection for your applications and to get access to all the features.
Files changed (606) hide show
  1. checksums.yaml +7 -0
  2. data/CHANGELOG.md +11 -0
  3. data/README.md +6 -6
  4. data/Rakefile +2 -1
  5. data/cache-lexers.rb +1 -1
  6. data/circle.yml +7 -0
  7. data/lib/pygments/lexer.rb +3 -3
  8. data/lib/pygments/popen.rb +67 -30
  9. data/lib/pygments/version.rb +1 -1
  10. data/pygments.rb.gemspec +2 -1
  11. data/test/test_pygments.rb +16 -16
  12. data/vendor/pygments-main/AUTHORS +54 -8
  13. data/vendor/pygments-main/CHANGES +247 -25
  14. data/vendor/pygments-main/LICENSE +1 -1
  15. data/vendor/pygments-main/Makefile +15 -6
  16. data/vendor/pygments-main/README.rst +39 -0
  17. data/vendor/pygments-main/REVISION +1 -1
  18. data/vendor/pygments-main/TODO +0 -3
  19. data/vendor/pygments-main/doc/_themes/pygments14/layout.html +1 -1
  20. data/vendor/pygments-main/doc/_themes/pygments14/static/pygments14.css_t +1 -1
  21. data/vendor/pygments-main/doc/conf.py +3 -11
  22. data/vendor/pygments-main/doc/docs/api.rst +15 -0
  23. data/vendor/pygments-main/doc/docs/cmdline.rst +6 -2
  24. data/vendor/pygments-main/doc/docs/filterdevelopment.rst +4 -3
  25. data/vendor/pygments-main/doc/docs/integrate.rst +11 -15
  26. data/vendor/pygments-main/doc/docs/java.rst +7 -7
  27. data/vendor/pygments-main/doc/docs/lexerdevelopment.rst +258 -171
  28. data/vendor/pygments-main/doc/docs/lexers.rst +2 -2
  29. data/vendor/pygments-main/doc/docs/styles.rst +58 -0
  30. data/vendor/pygments-main/doc/docs/tokens.rst +22 -2
  31. data/vendor/pygments-main/doc/docs/unicode.rst +15 -7
  32. data/vendor/pygments-main/doc/faq.rst +17 -21
  33. data/vendor/pygments-main/doc/index.rst +12 -11
  34. data/vendor/pygments-main/doc/languages.rst +10 -7
  35. data/vendor/pygments-main/external/autopygmentize +9 -6
  36. data/vendor/pygments-main/external/lasso-builtins-generator-9.lasso +70 -52
  37. data/vendor/pygments-main/external/markdown-processor.py +1 -1
  38. data/vendor/pygments-main/external/moin-parser.py +1 -1
  39. data/vendor/pygments-main/external/rst-directive.py +1 -1
  40. data/vendor/pygments-main/pygmentize +2 -1
  41. data/vendor/pygments-main/pygments/__init__.py +14 -15
  42. data/vendor/pygments-main/pygments/cmdline.py +188 -113
  43. data/vendor/pygments-main/pygments/console.py +13 -13
  44. data/vendor/pygments-main/pygments/filter.py +7 -7
  45. data/vendor/pygments-main/pygments/filters/__init__.py +24 -32
  46. data/vendor/pygments-main/pygments/formatter.py +5 -5
  47. data/vendor/pygments-main/pygments/formatters/__init__.py +92 -44
  48. data/vendor/pygments-main/pygments/formatters/_mapping.py +51 -69
  49. data/vendor/pygments-main/pygments/formatters/bbcode.py +1 -1
  50. data/vendor/pygments-main/pygments/formatters/html.py +63 -51
  51. data/vendor/pygments-main/pygments/formatters/img.py +25 -22
  52. data/vendor/pygments-main/pygments/formatters/irc.py +182 -0
  53. data/vendor/pygments-main/pygments/formatters/latex.py +34 -22
  54. data/vendor/pygments-main/pygments/formatters/other.py +5 -7
  55. data/vendor/pygments-main/pygments/formatters/rtf.py +28 -31
  56. data/vendor/pygments-main/pygments/formatters/svg.py +1 -2
  57. data/vendor/pygments-main/pygments/formatters/terminal.py +29 -45
  58. data/vendor/pygments-main/pygments/formatters/terminal256.py +118 -31
  59. data/vendor/pygments-main/pygments/lexer.py +120 -34
  60. data/vendor/pygments-main/pygments/lexers/__init__.py +85 -53
  61. data/vendor/pygments-main/pygments/lexers/{_asybuiltins.py → _asy_builtins.py} +7 -7
  62. data/vendor/pygments-main/pygments/lexers/{_clbuiltins.py → _cl_builtins.py} +17 -17
  63. data/vendor/pygments-main/pygments/lexers/_cocoa_builtins.py +72 -0
  64. data/vendor/pygments-main/pygments/lexers/_csound_builtins.py +1346 -0
  65. data/vendor/pygments-main/pygments/lexers/{_lassobuiltins.py → _lasso_builtins.py} +4699 -4561
  66. data/vendor/pygments-main/pygments/lexers/{_luabuiltins.py → _lua_builtins.py} +91 -51
  67. data/vendor/pygments-main/pygments/lexers/_mapping.py +342 -242
  68. data/vendor/pygments-main/pygments/lexers/_mql_builtins.py +1172 -0
  69. data/vendor/pygments-main/pygments/lexers/_openedge_builtins.py +2547 -0
  70. data/vendor/pygments-main/pygments/lexers/{_phpbuiltins.py → _php_builtins.py} +350 -353
  71. data/vendor/pygments-main/pygments/lexers/_postgres_builtins.py +600 -212
  72. data/vendor/pygments-main/pygments/lexers/_scilab_builtins.py +3084 -30
  73. data/vendor/pygments-main/pygments/lexers/_sourcemod_builtins.py +1163 -0
  74. data/vendor/pygments-main/pygments/lexers/_stan_builtins.py +91 -13
  75. data/vendor/pygments-main/pygments/lexers/_stata_builtins.py +419 -0
  76. data/vendor/pygments-main/pygments/lexers/_tsql_builtins.py +1004 -0
  77. data/vendor/pygments-main/pygments/lexers/_vim_builtins.py +1939 -0
  78. data/vendor/pygments-main/pygments/lexers/actionscript.py +240 -0
  79. data/vendor/pygments-main/pygments/lexers/agile.py +14 -2542
  80. data/vendor/pygments-main/pygments/lexers/algebra.py +221 -0
  81. data/vendor/pygments-main/pygments/lexers/ambient.py +76 -0
  82. data/vendor/pygments-main/pygments/lexers/ampl.py +87 -0
  83. data/vendor/pygments-main/pygments/lexers/apl.py +101 -0
  84. data/vendor/pygments-main/pygments/lexers/archetype.py +318 -0
  85. data/vendor/pygments-main/pygments/lexers/asm.py +276 -81
  86. data/vendor/pygments-main/pygments/lexers/automation.py +374 -0
  87. data/vendor/pygments-main/pygments/lexers/basic.py +500 -0
  88. data/vendor/pygments-main/pygments/lexers/bibtex.py +156 -0
  89. data/vendor/pygments-main/pygments/lexers/business.py +594 -0
  90. data/vendor/pygments-main/pygments/lexers/c_cpp.py +252 -0
  91. data/vendor/pygments-main/pygments/lexers/c_like.py +541 -0
  92. data/vendor/pygments-main/pygments/lexers/capnproto.py +79 -0
  93. data/vendor/pygments-main/pygments/lexers/chapel.py +101 -0
  94. data/vendor/pygments-main/pygments/lexers/clean.py +288 -0
  95. data/vendor/pygments-main/pygments/lexers/compiled.py +24 -5182
  96. data/vendor/pygments-main/pygments/lexers/configs.py +833 -0
  97. data/vendor/pygments-main/pygments/lexers/console.py +114 -0
  98. data/vendor/pygments-main/pygments/lexers/crystal.py +384 -0
  99. data/vendor/pygments-main/pygments/lexers/csound.py +366 -0
  100. data/vendor/pygments-main/pygments/lexers/css.py +689 -0
  101. data/vendor/pygments-main/pygments/lexers/d.py +251 -0
  102. data/vendor/pygments-main/pygments/lexers/dalvik.py +6 -6
  103. data/vendor/pygments-main/pygments/lexers/data.py +555 -0
  104. data/vendor/pygments-main/pygments/lexers/diff.py +165 -0
  105. data/vendor/pygments-main/pygments/lexers/dotnet.py +96 -88
  106. data/vendor/pygments-main/pygments/lexers/dsls.py +878 -0
  107. data/vendor/pygments-main/pygments/lexers/dylan.py +289 -0
  108. data/vendor/pygments-main/pygments/lexers/ecl.py +125 -0
  109. data/vendor/pygments-main/pygments/lexers/eiffel.py +65 -0
  110. data/vendor/pygments-main/pygments/lexers/elm.py +121 -0
  111. data/vendor/pygments-main/pygments/lexers/erlang.py +533 -0
  112. data/vendor/pygments-main/pygments/lexers/esoteric.py +275 -0
  113. data/vendor/pygments-main/pygments/lexers/ezhil.py +68 -0
  114. data/vendor/pygments-main/pygments/lexers/factor.py +344 -0
  115. data/vendor/pygments-main/pygments/lexers/fantom.py +250 -0
  116. data/vendor/pygments-main/pygments/lexers/felix.py +273 -0
  117. data/vendor/pygments-main/pygments/lexers/forth.py +177 -0
  118. data/vendor/pygments-main/pygments/lexers/fortran.py +205 -0
  119. data/vendor/pygments-main/pygments/lexers/foxpro.py +1 -1
  120. data/vendor/pygments-main/pygments/lexers/functional.py +11 -3661
  121. data/vendor/pygments-main/pygments/lexers/go.py +101 -0
  122. data/vendor/pygments-main/pygments/lexers/grammar_notation.py +213 -0
  123. data/vendor/pygments-main/pygments/lexers/graph.py +7 -8
  124. data/vendor/pygments-main/pygments/lexers/graphics.py +553 -0
  125. data/vendor/pygments-main/pygments/lexers/haskell.py +840 -0
  126. data/vendor/pygments-main/pygments/lexers/haxe.py +936 -0
  127. data/vendor/pygments-main/pygments/lexers/hdl.py +172 -145
  128. data/vendor/pygments-main/pygments/lexers/hexdump.py +97 -0
  129. data/vendor/pygments-main/pygments/lexers/html.py +602 -0
  130. data/vendor/pygments-main/pygments/lexers/idl.py +270 -0
  131. data/vendor/pygments-main/pygments/lexers/igor.py +288 -0
  132. data/vendor/pygments-main/pygments/lexers/inferno.py +3 -3
  133. data/vendor/pygments-main/pygments/lexers/installers.py +322 -0
  134. data/vendor/pygments-main/pygments/lexers/int_fiction.py +1343 -0
  135. data/vendor/pygments-main/pygments/lexers/iolang.py +63 -0
  136. data/vendor/pygments-main/pygments/lexers/j.py +146 -0
  137. data/vendor/pygments-main/pygments/lexers/javascript.py +1506 -0
  138. data/vendor/pygments-main/pygments/lexers/julia.py +333 -0
  139. data/vendor/pygments-main/pygments/lexers/jvm.py +232 -186
  140. data/vendor/pygments-main/pygments/lexers/lisp.py +2621 -0
  141. data/vendor/pygments-main/pygments/lexers/make.py +202 -0
  142. data/vendor/pygments-main/pygments/lexers/markup.py +595 -0
  143. data/vendor/pygments-main/pygments/lexers/math.py +11 -2276
  144. data/vendor/pygments-main/pygments/lexers/matlab.py +663 -0
  145. data/vendor/pygments-main/pygments/lexers/ml.py +769 -0
  146. data/vendor/pygments-main/pygments/lexers/modeling.py +358 -0
  147. data/vendor/pygments-main/pygments/lexers/modula2.py +1561 -0
  148. data/vendor/pygments-main/pygments/lexers/monte.py +203 -0
  149. data/vendor/pygments-main/pygments/lexers/ncl.py +1053 -0
  150. data/vendor/pygments-main/pygments/lexers/nimrod.py +159 -0
  151. data/vendor/pygments-main/pygments/lexers/nit.py +64 -0
  152. data/vendor/pygments-main/pygments/lexers/nix.py +136 -0
  153. data/vendor/pygments-main/pygments/lexers/oberon.py +105 -0
  154. data/vendor/pygments-main/pygments/lexers/objective.py +504 -0
  155. data/vendor/pygments-main/pygments/lexers/ooc.py +85 -0
  156. data/vendor/pygments-main/pygments/lexers/other.py +30 -4481
  157. data/vendor/pygments-main/pygments/lexers/parasail.py +79 -0
  158. data/vendor/pygments-main/pygments/lexers/parsers.py +171 -114
  159. data/vendor/pygments-main/pygments/lexers/pascal.py +644 -0
  160. data/vendor/pygments-main/pygments/lexers/pawn.py +199 -0
  161. data/vendor/pygments-main/pygments/lexers/perl.py +616 -0
  162. data/vendor/pygments-main/pygments/lexers/php.py +267 -0
  163. data/vendor/pygments-main/pygments/lexers/praat.py +294 -0
  164. data/vendor/pygments-main/pygments/lexers/prolog.py +306 -0
  165. data/vendor/pygments-main/pygments/lexers/python.py +938 -0
  166. data/vendor/pygments-main/pygments/lexers/qvt.py +152 -0
  167. data/vendor/pygments-main/pygments/lexers/r.py +453 -0
  168. data/vendor/pygments-main/pygments/lexers/rdf.py +195 -24
  169. data/vendor/pygments-main/pygments/lexers/rebol.py +431 -0
  170. data/vendor/pygments-main/pygments/lexers/resource.py +85 -0
  171. data/vendor/pygments-main/pygments/lexers/rnc.py +67 -0
  172. data/vendor/pygments-main/pygments/lexers/roboconf.py +82 -0
  173. data/vendor/pygments-main/pygments/lexers/{_robotframeworklexer.py → robotframework.py} +20 -18
  174. data/vendor/pygments-main/pygments/lexers/ruby.py +519 -0
  175. data/vendor/pygments-main/pygments/lexers/rust.py +209 -0
  176. data/vendor/pygments-main/pygments/lexers/sas.py +228 -0
  177. data/vendor/pygments-main/pygments/lexers/scripting.py +1222 -0
  178. data/vendor/pygments-main/pygments/lexers/shell.py +478 -115
  179. data/vendor/pygments-main/pygments/lexers/smalltalk.py +195 -0
  180. data/vendor/pygments-main/pygments/lexers/smv.py +75 -0
  181. data/vendor/pygments-main/pygments/lexers/snobol.py +83 -0
  182. data/vendor/pygments-main/pygments/lexers/special.py +6 -2
  183. data/vendor/pygments-main/pygments/lexers/sql.py +209 -120
  184. data/vendor/pygments-main/pygments/lexers/stata.py +106 -0
  185. data/vendor/pygments-main/pygments/lexers/supercollider.py +90 -0
  186. data/vendor/pygments-main/pygments/lexers/tcl.py +145 -0
  187. data/vendor/pygments-main/pygments/lexers/templates.py +282 -90
  188. data/vendor/pygments-main/pygments/lexers/testing.py +207 -0
  189. data/vendor/pygments-main/pygments/lexers/text.py +15 -2045
  190. data/vendor/pygments-main/pygments/lexers/textedit.py +169 -0
  191. data/vendor/pygments-main/pygments/lexers/textfmts.py +297 -0
  192. data/vendor/pygments-main/pygments/lexers/theorem.py +458 -0
  193. data/vendor/pygments-main/pygments/lexers/trafficscript.py +54 -0
  194. data/vendor/pygments-main/pygments/lexers/typoscript.py +225 -0
  195. data/vendor/pygments-main/pygments/lexers/urbi.py +133 -0
  196. data/vendor/pygments-main/pygments/lexers/varnish.py +190 -0
  197. data/vendor/pygments-main/pygments/lexers/verification.py +111 -0
  198. data/vendor/pygments-main/pygments/lexers/web.py +13 -4499
  199. data/vendor/pygments-main/pygments/lexers/webmisc.py +979 -0
  200. data/vendor/pygments-main/pygments/lexers/whiley.py +117 -0
  201. data/vendor/pygments-main/pygments/lexers/x10.py +69 -0
  202. data/vendor/pygments-main/pygments/modeline.py +5 -2
  203. data/vendor/pygments-main/pygments/plugin.py +1 -1
  204. data/vendor/pygments-main/pygments/regexopt.py +92 -0
  205. data/vendor/pygments-main/pygments/scanner.py +3 -2
  206. data/vendor/pygments-main/pygments/sphinxext.py +11 -6
  207. data/vendor/pygments-main/pygments/style.py +41 -4
  208. data/vendor/pygments-main/pygments/styles/__init__.py +5 -1
  209. data/vendor/pygments-main/pygments/styles/algol.py +63 -0
  210. data/vendor/pygments-main/pygments/styles/algol_nu.py +63 -0
  211. data/vendor/pygments-main/pygments/styles/arduino.py +98 -0
  212. data/vendor/pygments-main/pygments/styles/autumn.py +1 -1
  213. data/vendor/pygments-main/pygments/styles/borland.py +1 -1
  214. data/vendor/pygments-main/pygments/styles/bw.py +1 -1
  215. data/vendor/pygments-main/pygments/styles/colorful.py +1 -1
  216. data/vendor/pygments-main/pygments/styles/default.py +1 -1
  217. data/vendor/pygments-main/pygments/styles/emacs.py +1 -1
  218. data/vendor/pygments-main/pygments/styles/friendly.py +1 -1
  219. data/vendor/pygments-main/pygments/styles/fruity.py +1 -1
  220. data/vendor/pygments-main/pygments/styles/igor.py +1 -1
  221. data/vendor/pygments-main/pygments/styles/lovelace.py +97 -0
  222. data/vendor/pygments-main/pygments/styles/manni.py +1 -1
  223. data/vendor/pygments-main/pygments/styles/monokai.py +1 -1
  224. data/vendor/pygments-main/pygments/styles/murphy.py +1 -1
  225. data/vendor/pygments-main/pygments/styles/native.py +1 -1
  226. data/vendor/pygments-main/pygments/styles/paraiso_dark.py +1 -1
  227. data/vendor/pygments-main/pygments/styles/paraiso_light.py +1 -1
  228. data/vendor/pygments-main/pygments/styles/pastie.py +1 -1
  229. data/vendor/pygments-main/pygments/styles/perldoc.py +2 -2
  230. data/vendor/pygments-main/pygments/styles/rrt.py +1 -1
  231. data/vendor/pygments-main/pygments/styles/sas.py +41 -0
  232. data/vendor/pygments-main/pygments/styles/stata.py +37 -0
  233. data/vendor/pygments-main/pygments/styles/tango.py +1 -1
  234. data/vendor/pygments-main/pygments/styles/trac.py +1 -1
  235. data/vendor/pygments-main/pygments/styles/vim.py +1 -1
  236. data/vendor/pygments-main/pygments/styles/vs.py +1 -1
  237. data/vendor/pygments-main/pygments/styles/xcode.py +1 -1
  238. data/vendor/pygments-main/pygments/token.py +31 -16
  239. data/vendor/pygments-main/pygments/unistring.py +141 -65
  240. data/vendor/pygments-main/pygments/util.py +129 -33
  241. data/vendor/pygments-main/requirements.txt +5 -0
  242. data/vendor/pygments-main/scripts/check_sources.py +25 -40
  243. data/vendor/pygments-main/scripts/debug_lexer.py +246 -0
  244. data/vendor/pygments-main/scripts/find_error.py +1 -0
  245. data/vendor/pygments-main/scripts/get_vimkw.py +35 -4
  246. data/vendor/pygments-main/setup.cfg +3 -0
  247. data/vendor/pygments-main/setup.py +17 -30
  248. data/vendor/pygments-main/tox.ini +7 -0
  249. metadata +159 -387
  250. data/vendor/pygments-main/ez_setup.py +0 -382
  251. data/vendor/pygments-main/pygments/lexers/_cocoabuiltins.py +0 -73
  252. data/vendor/pygments-main/pygments/lexers/_openedgebuiltins.py +0 -562
  253. data/vendor/pygments-main/pygments/lexers/_sourcemodbuiltins.py +0 -1077
  254. data/vendor/pygments-main/pygments/lexers/_vimbuiltins.py +0 -13
  255. data/vendor/pygments-main/pygments/lexers/qbasic.py +0 -157
  256. data/vendor/pygments-main/scripts/find_codetags.py +0 -213
  257. data/vendor/pygments-main/scripts/find_error.py +0 -173
  258. data/vendor/pygments-main/tests/dtds/HTML4-f.dtd +0 -37
  259. data/vendor/pygments-main/tests/dtds/HTML4-s.dtd +0 -869
  260. data/vendor/pygments-main/tests/dtds/HTML4.dcl +0 -88
  261. data/vendor/pygments-main/tests/dtds/HTML4.dtd +0 -1092
  262. data/vendor/pygments-main/tests/dtds/HTML4.soc +0 -9
  263. data/vendor/pygments-main/tests/dtds/HTMLlat1.ent +0 -195
  264. data/vendor/pygments-main/tests/dtds/HTMLspec.ent +0 -77
  265. data/vendor/pygments-main/tests/dtds/HTMLsym.ent +0 -241
  266. data/vendor/pygments-main/tests/examplefiles/99_bottles_of_beer.chpl +0 -118
  267. data/vendor/pygments-main/tests/examplefiles/AcidStateAdvanced.hs +0 -209
  268. data/vendor/pygments-main/tests/examplefiles/AlternatingGroup.mu +0 -102
  269. data/vendor/pygments-main/tests/examplefiles/BOM.js +0 -1
  270. data/vendor/pygments-main/tests/examplefiles/CPDictionary.j +0 -611
  271. data/vendor/pygments-main/tests/examplefiles/Config.in.cache +0 -1973
  272. data/vendor/pygments-main/tests/examplefiles/Constants.mo +0 -158
  273. data/vendor/pygments-main/tests/examplefiles/DancingSudoku.lhs +0 -411
  274. data/vendor/pygments-main/tests/examplefiles/Deflate.fs +0 -578
  275. data/vendor/pygments-main/tests/examplefiles/Error.pmod +0 -38
  276. data/vendor/pygments-main/tests/examplefiles/Errors.scala +0 -18
  277. data/vendor/pygments-main/tests/examplefiles/FakeFile.pike +0 -360
  278. data/vendor/pygments-main/tests/examplefiles/Get-CommandDefinitionHtml.ps1 +0 -66
  279. data/vendor/pygments-main/tests/examplefiles/IPDispatchC.nc +0 -104
  280. data/vendor/pygments-main/tests/examplefiles/IPDispatchP.nc +0 -671
  281. data/vendor/pygments-main/tests/examplefiles/Intro.java +0 -1660
  282. data/vendor/pygments-main/tests/examplefiles/Makefile +0 -1131
  283. data/vendor/pygments-main/tests/examplefiles/Object.st +0 -4394
  284. data/vendor/pygments-main/tests/examplefiles/OrderedMap.hx +0 -584
  285. data/vendor/pygments-main/tests/examplefiles/RoleQ.pm6 +0 -23
  286. data/vendor/pygments-main/tests/examplefiles/SmallCheck.hs +0 -378
  287. data/vendor/pygments-main/tests/examplefiles/Sorting.mod +0 -470
  288. data/vendor/pygments-main/tests/examplefiles/Sudoku.lhs +0 -382
  289. data/vendor/pygments-main/tests/examplefiles/addressbook.proto +0 -30
  290. data/vendor/pygments-main/tests/examplefiles/antlr_ANTLRv3.g +0 -608
  291. data/vendor/pygments-main/tests/examplefiles/antlr_throws +0 -1
  292. data/vendor/pygments-main/tests/examplefiles/apache2.conf +0 -393
  293. data/vendor/pygments-main/tests/examplefiles/as3_test.as +0 -143
  294. data/vendor/pygments-main/tests/examplefiles/as3_test2.as +0 -46
  295. data/vendor/pygments-main/tests/examplefiles/as3_test3.as +0 -3
  296. data/vendor/pygments-main/tests/examplefiles/aspx-cs_example +0 -27
  297. data/vendor/pygments-main/tests/examplefiles/autoit_submit.au3 +0 -25
  298. data/vendor/pygments-main/tests/examplefiles/badcase.java +0 -2
  299. data/vendor/pygments-main/tests/examplefiles/batchfile.bat +0 -49
  300. data/vendor/pygments-main/tests/examplefiles/bigtest.nsi +0 -308
  301. data/vendor/pygments-main/tests/examplefiles/boot-9.scm +0 -1557
  302. data/vendor/pygments-main/tests/examplefiles/ca65_example +0 -284
  303. data/vendor/pygments-main/tests/examplefiles/cbmbas_example +0 -9
  304. data/vendor/pygments-main/tests/examplefiles/cells.ps +0 -515
  305. data/vendor/pygments-main/tests/examplefiles/ceval.c +0 -2604
  306. data/vendor/pygments-main/tests/examplefiles/cheetah_example.html +0 -13
  307. data/vendor/pygments-main/tests/examplefiles/classes.dylan +0 -125
  308. data/vendor/pygments-main/tests/examplefiles/clojure-weird-keywords.clj +0 -5
  309. data/vendor/pygments-main/tests/examplefiles/condensed_ruby.rb +0 -10
  310. data/vendor/pygments-main/tests/examplefiles/coq_RelationClasses +0 -447
  311. data/vendor/pygments-main/tests/examplefiles/core.cljs +0 -52
  312. data/vendor/pygments-main/tests/examplefiles/database.pytb +0 -20
  313. data/vendor/pygments-main/tests/examplefiles/de.MoinMoin.po +0 -2461
  314. data/vendor/pygments-main/tests/examplefiles/demo.ahk +0 -181
  315. data/vendor/pygments-main/tests/examplefiles/demo.cfm +0 -50
  316. data/vendor/pygments-main/tests/examplefiles/demo.hbs +0 -12
  317. data/vendor/pygments-main/tests/examplefiles/django_sample.html+django +0 -68
  318. data/vendor/pygments-main/tests/examplefiles/dwarf.cw +0 -17
  319. data/vendor/pygments-main/tests/examplefiles/ember.handlebars +0 -33
  320. data/vendor/pygments-main/tests/examplefiles/erl_session +0 -10
  321. data/vendor/pygments-main/tests/examplefiles/escape_semicolon.clj +0 -1
  322. data/vendor/pygments-main/tests/examplefiles/evil_regex.js +0 -48
  323. data/vendor/pygments-main/tests/examplefiles/example.Rd +0 -78
  324. data/vendor/pygments-main/tests/examplefiles/example.als +0 -217
  325. data/vendor/pygments-main/tests/examplefiles/example.bug +0 -54
  326. data/vendor/pygments-main/tests/examplefiles/example.c +0 -2080
  327. data/vendor/pygments-main/tests/examplefiles/example.ceylon +0 -52
  328. data/vendor/pygments-main/tests/examplefiles/example.chai +0 -6
  329. data/vendor/pygments-main/tests/examplefiles/example.clay +0 -33
  330. data/vendor/pygments-main/tests/examplefiles/example.cls +0 -15
  331. data/vendor/pygments-main/tests/examplefiles/example.cob +0 -3556
  332. data/vendor/pygments-main/tests/examplefiles/example.coffee +0 -27
  333. data/vendor/pygments-main/tests/examplefiles/example.cpp +0 -2363
  334. data/vendor/pygments-main/tests/examplefiles/example.e +0 -124
  335. data/vendor/pygments-main/tests/examplefiles/example.f90 +0 -8
  336. data/vendor/pygments-main/tests/examplefiles/example.feature +0 -16
  337. data/vendor/pygments-main/tests/examplefiles/example.gd +0 -23
  338. data/vendor/pygments-main/tests/examplefiles/example.gi +0 -64
  339. data/vendor/pygments-main/tests/examplefiles/example.groovy +0 -2
  340. data/vendor/pygments-main/tests/examplefiles/example.gs +0 -106
  341. data/vendor/pygments-main/tests/examplefiles/example.gst +0 -7
  342. data/vendor/pygments-main/tests/examplefiles/example.hs +0 -27
  343. data/vendor/pygments-main/tests/examplefiles/example.hx +0 -185
  344. data/vendor/pygments-main/tests/examplefiles/example.i6t +0 -32
  345. data/vendor/pygments-main/tests/examplefiles/example.i7x +0 -45
  346. data/vendor/pygments-main/tests/examplefiles/example.inf +0 -374
  347. data/vendor/pygments-main/tests/examplefiles/example.j +0 -564
  348. data/vendor/pygments-main/tests/examplefiles/example.jag +0 -48
  349. data/vendor/pygments-main/tests/examplefiles/example.java +0 -16
  350. data/vendor/pygments-main/tests/examplefiles/example.kal +0 -75
  351. data/vendor/pygments-main/tests/examplefiles/example.kt +0 -47
  352. data/vendor/pygments-main/tests/examplefiles/example.lagda +0 -19
  353. data/vendor/pygments-main/tests/examplefiles/example.liquid +0 -42
  354. data/vendor/pygments-main/tests/examplefiles/example.lua +0 -250
  355. data/vendor/pygments-main/tests/examplefiles/example.ma +0 -8
  356. data/vendor/pygments-main/tests/examplefiles/example.monkey +0 -152
  357. data/vendor/pygments-main/tests/examplefiles/example.moo +0 -26
  358. data/vendor/pygments-main/tests/examplefiles/example.moon +0 -629
  359. data/vendor/pygments-main/tests/examplefiles/example.mq4 +0 -187
  360. data/vendor/pygments-main/tests/examplefiles/example.mqh +0 -123
  361. data/vendor/pygments-main/tests/examplefiles/example.msc +0 -43
  362. data/vendor/pygments-main/tests/examplefiles/example.ni +0 -57
  363. data/vendor/pygments-main/tests/examplefiles/example.nim +0 -1010
  364. data/vendor/pygments-main/tests/examplefiles/example.nix +0 -80
  365. data/vendor/pygments-main/tests/examplefiles/example.ns2 +0 -69
  366. data/vendor/pygments-main/tests/examplefiles/example.pas +0 -2708
  367. data/vendor/pygments-main/tests/examplefiles/example.pp +0 -8
  368. data/vendor/pygments-main/tests/examplefiles/example.prg +0 -161
  369. data/vendor/pygments-main/tests/examplefiles/example.rb +0 -1852
  370. data/vendor/pygments-main/tests/examplefiles/example.red +0 -257
  371. data/vendor/pygments-main/tests/examplefiles/example.reds +0 -150
  372. data/vendor/pygments-main/tests/examplefiles/example.reg +0 -19
  373. data/vendor/pygments-main/tests/examplefiles/example.rexx +0 -50
  374. data/vendor/pygments-main/tests/examplefiles/example.rhtml +0 -561
  375. data/vendor/pygments-main/tests/examplefiles/example.rkt +0 -743
  376. data/vendor/pygments-main/tests/examplefiles/example.rpf +0 -4
  377. data/vendor/pygments-main/tests/examplefiles/example.sh +0 -22
  378. data/vendor/pygments-main/tests/examplefiles/example.sh-session +0 -19
  379. data/vendor/pygments-main/tests/examplefiles/example.shell-session +0 -45
  380. data/vendor/pygments-main/tests/examplefiles/example.slim +0 -31
  381. data/vendor/pygments-main/tests/examplefiles/example.sls +0 -51
  382. data/vendor/pygments-main/tests/examplefiles/example.sml +0 -156
  383. data/vendor/pygments-main/tests/examplefiles/example.snobol +0 -15
  384. data/vendor/pygments-main/tests/examplefiles/example.stan +0 -110
  385. data/vendor/pygments-main/tests/examplefiles/example.tea +0 -34
  386. data/vendor/pygments-main/tests/examplefiles/example.todotxt +0 -9
  387. data/vendor/pygments-main/tests/examplefiles/example.ts +0 -28
  388. data/vendor/pygments-main/tests/examplefiles/example.u +0 -548
  389. data/vendor/pygments-main/tests/examplefiles/example.weechatlog +0 -9
  390. data/vendor/pygments-main/tests/examplefiles/example.xhtml +0 -376
  391. data/vendor/pygments-main/tests/examplefiles/example.xtend +0 -34
  392. data/vendor/pygments-main/tests/examplefiles/example.yaml +0 -302
  393. data/vendor/pygments-main/tests/examplefiles/example2.aspx +0 -29
  394. data/vendor/pygments-main/tests/examplefiles/example2.msc +0 -79
  395. data/vendor/pygments-main/tests/examplefiles/exampleScript.cfc +0 -241
  396. data/vendor/pygments-main/tests/examplefiles/exampleTag.cfc +0 -18
  397. data/vendor/pygments-main/tests/examplefiles/example_elixir.ex +0 -205
  398. data/vendor/pygments-main/tests/examplefiles/example_file.fy +0 -128
  399. data/vendor/pygments-main/tests/examplefiles/firefox.mak +0 -586
  400. data/vendor/pygments-main/tests/examplefiles/flipflop.sv +0 -19
  401. data/vendor/pygments-main/tests/examplefiles/foo.sce +0 -6
  402. data/vendor/pygments-main/tests/examplefiles/format.ml +0 -1213
  403. data/vendor/pygments-main/tests/examplefiles/fucked_up.rb +0 -77
  404. data/vendor/pygments-main/tests/examplefiles/function.mu +0 -1
  405. data/vendor/pygments-main/tests/examplefiles/functional.rst +0 -1472
  406. data/vendor/pygments-main/tests/examplefiles/garcia-wachs.kk +0 -133
  407. data/vendor/pygments-main/tests/examplefiles/genclass.clj +0 -510
  408. data/vendor/pygments-main/tests/examplefiles/genshi_example.xml+genshi +0 -193
  409. data/vendor/pygments-main/tests/examplefiles/genshitext_example.genshitext +0 -33
  410. data/vendor/pygments-main/tests/examplefiles/glsl.frag +0 -7
  411. data/vendor/pygments-main/tests/examplefiles/glsl.vert +0 -13
  412. data/vendor/pygments-main/tests/examplefiles/grammar-test.p6 +0 -22
  413. data/vendor/pygments-main/tests/examplefiles/hash_syntax.rb +0 -5
  414. data/vendor/pygments-main/tests/examplefiles/hello.at +0 -6
  415. data/vendor/pygments-main/tests/examplefiles/hello.golo +0 -5
  416. data/vendor/pygments-main/tests/examplefiles/hello.lsl +0 -12
  417. data/vendor/pygments-main/tests/examplefiles/hello.smali +0 -40
  418. data/vendor/pygments-main/tests/examplefiles/hello.sp +0 -9
  419. data/vendor/pygments-main/tests/examplefiles/html+php_faulty.php +0 -1
  420. data/vendor/pygments-main/tests/examplefiles/http_request_example +0 -15
  421. data/vendor/pygments-main/tests/examplefiles/http_response_example +0 -29
  422. data/vendor/pygments-main/tests/examplefiles/hybris_File.hy +0 -174
  423. data/vendor/pygments-main/tests/examplefiles/idl_sample.pro +0 -73
  424. data/vendor/pygments-main/tests/examplefiles/iex_example +0 -23
  425. data/vendor/pygments-main/tests/examplefiles/inet_pton6.dg +0 -71
  426. data/vendor/pygments-main/tests/examplefiles/intro.ik +0 -24
  427. data/vendor/pygments-main/tests/examplefiles/ints.php +0 -10
  428. data/vendor/pygments-main/tests/examplefiles/intsyn.fun +0 -675
  429. data/vendor/pygments-main/tests/examplefiles/intsyn.sig +0 -286
  430. data/vendor/pygments-main/tests/examplefiles/irb_heredoc +0 -8
  431. data/vendor/pygments-main/tests/examplefiles/irc.lsp +0 -214
  432. data/vendor/pygments-main/tests/examplefiles/java.properties +0 -16
  433. data/vendor/pygments-main/tests/examplefiles/jbst_example1.jbst +0 -28
  434. data/vendor/pygments-main/tests/examplefiles/jbst_example2.jbst +0 -45
  435. data/vendor/pygments-main/tests/examplefiles/jinjadesignerdoc.rst +0 -713
  436. data/vendor/pygments-main/tests/examplefiles/json.lasso +0 -301
  437. data/vendor/pygments-main/tests/examplefiles/json.lasso9 +0 -213
  438. data/vendor/pygments-main/tests/examplefiles/language.hy +0 -165
  439. data/vendor/pygments-main/tests/examplefiles/lighttpd_config.conf +0 -13
  440. data/vendor/pygments-main/tests/examplefiles/limbo.b +0 -456
  441. data/vendor/pygments-main/tests/examplefiles/linecontinuation.py +0 -47
  442. data/vendor/pygments-main/tests/examplefiles/livescript-demo.ls +0 -43
  443. data/vendor/pygments-main/tests/examplefiles/logos_example.xm +0 -28
  444. data/vendor/pygments-main/tests/examplefiles/ltmain.sh +0 -2849
  445. data/vendor/pygments-main/tests/examplefiles/main.cmake +0 -44
  446. data/vendor/pygments-main/tests/examplefiles/markdown.lsp +0 -679
  447. data/vendor/pygments-main/tests/examplefiles/matlab_noreturn +0 -3
  448. data/vendor/pygments-main/tests/examplefiles/matlab_sample +0 -30
  449. data/vendor/pygments-main/tests/examplefiles/matlabsession_sample.txt +0 -37
  450. data/vendor/pygments-main/tests/examplefiles/metagrammar.treetop +0 -455
  451. data/vendor/pygments-main/tests/examplefiles/minehunt.qml +0 -112
  452. data/vendor/pygments-main/tests/examplefiles/minimal.ns2 +0 -4
  453. data/vendor/pygments-main/tests/examplefiles/moin_SyntaxReference.txt +0 -340
  454. data/vendor/pygments-main/tests/examplefiles/multiline_regexes.rb +0 -38
  455. data/vendor/pygments-main/tests/examplefiles/nanomsg.intr +0 -95
  456. data/vendor/pygments-main/tests/examplefiles/nasm_aoutso.asm +0 -96
  457. data/vendor/pygments-main/tests/examplefiles/nasm_objexe.asm +0 -30
  458. data/vendor/pygments-main/tests/examplefiles/nemerle_sample.n +0 -87
  459. data/vendor/pygments-main/tests/examplefiles/nginx_nginx.conf +0 -118
  460. data/vendor/pygments-main/tests/examplefiles/numbers.c +0 -12
  461. data/vendor/pygments-main/tests/examplefiles/objc_example.m +0 -179
  462. data/vendor/pygments-main/tests/examplefiles/openedge_example +0 -34
  463. data/vendor/pygments-main/tests/examplefiles/pawn_example +0 -25
  464. data/vendor/pygments-main/tests/examplefiles/perl_misc +0 -62
  465. data/vendor/pygments-main/tests/examplefiles/perl_perl5db +0 -998
  466. data/vendor/pygments-main/tests/examplefiles/perl_regex-delims +0 -120
  467. data/vendor/pygments-main/tests/examplefiles/perlfunc.1 +0 -856
  468. data/vendor/pygments-main/tests/examplefiles/phpMyAdmin.spec +0 -163
  469. data/vendor/pygments-main/tests/examplefiles/phpcomplete.vim +0 -567
  470. data/vendor/pygments-main/tests/examplefiles/pleac.in.rb +0 -1223
  471. data/vendor/pygments-main/tests/examplefiles/postgresql_test.txt +0 -47
  472. data/vendor/pygments-main/tests/examplefiles/pppoe.applescript +0 -10
  473. data/vendor/pygments-main/tests/examplefiles/psql_session.txt +0 -122
  474. data/vendor/pygments-main/tests/examplefiles/py3_test.txt +0 -2
  475. data/vendor/pygments-main/tests/examplefiles/py3tb_test.py3tb +0 -4
  476. data/vendor/pygments-main/tests/examplefiles/pycon_test.pycon +0 -14
  477. data/vendor/pygments-main/tests/examplefiles/pytb_test2.pytb +0 -2
  478. data/vendor/pygments-main/tests/examplefiles/pytb_test3.pytb +0 -4
  479. data/vendor/pygments-main/tests/examplefiles/python25-bsd.mak +0 -234
  480. data/vendor/pygments-main/tests/examplefiles/qbasic_example +0 -2
  481. data/vendor/pygments-main/tests/examplefiles/qsort.prolog +0 -13
  482. data/vendor/pygments-main/tests/examplefiles/r-console-transcript.Rout +0 -38
  483. data/vendor/pygments-main/tests/examplefiles/r6rs-comments.scm +0 -23
  484. data/vendor/pygments-main/tests/examplefiles/ragel-cpp_rlscan +0 -280
  485. data/vendor/pygments-main/tests/examplefiles/ragel-cpp_snippet +0 -2
  486. data/vendor/pygments-main/tests/examplefiles/regex.js +0 -22
  487. data/vendor/pygments-main/tests/examplefiles/reversi.lsp +0 -427
  488. data/vendor/pygments-main/tests/examplefiles/robotframework_test.txt +0 -39
  489. data/vendor/pygments-main/tests/examplefiles/rql-queries.rql +0 -34
  490. data/vendor/pygments-main/tests/examplefiles/ruby_func_def.rb +0 -11
  491. data/vendor/pygments-main/tests/examplefiles/rust_example.rs +0 -233
  492. data/vendor/pygments-main/tests/examplefiles/scilab.sci +0 -30
  493. data/vendor/pygments-main/tests/examplefiles/scope.cirru +0 -43
  494. data/vendor/pygments-main/tests/examplefiles/session.dylan-console +0 -9
  495. data/vendor/pygments-main/tests/examplefiles/sibling.prolog +0 -19
  496. data/vendor/pygments-main/tests/examplefiles/simple.md +0 -747
  497. data/vendor/pygments-main/tests/examplefiles/smarty_example.html +0 -209
  498. data/vendor/pygments-main/tests/examplefiles/source.lgt +0 -343
  499. data/vendor/pygments-main/tests/examplefiles/sources.list +0 -62
  500. data/vendor/pygments-main/tests/examplefiles/sparql.rq +0 -23
  501. data/vendor/pygments-main/tests/examplefiles/sphere.pov +0 -18
  502. data/vendor/pygments-main/tests/examplefiles/sqlite3.sqlite3-console +0 -27
  503. data/vendor/pygments-main/tests/examplefiles/squid.conf +0 -30
  504. data/vendor/pygments-main/tests/examplefiles/string.jl +0 -1031
  505. data/vendor/pygments-main/tests/examplefiles/string_delimiters.d +0 -21
  506. data/vendor/pygments-main/tests/examplefiles/stripheredoc.sh +0 -3
  507. data/vendor/pygments-main/tests/examplefiles/swig_java.swg +0 -1329
  508. data/vendor/pygments-main/tests/examplefiles/swig_std_vector.i +0 -225
  509. data/vendor/pygments-main/tests/examplefiles/test.R +0 -185
  510. data/vendor/pygments-main/tests/examplefiles/test.adb +0 -211
  511. data/vendor/pygments-main/tests/examplefiles/test.agda +0 -102
  512. data/vendor/pygments-main/tests/examplefiles/test.apl +0 -26
  513. data/vendor/pygments-main/tests/examplefiles/test.asy +0 -131
  514. data/vendor/pygments-main/tests/examplefiles/test.awk +0 -121
  515. data/vendor/pygments-main/tests/examplefiles/test.bb +0 -95
  516. data/vendor/pygments-main/tests/examplefiles/test.bmx +0 -145
  517. data/vendor/pygments-main/tests/examplefiles/test.boo +0 -39
  518. data/vendor/pygments-main/tests/examplefiles/test.bro +0 -250
  519. data/vendor/pygments-main/tests/examplefiles/test.cs +0 -374
  520. data/vendor/pygments-main/tests/examplefiles/test.css +0 -54
  521. data/vendor/pygments-main/tests/examplefiles/test.cu +0 -36
  522. data/vendor/pygments-main/tests/examplefiles/test.cyp +0 -123
  523. data/vendor/pygments-main/tests/examplefiles/test.d +0 -135
  524. data/vendor/pygments-main/tests/examplefiles/test.dart +0 -23
  525. data/vendor/pygments-main/tests/examplefiles/test.dtd +0 -89
  526. data/vendor/pygments-main/tests/examplefiles/test.ebnf +0 -31
  527. data/vendor/pygments-main/tests/examplefiles/test.ec +0 -605
  528. data/vendor/pygments-main/tests/examplefiles/test.ecl +0 -58
  529. data/vendor/pygments-main/tests/examplefiles/test.eh +0 -315
  530. data/vendor/pygments-main/tests/examplefiles/test.erl +0 -169
  531. data/vendor/pygments-main/tests/examplefiles/test.evoque +0 -33
  532. data/vendor/pygments-main/tests/examplefiles/test.fan +0 -818
  533. data/vendor/pygments-main/tests/examplefiles/test.flx +0 -57
  534. data/vendor/pygments-main/tests/examplefiles/test.gdc +0 -13
  535. data/vendor/pygments-main/tests/examplefiles/test.groovy +0 -97
  536. data/vendor/pygments-main/tests/examplefiles/test.html +0 -339
  537. data/vendor/pygments-main/tests/examplefiles/test.idr +0 -93
  538. data/vendor/pygments-main/tests/examplefiles/test.ini +0 -10
  539. data/vendor/pygments-main/tests/examplefiles/test.java +0 -653
  540. data/vendor/pygments-main/tests/examplefiles/test.jsp +0 -24
  541. data/vendor/pygments-main/tests/examplefiles/test.maql +0 -45
  542. data/vendor/pygments-main/tests/examplefiles/test.mask +0 -41
  543. data/vendor/pygments-main/tests/examplefiles/test.mod +0 -374
  544. data/vendor/pygments-main/tests/examplefiles/test.moo +0 -51
  545. data/vendor/pygments-main/tests/examplefiles/test.myt +0 -166
  546. data/vendor/pygments-main/tests/examplefiles/test.nim +0 -93
  547. data/vendor/pygments-main/tests/examplefiles/test.opa +0 -10
  548. data/vendor/pygments-main/tests/examplefiles/test.p6 +0 -252
  549. data/vendor/pygments-main/tests/examplefiles/test.pan +0 -54
  550. data/vendor/pygments-main/tests/examplefiles/test.pas +0 -743
  551. data/vendor/pygments-main/tests/examplefiles/test.php +0 -507
  552. data/vendor/pygments-main/tests/examplefiles/test.pig +0 -148
  553. data/vendor/pygments-main/tests/examplefiles/test.plot +0 -333
  554. data/vendor/pygments-main/tests/examplefiles/test.ps1 +0 -108
  555. data/vendor/pygments-main/tests/examplefiles/test.pwn +0 -253
  556. data/vendor/pygments-main/tests/examplefiles/test.pypylog +0 -1839
  557. data/vendor/pygments-main/tests/examplefiles/test.r3 +0 -114
  558. data/vendor/pygments-main/tests/examplefiles/test.rb +0 -177
  559. data/vendor/pygments-main/tests/examplefiles/test.rhtml +0 -43
  560. data/vendor/pygments-main/tests/examplefiles/test.rsl +0 -111
  561. data/vendor/pygments-main/tests/examplefiles/test.scaml +0 -8
  562. data/vendor/pygments-main/tests/examplefiles/test.ssp +0 -12
  563. data/vendor/pygments-main/tests/examplefiles/test.swift +0 -65
  564. data/vendor/pygments-main/tests/examplefiles/test.tcsh +0 -830
  565. data/vendor/pygments-main/tests/examplefiles/test.vb +0 -407
  566. data/vendor/pygments-main/tests/examplefiles/test.vhdl +0 -161
  567. data/vendor/pygments-main/tests/examplefiles/test.xqy +0 -138
  568. data/vendor/pygments-main/tests/examplefiles/test.xsl +0 -23
  569. data/vendor/pygments-main/tests/examplefiles/test.zep +0 -33
  570. data/vendor/pygments-main/tests/examplefiles/test2.pypylog +0 -120
  571. data/vendor/pygments-main/tests/examplefiles/truncated.pytb +0 -15
  572. data/vendor/pygments-main/tests/examplefiles/type.lisp +0 -1218
  573. data/vendor/pygments-main/tests/examplefiles/underscore.coffee +0 -603
  574. data/vendor/pygments-main/tests/examplefiles/unicode.applescript +0 -5
  575. data/vendor/pygments-main/tests/examplefiles/unicode.go +0 -10
  576. data/vendor/pygments-main/tests/examplefiles/unicodedoc.py +0 -11
  577. data/vendor/pygments-main/tests/examplefiles/unix-io.lid +0 -37
  578. data/vendor/pygments-main/tests/examplefiles/vbnet_test.bas +0 -29
  579. data/vendor/pygments-main/tests/examplefiles/vctreestatus_hg +0 -4
  580. data/vendor/pygments-main/tests/examplefiles/vimrc +0 -21
  581. data/vendor/pygments-main/tests/examplefiles/webkit-transition.css +0 -3
  582. data/vendor/pygments-main/tests/examplefiles/while.pov +0 -13
  583. data/vendor/pygments-main/tests/examplefiles/wiki.factor +0 -384
  584. data/vendor/pygments-main/tests/examplefiles/xml_example +0 -1897
  585. data/vendor/pygments-main/tests/examplefiles/zmlrpc.f90 +0 -798
  586. data/vendor/pygments-main/tests/run.py +0 -44
  587. data/vendor/pygments-main/tests/string_asserts.py +0 -22
  588. data/vendor/pygments-main/tests/support.py +0 -15
  589. data/vendor/pygments-main/tests/support/tags +0 -36
  590. data/vendor/pygments-main/tests/test_basic_api.py +0 -309
  591. data/vendor/pygments-main/tests/test_cfm.py +0 -46
  592. data/vendor/pygments-main/tests/test_clexer.py +0 -236
  593. data/vendor/pygments-main/tests/test_cmdline.py +0 -106
  594. data/vendor/pygments-main/tests/test_examplefiles.py +0 -110
  595. data/vendor/pygments-main/tests/test_html_formatter.py +0 -180
  596. data/vendor/pygments-main/tests/test_latex_formatter.py +0 -57
  597. data/vendor/pygments-main/tests/test_lexers_other.py +0 -68
  598. data/vendor/pygments-main/tests/test_objectiveclexer.py +0 -81
  599. data/vendor/pygments-main/tests/test_perllexer.py +0 -137
  600. data/vendor/pygments-main/tests/test_qbasiclexer.py +0 -43
  601. data/vendor/pygments-main/tests/test_regexlexer.py +0 -54
  602. data/vendor/pygments-main/tests/test_rtf_formatter.py +0 -109
  603. data/vendor/pygments-main/tests/test_string_asserts.py +0 -39
  604. data/vendor/pygments-main/tests/test_token.py +0 -46
  605. data/vendor/pygments-main/tests/test_using_api.py +0 -40
  606. data/vendor/pygments-main/tests/test_util.py +0 -135
@@ -0,0 +1,207 @@
1
+ # -*- coding: utf-8 -*-
2
+ """
3
+ pygments.lexers.testing
4
+ ~~~~~~~~~~~~~~~~~~~~~~~
5
+
6
+ Lexers for testing languages.
7
+
8
+ :copyright: Copyright 2006-2015 by the Pygments team, see AUTHORS.
9
+ :license: BSD, see LICENSE for details.
10
+ """
11
+
12
+ from pygments.lexer import RegexLexer, include, bygroups
13
+ from pygments.token import Comment, Keyword, Name, String, Number, Generic, Text
14
+
15
+ __all__ = ['GherkinLexer', 'TAPLexer']
16
+
17
+
18
+ class GherkinLexer(RegexLexer):
19
+ """
20
+ For `Gherkin <http://github.com/aslakhellesoy/gherkin/>` syntax.
21
+
22
+ .. versionadded:: 1.2
23
+ """
24
+ name = 'Gherkin'
25
+ aliases = ['cucumber', 'gherkin']
26
+ filenames = ['*.feature']
27
+ mimetypes = ['text/x-gherkin']
28
+
29
+ feature_keywords = u'^(기능|機能|功能|フィーチャ|خاصية|תכונה|Функціонал|Функционалност|Функционал|Фича|Особина|Могућност|Özellik|Właściwość|Tính năng|Trajto|Savybė|Požiadavka|Požadavek|Osobina|Ominaisuus|Omadus|OH HAI|Mogućnost|Mogucnost|Jellemző|Fīča|Funzionalità|Funktionalität|Funkcionalnost|Funkcionalitāte|Funcționalitate|Functionaliteit|Functionalitate|Funcionalitat|Funcionalidade|Fonctionnalité|Fitur|Feature|Egenskap|Egenskab|Crikey|Característica|Arwedd)(:)(.*)$'
30
+ feature_element_keywords = u'^(\\s*)(시나리오 개요|시나리오|배경|背景|場景大綱|場景|场景大纲|场景|劇本大綱|劇本|剧本大纲|剧本|テンプレ|シナリオテンプレート|シナリオテンプレ|シナリオアウトライン|シナリオ|سيناريو مخطط|سيناريو|الخلفية|תרחיש|תבנית תרחיש|רקע|Тарих|Сценарій|Сценарио|Сценарий структураси|Сценарий|Структура сценарію|Структура сценарија|Структура сценария|Скица|Рамка на сценарий|Пример|Предыстория|Предистория|Позадина|Передумова|Основа|Концепт|Контекст|Założenia|Wharrimean is|Tình huống|The thing of it is|Tausta|Taust|Tapausaihio|Tapaus|Szenariogrundriss|Szenario|Szablon scenariusza|Stsenaarium|Struktura scenarija|Skica|Skenario konsep|Skenario|Situācija|Senaryo taslağı|Senaryo|Scénář|Scénario|Schema dello scenario|Scenārijs pēc parauga|Scenārijs|Scenár|Scenaro|Scenariusz|Scenariul de şablon|Scenariul de sablon|Scenariu|Scenario Outline|Scenario Amlinellol|Scenario|Scenarijus|Scenarijaus šablonas|Scenarij|Scenarie|Rerefons|Raamstsenaarium|Primer|Pozadí|Pozadina|Pozadie|Plan du scénario|Plan du Scénario|Osnova scénáře|Osnova|Náčrt Scénáře|Náčrt Scenáru|Mate|MISHUN SRSLY|MISHUN|Kịch bản|Konturo de la scenaro|Kontext|Konteksts|Kontekstas|Kontekst|Koncept|Khung tình huống|Khung kịch bản|Háttér|Grundlage|Geçmiş|Forgatókönyv vázlat|Forgatókönyv|Fono|Esquema do Cenário|Esquema do Cenario|Esquema del escenario|Esquema de l\'escenari|Escenario|Escenari|Dis is what went down|Dasar|Contexto|Contexte|Contesto|Condiţii|Conditii|Cenário|Cenario|Cefndir|Bối cảnh|Blokes|Bakgrunn|Bakgrund|Baggrund|Background|B4|Antecedents|Antecedentes|All y\'all|Achtergrond|Abstrakt Scenario|Abstract Scenario)(:)(.*)$'
31
+ examples_keywords = u'^(\\s*)(예|例子|例|サンプル|امثلة|דוגמאות|Сценарији|Примери|Приклади|Мисоллар|Значения|Örnekler|Voorbeelden|Variantai|Tapaukset|Scenarios|Scenariji|Scenarijai|Příklady|Példák|Príklady|Przykłady|Primjeri|Primeri|Piemēri|Pavyzdžiai|Paraugs|Juhtumid|Exemplos|Exemples|Exemplele|Exempel|Examples|Esempi|Enghreifftiau|Ekzemploj|Eksempler|Ejemplos|EXAMPLZ|Dữ liệu|Contoh|Cobber|Beispiele)(:)(.*)$'
32
+ step_keywords = u'^(\\s*)(하지만|조건|먼저|만일|만약|단|그리고|그러면|那麼|那么|而且|當|当|前提|假設|假设|假如|假定|但是|但し|並且|并且|同時|同时|もし|ならば|ただし|しかし|かつ|و |متى |لكن |عندما |ثم |بفرض |اذاً |כאשר |וגם |בהינתן |אזי |אז |אבל |Якщо |Унда |То |Припустимо, що |Припустимо |Онда |Но |Нехай |Лекин |Когато |Када |Кад |К тому же |И |Задато |Задати |Задате |Если |Допустим |Дадено |Ва |Бирок |Аммо |Али |Але |Агар |А |І |Și |És |Zatati |Zakładając |Zadato |Zadate |Zadano |Zadani |Zadan |Youse know when youse got |Youse know like when |Yna |Ya know how |Ya gotta |Y |Wun |Wtedy |When y\'all |When |Wenn |WEN |Và |Ve |Und |Un |Thì |Then y\'all |Then |Tapi |Tak |Tada |Tad |Så |Stel |Soit |Siis |Si |Sed |Se |Quando |Quand |Quan |Pryd |Pokud |Pokiaľ |Però |Pero |Pak |Oraz |Onda |Ond |Oletetaan |Og |Och |O zaman |Når |När |Niin |Nhưng |N |Mutta |Men |Mas |Maka |Majd |Mais |Maar |Ma |Lorsque |Lorsqu\'|Kun |Kuid |Kui |Khi |Keď |Ketika |Když |Kaj |Kai |Kada |Kad |Jeżeli |Ja |Ir |I CAN HAZ |I |Ha |Givun |Givet |Given y\'all |Given |Gitt |Gegeven |Gegeben sei |Fakat |Eğer ki |Etant donné |Et |Então |Entonces |Entao |En |Eeldades |E |Duota |Dun |Donitaĵo |Donat |Donada |Do |Diyelim ki |Dengan |Den youse gotta |De |Dato |Dar |Dann |Dan |Dado |Dacă |Daca |DEN |Când |Cuando |Cho |Cept |Cand |Cal |But y\'all |But |Buh |Biết |Bet |BUT |Atès |Atunci |Atesa |Anrhegedig a |Angenommen |And y\'all |And |An |Ama |Als |Alors |Allora |Ali |Aleshores |Ale |Akkor |Aber |AN |A také |A |\* )'
33
+
34
+ tokens = {
35
+ 'comments': [
36
+ (r'^\s*#.*$', Comment),
37
+ ],
38
+ 'feature_elements': [
39
+ (step_keywords, Keyword, "step_content_stack"),
40
+ include('comments'),
41
+ (r"(\s|.)", Name.Function),
42
+ ],
43
+ 'feature_elements_on_stack': [
44
+ (step_keywords, Keyword, "#pop:2"),
45
+ include('comments'),
46
+ (r"(\s|.)", Name.Function),
47
+ ],
48
+ 'examples_table': [
49
+ (r"\s+\|", Keyword, 'examples_table_header'),
50
+ include('comments'),
51
+ (r"(\s|.)", Name.Function),
52
+ ],
53
+ 'examples_table_header': [
54
+ (r"\s+\|\s*$", Keyword, "#pop:2"),
55
+ include('comments'),
56
+ (r"\\\|", Name.Variable),
57
+ (r"\s*\|", Keyword),
58
+ (r"[^|]", Name.Variable),
59
+ ],
60
+ 'scenario_sections_on_stack': [
61
+ (feature_element_keywords,
62
+ bygroups(Name.Function, Keyword, Keyword, Name.Function),
63
+ "feature_elements_on_stack"),
64
+ ],
65
+ 'narrative': [
66
+ include('scenario_sections_on_stack'),
67
+ include('comments'),
68
+ (r"(\s|.)", Name.Function),
69
+ ],
70
+ 'table_vars': [
71
+ (r'(<[^>]+>)', Name.Variable),
72
+ ],
73
+ 'numbers': [
74
+ (r'(\d+\.?\d*|\d*\.\d+)([eE][+-]?[0-9]+)?', String),
75
+ ],
76
+ 'string': [
77
+ include('table_vars'),
78
+ (r'(\s|.)', String),
79
+ ],
80
+ 'py_string': [
81
+ (r'"""', Keyword, "#pop"),
82
+ include('string'),
83
+ ],
84
+ 'step_content_root': [
85
+ (r"$", Keyword, "#pop"),
86
+ include('step_content'),
87
+ ],
88
+ 'step_content_stack': [
89
+ (r"$", Keyword, "#pop:2"),
90
+ include('step_content'),
91
+ ],
92
+ 'step_content': [
93
+ (r'"', Name.Function, "double_string"),
94
+ include('table_vars'),
95
+ include('numbers'),
96
+ include('comments'),
97
+ (r'(\s|.)', Name.Function),
98
+ ],
99
+ 'table_content': [
100
+ (r"\s+\|\s*$", Keyword, "#pop"),
101
+ include('comments'),
102
+ (r"\\\|", String),
103
+ (r"\s*\|", Keyword),
104
+ include('string'),
105
+ ],
106
+ 'double_string': [
107
+ (r'"', Name.Function, "#pop"),
108
+ include('string'),
109
+ ],
110
+ 'root': [
111
+ (r'\n', Name.Function),
112
+ include('comments'),
113
+ (r'"""', Keyword, "py_string"),
114
+ (r'\s+\|', Keyword, 'table_content'),
115
+ (r'"', Name.Function, "double_string"),
116
+ include('table_vars'),
117
+ include('numbers'),
118
+ (r'(\s*)(@[^@\r\n\t ]+)', bygroups(Name.Function, Name.Tag)),
119
+ (step_keywords, bygroups(Name.Function, Keyword),
120
+ 'step_content_root'),
121
+ (feature_keywords, bygroups(Keyword, Keyword, Name.Function),
122
+ 'narrative'),
123
+ (feature_element_keywords,
124
+ bygroups(Name.Function, Keyword, Keyword, Name.Function),
125
+ 'feature_elements'),
126
+ (examples_keywords,
127
+ bygroups(Name.Function, Keyword, Keyword, Name.Function),
128
+ 'examples_table'),
129
+ (r'(\s|.)', Name.Function),
130
+ ]
131
+ }
132
+
133
+
134
+ class TAPLexer(RegexLexer):
135
+ """
136
+ For Test Anything Protocol (TAP) output.
137
+
138
+ .. versionadded:: 2.1
139
+ """
140
+ name = 'TAP'
141
+ aliases = ['tap']
142
+ filenames = ['*.tap']
143
+
144
+ tokens = {
145
+ 'root': [
146
+ # A TAP version may be specified.
147
+ (r'^TAP version \d+\n', Name.Namespace),
148
+
149
+ # Specify a plan with a plan line.
150
+ (r'^1\.\.\d+', Keyword.Declaration, 'plan'),
151
+
152
+ # A test failure
153
+ (r'^(not ok)([^\S\n]*)(\d*)',
154
+ bygroups(Generic.Error, Text, Number.Integer), 'test'),
155
+
156
+ # A test success
157
+ (r'^(ok)([^\S\n]*)(\d*)',
158
+ bygroups(Keyword.Reserved, Text, Number.Integer), 'test'),
159
+
160
+ # Diagnostics start with a hash.
161
+ (r'^#.*\n', Comment),
162
+
163
+ # TAP's version of an abort statement.
164
+ (r'^Bail out!.*\n', Generic.Error),
165
+
166
+ # TAP ignores any unrecognized lines.
167
+ (r'^.*\n', Text),
168
+ ],
169
+ 'plan': [
170
+ # Consume whitespace (but not newline).
171
+ (r'[^\S\n]+', Text),
172
+
173
+ # A plan may have a directive with it.
174
+ (r'#', Comment, 'directive'),
175
+
176
+ # Or it could just end.
177
+ (r'\n', Comment, '#pop'),
178
+
179
+ # Anything else is wrong.
180
+ (r'.*\n', Generic.Error, '#pop'),
181
+ ],
182
+ 'test': [
183
+ # Consume whitespace (but not newline).
184
+ (r'[^\S\n]+', Text),
185
+
186
+ # A test may have a directive with it.
187
+ (r'#', Comment, 'directive'),
188
+
189
+ (r'\S+', Text),
190
+
191
+ (r'\n', Text, '#pop'),
192
+ ],
193
+ 'directive': [
194
+ # Consume whitespace (but not newline).
195
+ (r'[^\S\n]+', Comment),
196
+
197
+ # Extract todo items.
198
+ (r'(?i)\bTODO\b', Comment.Preproc),
199
+
200
+ # Extract skip items.
201
+ (r'(?i)\bSKIP\S*', Comment.Preproc),
202
+
203
+ (r'\S+', Comment),
204
+
205
+ (r'\n', Comment, '#pop:2'),
206
+ ],
207
+ }
@@ -5,2051 +5,21 @@
5
5
 
6
6
  Lexers for non-source code file types.
7
7
 
8
- :copyright: Copyright 2006-2014 by the Pygments team, see AUTHORS.
8
+ :copyright: Copyright 2006-2015 by the Pygments team, see AUTHORS.
9
9
  :license: BSD, see LICENSE for details.
10
10
  """
11
11
 
12
- import re
13
- from bisect import bisect
14
-
15
- from pygments.lexer import Lexer, LexerContext, RegexLexer, ExtendedRegexLexer, \
16
- bygroups, include, using, this, do_insertions, default
17
- from pygments.token import Punctuation, Text, Comment, Keyword, Name, String, \
18
- Generic, Operator, Number, Whitespace, Literal
19
- from pygments.util import get_bool_opt, ClassNotFound
20
- from pygments.lexers.agile import PythonLexer
21
- from pygments.lexers.other import BashLexer
22
-
23
- __all__ = ['IniLexer', 'PropertiesLexer', 'SourcesListLexer', 'BaseMakefileLexer',
24
- 'MakefileLexer', 'DiffLexer', 'IrcLogsLexer', 'TexLexer',
25
- 'GroffLexer', 'ApacheConfLexer', 'BBCodeLexer', 'MoinWikiLexer',
26
- 'RstLexer', 'VimLexer', 'GettextLexer', 'SquidConfLexer',
27
- 'DebianControlLexer', 'DarcsPatchLexer', 'YamlLexer',
28
- 'LighttpdConfLexer', 'NginxConfLexer', 'CMakeLexer', 'HttpLexer',
29
- 'PyPyLogLexer', 'RegeditLexer', 'HxmlLexer', 'EbnfLexer',
30
- 'TodotxtLexer', 'DockerLexer']
31
-
32
-
33
- class IniLexer(RegexLexer):
34
- """
35
- Lexer for configuration files in INI style.
36
- """
37
-
38
- name = 'INI'
39
- aliases = ['ini', 'cfg', 'dosini']
40
- filenames = ['*.ini', '*.cfg']
41
- mimetypes = ['text/x-ini']
42
-
43
- tokens = {
44
- 'root': [
45
- (r'\s+', Text),
46
- (r'[;#].*', Comment.Single),
47
- (r'\[.*?\]$', Keyword),
48
- (r'(.*?)([ \t]*)(=)([ \t]*)(.*(?:\n[ \t].+)*)',
49
- bygroups(Name.Attribute, Text, Operator, Text, String))
50
- ]
51
- }
52
-
53
- def analyse_text(text):
54
- npos = text.find('\n')
55
- if npos < 3:
56
- return False
57
- return text[0] == '[' and text[npos-1] == ']'
58
-
59
-
60
- class RegeditLexer(RegexLexer):
61
- """
62
- Lexer for `Windows Registry
63
- <http://en.wikipedia.org/wiki/Windows_Registry#.REG_files>`_ files produced
64
- by regedit.
65
-
66
- .. versionadded:: 1.6
67
- """
68
-
69
- name = 'reg'
70
- aliases = ['registry']
71
- filenames = ['*.reg']
72
- mimetypes = ['text/x-windows-registry']
73
-
74
- tokens = {
75
- 'root': [
76
- (r'Windows Registry Editor.*', Text),
77
- (r'\s+', Text),
78
- (r'[;#].*', Comment.Single),
79
- (r'(\[)(-?)(HKEY_[A-Z_]+)(.*?\])$',
80
- bygroups(Keyword, Operator, Name.Builtin, Keyword)),
81
- # String keys, which obey somewhat normal escaping
82
- (r'("(?:\\"|\\\\|[^"])+")([ \t]*)(=)([ \t]*)',
83
- bygroups(Name.Attribute, Text, Operator, Text),
84
- 'value'),
85
- # Bare keys (includes @)
86
- (r'(.*?)([ \t]*)(=)([ \t]*)',
87
- bygroups(Name.Attribute, Text, Operator, Text),
88
- 'value'),
89
- ],
90
- 'value': [
91
- (r'-', Operator, '#pop'), # delete value
92
- (r'(dword|hex(?:\([0-9a-fA-F]\))?)(:)([0-9a-fA-F,]+)',
93
- bygroups(Name.Variable, Punctuation, Number), '#pop'),
94
- # As far as I know, .reg files do not support line continuation.
95
- (r'.*', String, '#pop'),
96
- ]
97
- }
98
-
99
- def analyse_text(text):
100
- return text.startswith('Windows Registry Editor')
101
-
102
-
103
- class PropertiesLexer(RegexLexer):
104
- """
105
- Lexer for configuration files in Java's properties format.
106
-
107
- .. versionadded:: 1.4
108
- """
109
-
110
- name = 'Properties'
111
- aliases = ['properties', 'jproperties']
112
- filenames = ['*.properties']
113
- mimetypes = ['text/x-java-properties']
114
-
115
- tokens = {
116
- 'root': [
117
- (r'\s+', Text),
118
- (r'(?:[;#]|//).*$', Comment),
119
- (r'(.*?)([ \t]*)([=:])([ \t]*)(.*(?:(?<=\\)\n.*)*)',
120
- bygroups(Name.Attribute, Text, Operator, Text, String)),
121
- ],
122
- }
123
-
124
-
125
- class SourcesListLexer(RegexLexer):
126
- """
127
- Lexer that highlights debian sources.list files.
128
-
129
- .. versionadded:: 0.7
130
- """
131
-
132
- name = 'Debian Sourcelist'
133
- aliases = ['sourceslist', 'sources.list', 'debsources']
134
- filenames = ['sources.list']
135
- mimetype = ['application/x-debian-sourceslist']
136
-
137
- tokens = {
138
- 'root': [
139
- (r'\s+', Text),
140
- (r'#.*?$', Comment),
141
- (r'^(deb(?:-src)?)(\s+)',
142
- bygroups(Keyword, Text), 'distribution')
143
- ],
144
- 'distribution': [
145
- (r'#.*?$', Comment, '#pop'),
146
- (r'\$\(ARCH\)', Name.Variable),
147
- (r'[^\s$[]+', String),
148
- (r'\[', String.Other, 'escaped-distribution'),
149
- (r'\$', String),
150
- (r'\s+', Text, 'components')
151
- ],
152
- 'escaped-distribution': [
153
- (r'\]', String.Other, '#pop'),
154
- (r'\$\(ARCH\)', Name.Variable),
155
- (r'[^\]$]+', String.Other),
156
- (r'\$', String.Other)
157
- ],
158
- 'components': [
159
- (r'#.*?$', Comment, '#pop:2'),
160
- (r'$', Text, '#pop:2'),
161
- (r'\s+', Text),
162
- (r'\S+', Keyword.Pseudo),
163
- ]
164
- }
165
-
166
- def analyse_text(text):
167
- for line in text.split('\n'):
168
- line = line.strip()
169
- if not (line.startswith('#') or line.startswith('deb ') or
170
- line.startswith('deb-src ') or not line):
171
- return False
172
- return True
173
-
174
-
175
- class MakefileLexer(Lexer):
176
- """
177
- Lexer for BSD and GNU make extensions (lenient enough to handle both in
178
- the same file even).
179
-
180
- *Rewritten in Pygments 0.10.*
181
- """
182
-
183
- name = 'Makefile'
184
- aliases = ['make', 'makefile', 'mf', 'bsdmake']
185
- filenames = ['*.mak', '*.mk', 'Makefile', 'makefile', 'Makefile.*', 'GNUmakefile']
186
- mimetypes = ['text/x-makefile']
187
-
188
- r_special = re.compile(r'^(?:'
189
- # BSD Make
190
- r'\.\s*(include|undef|error|warning|if|else|elif|endif|for|endfor)|'
191
- # GNU Make
192
- r'\s*(ifeq|ifneq|ifdef|ifndef|else|endif|-?include|define|endef|:))(?=\s)')
193
- r_comment = re.compile(r'^\s*@?#')
194
-
195
- def get_tokens_unprocessed(self, text):
196
- ins = []
197
- lines = text.splitlines(True)
198
- done = ''
199
- lex = BaseMakefileLexer(**self.options)
200
- backslashflag = False
201
- for line in lines:
202
- if self.r_special.match(line) or backslashflag:
203
- ins.append((len(done), [(0, Comment.Preproc, line)]))
204
- backslashflag = line.strip().endswith('\\')
205
- elif self.r_comment.match(line):
206
- ins.append((len(done), [(0, Comment, line)]))
207
- else:
208
- done += line
209
- for item in do_insertions(ins, lex.get_tokens_unprocessed(done)):
210
- yield item
211
-
212
- def analyse_text(text):
213
- # Many makefiles have $(BIG_CAPS) style variables
214
- if re.search(r'\$\([A-Z_]+\)', text):
215
- return 0.1
216
-
217
-
218
- class BaseMakefileLexer(RegexLexer):
219
- """
220
- Lexer for simple Makefiles (no preprocessing).
221
-
222
- .. versionadded:: 0.10
223
- """
224
-
225
- name = 'Base Makefile'
226
- aliases = ['basemake']
227
- filenames = []
228
- mimetypes = []
229
-
230
- tokens = {
231
- 'root': [
232
- # recipes (need to allow spaces because of expandtabs)
233
- (r'^(?:[\t ]+.*\n|\n)+', using(BashLexer)),
234
- # special variables
235
- (r'\$[<@$+%?|*]', Keyword),
236
- (r'\s+', Text),
237
- (r'#.*?\n', Comment),
238
- (r'(export)(\s+)(?=[\w${}\t -]+\n)',
239
- bygroups(Keyword, Text), 'export'),
240
- (r'export\s+', Keyword),
241
- # assignment
242
- (r'([\w${}.-]+)(\s*)([!?:+]?=)([ \t]*)((?:.*\\\n)+|.*\n)',
243
- bygroups(Name.Variable, Text, Operator, Text, using(BashLexer))),
244
- # strings
245
- (r'(?s)"(\\\\|\\.|[^"\\])*"', String.Double),
246
- (r"(?s)'(\\\\|\\.|[^'\\])*'", String.Single),
247
- # targets
248
- (r'([^\n:]+)(:+)([ \t]*)', bygroups(Name.Function, Operator, Text),
249
- 'block-header'),
250
- # expansions
251
- (r'\$\(', Keyword, 'expansion'),
252
- ],
253
- 'expansion': [
254
- (r'[^$a-zA-Z_)]+', Text),
255
- (r'[a-zA-Z_]+', Name.Variable),
256
- (r'\$', Keyword),
257
- (r'\(', Keyword, '#push'),
258
- (r'\)', Keyword, '#pop'),
259
- ],
260
- 'export': [
261
- (r'[\w${}-]+', Name.Variable),
262
- (r'\n', Text, '#pop'),
263
- (r'\s+', Text),
264
- ],
265
- 'block-header': [
266
- (r'[,|]', Punctuation),
267
- (r'#.*?\n', Comment, '#pop'),
268
- (r'\\\n', Text), # line continuation
269
- (r'\$\(', Keyword, 'expansion'),
270
- (r'[a-zA-Z_]+', Name),
271
- (r'\n', Text, '#pop'),
272
- (r'.', Text),
273
- ],
274
- }
275
-
276
-
277
- class DiffLexer(RegexLexer):
278
- """
279
- Lexer for unified or context-style diffs or patches.
280
- """
281
-
282
- name = 'Diff'
283
- aliases = ['diff', 'udiff']
284
- filenames = ['*.diff', '*.patch']
285
- mimetypes = ['text/x-diff', 'text/x-patch']
286
-
287
- tokens = {
288
- 'root': [
289
- (r' .*\n', Text),
290
- (r'\+.*\n', Generic.Inserted),
291
- (r'-.*\n', Generic.Deleted),
292
- (r'!.*\n', Generic.Strong),
293
- (r'@.*\n', Generic.Subheading),
294
- (r'([Ii]ndex|diff).*\n', Generic.Heading),
295
- (r'=.*\n', Generic.Heading),
296
- (r'.*\n', Text),
297
- ]
298
- }
299
-
300
- def analyse_text(text):
301
- if text[:7] == 'Index: ':
302
- return True
303
- if text[:5] == 'diff ':
304
- return True
305
- if text[:4] == '--- ':
306
- return 0.9
307
-
308
-
309
- DPATCH_KEYWORDS = ['hunk', 'addfile', 'adddir', 'rmfile', 'rmdir', 'move',
310
- 'replace']
311
-
312
- class DarcsPatchLexer(RegexLexer):
313
- """
314
- DarcsPatchLexer is a lexer for the various versions of the darcs patch
315
- format. Examples of this format are derived by commands such as
316
- ``darcs annotate --patch`` and ``darcs send``.
317
-
318
- .. versionadded:: 0.10
319
- """
320
- name = 'Darcs Patch'
321
- aliases = ['dpatch']
322
- filenames = ['*.dpatch', '*.darcspatch']
323
-
324
- tokens = {
325
- 'root': [
326
- (r'<', Operator),
327
- (r'>', Operator),
328
- (r'{', Operator),
329
- (r'}', Operator),
330
- (r'(\[)((?:TAG )?)(.*)(\n)(.*)(\*\*)(\d+)(\s?)(\])',
331
- bygroups(Operator, Keyword, Name, Text, Name, Operator,
332
- Literal.Date, Text, Operator)),
333
- (r'(\[)((?:TAG )?)(.*)(\n)(.*)(\*\*)(\d+)(\s?)',
334
- bygroups(Operator, Keyword, Name, Text, Name, Operator,
335
- Literal.Date, Text), 'comment'),
336
- (r'New patches:', Generic.Heading),
337
- (r'Context:', Generic.Heading),
338
- (r'Patch bundle hash:', Generic.Heading),
339
- (r'(\s*)(%s)(.*\n)' % '|'.join(DPATCH_KEYWORDS),
340
- bygroups(Text, Keyword, Text)),
341
- (r'\+', Generic.Inserted, "insert"),
342
- (r'-', Generic.Deleted, "delete"),
343
- (r'.*\n', Text),
344
- ],
345
- 'comment': [
346
- (r'[^\]].*\n', Comment),
347
- (r'\]', Operator, "#pop"),
348
- ],
349
- 'specialText': [ # darcs add [_CODE_] special operators for clarity
350
- (r'\n', Text, "#pop"), # line-based
351
- (r'\[_[^_]*_]', Operator),
352
- ],
353
- 'insert': [
354
- include('specialText'),
355
- (r'\[', Generic.Inserted),
356
- (r'[^\n\[]+', Generic.Inserted),
357
- ],
358
- 'delete': [
359
- include('specialText'),
360
- (r'\[', Generic.Deleted),
361
- (r'[^\n\[]+', Generic.Deleted),
362
- ],
363
- }
364
-
365
-
366
- class IrcLogsLexer(RegexLexer):
367
- """
368
- Lexer for IRC logs in *irssi*, *xchat* or *weechat* style.
369
- """
370
-
371
- name = 'IRC logs'
372
- aliases = ['irc']
373
- filenames = ['*.weechatlog']
374
- mimetypes = ['text/x-irclog']
375
-
376
- flags = re.VERBOSE | re.MULTILINE
377
- timestamp = r"""
378
- (
379
- # irssi / xchat and others
380
- (?: \[|\()? # Opening bracket or paren for the timestamp
381
- (?: # Timestamp
382
- (?: (?:\d{1,4} [-/]?)+ # Date as - or /-separated groups of digits
383
- [T ])? # Date/time separator: T or space
384
- (?: \d?\d [:.]?)+ # Time as :/.-separated groups of 1 or 2 digits
385
- )
386
- (?: \]|\))?\s+ # Closing bracket or paren for the timestamp
387
- |
388
- # weechat
389
- \d{4}\s\w{3}\s\d{2}\s # Date
390
- \d{2}:\d{2}:\d{2}\s+ # Time + Whitespace
391
- |
392
- # xchat
393
- \w{3}\s\d{2}\s # Date
394
- \d{2}:\d{2}:\d{2}\s+ # Time + Whitespace
395
- )?
396
- """
397
- tokens = {
398
- 'root': [
399
- # log start/end
400
- (r'^\*\*\*\*(.*)\*\*\*\*$', Comment),
401
- # hack
402
- ("^" + timestamp + r'(\s*<[^>]*>\s*)$', bygroups(Comment.Preproc, Name.Tag)),
403
- # normal msgs
404
- ("^" + timestamp + r"""
405
- (\s*<.*?>\s*) # Nick """,
406
- bygroups(Comment.Preproc, Name.Tag), 'msg'),
407
- # /me msgs
408
- ("^" + timestamp + r"""
409
- (\s*[*]\s+) # Star
410
- (\S+\s+.*?\n) # Nick + rest of message """,
411
- bygroups(Comment.Preproc, Keyword, Generic.Inserted)),
412
- # join/part msgs
413
- ("^" + timestamp + r"""
414
- (\s*(?:\*{3}|<?-[!@=P]?->?)\s*) # Star(s) or symbols
415
- (\S+\s+) # Nick + Space
416
- (.*?\n) # Rest of message """,
417
- bygroups(Comment.Preproc, Keyword, String, Comment)),
418
- (r"^.*?\n", Text),
419
- ],
420
- 'msg': [
421
- (r"\S+:(?!//)", Name.Attribute), # Prefix
422
- (r".*\n", Text, '#pop'),
423
- ],
424
- }
425
-
426
-
427
- class BBCodeLexer(RegexLexer):
428
- """
429
- A lexer that highlights BBCode(-like) syntax.
430
-
431
- .. versionadded:: 0.6
432
- """
433
-
434
- name = 'BBCode'
435
- aliases = ['bbcode']
436
- mimetypes = ['text/x-bbcode']
437
-
438
- tokens = {
439
- 'root': [
440
- (r'[^[]+', Text),
441
- # tag/end tag begin
442
- (r'\[/?\w+', Keyword, 'tag'),
443
- # stray bracket
444
- (r'\[', Text),
445
- ],
446
- 'tag': [
447
- (r'\s+', Text),
448
- # attribute with value
449
- (r'(\w+)(=)("?[^\s"\]]+"?)',
450
- bygroups(Name.Attribute, Operator, String)),
451
- # tag argument (a la [color=green])
452
- (r'(=)("?[^\s"\]]+"?)',
453
- bygroups(Operator, String)),
454
- # tag end
455
- (r'\]', Keyword, '#pop'),
456
- ],
457
- }
458
-
459
-
460
- class TexLexer(RegexLexer):
461
- """
462
- Lexer for the TeX and LaTeX typesetting languages.
463
- """
464
-
465
- name = 'TeX'
466
- aliases = ['tex', 'latex']
467
- filenames = ['*.tex', '*.aux', '*.toc']
468
- mimetypes = ['text/x-tex', 'text/x-latex']
469
-
470
- tokens = {
471
- 'general': [
472
- (r'%.*?\n', Comment),
473
- (r'[{}]', Name.Builtin),
474
- (r'[&_^]', Name.Builtin),
475
- ],
476
- 'root': [
477
- (r'\\\[', String.Backtick, 'displaymath'),
478
- (r'\\\(', String, 'inlinemath'),
479
- (r'\$\$', String.Backtick, 'displaymath'),
480
- (r'\$', String, 'inlinemath'),
481
- (r'\\([a-zA-Z]+|.)', Keyword, 'command'),
482
- include('general'),
483
- (r'[^\\$%&_^{}]+', Text),
484
- ],
485
- 'math': [
486
- (r'\\([a-zA-Z]+|.)', Name.Variable),
487
- include('general'),
488
- (r'[0-9]+', Number),
489
- (r'[-=!+*/()\[\]]', Operator),
490
- (r'[^=!+*/()\[\]\\$%&_^{}0-9-]+', Name.Builtin),
491
- ],
492
- 'inlinemath': [
493
- (r'\\\)', String, '#pop'),
494
- (r'\$', String, '#pop'),
495
- include('math'),
496
- ],
497
- 'displaymath': [
498
- (r'\\\]', String, '#pop'),
499
- (r'\$\$', String, '#pop'),
500
- (r'\$', Name.Builtin),
501
- include('math'),
502
- ],
503
- 'command': [
504
- (r'\[.*?\]', Name.Attribute),
505
- (r'\*', Keyword),
506
- default('#pop'),
507
- ],
508
- }
509
-
510
- def analyse_text(text):
511
- for start in ("\\documentclass", "\\input", "\\documentstyle",
512
- "\\relax"):
513
- if text[:len(start)] == start:
514
- return True
515
-
516
-
517
- class GroffLexer(RegexLexer):
518
- """
519
- Lexer for the (g)roff typesetting language, supporting groff
520
- extensions. Mainly useful for highlighting manpage sources.
521
-
522
- .. versionadded:: 0.6
523
- """
524
-
525
- name = 'Groff'
526
- aliases = ['groff', 'nroff', 'man']
527
- filenames = ['*.[1234567]', '*.man']
528
- mimetypes = ['application/x-troff', 'text/troff']
529
-
530
- tokens = {
531
- 'root': [
532
- (r'(\.)(\w+)', bygroups(Text, Keyword), 'request'),
533
- (r'\.', Punctuation, 'request'),
534
- # Regular characters, slurp till we find a backslash or newline
535
- (r'[^\\\n]*', Text, 'textline'),
536
- ],
537
- 'textline': [
538
- include('escapes'),
539
- (r'[^\\\n]+', Text),
540
- (r'\n', Text, '#pop'),
541
- ],
542
- 'escapes': [
543
- # groff has many ways to write escapes.
544
- (r'\\"[^\n]*', Comment),
545
- (r'\\[fn]\w', String.Escape),
546
- (r'\\\(.{2}', String.Escape),
547
- (r'\\.\[.*\]', String.Escape),
548
- (r'\\.', String.Escape),
549
- (r'\\\n', Text, 'request'),
550
- ],
551
- 'request': [
552
- (r'\n', Text, '#pop'),
553
- include('escapes'),
554
- (r'"[^\n"]+"', String.Double),
555
- (r'\d+', Number),
556
- (r'\S+', String),
557
- (r'\s+', Text),
558
- ],
559
- }
560
-
561
- def analyse_text(text):
562
- if text[:1] != '.':
563
- return False
564
- if text[:3] == '.\\"':
565
- return True
566
- if text[:4] == '.TH ':
567
- return True
568
- if text[1:3].isalnum() and text[3].isspace():
569
- return 0.9
570
-
571
-
572
- class ApacheConfLexer(RegexLexer):
573
- """
574
- Lexer for configuration files following the Apache config file
575
- format.
576
-
577
- .. versionadded:: 0.6
578
- """
579
-
580
- name = 'ApacheConf'
581
- aliases = ['apacheconf', 'aconf', 'apache']
582
- filenames = ['.htaccess', 'apache.conf', 'apache2.conf']
583
- mimetypes = ['text/x-apacheconf']
584
- flags = re.MULTILINE | re.IGNORECASE
585
-
586
- tokens = {
587
- 'root': [
588
- (r'\s+', Text),
589
- (r'(#.*?)$', Comment),
590
- (r'(<[^\s>]+)(?:(\s+)(.*?))?(>)',
591
- bygroups(Name.Tag, Text, String, Name.Tag)),
592
- (r'([a-z]\w*)(\s+)',
593
- bygroups(Name.Builtin, Text), 'value'),
594
- (r'\.+', Text),
595
- ],
596
- 'value': [
597
- (r'$', Text, '#pop'),
598
- (r'[^\S\n]+', Text),
599
- (r'\d+\.\d+\.\d+\.\d+(?:/\d+)?', Number),
600
- (r'\d+', Number),
601
- (r'/([a-z0-9][\w./-]+)', String.Other),
602
- (r'(on|off|none|any|all|double|email|dns|min|minimal|'
603
- r'os|productonly|full|emerg|alert|crit|error|warn|'
604
- r'notice|info|debug|registry|script|inetd|standalone|'
605
- r'user|group)\b', Keyword),
606
- (r'"([^"\\]*(?:\\.[^"\\]*)*)"', String.Double),
607
- (r'[^\s"]+', Text)
608
- ]
609
- }
610
-
611
-
612
- class MoinWikiLexer(RegexLexer):
613
- """
614
- For MoinMoin (and Trac) Wiki markup.
615
-
616
- .. versionadded:: 0.7
617
- """
618
-
619
- name = 'MoinMoin/Trac Wiki markup'
620
- aliases = ['trac-wiki', 'moin']
621
- filenames = []
622
- mimetypes = ['text/x-trac-wiki']
623
- flags = re.MULTILINE | re.IGNORECASE
624
-
625
- tokens = {
626
- 'root': [
627
- (r'^#.*$', Comment),
628
- (r'(!)(\S+)', bygroups(Keyword, Text)), # Ignore-next
629
- # Titles
630
- (r'^(=+)([^=]+)(=+)(\s*#.+)?$',
631
- bygroups(Generic.Heading, using(this), Generic.Heading, String)),
632
- # Literal code blocks, with optional shebang
633
- (r'({{{)(\n#!.+)?', bygroups(Name.Builtin, Name.Namespace), 'codeblock'),
634
- (r'(\'\'\'?|\|\||`|__|~~|\^|,,|::)', Comment), # Formatting
635
- # Lists
636
- (r'^( +)([.*-])( )', bygroups(Text, Name.Builtin, Text)),
637
- (r'^( +)([a-z]{1,5}\.)( )', bygroups(Text, Name.Builtin, Text)),
638
- # Other Formatting
639
- (r'\[\[\w+.*?\]\]', Keyword), # Macro
640
- (r'(\[[^\s\]]+)(\s+[^\]]+?)?(\])',
641
- bygroups(Keyword, String, Keyword)), # Link
642
- (r'^----+$', Keyword), # Horizontal rules
643
- (r'[^\n\'\[{!_~^,|]+', Text),
644
- (r'\n', Text),
645
- (r'.', Text),
646
- ],
647
- 'codeblock': [
648
- (r'}}}', Name.Builtin, '#pop'),
649
- # these blocks are allowed to be nested in Trac, but not MoinMoin
650
- (r'{{{', Text, '#push'),
651
- (r'[^{}]+', Comment.Preproc), # slurp boring text
652
- (r'.', Comment.Preproc), # allow loose { or }
653
- ],
654
- }
655
-
656
-
657
- class RstLexer(RegexLexer):
658
- """
659
- For `reStructuredText <http://docutils.sf.net/rst.html>`_ markup.
660
-
661
- .. versionadded:: 0.7
662
-
663
- Additional options accepted:
664
-
665
- `handlecodeblocks`
666
- Highlight the contents of ``.. sourcecode:: language``,
667
- ``.. code:: language`` and ``.. code-block:: language``
668
- directives with a lexer for the given language (default:
669
- ``True``).
670
-
671
- .. versionadded:: 0.8
672
- """
673
- name = 'reStructuredText'
674
- aliases = ['rst', 'rest', 'restructuredtext']
675
- filenames = ['*.rst', '*.rest']
676
- mimetypes = ["text/x-rst", "text/prs.fallenstein.rst"]
677
- flags = re.MULTILINE
678
-
679
- def _handle_sourcecode(self, match):
680
- from pygments.lexers import get_lexer_by_name
681
-
682
- # section header
683
- yield match.start(1), Punctuation, match.group(1)
684
- yield match.start(2), Text, match.group(2)
685
- yield match.start(3), Operator.Word, match.group(3)
686
- yield match.start(4), Punctuation, match.group(4)
687
- yield match.start(5), Text, match.group(5)
688
- yield match.start(6), Keyword, match.group(6)
689
- yield match.start(7), Text, match.group(7)
690
-
691
- # lookup lexer if wanted and existing
692
- lexer = None
693
- if self.handlecodeblocks:
694
- try:
695
- lexer = get_lexer_by_name(match.group(6).strip())
696
- except ClassNotFound:
697
- pass
698
- indention = match.group(8)
699
- indention_size = len(indention)
700
- code = (indention + match.group(9) + match.group(10) + match.group(11))
701
-
702
- # no lexer for this language. handle it like it was a code block
703
- if lexer is None:
704
- yield match.start(8), String, code
705
- return
706
-
707
- # highlight the lines with the lexer.
708
- ins = []
709
- codelines = code.splitlines(True)
710
- code = ''
711
- for line in codelines:
712
- if len(line) > indention_size:
713
- ins.append((len(code), [(0, Text, line[:indention_size])]))
714
- code += line[indention_size:]
715
- else:
716
- code += line
717
- for item in do_insertions(ins, lexer.get_tokens_unprocessed(code)):
718
- yield item
719
-
720
- # from docutils.parsers.rst.states
721
- closers = u'\'")]}>\u2019\u201d\xbb!?'
722
- unicode_delimiters = u'\u2010\u2011\u2012\u2013\u2014\u00a0'
723
- end_string_suffix = (r'((?=$)|(?=[-/:.,; \n\x00%s%s]))'
724
- % (re.escape(unicode_delimiters),
725
- re.escape(closers)))
726
-
727
- tokens = {
728
- 'root': [
729
- # Heading with overline
730
- (r'^(=+|-+|`+|:+|\.+|\'+|"+|~+|\^+|_+|\*+|\++|#+)([ \t]*\n)'
731
- r'(.+)(\n)(\1)(\n)',
732
- bygroups(Generic.Heading, Text, Generic.Heading,
733
- Text, Generic.Heading, Text)),
734
- # Plain heading
735
- (r'^(\S.*)(\n)(={3,}|-{3,}|`{3,}|:{3,}|\.{3,}|\'{3,}|"{3,}|'
736
- r'~{3,}|\^{3,}|_{3,}|\*{3,}|\+{3,}|#{3,})(\n)',
737
- bygroups(Generic.Heading, Text, Generic.Heading, Text)),
738
- # Bulleted lists
739
- (r'^(\s*)([-*+])( .+\n(?:\1 .+\n)*)',
740
- bygroups(Text, Number, using(this, state='inline'))),
741
- # Numbered lists
742
- (r'^(\s*)([0-9#ivxlcmIVXLCM]+\.)( .+\n(?:\1 .+\n)*)',
743
- bygroups(Text, Number, using(this, state='inline'))),
744
- (r'^(\s*)(\(?[0-9#ivxlcmIVXLCM]+\))( .+\n(?:\1 .+\n)*)',
745
- bygroups(Text, Number, using(this, state='inline'))),
746
- # Numbered, but keep words at BOL from becoming lists
747
- (r'^(\s*)([A-Z]+\.)( .+\n(?:\1 .+\n)+)',
748
- bygroups(Text, Number, using(this, state='inline'))),
749
- (r'^(\s*)(\(?[A-Za-z]+\))( .+\n(?:\1 .+\n)+)',
750
- bygroups(Text, Number, using(this, state='inline'))),
751
- # Line blocks
752
- (r'^(\s*)(\|)( .+\n(?:\| .+\n)*)',
753
- bygroups(Text, Operator, using(this, state='inline'))),
754
- # Sourcecode directives
755
- (r'^( *\.\.)(\s*)((?:source)?code(?:-block)?)(::)([ \t]*)([^\n]+)'
756
- r'(\n[ \t]*\n)([ \t]+)(.*)(\n)((?:(?:\8.*|)\n)+)',
757
- _handle_sourcecode),
758
- # A directive
759
- (r'^( *\.\.)(\s*)([\w:-]+?)(::)(?:([ \t]*)(.*))',
760
- bygroups(Punctuation, Text, Operator.Word, Punctuation, Text,
761
- using(this, state='inline'))),
762
- # A reference target
763
- (r'^( *\.\.)(\s*)(_(?:[^:\\]|\\.)+:)(.*?)$',
764
- bygroups(Punctuation, Text, Name.Tag, using(this, state='inline'))),
765
- # A footnote/citation target
766
- (r'^( *\.\.)(\s*)(\[.+\])(.*?)$',
767
- bygroups(Punctuation, Text, Name.Tag, using(this, state='inline'))),
768
- # A substitution def
769
- (r'^( *\.\.)(\s*)(\|.+\|)(\s*)([\w:-]+?)(::)(?:([ \t]*)(.*))',
770
- bygroups(Punctuation, Text, Name.Tag, Text, Operator.Word,
771
- Punctuation, Text, using(this, state='inline'))),
772
- # Comments
773
- (r'^ *\.\..*(\n( +.*\n|\n)+)?', Comment.Preproc),
774
- # Field list
775
- (r'^( *)(:[a-zA-Z-]+:)(\s*)$', bygroups(Text, Name.Class, Text)),
776
- (r'^( *)(:.*?:)([ \t]+)(.*?)$',
777
- bygroups(Text, Name.Class, Text, Name.Function)),
778
- # Definition list
779
- (r'^(\S.*(?<!::)\n)((?:(?: +.*)\n)+)',
780
- bygroups(using(this, state='inline'), using(this, state='inline'))),
781
- # Code blocks
782
- (r'(::)(\n[ \t]*\n)([ \t]+)(.*)(\n)((?:(?:\3.*|)\n)+)',
783
- bygroups(String.Escape, Text, String, String, Text, String)),
784
- include('inline'),
785
- ],
786
- 'inline': [
787
- (r'\\.', Text), # escape
788
- (r'``', String, 'literal'), # code
789
- (r'(`.+?)(<.+?>)(`__?)', # reference with inline target
790
- bygroups(String, String.Interpol, String)),
791
- (r'`.+?`__?', String), # reference
792
- (r'(`.+?`)(:[a-zA-Z0-9:-]+?:)?',
793
- bygroups(Name.Variable, Name.Attribute)), # role
794
- (r'(:[a-zA-Z0-9:-]+?:)(`.+?`)',
795
- bygroups(Name.Attribute, Name.Variable)), # role (content first)
796
- (r'\*\*.+?\*\*', Generic.Strong), # Strong emphasis
797
- (r'\*.+?\*', Generic.Emph), # Emphasis
798
- (r'\[.*?\]_', String), # Footnote or citation
799
- (r'<.+?>', Name.Tag), # Hyperlink
800
- (r'[^\\\n\[*`:]+', Text),
801
- (r'.', Text),
802
- ],
803
- 'literal': [
804
- (r'[^`]+', String),
805
- (r'``' + end_string_suffix, String, '#pop'),
806
- (r'`', String),
807
- ]
808
- }
809
-
810
- def __init__(self, **options):
811
- self.handlecodeblocks = get_bool_opt(options, 'handlecodeblocks', True)
812
- RegexLexer.__init__(self, **options)
813
-
814
- def analyse_text(text):
815
- if text[:2] == '..' and text[2:3] != '.':
816
- return 0.3
817
- p1 = text.find("\n")
818
- p2 = text.find("\n", p1 + 1)
819
- if (p2 > -1 and # has two lines
820
- p1 * 2 + 1 == p2 and # they are the same length
821
- text[p1+1] in '-=' and # the next line both starts and ends with
822
- text[p1+1] == text[p2-1]): # ...a sufficiently high header
823
- return 0.5
824
-
825
-
826
- class VimLexer(RegexLexer):
827
- """
828
- Lexer for VimL script files.
829
-
830
- .. versionadded:: 0.8
831
- """
832
- name = 'VimL'
833
- aliases = ['vim']
834
- filenames = ['*.vim', '.vimrc', '.exrc', '.gvimrc',
835
- '_vimrc', '_exrc', '_gvimrc', 'vimrc', 'gvimrc']
836
- mimetypes = ['text/x-vim']
837
- flags = re.MULTILINE
838
-
839
- _python = r'py(?:t(?:h(?:o(?:n)?)?)?)?'
840
-
841
- tokens = {
842
- 'root': [
843
- (r'^([ \t:]*)(' + _python + r')([ \t]*)(<<)([ \t]*)(.*)((?:\n|.)*)(\6)',
844
- bygroups(using(this), Keyword, Text, Operator, Text, Text,
845
- using(PythonLexer), Text)),
846
- (r'^([ \t:]*)(' + _python + r')([ \t])(.*)',
847
- bygroups(using(this), Keyword, Text, using(PythonLexer))),
848
-
849
- (r'^\s*".*', Comment),
850
-
851
- (r'[ \t]+', Text),
852
- # TODO: regexes can have other delims
853
- (r'/(\\\\|\\/|[^\n/])*/', String.Regex),
854
- (r'"(\\\\|\\"|[^\n"])*"', String.Double),
855
- (r"'(''|[^\n'])*'", String.Single),
856
-
857
- # Who decided that doublequote was a good comment character??
858
- (r'(?<=\s)"[^\-:.%#=*].*', Comment),
859
- (r'-?\d+', Number),
860
- (r'#[0-9a-f]{6}', Number.Hex),
861
- (r'^:', Punctuation),
862
- (r'[()<>+=!|,~-]', Punctuation), # Inexact list. Looks decent.
863
- (r'\b(let|if|else|endif|elseif|fun|function|endfunction)\b',
864
- Keyword),
865
- (r'\b(NONE|bold|italic|underline|dark|light)\b', Name.Builtin),
866
- (r'\b\w+\b', Name.Other), # These are postprocessed below
867
- (r'.', Text),
868
- ],
869
- }
870
- def __init__(self, **options):
871
- from pygments.lexers._vimbuiltins import command, option, auto
872
- self._cmd = command
873
- self._opt = option
874
- self._aut = auto
875
-
876
- RegexLexer.__init__(self, **options)
877
-
878
- def is_in(self, w, mapping):
879
- r"""
880
- It's kind of difficult to decide if something might be a keyword
881
- in VimL because it allows you to abbreviate them. In fact,
882
- 'ab[breviate]' is a good example. :ab, :abbre, or :abbreviate are
883
- valid ways to call it so rather than making really awful regexps
884
- like::
885
-
886
- \bab(?:b(?:r(?:e(?:v(?:i(?:a(?:t(?:e)?)?)?)?)?)?)?)?\b
887
-
888
- we match `\b\w+\b` and then call is_in() on those tokens. See
889
- `scripts/get_vimkw.py` for how the lists are extracted.
890
- """
891
- p = bisect(mapping, (w,))
892
- if p > 0:
893
- if mapping[p-1][0] == w[:len(mapping[p-1][0])] and \
894
- mapping[p-1][1][:len(w)] == w: return True
895
- if p < len(mapping):
896
- return mapping[p][0] == w[:len(mapping[p][0])] and \
897
- mapping[p][1][:len(w)] == w
898
- return False
899
-
900
- def get_tokens_unprocessed(self, text):
901
- # TODO: builtins are only subsequent tokens on lines
902
- # and 'keywords' only happen at the beginning except
903
- # for :au ones
904
- for index, token, value in \
905
- RegexLexer.get_tokens_unprocessed(self, text):
906
- if token is Name.Other:
907
- if self.is_in(value, self._cmd):
908
- yield index, Keyword, value
909
- elif self.is_in(value, self._opt) or \
910
- self.is_in(value, self._aut):
911
- yield index, Name.Builtin, value
912
- else:
913
- yield index, Text, value
914
- else:
915
- yield index, token, value
916
-
917
-
918
- class GettextLexer(RegexLexer):
919
- """
920
- Lexer for Gettext catalog files.
921
-
922
- .. versionadded:: 0.9
923
- """
924
- name = 'Gettext Catalog'
925
- aliases = ['pot', 'po']
926
- filenames = ['*.pot', '*.po']
927
- mimetypes = ['application/x-gettext', 'text/x-gettext', 'text/gettext']
928
-
929
- tokens = {
930
- 'root': [
931
- (r'^#,\s.*?$', Keyword.Type),
932
- (r'^#:\s.*?$', Keyword.Declaration),
933
- #(r'^#$', Comment),
934
- (r'^(#|#\.\s|#\|\s|#~\s|#\s).*$', Comment.Single),
935
- (r'^(")([A-Za-z-]+:)(.*")$',
936
- bygroups(String, Name.Property, String)),
937
- (r'^".*"$', String),
938
- (r'^(msgid|msgid_plural|msgstr)(\s+)(".*")$',
939
- bygroups(Name.Variable, Text, String)),
940
- (r'^(msgstr\[)(\d)(\])(\s+)(".*")$',
941
- bygroups(Name.Variable, Number.Integer, Name.Variable, Text, String)),
942
- ]
943
- }
944
-
945
-
946
- class SquidConfLexer(RegexLexer):
947
- """
948
- Lexer for `squid <http://www.squid-cache.org/>`_ configuration files.
949
-
950
- .. versionadded:: 0.9
951
- """
952
-
953
- name = 'SquidConf'
954
- aliases = ['squidconf', 'squid.conf', 'squid']
955
- filenames = ['squid.conf']
956
- mimetypes = ['text/x-squidconf']
957
- flags = re.IGNORECASE
958
-
959
- keywords = [
960
- "access_log", "acl", "always_direct", "announce_host",
961
- "announce_period", "announce_port", "announce_to", "anonymize_headers",
962
- "append_domain", "as_whois_server", "auth_param_basic",
963
- "authenticate_children", "authenticate_program", "authenticate_ttl",
964
- "broken_posts", "buffered_logs", "cache_access_log", "cache_announce",
965
- "cache_dir", "cache_dns_program", "cache_effective_group",
966
- "cache_effective_user", "cache_host", "cache_host_acl",
967
- "cache_host_domain", "cache_log", "cache_mem", "cache_mem_high",
968
- "cache_mem_low", "cache_mgr", "cachemgr_passwd", "cache_peer",
969
- "cache_peer_access", "cahce_replacement_policy", "cache_stoplist",
970
- "cache_stoplist_pattern", "cache_store_log", "cache_swap",
971
- "cache_swap_high", "cache_swap_log", "cache_swap_low", "client_db",
972
- "client_lifetime", "client_netmask", "connect_timeout", "coredump_dir",
973
- "dead_peer_timeout", "debug_options", "delay_access", "delay_class",
974
- "delay_initial_bucket_level", "delay_parameters", "delay_pools",
975
- "deny_info", "dns_children", "dns_defnames", "dns_nameservers",
976
- "dns_testnames", "emulate_httpd_log", "err_html_text",
977
- "fake_user_agent", "firewall_ip", "forwarded_for", "forward_snmpd_port",
978
- "fqdncache_size", "ftpget_options", "ftpget_program", "ftp_list_width",
979
- "ftp_passive", "ftp_user", "half_closed_clients", "header_access",
980
- "header_replace", "hierarchy_stoplist", "high_response_time_warning",
981
- "high_page_fault_warning", "hosts_file", "htcp_port", "http_access",
982
- "http_anonymizer", "httpd_accel", "httpd_accel_host",
983
- "httpd_accel_port", "httpd_accel_uses_host_header",
984
- "httpd_accel_with_proxy", "http_port", "http_reply_access",
985
- "icp_access", "icp_hit_stale", "icp_port", "icp_query_timeout",
986
- "ident_lookup", "ident_lookup_access", "ident_timeout",
987
- "incoming_http_average", "incoming_icp_average", "inside_firewall",
988
- "ipcache_high", "ipcache_low", "ipcache_size", "local_domain",
989
- "local_ip", "logfile_rotate", "log_fqdn", "log_icp_queries",
990
- "log_mime_hdrs", "maximum_object_size", "maximum_single_addr_tries",
991
- "mcast_groups", "mcast_icp_query_timeout", "mcast_miss_addr",
992
- "mcast_miss_encode_key", "mcast_miss_port", "memory_pools",
993
- "memory_pools_limit", "memory_replacement_policy", "mime_table",
994
- "min_http_poll_cnt", "min_icp_poll_cnt", "minimum_direct_hops",
995
- "minimum_object_size", "minimum_retry_timeout", "miss_access",
996
- "negative_dns_ttl", "negative_ttl", "neighbor_timeout",
997
- "neighbor_type_domain", "netdb_high", "netdb_low", "netdb_ping_period",
998
- "netdb_ping_rate", "never_direct", "no_cache", "passthrough_proxy",
999
- "pconn_timeout", "pid_filename", "pinger_program", "positive_dns_ttl",
1000
- "prefer_direct", "proxy_auth", "proxy_auth_realm", "query_icmp",
1001
- "quick_abort", "quick_abort", "quick_abort_max", "quick_abort_min",
1002
- "quick_abort_pct", "range_offset_limit", "read_timeout",
1003
- "redirect_children", "redirect_program",
1004
- "redirect_rewrites_host_header", "reference_age", "reference_age",
1005
- "refresh_pattern", "reload_into_ims", "request_body_max_size",
1006
- "request_size", "request_timeout", "shutdown_lifetime",
1007
- "single_parent_bypass", "siteselect_timeout", "snmp_access",
1008
- "snmp_incoming_address", "snmp_port", "source_ping", "ssl_proxy",
1009
- "store_avg_object_size", "store_objects_per_bucket",
1010
- "strip_query_terms", "swap_level1_dirs", "swap_level2_dirs",
1011
- "tcp_incoming_address", "tcp_outgoing_address", "tcp_recv_bufsize",
1012
- "test_reachability", "udp_hit_obj", "udp_hit_obj_size",
1013
- "udp_incoming_address", "udp_outgoing_address", "unique_hostname",
1014
- "unlinkd_program", "uri_whitespace", "useragent_log",
1015
- "visible_hostname", "wais_relay", "wais_relay_host", "wais_relay_port",
1016
- ]
1017
-
1018
- opts = [
1019
- "proxy-only", "weight", "ttl", "no-query", "default", "round-robin",
1020
- "multicast-responder", "on", "off", "all", "deny", "allow", "via",
1021
- "parent", "no-digest", "heap", "lru", "realm", "children", "q1", "q2",
1022
- "credentialsttl", "none", "disable", "offline_toggle", "diskd",
1023
- ]
1024
-
1025
- actions = [
1026
- "shutdown", "info", "parameter", "server_list", "client_list",
1027
- r'squid\.conf',
1028
- ]
1029
-
1030
- actions_stats = [
1031
- "objects", "vm_objects", "utilization", "ipcache", "fqdncache", "dns",
1032
- "redirector", "io", "reply_headers", "filedescriptors", "netdb",
1033
- ]
1034
-
1035
- actions_log = ["status", "enable", "disable", "clear"]
1036
-
1037
- acls = [
1038
- "url_regex", "urlpath_regex", "referer_regex", "port", "proto",
1039
- "req_mime_type", "rep_mime_type", "method", "browser", "user", "src",
1040
- "dst", "time", "dstdomain", "ident", "snmp_community",
1041
- ]
1042
-
1043
- ip_re = (
1044
- r'(?:(?:(?:[3-9]\d?|2(?:5[0-5]|[0-4]?\d)?|1\d{0,2}|0x0*[0-9a-f]{1,2}|'
1045
- r'0+[1-3]?[0-7]{0,2})(?:\.(?:[3-9]\d?|2(?:5[0-5]|[0-4]?\d)?|1\d{0,2}|'
1046
- r'0x0*[0-9a-f]{1,2}|0+[1-3]?[0-7]{0,2})){3})|(?!.*::.*::)(?:(?!:)|'
1047
- r':(?=:))(?:[0-9a-f]{0,4}(?:(?<=::)|(?<!::):)){6}(?:[0-9a-f]{0,4}'
1048
- r'(?:(?<=::)|(?<!::):)[0-9a-f]{0,4}(?:(?<=::)|(?<!:)|(?<=:)(?<!::):)|'
1049
- r'(?:25[0-4]|2[0-4]\d|1\d\d|[1-9]?\d)(?:\.(?:25[0-4]|2[0-4]\d|1\d\d|'
1050
- r'[1-9]?\d)){3}))'
1051
- )
1052
-
1053
- def makelistre(list):
1054
- return r'\b(?:' + '|'.join(list) + r')\b'
1055
-
1056
- tokens = {
1057
- 'root': [
1058
- (r'\s+', Whitespace),
1059
- (r'#', Comment, 'comment'),
1060
- (makelistre(keywords), Keyword),
1061
- (makelistre(opts), Name.Constant),
1062
- # Actions
1063
- (makelistre(actions), String),
1064
- (r'stats/'+makelistre(actions), String),
1065
- (r'log/'+makelistre(actions)+r'=', String),
1066
- (makelistre(acls), Keyword),
1067
- (ip_re + r'(?:/(?:' + ip_re + r'|\b\d+\b))?', Number.Float),
1068
- (r'(?:\b\d+\b(?:-\b\d+|%)?)', Number),
1069
- (r'\S+', Text),
1070
- ],
1071
- 'comment': [
1072
- (r'\s*TAG:.*', String.Escape, '#pop'),
1073
- (r'.*', Comment, '#pop'),
1074
- ],
1075
- }
1076
-
1077
-
1078
- class DebianControlLexer(RegexLexer):
1079
- """
1080
- Lexer for Debian ``control`` files and ``apt-cache show <pkg>`` outputs.
1081
-
1082
- .. versionadded:: 0.9
1083
- """
1084
- name = 'Debian Control file'
1085
- aliases = ['control', 'debcontrol']
1086
- filenames = ['control']
1087
-
1088
- tokens = {
1089
- 'root': [
1090
- (r'^(Description)', Keyword, 'description'),
1091
- (r'^(Maintainer)(:\s*)', bygroups(Keyword, Text), 'maintainer'),
1092
- (r'^((Build-)?Depends)', Keyword, 'depends'),
1093
- (r'^((?:Python-)?Version)(:\s*)(\S+)$',
1094
- bygroups(Keyword, Text, Number)),
1095
- (r'^((?:Installed-)?Size)(:\s*)(\S+)$',
1096
- bygroups(Keyword, Text, Number)),
1097
- (r'^(MD5Sum|SHA1|SHA256)(:\s*)(\S+)$',
1098
- bygroups(Keyword, Text, Number)),
1099
- (r'^([a-zA-Z\-0-9\.]*?)(:\s*)(.*?)$',
1100
- bygroups(Keyword, Whitespace, String)),
1101
- ],
1102
- 'maintainer': [
1103
- (r'<[^>]+>', Generic.Strong),
1104
- (r'<[^>]+>$', Generic.Strong, '#pop'),
1105
- (r',\n?', Text),
1106
- (r'.', Text),
1107
- ],
1108
- 'description': [
1109
- (r'(.*)(Homepage)(: )(\S+)',
1110
- bygroups(Text, String, Name, Name.Class)),
1111
- (r':.*\n', Generic.Strong),
1112
- (r' .*\n', Text),
1113
- ('', Text, '#pop'),
1114
- ],
1115
- 'depends': [
1116
- (r':\s*', Text),
1117
- (r'(\$)(\{)(\w+\s*:\s*\w+)', bygroups(Operator, Text, Name.Entity)),
1118
- (r'\(', Text, 'depend_vers'),
1119
- (r',', Text),
1120
- (r'\|', Operator),
1121
- (r'[\s]+', Text),
1122
- (r'[}\)]\s*$', Text, '#pop'),
1123
- (r'}', Text),
1124
- (r'[^,]$', Name.Function, '#pop'),
1125
- (r'([\+\.a-zA-Z0-9-])(\s*)', bygroups(Name.Function, Text)),
1126
- (r'\[.*?\]', Name.Entity),
1127
- ],
1128
- 'depend_vers': [
1129
- (r'\),', Text, '#pop'),
1130
- (r'\)[^,]', Text, '#pop:2'),
1131
- (r'([><=]+)(\s*)([^\)]+)', bygroups(Operator, Text, Number))
1132
- ]
1133
- }
1134
-
1135
-
1136
- class YamlLexerContext(LexerContext):
1137
- """Indentation context for the YAML lexer."""
1138
-
1139
- def __init__(self, *args, **kwds):
1140
- super(YamlLexerContext, self).__init__(*args, **kwds)
1141
- self.indent_stack = []
1142
- self.indent = -1
1143
- self.next_indent = 0
1144
- self.block_scalar_indent = None
1145
-
1146
-
1147
- class YamlLexer(ExtendedRegexLexer):
1148
- """
1149
- Lexer for `YAML <http://yaml.org/>`_, a human-friendly data serialization
1150
- language.
1151
-
1152
- .. versionadded:: 0.11
1153
- """
1154
-
1155
- name = 'YAML'
1156
- aliases = ['yaml']
1157
- filenames = ['*.yaml', '*.yml']
1158
- mimetypes = ['text/x-yaml']
1159
-
1160
-
1161
- def something(token_class):
1162
- """Do not produce empty tokens."""
1163
- def callback(lexer, match, context):
1164
- text = match.group()
1165
- if not text:
1166
- return
1167
- yield match.start(), token_class, text
1168
- context.pos = match.end()
1169
- return callback
1170
-
1171
- def reset_indent(token_class):
1172
- """Reset the indentation levels."""
1173
- def callback(lexer, match, context):
1174
- text = match.group()
1175
- context.indent_stack = []
1176
- context.indent = -1
1177
- context.next_indent = 0
1178
- context.block_scalar_indent = None
1179
- yield match.start(), token_class, text
1180
- context.pos = match.end()
1181
- return callback
1182
-
1183
- def save_indent(token_class, start=False):
1184
- """Save a possible indentation level."""
1185
- def callback(lexer, match, context):
1186
- text = match.group()
1187
- extra = ''
1188
- if start:
1189
- context.next_indent = len(text)
1190
- if context.next_indent < context.indent:
1191
- while context.next_indent < context.indent:
1192
- context.indent = context.indent_stack.pop()
1193
- if context.next_indent > context.indent:
1194
- extra = text[context.indent:]
1195
- text = text[:context.indent]
1196
- else:
1197
- context.next_indent += len(text)
1198
- if text:
1199
- yield match.start(), token_class, text
1200
- if extra:
1201
- yield match.start()+len(text), token_class.Error, extra
1202
- context.pos = match.end()
1203
- return callback
1204
-
1205
- def set_indent(token_class, implicit=False):
1206
- """Set the previously saved indentation level."""
1207
- def callback(lexer, match, context):
1208
- text = match.group()
1209
- if context.indent < context.next_indent:
1210
- context.indent_stack.append(context.indent)
1211
- context.indent = context.next_indent
1212
- if not implicit:
1213
- context.next_indent += len(text)
1214
- yield match.start(), token_class, text
1215
- context.pos = match.end()
1216
- return callback
1217
-
1218
- def set_block_scalar_indent(token_class):
1219
- """Set an explicit indentation level for a block scalar."""
1220
- def callback(lexer, match, context):
1221
- text = match.group()
1222
- context.block_scalar_indent = None
1223
- if not text:
1224
- return
1225
- increment = match.group(1)
1226
- if increment:
1227
- current_indent = max(context.indent, 0)
1228
- increment = int(increment)
1229
- context.block_scalar_indent = current_indent + increment
1230
- if text:
1231
- yield match.start(), token_class, text
1232
- context.pos = match.end()
1233
- return callback
1234
-
1235
- def parse_block_scalar_empty_line(indent_token_class, content_token_class):
1236
- """Process an empty line in a block scalar."""
1237
- def callback(lexer, match, context):
1238
- text = match.group()
1239
- if (context.block_scalar_indent is None or
1240
- len(text) <= context.block_scalar_indent):
1241
- if text:
1242
- yield match.start(), indent_token_class, text
1243
- else:
1244
- indentation = text[:context.block_scalar_indent]
1245
- content = text[context.block_scalar_indent:]
1246
- yield match.start(), indent_token_class, indentation
1247
- yield (match.start()+context.block_scalar_indent,
1248
- content_token_class, content)
1249
- context.pos = match.end()
1250
- return callback
1251
-
1252
- def parse_block_scalar_indent(token_class):
1253
- """Process indentation spaces in a block scalar."""
1254
- def callback(lexer, match, context):
1255
- text = match.group()
1256
- if context.block_scalar_indent is None:
1257
- if len(text) <= max(context.indent, 0):
1258
- context.stack.pop()
1259
- context.stack.pop()
1260
- return
1261
- context.block_scalar_indent = len(text)
1262
- else:
1263
- if len(text) < context.block_scalar_indent:
1264
- context.stack.pop()
1265
- context.stack.pop()
1266
- return
1267
- if text:
1268
- yield match.start(), token_class, text
1269
- context.pos = match.end()
1270
- return callback
1271
-
1272
- def parse_plain_scalar_indent(token_class):
1273
- """Process indentation spaces in a plain scalar."""
1274
- def callback(lexer, match, context):
1275
- text = match.group()
1276
- if len(text) <= context.indent:
1277
- context.stack.pop()
1278
- context.stack.pop()
1279
- return
1280
- if text:
1281
- yield match.start(), token_class, text
1282
- context.pos = match.end()
1283
- return callback
1284
-
1285
-
1286
-
1287
- tokens = {
1288
- # the root rules
1289
- 'root': [
1290
- # ignored whitespaces
1291
- (r'[ ]+(?=#|$)', Text),
1292
- # line breaks
1293
- (r'\n+', Text),
1294
- # a comment
1295
- (r'#[^\n]*', Comment.Single),
1296
- # the '%YAML' directive
1297
- (r'^%YAML(?=[ ]|$)', reset_indent(Name.Tag), 'yaml-directive'),
1298
- # the %TAG directive
1299
- (r'^%TAG(?=[ ]|$)', reset_indent(Name.Tag), 'tag-directive'),
1300
- # document start and document end indicators
1301
- (r'^(?:---|\.\.\.)(?=[ ]|$)', reset_indent(Name.Namespace),
1302
- 'block-line'),
1303
- # indentation spaces
1304
- (r'[ ]*(?![ \t\n\r\f\v]|$)', save_indent(Text, start=True),
1305
- ('block-line', 'indentation')),
1306
- ],
1307
-
1308
- # trailing whitespaces after directives or a block scalar indicator
1309
- 'ignored-line': [
1310
- # ignored whitespaces
1311
- (r'[ ]+(?=#|$)', Text),
1312
- # a comment
1313
- (r'#[^\n]*', Comment.Single),
1314
- # line break
1315
- (r'\n', Text, '#pop:2'),
1316
- ],
1317
-
1318
- # the %YAML directive
1319
- 'yaml-directive': [
1320
- # the version number
1321
- (r'([ ]+)([0-9]+\.[0-9]+)',
1322
- bygroups(Text, Number), 'ignored-line'),
1323
- ],
1324
-
1325
- # the %YAG directive
1326
- 'tag-directive': [
1327
- # a tag handle and the corresponding prefix
1328
- (r'([ ]+)(!|![0-9A-Za-z_-]*!)'
1329
- r'([ ]+)(!|!?[0-9A-Za-z;/?:@&=+$,_.!~*\'()\[\]%-]+)',
1330
- bygroups(Text, Keyword.Type, Text, Keyword.Type),
1331
- 'ignored-line'),
1332
- ],
1333
-
1334
- # block scalar indicators and indentation spaces
1335
- 'indentation': [
1336
- # trailing whitespaces are ignored
1337
- (r'[ ]*$', something(Text), '#pop:2'),
1338
- # whitespaces preceeding block collection indicators
1339
- (r'[ ]+(?=[?:-](?:[ ]|$))', save_indent(Text)),
1340
- # block collection indicators
1341
- (r'[?:-](?=[ ]|$)', set_indent(Punctuation.Indicator)),
1342
- # the beginning a block line
1343
- (r'[ ]*', save_indent(Text), '#pop'),
1344
- ],
1345
-
1346
- # an indented line in the block context
1347
- 'block-line': [
1348
- # the line end
1349
- (r'[ ]*(?=#|$)', something(Text), '#pop'),
1350
- # whitespaces separating tokens
1351
- (r'[ ]+', Text),
1352
- # tags, anchors and aliases,
1353
- include('descriptors'),
1354
- # block collections and scalars
1355
- include('block-nodes'),
1356
- # flow collections and quoted scalars
1357
- include('flow-nodes'),
1358
- # a plain scalar
1359
- (r'(?=[^ \t\n\r\f\v?:,\[\]{}#&*!|>\'"%@`-]|[?:-][^ \t\n\r\f\v])',
1360
- something(Name.Variable),
1361
- 'plain-scalar-in-block-context'),
1362
- ],
1363
-
1364
- # tags, anchors, aliases
1365
- 'descriptors' : [
1366
- # a full-form tag
1367
- (r'!<[0-9A-Za-z;/?:@&=+$,_.!~*\'()\[\]%-]+>', Keyword.Type),
1368
- # a tag in the form '!', '!suffix' or '!handle!suffix'
1369
- (r'!(?:[0-9A-Za-z_-]+)?'
1370
- r'(?:![0-9A-Za-z;/?:@&=+$,_.!~*\'()\[\]%-]+)?', Keyword.Type),
1371
- # an anchor
1372
- (r'&[0-9A-Za-z_-]+', Name.Label),
1373
- # an alias
1374
- (r'\*[0-9A-Za-z_-]+', Name.Variable),
1375
- ],
1376
-
1377
- # block collections and scalars
1378
- 'block-nodes': [
1379
- # implicit key
1380
- (r':(?=[ ]|$)', set_indent(Punctuation.Indicator, implicit=True)),
1381
- # literal and folded scalars
1382
- (r'[|>]', Punctuation.Indicator,
1383
- ('block-scalar-content', 'block-scalar-header')),
1384
- ],
1385
-
1386
- # flow collections and quoted scalars
1387
- 'flow-nodes': [
1388
- # a flow sequence
1389
- (r'\[', Punctuation.Indicator, 'flow-sequence'),
1390
- # a flow mapping
1391
- (r'\{', Punctuation.Indicator, 'flow-mapping'),
1392
- # a single-quoted scalar
1393
- (r'\'', String, 'single-quoted-scalar'),
1394
- # a double-quoted scalar
1395
- (r'\"', String, 'double-quoted-scalar'),
1396
- ],
1397
-
1398
- # the content of a flow collection
1399
- 'flow-collection': [
1400
- # whitespaces
1401
- (r'[ ]+', Text),
1402
- # line breaks
1403
- (r'\n+', Text),
1404
- # a comment
1405
- (r'#[^\n]*', Comment.Single),
1406
- # simple indicators
1407
- (r'[?:,]', Punctuation.Indicator),
1408
- # tags, anchors and aliases
1409
- include('descriptors'),
1410
- # nested collections and quoted scalars
1411
- include('flow-nodes'),
1412
- # a plain scalar
1413
- (r'(?=[^ \t\n\r\f\v?:,\[\]{}#&*!|>\'"%@`])',
1414
- something(Name.Variable),
1415
- 'plain-scalar-in-flow-context'),
1416
- ],
1417
-
1418
- # a flow sequence indicated by '[' and ']'
1419
- 'flow-sequence': [
1420
- # include flow collection rules
1421
- include('flow-collection'),
1422
- # the closing indicator
1423
- (r'\]', Punctuation.Indicator, '#pop'),
1424
- ],
1425
-
1426
- # a flow mapping indicated by '{' and '}'
1427
- 'flow-mapping': [
1428
- # include flow collection rules
1429
- include('flow-collection'),
1430
- # the closing indicator
1431
- (r'\}', Punctuation.Indicator, '#pop'),
1432
- ],
1433
-
1434
- # block scalar lines
1435
- 'block-scalar-content': [
1436
- # line break
1437
- (r'\n', Text),
1438
- # empty line
1439
- (r'^[ ]+$',
1440
- parse_block_scalar_empty_line(Text, Name.Constant)),
1441
- # indentation spaces (we may leave the state here)
1442
- (r'^[ ]*', parse_block_scalar_indent(Text)),
1443
- # line content
1444
- (r'[^\n\r\f\v]+', Name.Constant),
1445
- ],
1446
-
1447
- # the content of a literal or folded scalar
1448
- 'block-scalar-header': [
1449
- # indentation indicator followed by chomping flag
1450
- (r'([1-9])?[+-]?(?=[ ]|$)',
1451
- set_block_scalar_indent(Punctuation.Indicator),
1452
- 'ignored-line'),
1453
- # chomping flag followed by indentation indicator
1454
- (r'[+-]?([1-9])?(?=[ ]|$)',
1455
- set_block_scalar_indent(Punctuation.Indicator),
1456
- 'ignored-line'),
1457
- ],
1458
-
1459
- # ignored and regular whitespaces in quoted scalars
1460
- 'quoted-scalar-whitespaces': [
1461
- # leading and trailing whitespaces are ignored
1462
- (r'^[ ]+', Text),
1463
- (r'[ ]+$', Text),
1464
- # line breaks are ignored
1465
- (r'\n+', Text),
1466
- # other whitespaces are a part of the value
1467
- (r'[ ]+', Name.Variable),
1468
- ],
1469
-
1470
- # single-quoted scalars
1471
- 'single-quoted-scalar': [
1472
- # include whitespace and line break rules
1473
- include('quoted-scalar-whitespaces'),
1474
- # escaping of the quote character
1475
- (r'\'\'', String.Escape),
1476
- # regular non-whitespace characters
1477
- (r'[^ \t\n\r\f\v\']+', String),
1478
- # the closing quote
1479
- (r'\'', String, '#pop'),
1480
- ],
1481
-
1482
- # double-quoted scalars
1483
- 'double-quoted-scalar': [
1484
- # include whitespace and line break rules
1485
- include('quoted-scalar-whitespaces'),
1486
- # escaping of special characters
1487
- (r'\\[0abt\tn\nvfre "\\N_LP]', String),
1488
- # escape codes
1489
- (r'\\(?:x[0-9A-Fa-f]{2}|u[0-9A-Fa-f]{4}|U[0-9A-Fa-f]{8})',
1490
- String.Escape),
1491
- # regular non-whitespace characters
1492
- (r'[^ \t\n\r\f\v\"\\]+', String),
1493
- # the closing quote
1494
- (r'"', String, '#pop'),
1495
- ],
1496
-
1497
- # the beginning of a new line while scanning a plain scalar
1498
- 'plain-scalar-in-block-context-new-line': [
1499
- # empty lines
1500
- (r'^[ ]+$', Text),
1501
- # line breaks
1502
- (r'\n+', Text),
1503
- # document start and document end indicators
1504
- (r'^(?=---|\.\.\.)', something(Name.Namespace), '#pop:3'),
1505
- # indentation spaces (we may leave the block line state here)
1506
- (r'^[ ]*', parse_plain_scalar_indent(Text), '#pop'),
1507
- ],
1508
-
1509
- # a plain scalar in the block context
1510
- 'plain-scalar-in-block-context': [
1511
- # the scalar ends with the ':' indicator
1512
- (r'[ ]*(?=:[ ]|:$)', something(Text), '#pop'),
1513
- # the scalar ends with whitespaces followed by a comment
1514
- (r'[ ]+(?=#)', Text, '#pop'),
1515
- # trailing whitespaces are ignored
1516
- (r'[ ]+$', Text),
1517
- # line breaks are ignored
1518
- (r'\n+', Text, 'plain-scalar-in-block-context-new-line'),
1519
- # other whitespaces are a part of the value
1520
- (r'[ ]+', Literal.Scalar.Plain),
1521
- # regular non-whitespace characters
1522
- (r'(?::(?![ \t\n\r\f\v])|[^ \t\n\r\f\v:])+', Literal.Scalar.Plain),
1523
- ],
1524
-
1525
- # a plain scalar is the flow context
1526
- 'plain-scalar-in-flow-context': [
1527
- # the scalar ends with an indicator character
1528
- (r'[ ]*(?=[,:?\[\]{}])', something(Text), '#pop'),
1529
- # the scalar ends with a comment
1530
- (r'[ ]+(?=#)', Text, '#pop'),
1531
- # leading and trailing whitespaces are ignored
1532
- (r'^[ ]+', Text),
1533
- (r'[ ]+$', Text),
1534
- # line breaks are ignored
1535
- (r'\n+', Text),
1536
- # other whitespaces are a part of the value
1537
- (r'[ ]+', Name.Variable),
1538
- # regular non-whitespace characters
1539
- (r'[^ \t\n\r\f\v,:?\[\]{}]+', Name.Variable),
1540
- ],
1541
-
1542
- }
1543
-
1544
- def get_tokens_unprocessed(self, text=None, context=None):
1545
- if context is None:
1546
- context = YamlLexerContext(text, 0)
1547
- return super(YamlLexer, self).get_tokens_unprocessed(text, context)
1548
-
1549
-
1550
- class LighttpdConfLexer(RegexLexer):
1551
- """
1552
- Lexer for `Lighttpd <http://lighttpd.net/>`_ configuration files.
1553
-
1554
- .. versionadded:: 0.11
1555
- """
1556
- name = 'Lighttpd configuration file'
1557
- aliases = ['lighty', 'lighttpd']
1558
- filenames = []
1559
- mimetypes = ['text/x-lighttpd-conf']
1560
-
1561
- tokens = {
1562
- 'root': [
1563
- (r'#.*\n', Comment.Single),
1564
- (r'/\S*', Name), # pathname
1565
- (r'[a-zA-Z._-]+', Keyword),
1566
- (r'\d+\.\d+\.\d+\.\d+(?:/\d+)?', Number),
1567
- (r'[0-9]+', Number),
1568
- (r'=>|=~|\+=|==|=|\+', Operator),
1569
- (r'\$[A-Z]+', Name.Builtin),
1570
- (r'[(){}\[\],]', Punctuation),
1571
- (r'"([^"\\]*(?:\\.[^"\\]*)*)"', String.Double),
1572
- (r'\s+', Text),
1573
- ],
1574
-
1575
- }
1576
-
1577
-
1578
- class NginxConfLexer(RegexLexer):
1579
- """
1580
- Lexer for `Nginx <http://nginx.net/>`_ configuration files.
1581
-
1582
- .. versionadded:: 0.11
1583
- """
1584
- name = 'Nginx configuration file'
1585
- aliases = ['nginx']
1586
- filenames = []
1587
- mimetypes = ['text/x-nginx-conf']
1588
-
1589
- tokens = {
1590
- 'root': [
1591
- (r'(include)(\s+)([^\s;]+)', bygroups(Keyword, Text, Name)),
1592
- (r'[^\s;#]+', Keyword, 'stmt'),
1593
- include('base'),
1594
- ],
1595
- 'block': [
1596
- (r'}', Punctuation, '#pop:2'),
1597
- (r'[^\s;#]+', Keyword.Namespace, 'stmt'),
1598
- include('base'),
1599
- ],
1600
- 'stmt': [
1601
- (r'{', Punctuation, 'block'),
1602
- (r';', Punctuation, '#pop'),
1603
- include('base'),
1604
- ],
1605
- 'base': [
1606
- (r'#.*\n', Comment.Single),
1607
- (r'on|off', Name.Constant),
1608
- (r'\$[^\s;#()]+', Name.Variable),
1609
- (r'([a-z0-9.-]+)(:)([0-9]+)',
1610
- bygroups(Name, Punctuation, Number.Integer)),
1611
- (r'[a-z-]+/[a-z-+]+', String), # mimetype
1612
- #(r'[a-zA-Z._-]+', Keyword),
1613
- (r'[0-9]+[km]?\b', Number.Integer),
1614
- (r'(~)(\s*)([^\s{]+)', bygroups(Punctuation, Text, String.Regex)),
1615
- (r'[:=~]', Punctuation),
1616
- (r'[^\s;#{}$]+', String), # catch all
1617
- (r'/[^\s;#]*', Name), # pathname
1618
- (r'\s+', Text),
1619
- (r'[$;]', Text), # leftover characters
1620
- ],
1621
- }
1622
-
1623
-
1624
- class CMakeLexer(RegexLexer):
1625
- """
1626
- Lexer for `CMake <http://cmake.org/Wiki/CMake>`_ files.
1627
-
1628
- .. versionadded:: 1.2
1629
- """
1630
- name = 'CMake'
1631
- aliases = ['cmake']
1632
- filenames = ['*.cmake', 'CMakeLists.txt']
1633
- mimetypes = ['text/x-cmake']
1634
-
1635
- tokens = {
1636
- 'root': [
1637
- #(r'(ADD_CUSTOM_COMMAND|ADD_CUSTOM_TARGET|ADD_DEFINITIONS|'
1638
- # r'ADD_DEPENDENCIES|ADD_EXECUTABLE|ADD_LIBRARY|ADD_SUBDIRECTORY|'
1639
- # r'ADD_TEST|AUX_SOURCE_DIRECTORY|BUILD_COMMAND|BUILD_NAME|'
1640
- # r'CMAKE_MINIMUM_REQUIRED|CONFIGURE_FILE|CREATE_TEST_SOURCELIST|'
1641
- # r'ELSE|ELSEIF|ENABLE_LANGUAGE|ENABLE_TESTING|ENDFOREACH|'
1642
- # r'ENDFUNCTION|ENDIF|ENDMACRO|ENDWHILE|EXEC_PROGRAM|'
1643
- # r'EXECUTE_PROCESS|EXPORT_LIBRARY_DEPENDENCIES|FILE|FIND_FILE|'
1644
- # r'FIND_LIBRARY|FIND_PACKAGE|FIND_PATH|FIND_PROGRAM|FLTK_WRAP_UI|'
1645
- # r'FOREACH|FUNCTION|GET_CMAKE_PROPERTY|GET_DIRECTORY_PROPERTY|'
1646
- # r'GET_FILENAME_COMPONENT|GET_SOURCE_FILE_PROPERTY|'
1647
- # r'GET_TARGET_PROPERTY|GET_TEST_PROPERTY|IF|INCLUDE|'
1648
- # r'INCLUDE_DIRECTORIES|INCLUDE_EXTERNAL_MSPROJECT|'
1649
- # r'INCLUDE_REGULAR_EXPRESSION|INSTALL|INSTALL_FILES|'
1650
- # r'INSTALL_PROGRAMS|INSTALL_TARGETS|LINK_DIRECTORIES|'
1651
- # r'LINK_LIBRARIES|LIST|LOAD_CACHE|LOAD_COMMAND|MACRO|'
1652
- # r'MAKE_DIRECTORY|MARK_AS_ADVANCED|MATH|MESSAGE|OPTION|'
1653
- # r'OUTPUT_REQUIRED_FILES|PROJECT|QT_WRAP_CPP|QT_WRAP_UI|REMOVE|'
1654
- # r'REMOVE_DEFINITIONS|SEPARATE_ARGUMENTS|SET|'
1655
- # r'SET_DIRECTORY_PROPERTIES|SET_SOURCE_FILES_PROPERTIES|'
1656
- # r'SET_TARGET_PROPERTIES|SET_TESTS_PROPERTIES|SITE_NAME|'
1657
- # r'SOURCE_GROUP|STRING|SUBDIR_DEPENDS|SUBDIRS|'
1658
- # r'TARGET_LINK_LIBRARIES|TRY_COMPILE|TRY_RUN|UNSET|'
1659
- # r'USE_MANGLED_MESA|UTILITY_SOURCE|VARIABLE_REQUIRES|'
1660
- # r'VTK_MAKE_INSTANTIATOR|VTK_WRAP_JAVA|VTK_WRAP_PYTHON|'
1661
- # r'VTK_WRAP_TCL|WHILE|WRITE_FILE|'
1662
- # r'COUNTARGS)\b', Name.Builtin, 'args'),
1663
- (r'\b(\w+)([ \t]*)(\()', bygroups(Name.Builtin, Text,
1664
- Punctuation), 'args'),
1665
- include('keywords'),
1666
- include('ws')
1667
- ],
1668
- 'args': [
1669
- (r'\(', Punctuation, '#push'),
1670
- (r'\)', Punctuation, '#pop'),
1671
- (r'(\${)(.+?)(})', bygroups(Operator, Name.Variable, Operator)),
1672
- (r'(\$<)(.+?)(>)', bygroups(Operator, Name.Variable, Operator)),
1673
- (r'(?s)".*?"', String.Double),
1674
- (r'\\\S+', String),
1675
- (r'[^\)$"# \t\n]+', String),
1676
- (r'\n', Text), # explicitly legal
1677
- include('keywords'),
1678
- include('ws')
1679
- ],
1680
- 'string': [
1681
-
1682
- ],
1683
- 'keywords': [
1684
- (r'\b(WIN32|UNIX|APPLE|CYGWIN|BORLAND|MINGW|MSVC|MSVC_IDE|MSVC60|'
1685
- r'MSVC70|MSVC71|MSVC80|MSVC90)\b', Keyword),
1686
- ],
1687
- 'ws': [
1688
- (r'[ \t]+', Text),
1689
- (r'#.*\n', Comment),
1690
- ]
1691
- }
1692
-
1693
- def analyse_text(text):
1694
- exp = r'^ *CMAKE_MINIMUM_REQUIRED *\( *VERSION *\d(\.\d)* *( FATAL_ERROR)? *\) *$'
1695
- if re.search(exp, text, flags=re.MULTILINE | re.IGNORECASE):
1696
- return 0.8
1697
- return 0.0
1698
-
1699
-
1700
- class HttpLexer(RegexLexer):
1701
- """
1702
- Lexer for HTTP sessions.
1703
-
1704
- .. versionadded:: 1.5
1705
- """
1706
-
1707
- name = 'HTTP'
1708
- aliases = ['http']
1709
-
1710
- flags = re.DOTALL
1711
-
1712
- def header_callback(self, match):
1713
- if match.group(1).lower() == 'content-type':
1714
- content_type = match.group(5).strip()
1715
- if ';' in content_type:
1716
- content_type = content_type[:content_type.find(';')].strip()
1717
- self.content_type = content_type
1718
- yield match.start(1), Name.Attribute, match.group(1)
1719
- yield match.start(2), Text, match.group(2)
1720
- yield match.start(3), Operator, match.group(3)
1721
- yield match.start(4), Text, match.group(4)
1722
- yield match.start(5), Literal, match.group(5)
1723
- yield match.start(6), Text, match.group(6)
1724
-
1725
- def continuous_header_callback(self, match):
1726
- yield match.start(1), Text, match.group(1)
1727
- yield match.start(2), Literal, match.group(2)
1728
- yield match.start(3), Text, match.group(3)
1729
-
1730
- def content_callback(self, match):
1731
- content_type = getattr(self, 'content_type', None)
1732
- content = match.group()
1733
- offset = match.start()
1734
- if content_type:
1735
- from pygments.lexers import get_lexer_for_mimetype
1736
- try:
1737
- lexer = get_lexer_for_mimetype(content_type)
1738
- except ClassNotFound:
1739
- pass
1740
- else:
1741
- for idx, token, value in lexer.get_tokens_unprocessed(content):
1742
- yield offset + idx, token, value
1743
- return
1744
- yield offset, Text, content
1745
-
1746
- tokens = {
1747
- 'root': [
1748
- (r'(GET|POST|PUT|DELETE|HEAD|OPTIONS|TRACE|PATCH)( +)([^ ]+)( +)'
1749
- r'(HTTP)(/)(1\.[01])(\r?\n|$)',
1750
- bygroups(Name.Function, Text, Name.Namespace, Text,
1751
- Keyword.Reserved, Operator, Number, Text),
1752
- 'headers'),
1753
- (r'(HTTP)(/)(1\.[01])( +)(\d{3})( +)([^\r\n]+)(\r?\n|$)',
1754
- bygroups(Keyword.Reserved, Operator, Number, Text, Number,
1755
- Text, Name.Exception, Text),
1756
- 'headers'),
1757
- ],
1758
- 'headers': [
1759
- (r'([^\s:]+)( *)(:)( *)([^\r\n]+)(\r?\n|$)', header_callback),
1760
- (r'([\t ]+)([^\r\n]+)(\r?\n|$)', continuous_header_callback),
1761
- (r'\r?\n', Text, 'content')
1762
- ],
1763
- 'content': [
1764
- (r'.+', content_callback)
1765
- ]
1766
- }
1767
-
1768
-
1769
- class PyPyLogLexer(RegexLexer):
1770
- """
1771
- Lexer for PyPy log files.
1772
-
1773
- .. versionadded:: 1.5
1774
- """
1775
- name = "PyPy Log"
1776
- aliases = ["pypylog", "pypy"]
1777
- filenames = ["*.pypylog"]
1778
- mimetypes = ['application/x-pypylog']
1779
-
1780
- tokens = {
1781
- "root": [
1782
- (r"\[\w+\] {jit-log-.*?$", Keyword, "jit-log"),
1783
- (r"\[\w+\] {jit-backend-counts$", Keyword, "jit-backend-counts"),
1784
- include("extra-stuff"),
1785
- ],
1786
- "jit-log": [
1787
- (r"\[\w+\] jit-log-.*?}$", Keyword, "#pop"),
1788
- (r"^\+\d+: ", Comment),
1789
- (r"--end of the loop--", Comment),
1790
- (r"[ifp]\d+", Name),
1791
- (r"ptr\d+", Name),
1792
- (r"(\()(\w+(?:\.\w+)?)(\))",
1793
- bygroups(Punctuation, Name.Builtin, Punctuation)),
1794
- (r"[\[\]=,()]", Punctuation),
1795
- (r"(\d+\.\d+|inf|-inf)", Number.Float),
1796
- (r"-?\d+", Number.Integer),
1797
- (r"'.*'", String),
1798
- (r"(None|descr|ConstClass|ConstPtr|TargetToken)", Name),
1799
- (r"<.*?>+", Name.Builtin),
1800
- (r"(label|debug_merge_point|jump|finish)", Name.Class),
1801
- (r"(int_add_ovf|int_add|int_sub_ovf|int_sub|int_mul_ovf|int_mul|"
1802
- r"int_floordiv|int_mod|int_lshift|int_rshift|int_and|int_or|"
1803
- r"int_xor|int_eq|int_ne|int_ge|int_gt|int_le|int_lt|int_is_zero|"
1804
- r"int_is_true|"
1805
- r"uint_floordiv|uint_ge|uint_lt|"
1806
- r"float_add|float_sub|float_mul|float_truediv|float_neg|"
1807
- r"float_eq|float_ne|float_ge|float_gt|float_le|float_lt|float_abs|"
1808
- r"ptr_eq|ptr_ne|instance_ptr_eq|instance_ptr_ne|"
1809
- r"cast_int_to_float|cast_float_to_int|"
1810
- r"force_token|quasiimmut_field|same_as|virtual_ref_finish|"
1811
- r"virtual_ref|mark_opaque_ptr|"
1812
- r"call_may_force|call_assembler|call_loopinvariant|"
1813
- r"call_release_gil|call_pure|call|"
1814
- r"new_with_vtable|new_array|newstr|newunicode|new|"
1815
- r"arraylen_gc|"
1816
- r"getarrayitem_gc_pure|getarrayitem_gc|setarrayitem_gc|"
1817
- r"getarrayitem_raw|setarrayitem_raw|getfield_gc_pure|"
1818
- r"getfield_gc|getinteriorfield_gc|setinteriorfield_gc|"
1819
- r"getfield_raw|setfield_gc|setfield_raw|"
1820
- r"strgetitem|strsetitem|strlen|copystrcontent|"
1821
- r"unicodegetitem|unicodesetitem|unicodelen|"
1822
- r"guard_true|guard_false|guard_value|guard_isnull|"
1823
- r"guard_nonnull_class|guard_nonnull|guard_class|guard_no_overflow|"
1824
- r"guard_not_forced|guard_no_exception|guard_not_invalidated)",
1825
- Name.Builtin),
1826
- include("extra-stuff"),
1827
- ],
1828
- "jit-backend-counts": [
1829
- (r"\[\w+\] jit-backend-counts}$", Keyword, "#pop"),
1830
- (r":", Punctuation),
1831
- (r"\d+", Number),
1832
- include("extra-stuff"),
1833
- ],
1834
- "extra-stuff": [
1835
- (r"\s+", Text),
1836
- (r"#.*?$", Comment),
1837
- ],
1838
- }
1839
-
1840
-
1841
- class HxmlLexer(RegexLexer):
1842
- """
1843
- Lexer for `haXe build <http://haxe.org/doc/compiler>`_ files.
1844
-
1845
- .. versionadded:: 1.6
1846
- """
1847
- name = 'Hxml'
1848
- aliases = ['haxeml', 'hxml']
1849
- filenames = ['*.hxml']
1850
-
1851
- tokens = {
1852
- 'root': [
1853
- # Seperator
1854
- (r'(--)(next)', bygroups(Punctuation, Generic.Heading)),
1855
- # Compiler switches with one dash
1856
- (r'(-)(prompt|debug|v)', bygroups(Punctuation, Keyword.Keyword)),
1857
- # Compilerswitches with two dashes
1858
- (r'(--)(neko-source|flash-strict|flash-use-stage|no-opt|no-traces|'
1859
- r'no-inline|times|no-output)', bygroups(Punctuation, Keyword)),
1860
- # Targets and other options that take an argument
1861
- (r'(-)(cpp|js|neko|x|as3|swf9?|swf-lib|php|xml|main|lib|D|resource|'
1862
- r'cp|cmd)( +)(.+)',
1863
- bygroups(Punctuation, Keyword, Whitespace, String)),
1864
- # Options that take only numerical arguments
1865
- (r'(-)(swf-version)( +)(\d+)',
1866
- bygroups(Punctuation, Keyword, Number.Integer)),
1867
- # An Option that defines the size, the fps and the background
1868
- # color of an flash movie
1869
- (r'(-)(swf-header)( +)(\d+)(:)(\d+)(:)(\d+)(:)([A-Fa-f0-9]{6})',
1870
- bygroups(Punctuation, Keyword, Whitespace, Number.Integer,
1871
- Punctuation, Number.Integer, Punctuation, Number.Integer,
1872
- Punctuation, Number.Hex)),
1873
- # options with two dashes that takes arguments
1874
- (r'(--)(js-namespace|php-front|php-lib|remap|gen-hx-classes)( +)'
1875
- r'(.+)', bygroups(Punctuation, Keyword, Whitespace, String)),
1876
- # Single line comment, multiline ones are not allowed.
1877
- (r'#.*', Comment.Single)
1878
- ]
1879
- }
1880
-
1881
-
1882
- class EbnfLexer(RegexLexer):
1883
- """
1884
- Lexer for `ISO/IEC 14977 EBNF
1885
- <http://en.wikipedia.org/wiki/Extended_Backus%E2%80%93Naur_Form>`_
1886
- grammars.
1887
-
1888
- .. versionadded:: 2.0
1889
- """
1890
-
1891
- name = 'EBNF'
1892
- aliases = ['ebnf']
1893
- filenames = ['*.ebnf']
1894
- mimetypes = ['text/x-ebnf']
1895
-
1896
- tokens = {
1897
- 'root': [
1898
- include('whitespace'),
1899
- include('comment_start'),
1900
- include('identifier'),
1901
- (r'=', Operator, 'production'),
1902
- ],
1903
- 'production': [
1904
- include('whitespace'),
1905
- include('comment_start'),
1906
- include('identifier'),
1907
- (r'"[^"]*"', String.Double),
1908
- (r"'[^']*'", String.Single),
1909
- (r'(\?[^?]*\?)', Name.Entity),
1910
- (r'[\[\]{}(),|]', Punctuation),
1911
- (r'-', Operator),
1912
- (r';', Punctuation, '#pop'),
1913
- ],
1914
- 'whitespace': [
1915
- (r'\s+', Text),
1916
- ],
1917
- 'comment_start': [
1918
- (r'\(\*', Comment.Multiline, 'comment'),
1919
- ],
1920
- 'comment': [
1921
- (r'[^*)]', Comment.Multiline),
1922
- include('comment_start'),
1923
- (r'\*\)', Comment.Multiline, '#pop'),
1924
- (r'[*)]', Comment.Multiline),
1925
- ],
1926
- 'identifier': [
1927
- (r'([a-zA-Z][a-zA-Z0-9 \-]*)', Keyword),
1928
- ],
1929
- }
1930
-
1931
- class TodotxtLexer(RegexLexer):
1932
- """
1933
- Lexer for `Todo.txt <http://todotxt.com/>`_ todo list format.
1934
-
1935
- .. versionadded:: 2.0
1936
- """
1937
-
1938
- name = 'Todotxt'
1939
- aliases = ['todotxt']
1940
- # *.todotxt is not a standard extension for Todo.txt files; including it
1941
- # makes testing easier, and also makes autodetecting file type easier.
1942
- filenames = ['todo.txt', '*.todotxt']
1943
- mimetypes = ['text/x-todo']
1944
-
1945
- ## Aliases mapping standard token types of Todo.txt format concepts
1946
- CompleteTaskText = Operator # Chosen to de-emphasize complete tasks
1947
- IncompleteTaskText = Text # Incomplete tasks should look like plain text
1948
-
1949
- # Priority should have most emphasis to indicate importance of tasks
1950
- Priority = Generic.Heading
1951
- # Dates should have next most emphasis because time is important
1952
- Date = Generic.Subheading
1953
-
1954
- # Project and context should have equal weight, and be in different colors
1955
- Project = Generic.Error
1956
- Context = String
1957
-
1958
- # If tag functionality is added, it should have the same weight as Project
1959
- # and Context, and a different color. Generic.Traceback would work well.
1960
-
1961
- # Regex patterns for building up rules; dates, priorities, projects, and
1962
- # contexts are all atomic
1963
- # TODO: Make date regex more ISO 8601 compliant
1964
- date_regex = r'\d{4,}-\d{2}-\d{2}'
1965
- priority_regex = r'\([A-Z]\)'
1966
- project_regex = r'\+\S+'
1967
- context_regex = r'@\S+'
1968
-
1969
- # Compound regex expressions
1970
- complete_one_date_regex = r'(x )(' + date_regex + r')'
1971
- complete_two_date_regex = (complete_one_date_regex + r'( )(' +
1972
- date_regex + r')')
1973
- priority_date_regex = r'(' + priority_regex + r')( )(' + date_regex + r')'
1974
-
1975
- tokens = {
1976
- # Should parse starting at beginning of line; each line is a task
1977
- 'root': [
1978
- ## Complete task entry points: two total:
1979
- # 1. Complete task with two dates
1980
- (complete_two_date_regex, bygroups(CompleteTaskText, Date,
1981
- CompleteTaskText, Date),
1982
- 'complete'),
1983
- # 2. Complete task with one date
1984
- (complete_one_date_regex, bygroups(CompleteTaskText, Date),
1985
- 'complete'),
1986
-
1987
- ## Incomplete task entry points: six total:
1988
- # 1. Priority plus date
1989
- (priority_date_regex, bygroups(Priority, IncompleteTaskText, Date),
1990
- 'incomplete'),
1991
- # 2. Priority only
1992
- (priority_regex, Priority, 'incomplete'),
1993
- # 3. Leading date
1994
- (date_regex, Date, 'incomplete'),
1995
- # 4. Leading context
1996
- (context_regex, Context, 'incomplete'),
1997
- # 5. Leading project
1998
- (project_regex, Project, 'incomplete'),
1999
- # 6. Non-whitespace catch-all
2000
- ('\S+', IncompleteTaskText, 'incomplete'),
2001
- ],
2002
-
2003
- # Parse a complete task
2004
- 'complete': [
2005
- # Newline indicates end of task, should return to root
2006
- (r'\s*\n', CompleteTaskText, '#pop'),
2007
- # Tokenize contexts and projects
2008
- (context_regex, Context),
2009
- (project_regex, Project),
2010
- # Tokenize non-whitespace text
2011
- ('\S+', CompleteTaskText),
2012
- # Tokenize whitespace not containing a newline
2013
- ('\s+', CompleteTaskText),
2014
- ],
2015
-
2016
- # Parse an incomplete task
2017
- 'incomplete': [
2018
- # Newline indicates end of task, should return to root
2019
- (r'\s*\n', IncompleteTaskText, '#pop'),
2020
- # Tokenize contexts and projects
2021
- (context_regex, Context),
2022
- (project_regex, Project),
2023
- # Tokenize non-whitespace text
2024
- ('\S+', IncompleteTaskText),
2025
- # Tokenize whitespace not containing a newline
2026
- ('\s+', IncompleteTaskText),
2027
- ],
2028
- }
2029
-
2030
-
2031
- class DockerLexer(RegexLexer):
2032
- """
2033
- Lexer for `Docker <http://docker.io>`_ configuration files.
2034
-
2035
- .. versionadded:: 2.0
2036
- """
2037
- name = 'Docker'
2038
- aliases = ['docker', 'dockerfile']
2039
- filenames = ['Dockerfile', '*.docker']
2040
- mimetypes = ['text/x-dockerfile-config']
2041
-
2042
- _keywords = (r'(?:FROM|MAINTAINER|RUN|CMD|EXPOSE|ENV|ADD|ENTRYPOINT|'
2043
- r'VOLUME|WORKDIR)')
2044
-
2045
- flags = re.IGNORECASE | re.MULTILINE
2046
-
2047
- tokens = {
2048
- 'root': [
2049
- (r'^(ONBUILD)(\s+)(%s)\b' % (_keywords,),
2050
- bygroups(Name.Keyword, Whitespace, Keyword)),
2051
- (_keywords + r'\b', Keyword),
2052
- (r'#.*', Comment),
2053
- (r'.+', using(BashLexer)),
2054
- ],
2055
- }
12
+ from pygments.lexers.configs import ApacheConfLexer, NginxConfLexer, \
13
+ SquidConfLexer, LighttpdConfLexer, IniLexer, RegeditLexer, PropertiesLexer
14
+ from pygments.lexers.console import PyPyLogLexer
15
+ from pygments.lexers.textedit import VimLexer
16
+ from pygments.lexers.markup import BBCodeLexer, MoinWikiLexer, RstLexer, \
17
+ TexLexer, GroffLexer
18
+ from pygments.lexers.installers import DebianControlLexer, SourcesListLexer
19
+ from pygments.lexers.make import MakefileLexer, BaseMakefileLexer, CMakeLexer
20
+ from pygments.lexers.haxe import HxmlLexer
21
+ from pygments.lexers.diff import DiffLexer, DarcsPatchLexer
22
+ from pygments.lexers.data import YamlLexer
23
+ from pygments.lexers.textfmts import IrcLogsLexer, GettextLexer, HttpLexer
24
+
25
+ __all__ = []