therubyracer 0.8.1.pre2 → 0.8.1

Sign up to get free protection for your applications and to get access to all the features.

Potentially problematic release.


This version of therubyracer might be problematic. Click here for more details.

Files changed (877) hide show
  1. data/Changelog.md +2 -1
  2. data/README.md +6 -3
  3. data/ext/v8/upstream/3.1.8/.gitignore +31 -0
  4. data/ext/v8/upstream/3.1.8/AUTHORS +40 -0
  5. data/ext/v8/upstream/3.1.8/ChangeLog +2566 -0
  6. data/ext/v8/upstream/3.1.8/LICENSE +52 -0
  7. data/ext/v8/upstream/3.1.8/LICENSE.strongtalk +29 -0
  8. data/ext/v8/upstream/3.1.8/LICENSE.v8 +26 -0
  9. data/ext/v8/upstream/3.1.8/LICENSE.valgrind +45 -0
  10. data/ext/v8/upstream/3.1.8/SConstruct +1192 -0
  11. data/ext/v8/upstream/3.1.8/build/README.txt +25 -0
  12. data/ext/v8/upstream/3.1.8/build/all.gyp +18 -0
  13. data/ext/v8/upstream/3.1.8/build/armu.gypi +32 -0
  14. data/ext/v8/upstream/3.1.8/build/common.gypi +82 -0
  15. data/ext/v8/upstream/3.1.8/build/gyp_v8 +145 -0
  16. data/ext/v8/upstream/3.1.8/include/v8-debug.h +384 -0
  17. data/ext/v8/upstream/3.1.8/include/v8-preparser.h +116 -0
  18. data/ext/v8/upstream/3.1.8/include/v8-profiler.h +426 -0
  19. data/ext/v8/upstream/3.1.8/include/v8-testing.h +99 -0
  20. data/ext/v8/upstream/3.1.8/include/v8.h +3846 -0
  21. data/ext/v8/upstream/3.1.8/include/v8stdint.h +53 -0
  22. data/ext/v8/upstream/3.1.8/preparser/preparser-process.cc +206 -0
  23. data/ext/v8/upstream/3.1.8/src/SConscript +356 -0
  24. data/ext/v8/upstream/3.1.8/src/accessors.cc +907 -0
  25. data/ext/v8/upstream/3.1.8/src/accessors.h +121 -0
  26. data/ext/v8/upstream/3.1.8/src/allocation.cc +204 -0
  27. data/ext/v8/upstream/3.1.8/src/allocation.h +176 -0
  28. data/ext/v8/upstream/3.1.8/src/api.cc +5191 -0
  29. data/ext/v8/upstream/3.1.8/src/api.h +508 -0
  30. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/apinatives.js +0 -0
  31. data/ext/v8/upstream/3.1.8/src/apiutils.h +80 -0
  32. data/ext/v8/upstream/3.1.8/src/arguments.h +105 -0
  33. data/ext/v8/upstream/3.1.8/src/arm/assembler-arm-inl.h +352 -0
  34. data/ext/v8/upstream/3.1.8/src/arm/assembler-arm.cc +2756 -0
  35. data/ext/v8/upstream/3.1.8/src/arm/assembler-arm.h +1294 -0
  36. data/ext/v8/upstream/3.1.8/src/arm/builtins-arm.cc +1628 -0
  37. data/ext/v8/upstream/3.1.8/src/arm/code-stubs-arm.cc +6783 -0
  38. data/ext/v8/upstream/3.1.8/src/arm/code-stubs-arm.h +657 -0
  39. data/ext/v8/upstream/3.1.8/src/arm/codegen-arm-inl.h +48 -0
  40. data/ext/v8/upstream/3.1.8/src/arm/codegen-arm.cc +7403 -0
  41. data/ext/v8/upstream/3.1.8/src/arm/codegen-arm.h +595 -0
  42. data/ext/v8/upstream/3.1.8/src/arm/constants-arm.cc +152 -0
  43. data/ext/v8/upstream/3.1.8/src/arm/constants-arm.h +769 -0
  44. data/ext/v8/upstream/3.1.8/src/arm/cpu-arm.cc +147 -0
  45. data/ext/v8/upstream/3.1.8/src/arm/debug-arm.cc +315 -0
  46. data/ext/v8/upstream/3.1.8/src/arm/deoptimizer-arm.cc +700 -0
  47. data/ext/v8/upstream/3.1.8/src/arm/disasm-arm.cc +1439 -0
  48. data/ext/v8/upstream/3.1.8/src/arm/frames-arm.cc +45 -0
  49. data/ext/v8/upstream/3.1.8/src/arm/frames-arm.h +168 -0
  50. data/ext/v8/upstream/3.1.8/src/arm/full-codegen-arm.cc +4230 -0
  51. data/ext/v8/upstream/3.1.8/src/arm/ic-arm.cc +1799 -0
  52. data/ext/v8/upstream/3.1.8/src/arm/jump-target-arm.cc +174 -0
  53. data/ext/v8/upstream/3.1.8/src/arm/lithium-arm.cc +2041 -0
  54. data/ext/v8/upstream/3.1.8/src/arm/lithium-arm.h +2046 -0
  55. data/ext/v8/upstream/3.1.8/src/arm/lithium-codegen-arm.cc +3822 -0
  56. data/ext/v8/upstream/3.1.8/src/arm/lithium-codegen-arm.h +312 -0
  57. data/ext/v8/upstream/3.1.8/src/arm/lithium-gap-resolver-arm.cc +303 -0
  58. data/ext/v8/upstream/3.1.8/src/arm/lithium-gap-resolver-arm.h +84 -0
  59. data/ext/v8/upstream/3.1.8/src/arm/macro-assembler-arm.cc +2701 -0
  60. data/ext/v8/upstream/3.1.8/src/arm/macro-assembler-arm.h +1015 -0
  61. data/ext/v8/upstream/3.1.8/src/arm/regexp-macro-assembler-arm.cc +1280 -0
  62. data/ext/v8/upstream/3.1.8/src/arm/regexp-macro-assembler-arm.h +252 -0
  63. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/arm/register-allocator-arm-inl.h +0 -0
  64. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/arm/register-allocator-arm.cc +0 -0
  65. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/arm/register-allocator-arm.h +0 -0
  66. data/ext/v8/upstream/3.1.8/src/arm/simulator-arm.cc +3165 -0
  67. data/ext/v8/upstream/3.1.8/src/arm/simulator-arm.h +402 -0
  68. data/ext/v8/upstream/3.1.8/src/arm/stub-cache-arm.cc +4077 -0
  69. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/arm/virtual-frame-arm-inl.h +0 -0
  70. data/ext/v8/upstream/3.1.8/src/arm/virtual-frame-arm.cc +843 -0
  71. data/ext/v8/upstream/3.1.8/src/arm/virtual-frame-arm.h +520 -0
  72. data/ext/v8/upstream/3.1.8/src/array.js +1231 -0
  73. data/ext/v8/upstream/3.1.8/src/assembler.cc +973 -0
  74. data/ext/v8/upstream/3.1.8/src/assembler.h +787 -0
  75. data/ext/v8/upstream/3.1.8/src/ast-inl.h +107 -0
  76. data/ext/v8/upstream/3.1.8/src/ast.cc +1067 -0
  77. data/ext/v8/upstream/3.1.8/src/ast.h +2177 -0
  78. data/ext/v8/upstream/3.1.8/src/atomicops.h +165 -0
  79. data/ext/v8/upstream/3.1.8/src/atomicops_internals_arm_gcc.h +145 -0
  80. data/ext/v8/upstream/3.1.8/src/atomicops_internals_x86_gcc.cc +126 -0
  81. data/ext/v8/upstream/3.1.8/src/atomicops_internals_x86_gcc.h +287 -0
  82. data/ext/v8/upstream/3.1.8/src/atomicops_internals_x86_macosx.h +301 -0
  83. data/ext/v8/upstream/3.1.8/src/atomicops_internals_x86_msvc.h +203 -0
  84. data/ext/v8/upstream/3.1.8/src/bignum-dtoa.cc +655 -0
  85. data/ext/v8/upstream/3.1.8/src/bignum-dtoa.h +81 -0
  86. data/ext/v8/upstream/3.1.8/src/bignum.cc +768 -0
  87. data/ext/v8/upstream/3.1.8/src/bignum.h +140 -0
  88. data/ext/v8/upstream/3.1.8/src/bootstrapper.cc +1888 -0
  89. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/bootstrapper.h +0 -0
  90. data/ext/v8/upstream/3.1.8/src/builtins.cc +1586 -0
  91. data/ext/v8/upstream/3.1.8/src/builtins.h +339 -0
  92. data/ext/v8/upstream/3.1.8/src/bytecodes-irregexp.h +105 -0
  93. data/ext/v8/upstream/3.1.8/src/cached-powers.cc +177 -0
  94. data/ext/v8/upstream/3.1.8/src/cached-powers.h +65 -0
  95. data/ext/v8/upstream/3.1.8/src/char-predicates-inl.h +94 -0
  96. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/char-predicates.h +0 -0
  97. data/ext/v8/upstream/3.1.8/src/checks.cc +110 -0
  98. data/ext/v8/upstream/3.1.8/src/checks.h +292 -0
  99. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/circular-queue-inl.h +0 -0
  100. data/ext/v8/upstream/3.1.8/src/circular-queue.cc +122 -0
  101. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/circular-queue.h +0 -0
  102. data/ext/v8/upstream/3.1.8/src/code-stubs.cc +230 -0
  103. data/ext/v8/upstream/3.1.8/src/code-stubs.h +950 -0
  104. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/code.h +0 -0
  105. data/ext/v8/upstream/3.1.8/src/codegen-inl.h +64 -0
  106. data/ext/v8/upstream/3.1.8/src/codegen.cc +495 -0
  107. data/ext/v8/upstream/3.1.8/src/codegen.h +245 -0
  108. data/ext/v8/upstream/3.1.8/src/compilation-cache.cc +654 -0
  109. data/ext/v8/upstream/3.1.8/src/compilation-cache.h +112 -0
  110. data/ext/v8/upstream/3.1.8/src/compiler.cc +806 -0
  111. data/ext/v8/upstream/3.1.8/src/compiler.h +290 -0
  112. data/ext/v8/upstream/3.1.8/src/contexts.cc +320 -0
  113. data/ext/v8/upstream/3.1.8/src/contexts.h +376 -0
  114. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/conversions-inl.h +0 -0
  115. data/ext/v8/upstream/3.1.8/src/conversions.cc +1069 -0
  116. data/ext/v8/upstream/3.1.8/src/conversions.h +122 -0
  117. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/counters.cc +0 -0
  118. data/ext/v8/upstream/3.1.8/src/counters.h +242 -0
  119. data/ext/v8/upstream/3.1.8/src/cpu-profiler-inl.h +100 -0
  120. data/ext/v8/upstream/3.1.8/src/cpu-profiler.cc +554 -0
  121. data/ext/v8/upstream/3.1.8/src/cpu-profiler.h +291 -0
  122. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/cpu.h +0 -0
  123. data/ext/v8/upstream/3.1.8/src/d8-debug.cc +367 -0
  124. data/ext/v8/upstream/3.1.8/src/d8-debug.h +157 -0
  125. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/d8-posix.cc +0 -0
  126. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/d8-readline.cc +0 -0
  127. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/d8-windows.cc +0 -0
  128. data/ext/v8/upstream/3.1.8/src/d8.cc +792 -0
  129. data/ext/v8/upstream/3.1.8/src/d8.gyp +85 -0
  130. data/ext/v8/upstream/3.1.8/src/d8.h +231 -0
  131. data/ext/v8/upstream/3.1.8/src/d8.js +2798 -0
  132. data/ext/v8/upstream/3.1.8/src/data-flow.cc +545 -0
  133. data/ext/v8/upstream/3.1.8/src/data-flow.h +379 -0
  134. data/ext/v8/upstream/3.1.8/src/date.js +1103 -0
  135. data/ext/v8/upstream/3.1.8/src/dateparser-inl.h +125 -0
  136. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/dateparser.cc +0 -0
  137. data/ext/v8/upstream/3.1.8/src/dateparser.h +263 -0
  138. data/ext/v8/upstream/3.1.8/src/debug-agent.cc +446 -0
  139. data/ext/v8/upstream/3.1.8/src/debug-agent.h +131 -0
  140. data/ext/v8/upstream/3.1.8/src/debug-debugger.js +2569 -0
  141. data/ext/v8/upstream/3.1.8/src/debug.cc +3085 -0
  142. data/ext/v8/upstream/3.1.8/src/debug.h +1025 -0
  143. data/ext/v8/upstream/3.1.8/src/deoptimizer.cc +1185 -0
  144. data/ext/v8/upstream/3.1.8/src/deoptimizer.h +529 -0
  145. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/disasm.h +0 -0
  146. data/ext/v8/upstream/3.1.8/src/disassembler.cc +338 -0
  147. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/disassembler.h +0 -0
  148. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/diy-fp.cc +0 -0
  149. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/diy-fp.h +0 -0
  150. data/ext/v8/upstream/3.1.8/src/double.h +238 -0
  151. data/ext/v8/upstream/3.1.8/src/dtoa.cc +103 -0
  152. data/ext/v8/upstream/3.1.8/src/dtoa.h +85 -0
  153. data/ext/v8/upstream/3.1.8/src/execution.cc +735 -0
  154. data/ext/v8/upstream/3.1.8/src/execution.h +322 -0
  155. data/ext/v8/upstream/3.1.8/src/extensions/experimental/experimental.gyp +53 -0
  156. data/ext/v8/upstream/3.1.8/src/extensions/experimental/i18n-extension.cc +264 -0
  157. data/ext/v8/upstream/3.1.8/src/extensions/experimental/i18n-extension.h +64 -0
  158. data/ext/v8/upstream/3.1.8/src/extensions/externalize-string-extension.cc +141 -0
  159. data/ext/v8/upstream/3.1.8/src/extensions/externalize-string-extension.h +50 -0
  160. data/ext/v8/upstream/3.1.8/src/extensions/gc-extension.cc +58 -0
  161. data/ext/v8/upstream/3.1.8/src/extensions/gc-extension.h +49 -0
  162. data/ext/v8/upstream/3.1.8/src/factory.cc +1087 -0
  163. data/ext/v8/upstream/3.1.8/src/factory.h +432 -0
  164. data/ext/v8/upstream/3.1.8/src/fast-dtoa.cc +736 -0
  165. data/ext/v8/upstream/3.1.8/src/fast-dtoa.h +83 -0
  166. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/fixed-dtoa.cc +0 -0
  167. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/fixed-dtoa.h +0 -0
  168. data/ext/v8/upstream/3.1.8/src/flag-definitions.h +552 -0
  169. data/ext/v8/upstream/3.1.8/src/flags.cc +551 -0
  170. data/ext/v8/upstream/3.1.8/src/flags.h +79 -0
  171. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/frame-element.cc +0 -0
  172. data/ext/v8/upstream/3.1.8/src/frame-element.h +277 -0
  173. data/ext/v8/upstream/3.1.8/src/frames-inl.h +210 -0
  174. data/ext/v8/upstream/3.1.8/src/frames.cc +1232 -0
  175. data/ext/v8/upstream/3.1.8/src/frames.h +826 -0
  176. data/ext/v8/upstream/3.1.8/src/full-codegen.cc +1382 -0
  177. data/ext/v8/upstream/3.1.8/src/full-codegen.h +751 -0
  178. data/ext/v8/upstream/3.1.8/src/func-name-inferrer.cc +90 -0
  179. data/ext/v8/upstream/3.1.8/src/func-name-inferrer.h +111 -0
  180. data/ext/v8/upstream/3.1.8/src/gdb-jit.cc +1547 -0
  181. data/ext/v8/upstream/3.1.8/src/gdb-jit.h +138 -0
  182. data/ext/v8/upstream/3.1.8/src/global-handles.cc +534 -0
  183. data/ext/v8/upstream/3.1.8/src/global-handles.h +181 -0
  184. data/ext/v8/upstream/3.1.8/src/globals.h +325 -0
  185. data/ext/v8/upstream/3.1.8/src/handles-inl.h +80 -0
  186. data/ext/v8/upstream/3.1.8/src/handles.cc +910 -0
  187. data/ext/v8/upstream/3.1.8/src/handles.h +424 -0
  188. data/ext/v8/upstream/3.1.8/src/hashmap.cc +230 -0
  189. data/ext/v8/upstream/3.1.8/src/hashmap.h +121 -0
  190. data/ext/v8/upstream/3.1.8/src/heap-inl.h +587 -0
  191. data/ext/v8/upstream/3.1.8/src/heap-profiler.cc +1128 -0
  192. data/ext/v8/upstream/3.1.8/src/heap-profiler.h +381 -0
  193. data/ext/v8/upstream/3.1.8/src/heap.cc +5610 -0
  194. data/ext/v8/upstream/3.1.8/src/heap.h +2218 -0
  195. data/ext/v8/upstream/3.1.8/src/hydrogen-instructions.cc +1490 -0
  196. data/ext/v8/upstream/3.1.8/src/hydrogen-instructions.h +3493 -0
  197. data/ext/v8/upstream/3.1.8/src/hydrogen.cc +6056 -0
  198. data/ext/v8/upstream/3.1.8/src/hydrogen.h +1091 -0
  199. data/ext/v8/upstream/3.1.8/src/ia32/assembler-ia32-inl.h +429 -0
  200. data/ext/v8/upstream/3.1.8/src/ia32/assembler-ia32.cc +2800 -0
  201. data/ext/v8/upstream/3.1.8/src/ia32/assembler-ia32.h +1093 -0
  202. data/ext/v8/upstream/3.1.8/src/ia32/builtins-ia32.cc +1590 -0
  203. data/ext/v8/upstream/3.1.8/src/ia32/code-stubs-ia32.cc +6624 -0
  204. data/ext/v8/upstream/3.1.8/src/ia32/code-stubs-ia32.h +536 -0
  205. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/ia32/codegen-ia32-inl.h +0 -0
  206. data/ext/v8/upstream/3.1.8/src/ia32/codegen-ia32.cc +10354 -0
  207. data/ext/v8/upstream/3.1.8/src/ia32/codegen-ia32.h +798 -0
  208. data/ext/v8/upstream/3.1.8/src/ia32/cpu-ia32.cc +87 -0
  209. data/ext/v8/upstream/3.1.8/src/ia32/debug-ia32.cc +309 -0
  210. data/ext/v8/upstream/3.1.8/src/ia32/deoptimizer-ia32.cc +664 -0
  211. data/ext/v8/upstream/3.1.8/src/ia32/disasm-ia32.cc +1597 -0
  212. data/ext/v8/upstream/3.1.8/src/ia32/frames-ia32.cc +45 -0
  213. data/ext/v8/upstream/3.1.8/src/ia32/frames-ia32.h +140 -0
  214. data/ext/v8/upstream/3.1.8/src/ia32/full-codegen-ia32.cc +4278 -0
  215. data/ext/v8/upstream/3.1.8/src/ia32/ic-ia32.cc +1786 -0
  216. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/ia32/jump-target-ia32.cc +0 -0
  217. data/ext/v8/upstream/3.1.8/src/ia32/lithium-codegen-ia32.cc +3880 -0
  218. data/ext/v8/upstream/3.1.8/src/ia32/lithium-codegen-ia32.h +309 -0
  219. data/ext/v8/upstream/3.1.8/src/ia32/lithium-gap-resolver-ia32.cc +460 -0
  220. data/ext/v8/upstream/3.1.8/src/ia32/lithium-gap-resolver-ia32.h +110 -0
  221. data/ext/v8/upstream/3.1.8/src/ia32/lithium-ia32.cc +2095 -0
  222. data/ext/v8/upstream/3.1.8/src/ia32/lithium-ia32.h +2127 -0
  223. data/ext/v8/upstream/3.1.8/src/ia32/macro-assembler-ia32.cc +2031 -0
  224. data/ext/v8/upstream/3.1.8/src/ia32/macro-assembler-ia32.h +798 -0
  225. data/ext/v8/upstream/3.1.8/src/ia32/regexp-macro-assembler-ia32.cc +1253 -0
  226. data/ext/v8/upstream/3.1.8/src/ia32/regexp-macro-assembler-ia32.h +215 -0
  227. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/ia32/register-allocator-ia32-inl.h +0 -0
  228. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/ia32/register-allocator-ia32.cc +0 -0
  229. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/ia32/register-allocator-ia32.h +0 -0
  230. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/ia32/simulator-ia32.cc +0 -0
  231. data/ext/v8/upstream/3.1.8/src/ia32/simulator-ia32.h +72 -0
  232. data/ext/v8/upstream/3.1.8/src/ia32/stub-cache-ia32.cc +3732 -0
  233. data/ext/v8/upstream/3.1.8/src/ia32/virtual-frame-ia32.cc +1360 -0
  234. data/ext/v8/upstream/3.1.8/src/ia32/virtual-frame-ia32.h +646 -0
  235. data/ext/v8/upstream/3.1.8/src/ic-inl.h +129 -0
  236. data/ext/v8/upstream/3.1.8/src/ic.cc +2333 -0
  237. data/ext/v8/upstream/3.1.8/src/ic.h +639 -0
  238. data/ext/v8/upstream/3.1.8/src/inspector.cc +63 -0
  239. data/ext/v8/upstream/3.1.8/src/inspector.h +62 -0
  240. data/ext/v8/upstream/3.1.8/src/interpreter-irregexp.cc +655 -0
  241. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/interpreter-irregexp.h +0 -0
  242. data/ext/v8/upstream/3.1.8/src/json.js +342 -0
  243. data/ext/v8/upstream/3.1.8/src/jsregexp.cc +5340 -0
  244. data/ext/v8/upstream/3.1.8/src/jsregexp.h +1484 -0
  245. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/jump-target-heavy-inl.h +0 -0
  246. data/ext/v8/upstream/3.1.8/src/jump-target-heavy.cc +430 -0
  247. data/ext/v8/upstream/3.1.8/src/jump-target-heavy.h +244 -0
  248. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/jump-target-inl.h +0 -0
  249. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/jump-target-light-inl.h +0 -0
  250. data/ext/v8/upstream/3.1.8/src/jump-target-light.cc +111 -0
  251. data/ext/v8/upstream/3.1.8/src/jump-target-light.h +193 -0
  252. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/jump-target.cc +0 -0
  253. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/jump-target.h +0 -0
  254. data/ext/v8/upstream/3.1.8/src/list-inl.h +206 -0
  255. data/ext/v8/upstream/3.1.8/src/list.h +164 -0
  256. data/ext/v8/upstream/3.1.8/src/lithium-allocator-inl.h +140 -0
  257. data/ext/v8/upstream/3.1.8/src/lithium-allocator.cc +2093 -0
  258. data/ext/v8/upstream/3.1.8/src/lithium-allocator.h +644 -0
  259. data/ext/v8/upstream/3.1.8/src/lithium.cc +168 -0
  260. data/ext/v8/upstream/3.1.8/src/lithium.h +592 -0
  261. data/ext/v8/upstream/3.1.8/src/liveedit-debugger.js +1082 -0
  262. data/ext/v8/upstream/3.1.8/src/liveedit.cc +1650 -0
  263. data/ext/v8/upstream/3.1.8/src/liveedit.h +174 -0
  264. data/ext/v8/upstream/3.1.8/src/liveobjectlist-inl.h +126 -0
  265. data/ext/v8/upstream/3.1.8/src/liveobjectlist.cc +2527 -0
  266. data/ext/v8/upstream/3.1.8/src/liveobjectlist.h +322 -0
  267. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/log-inl.h +0 -0
  268. data/ext/v8/upstream/3.1.8/src/log-utils.cc +336 -0
  269. data/ext/v8/upstream/3.1.8/src/log-utils.h +232 -0
  270. data/ext/v8/upstream/3.1.8/src/log.cc +1608 -0
  271. data/ext/v8/upstream/3.1.8/src/log.h +379 -0
  272. data/ext/v8/upstream/3.1.8/src/macro-assembler.h +120 -0
  273. data/ext/v8/upstream/3.1.8/src/macros.py +178 -0
  274. data/ext/v8/upstream/3.1.8/src/mark-compact.cc +2957 -0
  275. data/ext/v8/upstream/3.1.8/src/mark-compact.h +433 -0
  276. data/ext/v8/upstream/3.1.8/src/math.js +264 -0
  277. data/ext/v8/upstream/3.1.8/src/memory.h +82 -0
  278. data/ext/v8/upstream/3.1.8/src/messages.cc +164 -0
  279. data/ext/v8/upstream/3.1.8/src/messages.h +114 -0
  280. data/ext/v8/upstream/3.1.8/src/messages.js +1071 -0
  281. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/mips/assembler-mips-inl.h +0 -0
  282. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/mips/assembler-mips.cc +0 -0
  283. data/ext/v8/upstream/3.1.8/src/mips/assembler-mips.h +667 -0
  284. data/ext/v8/upstream/3.1.8/src/mips/builtins-mips.cc +205 -0
  285. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/mips/codegen-mips-inl.h +0 -0
  286. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/mips/codegen-mips.cc +0 -0
  287. data/ext/v8/upstream/3.1.8/src/mips/codegen-mips.h +431 -0
  288. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/mips/constants-mips.cc +0 -0
  289. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/mips/constants-mips.h +0 -0
  290. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/mips/cpu-mips.cc +0 -0
  291. data/ext/v8/upstream/3.1.8/src/mips/debug-mips.cc +127 -0
  292. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/mips/disasm-mips.cc +0 -0
  293. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/mips/fast-codegen-mips.cc +0 -0
  294. data/ext/v8/upstream/3.1.8/src/mips/frames-mips.cc +96 -0
  295. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/mips/frames-mips.h +0 -0
  296. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/mips/full-codegen-mips.cc +0 -0
  297. data/ext/v8/upstream/3.1.8/src/mips/ic-mips.cc +208 -0
  298. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/mips/jump-target-mips.cc +0 -0
  299. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/mips/macro-assembler-mips.cc +0 -0
  300. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/mips/macro-assembler-mips.h +0 -0
  301. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/mips/register-allocator-mips-inl.h +0 -0
  302. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/mips/register-allocator-mips.cc +0 -0
  303. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/mips/register-allocator-mips.h +0 -0
  304. data/ext/v8/upstream/3.1.8/src/mips/simulator-mips.cc +1650 -0
  305. data/ext/v8/upstream/3.1.8/src/mips/simulator-mips.h +311 -0
  306. data/ext/v8/upstream/3.1.8/src/mips/stub-cache-mips.cc +418 -0
  307. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/mips/virtual-frame-mips.cc +0 -0
  308. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/mips/virtual-frame-mips.h +0 -0
  309. data/ext/v8/upstream/3.1.8/src/mirror-debugger.js +2380 -0
  310. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/mksnapshot.cc +0 -0
  311. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/natives.h +0 -0
  312. data/ext/v8/upstream/3.1.8/src/objects-debug.cc +722 -0
  313. data/ext/v8/upstream/3.1.8/src/objects-inl.h +3946 -0
  314. data/ext/v8/upstream/3.1.8/src/objects-printer.cc +801 -0
  315. data/ext/v8/upstream/3.1.8/src/objects-visiting.cc +142 -0
  316. data/ext/v8/upstream/3.1.8/src/objects-visiting.h +401 -0
  317. data/ext/v8/upstream/3.1.8/src/objects.cc +10044 -0
  318. data/ext/v8/upstream/3.1.8/src/objects.h +6571 -0
  319. data/ext/v8/upstream/3.1.8/src/parser.cc +5165 -0
  320. data/ext/v8/upstream/3.1.8/src/parser.h +802 -0
  321. data/ext/v8/upstream/3.1.8/src/platform-cygwin.cc +745 -0
  322. data/ext/v8/upstream/3.1.8/src/platform-freebsd.cc +702 -0
  323. data/ext/v8/upstream/3.1.8/src/platform-linux.cc +981 -0
  324. data/ext/v8/upstream/3.1.8/src/platform-macos.cc +732 -0
  325. data/ext/v8/upstream/3.1.8/src/platform-nullos.cc +498 -0
  326. data/ext/v8/upstream/3.1.8/src/platform-openbsd.cc +657 -0
  327. data/ext/v8/upstream/3.1.8/src/platform-posix.cc +399 -0
  328. data/ext/v8/upstream/3.1.8/src/platform-solaris.cc +714 -0
  329. data/ext/v8/upstream/3.1.8/src/platform-win32.cc +1974 -0
  330. data/ext/v8/upstream/3.1.8/src/platform.h +636 -0
  331. data/ext/v8/upstream/3.1.8/src/preparse-data.cc +183 -0
  332. data/ext/v8/upstream/3.1.8/src/preparse-data.h +249 -0
  333. data/ext/v8/upstream/3.1.8/src/preparser-api.cc +213 -0
  334. data/ext/v8/upstream/3.1.8/src/preparser.cc +1205 -0
  335. data/ext/v8/upstream/3.1.8/src/preparser.h +278 -0
  336. data/ext/v8/upstream/3.1.8/src/prettyprinter.cc +1539 -0
  337. data/ext/v8/upstream/3.1.8/src/prettyprinter.h +223 -0
  338. data/ext/v8/upstream/3.1.8/src/profile-generator-inl.h +128 -0
  339. data/ext/v8/upstream/3.1.8/src/profile-generator.cc +2899 -0
  340. data/ext/v8/upstream/3.1.8/src/profile-generator.h +1151 -0
  341. data/ext/v8/upstream/3.1.8/src/property.cc +96 -0
  342. data/ext/v8/upstream/3.1.8/src/property.h +337 -0
  343. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/regexp-macro-assembler-irregexp-inl.h +0 -0
  344. data/ext/v8/upstream/3.1.8/src/regexp-macro-assembler-irregexp.cc +470 -0
  345. data/ext/v8/upstream/3.1.8/src/regexp-macro-assembler-irregexp.h +142 -0
  346. data/ext/v8/upstream/3.1.8/src/regexp-macro-assembler-tracer.cc +373 -0
  347. data/ext/v8/upstream/3.1.8/src/regexp-macro-assembler-tracer.h +104 -0
  348. data/ext/v8/upstream/3.1.8/src/regexp-macro-assembler.cc +257 -0
  349. data/ext/v8/upstream/3.1.8/src/regexp-macro-assembler.h +231 -0
  350. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/regexp-stack.cc +0 -0
  351. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/regexp-stack.h +0 -0
  352. data/ext/v8/upstream/3.1.8/src/regexp.js +483 -0
  353. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/register-allocator-inl.h +0 -0
  354. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/register-allocator.cc +0 -0
  355. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/register-allocator.h +0 -0
  356. data/ext/v8/upstream/3.1.8/src/rewriter.cc +1023 -0
  357. data/ext/v8/upstream/3.1.8/src/rewriter.h +59 -0
  358. data/ext/v8/upstream/3.1.8/src/runtime-profiler.cc +443 -0
  359. data/ext/v8/upstream/3.1.8/src/runtime-profiler.h +77 -0
  360. data/ext/v8/upstream/3.1.8/src/runtime.cc +11592 -0
  361. data/ext/v8/upstream/3.1.8/src/runtime.h +582 -0
  362. data/ext/v8/upstream/3.1.8/src/runtime.js +643 -0
  363. data/ext/v8/upstream/3.1.8/src/safepoint-table.cc +253 -0
  364. data/ext/v8/upstream/3.1.8/src/safepoint-table.h +263 -0
  365. data/ext/v8/upstream/3.1.8/src/scanner-base.cc +971 -0
  366. data/ext/v8/upstream/3.1.8/src/scanner-base.h +653 -0
  367. data/ext/v8/upstream/3.1.8/src/scanner.cc +586 -0
  368. data/ext/v8/upstream/3.1.8/src/scanner.h +194 -0
  369. data/ext/v8/upstream/3.1.8/src/scopeinfo.cc +636 -0
  370. data/ext/v8/upstream/3.1.8/src/scopeinfo.h +238 -0
  371. data/ext/v8/upstream/3.1.8/src/scopes.cc +1063 -0
  372. data/ext/v8/upstream/3.1.8/src/scopes.h +494 -0
  373. data/ext/v8/upstream/3.1.8/src/serialize.cc +1535 -0
  374. data/ext/v8/upstream/3.1.8/src/serialize.h +584 -0
  375. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/shell.h +0 -0
  376. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/simulator.h +0 -0
  377. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/smart-pointer.h +0 -0
  378. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/snapshot-common.cc +0 -0
  379. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/snapshot-empty.cc +0 -0
  380. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/snapshot.h +0 -0
  381. data/ext/v8/upstream/3.1.8/src/spaces-inl.h +524 -0
  382. data/ext/v8/upstream/3.1.8/src/spaces.cc +3254 -0
  383. data/ext/v8/upstream/3.1.8/src/spaces.h +2362 -0
  384. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/splay-tree-inl.h +0 -0
  385. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/splay-tree.h +0 -0
  386. data/ext/v8/upstream/3.1.8/src/string-search.cc +40 -0
  387. data/ext/v8/upstream/3.1.8/src/string-search.h +567 -0
  388. data/ext/v8/upstream/3.1.8/src/string-stream.cc +584 -0
  389. data/ext/v8/upstream/3.1.8/src/string-stream.h +191 -0
  390. data/ext/v8/upstream/3.1.8/src/string.js +915 -0
  391. data/ext/v8/upstream/3.1.8/src/strtod.cc +440 -0
  392. data/ext/v8/upstream/3.1.8/src/strtod.h +40 -0
  393. data/ext/v8/upstream/3.1.8/src/stub-cache.cc +1878 -0
  394. data/ext/v8/upstream/3.1.8/src/stub-cache.h +849 -0
  395. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/third_party/valgrind/valgrind.h +0 -0
  396. data/ext/v8/upstream/3.1.8/src/token.cc +63 -0
  397. data/ext/v8/upstream/3.1.8/src/token.h +288 -0
  398. data/ext/v8/upstream/3.1.8/src/top.cc +1152 -0
  399. data/ext/v8/upstream/3.1.8/src/top.h +608 -0
  400. data/ext/v8/upstream/3.1.8/src/type-info.cc +406 -0
  401. data/ext/v8/upstream/3.1.8/src/type-info.h +283 -0
  402. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/unbound-queue-inl.h +0 -0
  403. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/unbound-queue.h +0 -0
  404. data/ext/v8/upstream/3.1.8/src/unicode-inl.h +238 -0
  405. data/ext/v8/upstream/3.1.8/src/unicode.cc +1624 -0
  406. data/ext/v8/upstream/3.1.8/src/unicode.h +280 -0
  407. data/ext/v8/upstream/3.1.8/src/uri.js +402 -0
  408. data/ext/v8/upstream/3.1.8/src/utils.cc +371 -0
  409. data/ext/v8/upstream/3.1.8/src/utils.h +793 -0
  410. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/v8-counters.cc +0 -0
  411. data/ext/v8/upstream/3.1.8/src/v8-counters.h +290 -0
  412. data/ext/v8/upstream/3.1.8/src/v8.cc +270 -0
  413. data/ext/v8/upstream/3.1.8/src/v8.h +127 -0
  414. data/ext/v8/upstream/3.1.8/src/v8checks.h +64 -0
  415. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/v8dll-main.cc +0 -0
  416. data/ext/v8/upstream/3.1.8/src/v8globals.h +480 -0
  417. data/ext/v8/upstream/3.1.8/src/v8natives.js +1252 -0
  418. data/ext/v8/upstream/3.1.8/src/v8preparserdll-main.cc +39 -0
  419. data/ext/v8/upstream/3.1.8/src/v8threads.cc +440 -0
  420. data/ext/v8/upstream/3.1.8/src/v8threads.h +157 -0
  421. data/ext/v8/upstream/3.1.8/src/v8utils.h +354 -0
  422. data/ext/v8/upstream/3.1.8/src/variables.cc +132 -0
  423. data/ext/v8/upstream/3.1.8/src/variables.h +212 -0
  424. data/ext/v8/upstream/3.1.8/src/version.cc +95 -0
  425. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/version.h +0 -0
  426. data/ext/v8/upstream/3.1.8/src/virtual-frame-heavy-inl.h +190 -0
  427. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/virtual-frame-heavy.cc +0 -0
  428. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/virtual-frame-inl.h +0 -0
  429. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/virtual-frame-light-inl.h +0 -0
  430. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/virtual-frame-light.cc +0 -0
  431. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/virtual-frame.cc +0 -0
  432. data/ext/v8/upstream/3.1.8/src/virtual-frame.h +59 -0
  433. data/ext/v8/upstream/3.1.8/src/vm-state-inl.h +134 -0
  434. data/ext/v8/upstream/3.1.8/src/vm-state.h +68 -0
  435. data/ext/v8/upstream/3.1.8/src/win32-headers.h +95 -0
  436. data/ext/v8/upstream/3.1.8/src/x64/assembler-x64-inl.h +455 -0
  437. data/ext/v8/upstream/3.1.8/src/x64/assembler-x64.cc +3162 -0
  438. data/ext/v8/upstream/3.1.8/src/x64/assembler-x64.h +1584 -0
  439. data/ext/v8/upstream/3.1.8/src/x64/builtins-x64.cc +1492 -0
  440. data/ext/v8/upstream/3.1.8/src/x64/code-stubs-x64.cc +5150 -0
  441. data/ext/v8/upstream/3.1.8/src/x64/code-stubs-x64.h +519 -0
  442. data/ext/v8/upstream/3.1.8/src/x64/codegen-x64-inl.h +46 -0
  443. data/ext/v8/upstream/3.1.8/src/x64/codegen-x64.cc +8835 -0
  444. data/ext/v8/upstream/3.1.8/src/x64/codegen-x64.h +750 -0
  445. data/ext/v8/upstream/3.1.8/src/x64/cpu-x64.cc +86 -0
  446. data/ext/v8/upstream/3.1.8/src/x64/debug-x64.cc +316 -0
  447. data/ext/v8/upstream/3.1.8/src/x64/deoptimizer-x64.cc +781 -0
  448. data/ext/v8/upstream/3.1.8/src/x64/disasm-x64.cc +1737 -0
  449. data/ext/v8/upstream/3.1.8/src/x64/frames-x64.cc +45 -0
  450. data/ext/v8/upstream/3.1.8/src/x64/frames-x64.h +130 -0
  451. data/ext/v8/upstream/3.1.8/src/x64/full-codegen-x64.cc +3984 -0
  452. data/ext/v8/upstream/3.1.8/src/x64/ic-x64.cc +1761 -0
  453. data/ext/v8/upstream/3.1.8/src/x64/jump-target-x64.cc +437 -0
  454. data/ext/v8/upstream/3.1.8/src/x64/lithium-codegen-x64.cc +3639 -0
  455. data/ext/v8/upstream/3.1.8/src/x64/lithium-codegen-x64.h +305 -0
  456. data/ext/v8/upstream/3.1.8/src/x64/lithium-gap-resolver-x64.cc +320 -0
  457. data/ext/v8/upstream/3.1.8/src/x64/lithium-gap-resolver-x64.h +74 -0
  458. data/ext/v8/upstream/3.1.8/src/x64/lithium-x64.cc +2044 -0
  459. data/ext/v8/upstream/3.1.8/src/x64/lithium-x64.h +2052 -0
  460. data/ext/v8/upstream/3.1.8/src/x64/macro-assembler-x64.cc +2660 -0
  461. data/ext/v8/upstream/3.1.8/src/x64/macro-assembler-x64.h +1852 -0
  462. data/ext/v8/upstream/3.1.8/src/x64/regexp-macro-assembler-x64.cc +1382 -0
  463. data/ext/v8/upstream/3.1.8/src/x64/regexp-macro-assembler-x64.h +278 -0
  464. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/x64/register-allocator-x64-inl.h +0 -0
  465. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/x64/register-allocator-x64.cc +0 -0
  466. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/x64/register-allocator-x64.h +0 -0
  467. data/ext/v8/upstream/{2.3.3 → 3.1.8}/src/x64/simulator-x64.cc +0 -0
  468. data/ext/v8/upstream/3.1.8/src/x64/simulator-x64.h +71 -0
  469. data/ext/v8/upstream/3.1.8/src/x64/stub-cache-x64.cc +3509 -0
  470. data/ext/v8/upstream/3.1.8/src/x64/virtual-frame-x64.cc +1292 -0
  471. data/ext/v8/upstream/3.1.8/src/x64/virtual-frame-x64.h +593 -0
  472. data/ext/v8/upstream/3.1.8/src/zone-inl.h +83 -0
  473. data/ext/v8/upstream/3.1.8/src/zone.cc +195 -0
  474. data/ext/v8/upstream/3.1.8/src/zone.h +233 -0
  475. data/ext/v8/upstream/3.1.8/tools/codemap.js +265 -0
  476. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/consarray.js +0 -0
  477. data/ext/v8/upstream/3.1.8/tools/csvparser.js +78 -0
  478. data/ext/v8/upstream/3.1.8/tools/disasm.py +92 -0
  479. data/ext/v8/upstream/3.1.8/tools/gc-nvp-trace-processor.py +328 -0
  480. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/generate-ten-powers.scm +0 -0
  481. data/ext/v8/upstream/3.1.8/tools/grokdump.py +840 -0
  482. data/ext/v8/upstream/3.1.8/tools/gyp/v8.gyp +869 -0
  483. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/js2c.py +0 -0
  484. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/jsmin.py +0 -0
  485. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/linux-tick-processor +0 -0
  486. data/ext/v8/upstream/3.1.8/tools/ll_prof.py +919 -0
  487. data/ext/v8/upstream/3.1.8/tools/logreader.js +185 -0
  488. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/mac-nm +0 -0
  489. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/mac-tick-processor +0 -0
  490. data/ext/v8/upstream/3.1.8/tools/oom_dump/README +31 -0
  491. data/ext/v8/upstream/3.1.8/tools/oom_dump/SConstruct +42 -0
  492. data/ext/v8/upstream/3.1.8/tools/oom_dump/oom_dump.cc +288 -0
  493. data/ext/v8/upstream/3.1.8/tools/presubmit.py +305 -0
  494. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/process-heap-prof.py +0 -0
  495. data/ext/v8/upstream/3.1.8/tools/profile.js +751 -0
  496. data/ext/v8/upstream/3.1.8/tools/profile_view.js +219 -0
  497. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/run-valgrind.py +0 -0
  498. data/ext/v8/upstream/3.1.8/tools/splaytree.js +316 -0
  499. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/stats-viewer.py +0 -0
  500. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/tickprocessor-driver.js +0 -0
  501. data/ext/v8/upstream/3.1.8/tools/tickprocessor.js +863 -0
  502. data/ext/v8/upstream/3.1.8/tools/utils.py +96 -0
  503. data/ext/v8/upstream/3.1.8/tools/visual_studio/README.txt +70 -0
  504. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/visual_studio/arm.vsprops +0 -0
  505. data/ext/v8/upstream/3.1.8/tools/visual_studio/common.vsprops +34 -0
  506. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/visual_studio/d8.vcproj +0 -0
  507. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/visual_studio/d8_arm.vcproj +0 -0
  508. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/visual_studio/d8_x64.vcproj +0 -0
  509. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/visual_studio/d8js2c.cmd +0 -0
  510. data/ext/v8/upstream/3.1.8/tools/visual_studio/debug.vsprops +17 -0
  511. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/visual_studio/ia32.vsprops +0 -0
  512. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/visual_studio/js2c.cmd +0 -0
  513. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/visual_studio/release.vsprops +0 -0
  514. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/visual_studio/v8.sln +0 -0
  515. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/visual_studio/v8.vcproj +0 -0
  516. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/visual_studio/v8_arm.sln +0 -0
  517. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/visual_studio/v8_arm.vcproj +0 -0
  518. data/ext/v8/upstream/3.1.8/tools/visual_studio/v8_base.vcproj +1296 -0
  519. data/ext/v8/upstream/3.1.8/tools/visual_studio/v8_base_arm.vcproj +1234 -0
  520. data/ext/v8/upstream/3.1.8/tools/visual_studio/v8_base_x64.vcproj +1296 -0
  521. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/visual_studio/v8_cctest.vcproj +0 -0
  522. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/visual_studio/v8_cctest_arm.vcproj +0 -0
  523. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/visual_studio/v8_cctest_x64.vcproj +0 -0
  524. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/visual_studio/v8_mksnapshot.vcproj +0 -0
  525. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/visual_studio/v8_mksnapshot_x64.vcproj +0 -0
  526. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/visual_studio/v8_process_sample.vcproj +0 -0
  527. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/visual_studio/v8_process_sample_arm.vcproj +0 -0
  528. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/visual_studio/v8_process_sample_x64.vcproj +0 -0
  529. data/ext/v8/upstream/3.1.8/tools/visual_studio/v8_shell_sample.vcproj +147 -0
  530. data/ext/v8/upstream/3.1.8/tools/visual_studio/v8_shell_sample_arm.vcproj +147 -0
  531. data/ext/v8/upstream/3.1.8/tools/visual_studio/v8_shell_sample_x64.vcproj +163 -0
  532. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/visual_studio/v8_snapshot.vcproj +0 -0
  533. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/visual_studio/v8_snapshot_cc.vcproj +0 -0
  534. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/visual_studio/v8_snapshot_cc_x64.vcproj +0 -0
  535. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/visual_studio/v8_snapshot_x64.vcproj +0 -0
  536. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/visual_studio/v8_x64.sln +0 -0
  537. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/visual_studio/v8_x64.vcproj +0 -0
  538. data/ext/v8/upstream/{2.3.3 → 3.1.8}/tools/visual_studio/x64.vsprops +0 -0
  539. data/ext/v8/upstream/3.1.8/tools/windows-tick-processor.bat +30 -0
  540. data/ext/v8/upstream/Makefile +2 -1
  541. data/ext/v8/v8_template.cpp +2 -2
  542. data/lib/v8/version.rb +1 -1
  543. data/spec/redjs/jsapi_spec.rb +2 -2
  544. metadata +552 -490
  545. data/ext/v8/upstream/2.3.3/.gitignore +0 -26
  546. data/ext/v8/upstream/2.3.3/AUTHORS +0 -31
  547. data/ext/v8/upstream/2.3.3/ChangeLog +0 -1916
  548. data/ext/v8/upstream/2.3.3/LICENSE +0 -55
  549. data/ext/v8/upstream/2.3.3/SConstruct +0 -1154
  550. data/ext/v8/upstream/2.3.3/include/v8-debug.h +0 -381
  551. data/ext/v8/upstream/2.3.3/include/v8-profiler.h +0 -353
  552. data/ext/v8/upstream/2.3.3/include/v8.h +0 -3616
  553. data/ext/v8/upstream/2.3.3/src/SConscript +0 -330
  554. data/ext/v8/upstream/2.3.3/src/accessors.cc +0 -661
  555. data/ext/v8/upstream/2.3.3/src/accessors.h +0 -114
  556. data/ext/v8/upstream/2.3.3/src/allocation.cc +0 -198
  557. data/ext/v8/upstream/2.3.3/src/allocation.h +0 -169
  558. data/ext/v8/upstream/2.3.3/src/api.cc +0 -4795
  559. data/ext/v8/upstream/2.3.3/src/api.h +0 -485
  560. data/ext/v8/upstream/2.3.3/src/apiutils.h +0 -69
  561. data/ext/v8/upstream/2.3.3/src/arguments.h +0 -96
  562. data/ext/v8/upstream/2.3.3/src/arm/assembler-arm-inl.h +0 -305
  563. data/ext/v8/upstream/2.3.3/src/arm/assembler-arm.cc +0 -2580
  564. data/ext/v8/upstream/2.3.3/src/arm/assembler-arm.h +0 -1275
  565. data/ext/v8/upstream/2.3.3/src/arm/builtins-arm.cc +0 -1320
  566. data/ext/v8/upstream/2.3.3/src/arm/codegen-arm-inl.h +0 -48
  567. data/ext/v8/upstream/2.3.3/src/arm/codegen-arm.cc +0 -11398
  568. data/ext/v8/upstream/2.3.3/src/arm/codegen-arm.h +0 -1102
  569. data/ext/v8/upstream/2.3.3/src/arm/constants-arm.cc +0 -154
  570. data/ext/v8/upstream/2.3.3/src/arm/constants-arm.h +0 -388
  571. data/ext/v8/upstream/2.3.3/src/arm/cpu-arm.cc +0 -142
  572. data/ext/v8/upstream/2.3.3/src/arm/debug-arm.cc +0 -309
  573. data/ext/v8/upstream/2.3.3/src/arm/disasm-arm.cc +0 -1459
  574. data/ext/v8/upstream/2.3.3/src/arm/fast-codegen-arm.cc +0 -241
  575. data/ext/v8/upstream/2.3.3/src/arm/frames-arm.cc +0 -123
  576. data/ext/v8/upstream/2.3.3/src/arm/frames-arm.h +0 -162
  577. data/ext/v8/upstream/2.3.3/src/arm/full-codegen-arm.cc +0 -3178
  578. data/ext/v8/upstream/2.3.3/src/arm/ic-arm.cc +0 -2258
  579. data/ext/v8/upstream/2.3.3/src/arm/jump-target-arm.cc +0 -164
  580. data/ext/v8/upstream/2.3.3/src/arm/macro-assembler-arm.cc +0 -1892
  581. data/ext/v8/upstream/2.3.3/src/arm/macro-assembler-arm.h +0 -727
  582. data/ext/v8/upstream/2.3.3/src/arm/regexp-macro-assembler-arm.cc +0 -1261
  583. data/ext/v8/upstream/2.3.3/src/arm/regexp-macro-assembler-arm.h +0 -266
  584. data/ext/v8/upstream/2.3.3/src/arm/simulator-arm.cc +0 -2822
  585. data/ext/v8/upstream/2.3.3/src/arm/simulator-arm.h +0 -361
  586. data/ext/v8/upstream/2.3.3/src/arm/stub-cache-arm.cc +0 -2387
  587. data/ext/v8/upstream/2.3.3/src/arm/virtual-frame-arm.cc +0 -834
  588. data/ext/v8/upstream/2.3.3/src/arm/virtual-frame-arm.h +0 -519
  589. data/ext/v8/upstream/2.3.3/src/array.js +0 -1127
  590. data/ext/v8/upstream/2.3.3/src/assembler.cc +0 -801
  591. data/ext/v8/upstream/2.3.3/src/assembler.h +0 -573
  592. data/ext/v8/upstream/2.3.3/src/ast-inl.h +0 -81
  593. data/ext/v8/upstream/2.3.3/src/ast.cc +0 -1152
  594. data/ext/v8/upstream/2.3.3/src/ast.h +0 -2106
  595. data/ext/v8/upstream/2.3.3/src/bootstrapper.cc +0 -1819
  596. data/ext/v8/upstream/2.3.3/src/builtins.cc +0 -1529
  597. data/ext/v8/upstream/2.3.3/src/builtins.h +0 -263
  598. data/ext/v8/upstream/2.3.3/src/bytecodes-irregexp.h +0 -104
  599. data/ext/v8/upstream/2.3.3/src/cached-powers.h +0 -119
  600. data/ext/v8/upstream/2.3.3/src/char-predicates-inl.h +0 -86
  601. data/ext/v8/upstream/2.3.3/src/checks.cc +0 -100
  602. data/ext/v8/upstream/2.3.3/src/checks.h +0 -310
  603. data/ext/v8/upstream/2.3.3/src/circular-queue.cc +0 -121
  604. data/ext/v8/upstream/2.3.3/src/code-stubs.cc +0 -177
  605. data/ext/v8/upstream/2.3.3/src/code-stubs.h +0 -177
  606. data/ext/v8/upstream/2.3.3/src/codegen-inl.h +0 -60
  607. data/ext/v8/upstream/2.3.3/src/codegen.cc +0 -516
  608. data/ext/v8/upstream/2.3.3/src/codegen.h +0 -897
  609. data/ext/v8/upstream/2.3.3/src/compilation-cache.cc +0 -562
  610. data/ext/v8/upstream/2.3.3/src/compilation-cache.h +0 -102
  611. data/ext/v8/upstream/2.3.3/src/compiler.cc +0 -654
  612. data/ext/v8/upstream/2.3.3/src/compiler.h +0 -299
  613. data/ext/v8/upstream/2.3.3/src/contexts.cc +0 -256
  614. data/ext/v8/upstream/2.3.3/src/contexts.h +0 -342
  615. data/ext/v8/upstream/2.3.3/src/conversions.cc +0 -1119
  616. data/ext/v8/upstream/2.3.3/src/conversions.h +0 -123
  617. data/ext/v8/upstream/2.3.3/src/counters.h +0 -239
  618. data/ext/v8/upstream/2.3.3/src/cpu-profiler-inl.h +0 -100
  619. data/ext/v8/upstream/2.3.3/src/cpu-profiler.cc +0 -538
  620. data/ext/v8/upstream/2.3.3/src/cpu-profiler.h +0 -285
  621. data/ext/v8/upstream/2.3.3/src/d8-debug.cc +0 -356
  622. data/ext/v8/upstream/2.3.3/src/d8-debug.h +0 -155
  623. data/ext/v8/upstream/2.3.3/src/d8.cc +0 -783
  624. data/ext/v8/upstream/2.3.3/src/d8.h +0 -227
  625. data/ext/v8/upstream/2.3.3/src/d8.js +0 -1683
  626. data/ext/v8/upstream/2.3.3/src/data-flow.cc +0 -758
  627. data/ext/v8/upstream/2.3.3/src/data-flow.h +0 -278
  628. data/ext/v8/upstream/2.3.3/src/date.js +0 -1059
  629. data/ext/v8/upstream/2.3.3/src/dateparser-inl.h +0 -123
  630. data/ext/v8/upstream/2.3.3/src/dateparser.h +0 -244
  631. data/ext/v8/upstream/2.3.3/src/debug-agent.cc +0 -427
  632. data/ext/v8/upstream/2.3.3/src/debug-agent.h +0 -129
  633. data/ext/v8/upstream/2.3.3/src/debug-debugger.js +0 -2227
  634. data/ext/v8/upstream/2.3.3/src/debug.cc +0 -3005
  635. data/ext/v8/upstream/2.3.3/src/debug.h +0 -993
  636. data/ext/v8/upstream/2.3.3/src/disassembler.cc +0 -312
  637. data/ext/v8/upstream/2.3.3/src/double.h +0 -169
  638. data/ext/v8/upstream/2.3.3/src/dtoa-config.c +0 -92
  639. data/ext/v8/upstream/2.3.3/src/dtoa.cc +0 -77
  640. data/ext/v8/upstream/2.3.3/src/dtoa.h +0 -81
  641. data/ext/v8/upstream/2.3.3/src/execution.cc +0 -809
  642. data/ext/v8/upstream/2.3.3/src/execution.h +0 -336
  643. data/ext/v8/upstream/2.3.3/src/factory.cc +0 -1003
  644. data/ext/v8/upstream/2.3.3/src/factory.h +0 -410
  645. data/ext/v8/upstream/2.3.3/src/fast-codegen.cc +0 -746
  646. data/ext/v8/upstream/2.3.3/src/fast-codegen.h +0 -161
  647. data/ext/v8/upstream/2.3.3/src/fast-dtoa.cc +0 -505
  648. data/ext/v8/upstream/2.3.3/src/fast-dtoa.h +0 -58
  649. data/ext/v8/upstream/2.3.3/src/flag-definitions.h +0 -455
  650. data/ext/v8/upstream/2.3.3/src/flags.cc +0 -551
  651. data/ext/v8/upstream/2.3.3/src/flags.h +0 -81
  652. data/ext/v8/upstream/2.3.3/src/flow-graph.cc +0 -763
  653. data/ext/v8/upstream/2.3.3/src/flow-graph.h +0 -180
  654. data/ext/v8/upstream/2.3.3/src/frame-element.h +0 -273
  655. data/ext/v8/upstream/2.3.3/src/frames-inl.h +0 -217
  656. data/ext/v8/upstream/2.3.3/src/frames.cc +0 -826
  657. data/ext/v8/upstream/2.3.3/src/frames.h +0 -682
  658. data/ext/v8/upstream/2.3.3/src/full-codegen.cc +0 -1443
  659. data/ext/v8/upstream/2.3.3/src/full-codegen.h +0 -548
  660. data/ext/v8/upstream/2.3.3/src/func-name-inferrer.cc +0 -76
  661. data/ext/v8/upstream/2.3.3/src/func-name-inferrer.h +0 -135
  662. data/ext/v8/upstream/2.3.3/src/global-handles.cc +0 -520
  663. data/ext/v8/upstream/2.3.3/src/global-handles.h +0 -180
  664. data/ext/v8/upstream/2.3.3/src/globals.h +0 -669
  665. data/ext/v8/upstream/2.3.3/src/handles-inl.h +0 -76
  666. data/ext/v8/upstream/2.3.3/src/handles.cc +0 -825
  667. data/ext/v8/upstream/2.3.3/src/handles.h +0 -393
  668. data/ext/v8/upstream/2.3.3/src/hashmap.cc +0 -226
  669. data/ext/v8/upstream/2.3.3/src/hashmap.h +0 -120
  670. data/ext/v8/upstream/2.3.3/src/heap-inl.h +0 -493
  671. data/ext/v8/upstream/2.3.3/src/heap-profiler.cc +0 -779
  672. data/ext/v8/upstream/2.3.3/src/heap-profiler.h +0 -323
  673. data/ext/v8/upstream/2.3.3/src/heap.cc +0 -4994
  674. data/ext/v8/upstream/2.3.3/src/heap.h +0 -1984
  675. data/ext/v8/upstream/2.3.3/src/ia32/assembler-ia32-inl.h +0 -360
  676. data/ext/v8/upstream/2.3.3/src/ia32/assembler-ia32.cc +0 -2600
  677. data/ext/v8/upstream/2.3.3/src/ia32/assembler-ia32.h +0 -969
  678. data/ext/v8/upstream/2.3.3/src/ia32/builtins-ia32.cc +0 -1261
  679. data/ext/v8/upstream/2.3.3/src/ia32/codegen-ia32.cc +0 -13968
  680. data/ext/v8/upstream/2.3.3/src/ia32/codegen-ia32.h +0 -1097
  681. data/ext/v8/upstream/2.3.3/src/ia32/cpu-ia32.cc +0 -83
  682. data/ext/v8/upstream/2.3.3/src/ia32/debug-ia32.cc +0 -309
  683. data/ext/v8/upstream/2.3.3/src/ia32/disasm-ia32.cc +0 -1471
  684. data/ext/v8/upstream/2.3.3/src/ia32/fast-codegen-ia32.cc +0 -954
  685. data/ext/v8/upstream/2.3.3/src/ia32/fast-codegen-ia32.h +0 -155
  686. data/ext/v8/upstream/2.3.3/src/ia32/frames-ia32.cc +0 -115
  687. data/ext/v8/upstream/2.3.3/src/ia32/frames-ia32.h +0 -135
  688. data/ext/v8/upstream/2.3.3/src/ia32/full-codegen-ia32.cc +0 -3281
  689. data/ext/v8/upstream/2.3.3/src/ia32/ic-ia32.cc +0 -1966
  690. data/ext/v8/upstream/2.3.3/src/ia32/macro-assembler-ia32.cc +0 -1610
  691. data/ext/v8/upstream/2.3.3/src/ia32/macro-assembler-ia32.h +0 -610
  692. data/ext/v8/upstream/2.3.3/src/ia32/regexp-macro-assembler-ia32.cc +0 -1247
  693. data/ext/v8/upstream/2.3.3/src/ia32/regexp-macro-assembler-ia32.h +0 -214
  694. data/ext/v8/upstream/2.3.3/src/ia32/simulator-ia32.h +0 -62
  695. data/ext/v8/upstream/2.3.3/src/ia32/stub-cache-ia32.cc +0 -2750
  696. data/ext/v8/upstream/2.3.3/src/ia32/virtual-frame-ia32.cc +0 -1334
  697. data/ext/v8/upstream/2.3.3/src/ia32/virtual-frame-ia32.h +0 -627
  698. data/ext/v8/upstream/2.3.3/src/ic-inl.h +0 -120
  699. data/ext/v8/upstream/2.3.3/src/ic.cc +0 -1827
  700. data/ext/v8/upstream/2.3.3/src/ic.h +0 -515
  701. data/ext/v8/upstream/2.3.3/src/interpreter-irregexp.cc +0 -646
  702. data/ext/v8/upstream/2.3.3/src/json.js +0 -268
  703. data/ext/v8/upstream/2.3.3/src/jsregexp.cc +0 -5283
  704. data/ext/v8/upstream/2.3.3/src/jsregexp.h +0 -1463
  705. data/ext/v8/upstream/2.3.3/src/jump-target-heavy.cc +0 -429
  706. data/ext/v8/upstream/2.3.3/src/jump-target-heavy.h +0 -244
  707. data/ext/v8/upstream/2.3.3/src/jump-target-light.cc +0 -110
  708. data/ext/v8/upstream/2.3.3/src/jump-target-light.h +0 -192
  709. data/ext/v8/upstream/2.3.3/src/list-inl.h +0 -166
  710. data/ext/v8/upstream/2.3.3/src/list.h +0 -159
  711. data/ext/v8/upstream/2.3.3/src/liveedit-debugger.js +0 -1057
  712. data/ext/v8/upstream/2.3.3/src/liveedit.cc +0 -1480
  713. data/ext/v8/upstream/2.3.3/src/liveedit.h +0 -170
  714. data/ext/v8/upstream/2.3.3/src/log-utils.cc +0 -497
  715. data/ext/v8/upstream/2.3.3/src/log-utils.h +0 -289
  716. data/ext/v8/upstream/2.3.3/src/log.cc +0 -1561
  717. data/ext/v8/upstream/2.3.3/src/log.h +0 -384
  718. data/ext/v8/upstream/2.3.3/src/macro-assembler.h +0 -86
  719. data/ext/v8/upstream/2.3.3/src/macros.py +0 -177
  720. data/ext/v8/upstream/2.3.3/src/mark-compact.cc +0 -2330
  721. data/ext/v8/upstream/2.3.3/src/mark-compact.h +0 -451
  722. data/ext/v8/upstream/2.3.3/src/math.js +0 -264
  723. data/ext/v8/upstream/2.3.3/src/memory.h +0 -74
  724. data/ext/v8/upstream/2.3.3/src/messages.cc +0 -183
  725. data/ext/v8/upstream/2.3.3/src/messages.h +0 -113
  726. data/ext/v8/upstream/2.3.3/src/messages.js +0 -982
  727. data/ext/v8/upstream/2.3.3/src/mips/assembler-mips.h +0 -668
  728. data/ext/v8/upstream/2.3.3/src/mips/builtins-mips.cc +0 -205
  729. data/ext/v8/upstream/2.3.3/src/mips/codegen-mips.h +0 -434
  730. data/ext/v8/upstream/2.3.3/src/mips/debug-mips.cc +0 -131
  731. data/ext/v8/upstream/2.3.3/src/mips/frames-mips.cc +0 -102
  732. data/ext/v8/upstream/2.3.3/src/mips/ic-mips.cc +0 -220
  733. data/ext/v8/upstream/2.3.3/src/mips/simulator-mips.cc +0 -1651
  734. data/ext/v8/upstream/2.3.3/src/mips/simulator-mips.h +0 -311
  735. data/ext/v8/upstream/2.3.3/src/mips/stub-cache-mips.cc +0 -403
  736. data/ext/v8/upstream/2.3.3/src/mirror-debugger.js +0 -2380
  737. data/ext/v8/upstream/2.3.3/src/objects-debug.cc +0 -1366
  738. data/ext/v8/upstream/2.3.3/src/objects-inl.h +0 -3333
  739. data/ext/v8/upstream/2.3.3/src/objects.cc +0 -8820
  740. data/ext/v8/upstream/2.3.3/src/objects.h +0 -5373
  741. data/ext/v8/upstream/2.3.3/src/oprofile-agent.cc +0 -108
  742. data/ext/v8/upstream/2.3.3/src/oprofile-agent.h +0 -77
  743. data/ext/v8/upstream/2.3.3/src/parser.cc +0 -5207
  744. data/ext/v8/upstream/2.3.3/src/parser.h +0 -197
  745. data/ext/v8/upstream/2.3.3/src/platform-freebsd.cc +0 -667
  746. data/ext/v8/upstream/2.3.3/src/platform-linux.cc +0 -862
  747. data/ext/v8/upstream/2.3.3/src/platform-macos.cc +0 -665
  748. data/ext/v8/upstream/2.3.3/src/platform-nullos.cc +0 -454
  749. data/ext/v8/upstream/2.3.3/src/platform-openbsd.cc +0 -622
  750. data/ext/v8/upstream/2.3.3/src/platform-posix.cc +0 -362
  751. data/ext/v8/upstream/2.3.3/src/platform-solaris.cc +0 -653
  752. data/ext/v8/upstream/2.3.3/src/platform-win32.cc +0 -1911
  753. data/ext/v8/upstream/2.3.3/src/platform.h +0 -577
  754. data/ext/v8/upstream/2.3.3/src/powers-ten.h +0 -2461
  755. data/ext/v8/upstream/2.3.3/src/prettyprinter.cc +0 -1531
  756. data/ext/v8/upstream/2.3.3/src/prettyprinter.h +0 -221
  757. data/ext/v8/upstream/2.3.3/src/profile-generator-inl.h +0 -148
  758. data/ext/v8/upstream/2.3.3/src/profile-generator.cc +0 -1830
  759. data/ext/v8/upstream/2.3.3/src/profile-generator.h +0 -853
  760. data/ext/v8/upstream/2.3.3/src/property.cc +0 -96
  761. data/ext/v8/upstream/2.3.3/src/property.h +0 -315
  762. data/ext/v8/upstream/2.3.3/src/regexp-macro-assembler-irregexp.cc +0 -464
  763. data/ext/v8/upstream/2.3.3/src/regexp-macro-assembler-irregexp.h +0 -141
  764. data/ext/v8/upstream/2.3.3/src/regexp-macro-assembler-tracer.cc +0 -356
  765. data/ext/v8/upstream/2.3.3/src/regexp-macro-assembler-tracer.h +0 -103
  766. data/ext/v8/upstream/2.3.3/src/regexp-macro-assembler.cc +0 -261
  767. data/ext/v8/upstream/2.3.3/src/regexp-macro-assembler.h +0 -228
  768. data/ext/v8/upstream/2.3.3/src/regexp.js +0 -549
  769. data/ext/v8/upstream/2.3.3/src/rewriter.cc +0 -1038
  770. data/ext/v8/upstream/2.3.3/src/rewriter.h +0 -54
  771. data/ext/v8/upstream/2.3.3/src/runtime.cc +0 -10599
  772. data/ext/v8/upstream/2.3.3/src/runtime.h +0 -459
  773. data/ext/v8/upstream/2.3.3/src/runtime.js +0 -629
  774. data/ext/v8/upstream/2.3.3/src/scanner.cc +0 -1346
  775. data/ext/v8/upstream/2.3.3/src/scanner.h +0 -503
  776. data/ext/v8/upstream/2.3.3/src/scopeinfo.cc +0 -637
  777. data/ext/v8/upstream/2.3.3/src/scopeinfo.h +0 -233
  778. data/ext/v8/upstream/2.3.3/src/scopes.cc +0 -962
  779. data/ext/v8/upstream/2.3.3/src/scopes.h +0 -400
  780. data/ext/v8/upstream/2.3.3/src/serialize.cc +0 -1461
  781. data/ext/v8/upstream/2.3.3/src/serialize.h +0 -581
  782. data/ext/v8/upstream/2.3.3/src/spaces-inl.h +0 -483
  783. data/ext/v8/upstream/2.3.3/src/spaces.cc +0 -2901
  784. data/ext/v8/upstream/2.3.3/src/spaces.h +0 -2197
  785. data/ext/v8/upstream/2.3.3/src/string-stream.cc +0 -584
  786. data/ext/v8/upstream/2.3.3/src/string-stream.h +0 -189
  787. data/ext/v8/upstream/2.3.3/src/string.js +0 -1006
  788. data/ext/v8/upstream/2.3.3/src/stub-cache.cc +0 -1379
  789. data/ext/v8/upstream/2.3.3/src/stub-cache.h +0 -756
  790. data/ext/v8/upstream/2.3.3/src/third_party/dtoa/COPYING +0 -15
  791. data/ext/v8/upstream/2.3.3/src/third_party/dtoa/dtoa.c +0 -3334
  792. data/ext/v8/upstream/2.3.3/src/token.cc +0 -56
  793. data/ext/v8/upstream/2.3.3/src/token.h +0 -270
  794. data/ext/v8/upstream/2.3.3/src/top.cc +0 -1067
  795. data/ext/v8/upstream/2.3.3/src/top.h +0 -463
  796. data/ext/v8/upstream/2.3.3/src/type-info.cc +0 -53
  797. data/ext/v8/upstream/2.3.3/src/type-info.h +0 -244
  798. data/ext/v8/upstream/2.3.3/src/unicode-inl.h +0 -238
  799. data/ext/v8/upstream/2.3.3/src/unicode.cc +0 -749
  800. data/ext/v8/upstream/2.3.3/src/unicode.h +0 -279
  801. data/ext/v8/upstream/2.3.3/src/uri.js +0 -415
  802. data/ext/v8/upstream/2.3.3/src/utils.cc +0 -285
  803. data/ext/v8/upstream/2.3.3/src/utils.h +0 -745
  804. data/ext/v8/upstream/2.3.3/src/v8-counters.h +0 -250
  805. data/ext/v8/upstream/2.3.3/src/v8.cc +0 -228
  806. data/ext/v8/upstream/2.3.3/src/v8.h +0 -121
  807. data/ext/v8/upstream/2.3.3/src/v8natives.js +0 -1188
  808. data/ext/v8/upstream/2.3.3/src/v8threads.cc +0 -461
  809. data/ext/v8/upstream/2.3.3/src/v8threads.h +0 -159
  810. data/ext/v8/upstream/2.3.3/src/variables.cc +0 -119
  811. data/ext/v8/upstream/2.3.3/src/variables.h +0 -205
  812. data/ext/v8/upstream/2.3.3/src/version.cc +0 -88
  813. data/ext/v8/upstream/2.3.3/src/virtual-frame-heavy-inl.h +0 -192
  814. data/ext/v8/upstream/2.3.3/src/virtual-frame.h +0 -46
  815. data/ext/v8/upstream/2.3.3/src/vm-state-inl.h +0 -137
  816. data/ext/v8/upstream/2.3.3/src/vm-state.cc +0 -39
  817. data/ext/v8/upstream/2.3.3/src/vm-state.h +0 -77
  818. data/ext/v8/upstream/2.3.3/src/x64/assembler-x64-inl.h +0 -400
  819. data/ext/v8/upstream/2.3.3/src/x64/assembler-x64.cc +0 -2963
  820. data/ext/v8/upstream/2.3.3/src/x64/assembler-x64.h +0 -1438
  821. data/ext/v8/upstream/2.3.3/src/x64/builtins-x64.cc +0 -1296
  822. data/ext/v8/upstream/2.3.3/src/x64/codegen-x64-inl.h +0 -46
  823. data/ext/v8/upstream/2.3.3/src/x64/codegen-x64.cc +0 -12491
  824. data/ext/v8/upstream/2.3.3/src/x64/codegen-x64.h +0 -1090
  825. data/ext/v8/upstream/2.3.3/src/x64/cpu-x64.cc +0 -83
  826. data/ext/v8/upstream/2.3.3/src/x64/debug-x64.cc +0 -267
  827. data/ext/v8/upstream/2.3.3/src/x64/disasm-x64.cc +0 -1696
  828. data/ext/v8/upstream/2.3.3/src/x64/fast-codegen-x64.cc +0 -250
  829. data/ext/v8/upstream/2.3.3/src/x64/frames-x64.cc +0 -113
  830. data/ext/v8/upstream/2.3.3/src/x64/frames-x64.h +0 -125
  831. data/ext/v8/upstream/2.3.3/src/x64/full-codegen-x64.cc +0 -3270
  832. data/ext/v8/upstream/2.3.3/src/x64/ic-x64.cc +0 -1907
  833. data/ext/v8/upstream/2.3.3/src/x64/jump-target-x64.cc +0 -437
  834. data/ext/v8/upstream/2.3.3/src/x64/macro-assembler-x64.cc +0 -2793
  835. data/ext/v8/upstream/2.3.3/src/x64/macro-assembler-x64.h +0 -916
  836. data/ext/v8/upstream/2.3.3/src/x64/regexp-macro-assembler-x64.cc +0 -1374
  837. data/ext/v8/upstream/2.3.3/src/x64/regexp-macro-assembler-x64.h +0 -277
  838. data/ext/v8/upstream/2.3.3/src/x64/simulator-x64.h +0 -63
  839. data/ext/v8/upstream/2.3.3/src/x64/stub-cache-x64.cc +0 -2560
  840. data/ext/v8/upstream/2.3.3/src/x64/virtual-frame-x64.cc +0 -1264
  841. data/ext/v8/upstream/2.3.3/src/x64/virtual-frame-x64.h +0 -590
  842. data/ext/v8/upstream/2.3.3/src/zone-inl.h +0 -82
  843. data/ext/v8/upstream/2.3.3/src/zone.cc +0 -194
  844. data/ext/v8/upstream/2.3.3/src/zone.h +0 -221
  845. data/ext/v8/upstream/2.3.3/tools/codemap.js +0 -270
  846. data/ext/v8/upstream/2.3.3/tools/csvparser.js +0 -83
  847. data/ext/v8/upstream/2.3.3/tools/gc-nvp-trace-processor.py +0 -317
  848. data/ext/v8/upstream/2.3.3/tools/gyp/v8.gyp +0 -749
  849. data/ext/v8/upstream/2.3.3/tools/linux-tick-processor.py +0 -78
  850. data/ext/v8/upstream/2.3.3/tools/logreader.js +0 -338
  851. data/ext/v8/upstream/2.3.3/tools/oprofile/annotate +0 -7
  852. data/ext/v8/upstream/2.3.3/tools/oprofile/common +0 -19
  853. data/ext/v8/upstream/2.3.3/tools/oprofile/dump +0 -7
  854. data/ext/v8/upstream/2.3.3/tools/oprofile/report +0 -7
  855. data/ext/v8/upstream/2.3.3/tools/oprofile/reset +0 -7
  856. data/ext/v8/upstream/2.3.3/tools/oprofile/run +0 -14
  857. data/ext/v8/upstream/2.3.3/tools/oprofile/shutdown +0 -7
  858. data/ext/v8/upstream/2.3.3/tools/oprofile/start +0 -7
  859. data/ext/v8/upstream/2.3.3/tools/presubmit.py +0 -299
  860. data/ext/v8/upstream/2.3.3/tools/profile.js +0 -691
  861. data/ext/v8/upstream/2.3.3/tools/profile_view.js +0 -224
  862. data/ext/v8/upstream/2.3.3/tools/splaytree.js +0 -322
  863. data/ext/v8/upstream/2.3.3/tools/splaytree.py +0 -226
  864. data/ext/v8/upstream/2.3.3/tools/tickprocessor.js +0 -862
  865. data/ext/v8/upstream/2.3.3/tools/tickprocessor.py +0 -571
  866. data/ext/v8/upstream/2.3.3/tools/utils.py +0 -88
  867. data/ext/v8/upstream/2.3.3/tools/visual_studio/README.txt +0 -71
  868. data/ext/v8/upstream/2.3.3/tools/visual_studio/common.vsprops +0 -34
  869. data/ext/v8/upstream/2.3.3/tools/visual_studio/debug.vsprops +0 -17
  870. data/ext/v8/upstream/2.3.3/tools/visual_studio/v8_base.vcproj +0 -1143
  871. data/ext/v8/upstream/2.3.3/tools/visual_studio/v8_base_arm.vcproj +0 -1115
  872. data/ext/v8/upstream/2.3.3/tools/visual_studio/v8_base_x64.vcproj +0 -1096
  873. data/ext/v8/upstream/2.3.3/tools/visual_studio/v8_shell_sample.vcproj +0 -145
  874. data/ext/v8/upstream/2.3.3/tools/visual_studio/v8_shell_sample_arm.vcproj +0 -145
  875. data/ext/v8/upstream/2.3.3/tools/visual_studio/v8_shell_sample_x64.vcproj +0 -161
  876. data/ext/v8/upstream/2.3.3/tools/windows-tick-processor.bat +0 -29
  877. data/ext/v8/upstream/2.3.3/tools/windows-tick-processor.py +0 -137
@@ -0,0 +1,2362 @@
1
+ // Copyright 2006-2010 the V8 project authors. All rights reserved.
2
+ // Redistribution and use in source and binary forms, with or without
3
+ // modification, are permitted provided that the following conditions are
4
+ // met:
5
+ //
6
+ // * Redistributions of source code must retain the above copyright
7
+ // notice, this list of conditions and the following disclaimer.
8
+ // * Redistributions in binary form must reproduce the above
9
+ // copyright notice, this list of conditions and the following
10
+ // disclaimer in the documentation and/or other materials provided
11
+ // with the distribution.
12
+ // * Neither the name of Google Inc. nor the names of its
13
+ // contributors may be used to endorse or promote products derived
14
+ // from this software without specific prior written permission.
15
+ //
16
+ // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
17
+ // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
18
+ // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
19
+ // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
20
+ // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
21
+ // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
22
+ // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
23
+ // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
24
+ // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
25
+ // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
26
+ // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
27
+
28
+ #ifndef V8_SPACES_H_
29
+ #define V8_SPACES_H_
30
+
31
+ #include "list-inl.h"
32
+ #include "log.h"
33
+
34
+ namespace v8 {
35
+ namespace internal {
36
+
37
+ // -----------------------------------------------------------------------------
38
+ // Heap structures:
39
+ //
40
+ // A JS heap consists of a young generation, an old generation, and a large
41
+ // object space. The young generation is divided into two semispaces. A
42
+ // scavenger implements Cheney's copying algorithm. The old generation is
43
+ // separated into a map space and an old object space. The map space contains
44
+ // all (and only) map objects, the rest of old objects go into the old space.
45
+ // The old generation is collected by a mark-sweep-compact collector.
46
+ //
47
+ // The semispaces of the young generation are contiguous. The old and map
48
+ // spaces consists of a list of pages. A page has a page header and an object
49
+ // area. A page size is deliberately chosen as 8K bytes.
50
+ // The first word of a page is an opaque page header that has the
51
+ // address of the next page and its ownership information. The second word may
52
+ // have the allocation top address of this page. Heap objects are aligned to the
53
+ // pointer size.
54
+ //
55
+ // There is a separate large object space for objects larger than
56
+ // Page::kMaxHeapObjectSize, so that they do not have to move during
57
+ // collection. The large object space is paged. Pages in large object space
58
+ // may be larger than 8K.
59
+ //
60
+ // A card marking write barrier is used to keep track of intergenerational
61
+ // references. Old space pages are divided into regions of Page::kRegionSize
62
+ // size. Each region has a corresponding dirty bit in the page header which is
63
+ // set if the region might contain pointers to new space. For details about
64
+ // dirty bits encoding see comments in the Page::GetRegionNumberForAddress()
65
+ // method body.
66
+ //
67
+ // During scavenges and mark-sweep collections we iterate intergenerational
68
+ // pointers without decoding heap object maps so if the page belongs to old
69
+ // pointer space or large object space it is essential to guarantee that
70
+ // the page does not contain any garbage pointers to new space: every pointer
71
+ // aligned word which satisfies the Heap::InNewSpace() predicate must be a
72
+ // pointer to a live heap object in new space. Thus objects in old pointer
73
+ // and large object spaces should have a special layout (e.g. no bare integer
74
+ // fields). This requirement does not apply to map space which is iterated in
75
+ // a special fashion. However we still require pointer fields of dead maps to
76
+ // be cleaned.
77
+ //
78
+ // To enable lazy cleaning of old space pages we use a notion of allocation
79
+ // watermark. Every pointer under watermark is considered to be well formed.
80
+ // Page allocation watermark is not necessarily equal to page allocation top but
81
+ // all alive objects on page should reside under allocation watermark.
82
+ // During scavenge allocation watermark might be bumped and invalid pointers
83
+ // might appear below it. To avoid following them we store a valid watermark
84
+ // into special field in the page header and set a page WATERMARK_INVALIDATED
85
+ // flag. For details see comments in the Page::SetAllocationWatermark() method
86
+ // body.
87
+ //
88
+
89
+ // Some assertion macros used in the debugging mode.
90
+
91
+ #define ASSERT_PAGE_ALIGNED(address) \
92
+ ASSERT((OffsetFrom(address) & Page::kPageAlignmentMask) == 0)
93
+
94
+ #define ASSERT_OBJECT_ALIGNED(address) \
95
+ ASSERT((OffsetFrom(address) & kObjectAlignmentMask) == 0)
96
+
97
+ #define ASSERT_MAP_ALIGNED(address) \
98
+ ASSERT((OffsetFrom(address) & kMapAlignmentMask) == 0)
99
+
100
+ #define ASSERT_OBJECT_SIZE(size) \
101
+ ASSERT((0 < size) && (size <= Page::kMaxHeapObjectSize))
102
+
103
+ #define ASSERT_PAGE_OFFSET(offset) \
104
+ ASSERT((Page::kObjectStartOffset <= offset) \
105
+ && (offset <= Page::kPageSize))
106
+
107
+ #define ASSERT_MAP_PAGE_INDEX(index) \
108
+ ASSERT((0 <= index) && (index <= MapSpace::kMaxMapPageIndex))
109
+
110
+
111
+ class PagedSpace;
112
+ class MemoryAllocator;
113
+ class AllocationInfo;
114
+
115
+ // -----------------------------------------------------------------------------
116
+ // A page normally has 8K bytes. Large object pages may be larger. A page
117
+ // address is always aligned to the 8K page size.
118
+ //
119
+ // Each page starts with a header of Page::kPageHeaderSize size which contains
120
+ // bookkeeping data.
121
+ //
122
+ // The mark-compact collector transforms a map pointer into a page index and a
123
+ // page offset. The exact encoding is described in the comments for
124
+ // class MapWord in objects.h.
125
+ //
126
+ // The only way to get a page pointer is by calling factory methods:
127
+ // Page* p = Page::FromAddress(addr); or
128
+ // Page* p = Page::FromAllocationTop(top);
129
+ class Page {
130
+ public:
131
+ // Returns the page containing a given address. The address ranges
132
+ // from [page_addr .. page_addr + kPageSize[
133
+ //
134
+ // Note that this function only works for addresses in normal paged
135
+ // spaces and addresses in the first 8K of large object pages (i.e.,
136
+ // the start of large objects but not necessarily derived pointers
137
+ // within them).
138
+ INLINE(static Page* FromAddress(Address a)) {
139
+ return reinterpret_cast<Page*>(OffsetFrom(a) & ~kPageAlignmentMask);
140
+ }
141
+
142
+ // Returns the page containing an allocation top. Because an allocation
143
+ // top address can be the upper bound of the page, we need to subtract
144
+ // it with kPointerSize first. The address ranges from
145
+ // [page_addr + kObjectStartOffset .. page_addr + kPageSize].
146
+ INLINE(static Page* FromAllocationTop(Address top)) {
147
+ Page* p = FromAddress(top - kPointerSize);
148
+ ASSERT_PAGE_OFFSET(p->Offset(top));
149
+ return p;
150
+ }
151
+
152
+ // Returns the start address of this page.
153
+ Address address() { return reinterpret_cast<Address>(this); }
154
+
155
+ // Checks whether this is a valid page address.
156
+ bool is_valid() { return address() != NULL; }
157
+
158
+ // Returns the next page of this page.
159
+ inline Page* next_page();
160
+
161
+ // Return the end of allocation in this page. Undefined for unused pages.
162
+ inline Address AllocationTop();
163
+
164
+ // Return the allocation watermark for the page.
165
+ // For old space pages it is guaranteed that the area under the watermark
166
+ // does not contain any garbage pointers to new space.
167
+ inline Address AllocationWatermark();
168
+
169
+ // Return the allocation watermark offset from the beginning of the page.
170
+ inline uint32_t AllocationWatermarkOffset();
171
+
172
+ inline void SetAllocationWatermark(Address allocation_watermark);
173
+
174
+ inline void SetCachedAllocationWatermark(Address allocation_watermark);
175
+ inline Address CachedAllocationWatermark();
176
+
177
+ // Returns the start address of the object area in this page.
178
+ Address ObjectAreaStart() { return address() + kObjectStartOffset; }
179
+
180
+ // Returns the end address (exclusive) of the object area in this page.
181
+ Address ObjectAreaEnd() { return address() + Page::kPageSize; }
182
+
183
+ // Checks whether an address is page aligned.
184
+ static bool IsAlignedToPageSize(Address a) {
185
+ return 0 == (OffsetFrom(a) & kPageAlignmentMask);
186
+ }
187
+
188
+ // True if this page was in use before current compaction started.
189
+ // Result is valid only for pages owned by paged spaces and
190
+ // only after PagedSpace::PrepareForMarkCompact was called.
191
+ inline bool WasInUseBeforeMC();
192
+
193
+ inline void SetWasInUseBeforeMC(bool was_in_use);
194
+
195
+ // True if this page is a large object page.
196
+ inline bool IsLargeObjectPage();
197
+
198
+ inline void SetIsLargeObjectPage(bool is_large_object_page);
199
+
200
+ inline bool IsPageExecutable();
201
+
202
+ inline void SetIsPageExecutable(bool is_page_executable);
203
+
204
+ // Returns the offset of a given address to this page.
205
+ INLINE(int Offset(Address a)) {
206
+ int offset = static_cast<int>(a - address());
207
+ ASSERT_PAGE_OFFSET(offset);
208
+ return offset;
209
+ }
210
+
211
+ // Returns the address for a given offset to the this page.
212
+ Address OffsetToAddress(int offset) {
213
+ ASSERT_PAGE_OFFSET(offset);
214
+ return address() + offset;
215
+ }
216
+
217
+ // ---------------------------------------------------------------------
218
+ // Card marking support
219
+
220
+ static const uint32_t kAllRegionsCleanMarks = 0x0;
221
+ static const uint32_t kAllRegionsDirtyMarks = 0xFFFFFFFF;
222
+
223
+ inline uint32_t GetRegionMarks();
224
+ inline void SetRegionMarks(uint32_t dirty);
225
+
226
+ inline uint32_t GetRegionMaskForAddress(Address addr);
227
+ inline uint32_t GetRegionMaskForSpan(Address start, int length_in_bytes);
228
+ inline int GetRegionNumberForAddress(Address addr);
229
+
230
+ inline void MarkRegionDirty(Address addr);
231
+ inline bool IsRegionDirty(Address addr);
232
+
233
+ inline void ClearRegionMarks(Address start,
234
+ Address end,
235
+ bool reaches_limit);
236
+
237
+ // Page size in bytes. This must be a multiple of the OS page size.
238
+ static const int kPageSize = 1 << kPageSizeBits;
239
+
240
+ // Page size mask.
241
+ static const intptr_t kPageAlignmentMask = (1 << kPageSizeBits) - 1;
242
+
243
+ static const int kPageHeaderSize = kPointerSize + kPointerSize + kIntSize +
244
+ kIntSize + kPointerSize;
245
+
246
+ // The start offset of the object area in a page. Aligned to both maps and
247
+ // code alignment to be suitable for both.
248
+ static const int kObjectStartOffset =
249
+ CODE_POINTER_ALIGN(MAP_POINTER_ALIGN(kPageHeaderSize));
250
+
251
+ // Object area size in bytes.
252
+ static const int kObjectAreaSize = kPageSize - kObjectStartOffset;
253
+
254
+ // Maximum object size that fits in a page.
255
+ static const int kMaxHeapObjectSize = kObjectAreaSize;
256
+
257
+ static const int kDirtyFlagOffset = 2 * kPointerSize;
258
+ static const int kRegionSizeLog2 = 8;
259
+ static const int kRegionSize = 1 << kRegionSizeLog2;
260
+ static const intptr_t kRegionAlignmentMask = (kRegionSize - 1);
261
+
262
+ STATIC_CHECK(kRegionSize == kPageSize / kBitsPerInt);
263
+
264
+ enum PageFlag {
265
+ IS_NORMAL_PAGE = 0,
266
+ WAS_IN_USE_BEFORE_MC,
267
+
268
+ // Page allocation watermark was bumped by preallocation during scavenge.
269
+ // Correct watermark can be retrieved by CachedAllocationWatermark() method
270
+ WATERMARK_INVALIDATED,
271
+ IS_EXECUTABLE,
272
+ NUM_PAGE_FLAGS // Must be last
273
+ };
274
+ static const int kPageFlagMask = (1 << NUM_PAGE_FLAGS) - 1;
275
+
276
+ // To avoid an additional WATERMARK_INVALIDATED flag clearing pass during
277
+ // scavenge we just invalidate the watermark on each old space page after
278
+ // processing it. And then we flip the meaning of the WATERMARK_INVALIDATED
279
+ // flag at the beginning of the next scavenge and each page becomes marked as
280
+ // having a valid watermark.
281
+ //
282
+ // The following invariant must hold for pages in old pointer and map spaces:
283
+ // If page is in use then page is marked as having invalid watermark at
284
+ // the beginning and at the end of any GC.
285
+ //
286
+ // This invariant guarantees that after flipping flag meaning at the
287
+ // beginning of scavenge all pages in use will be marked as having valid
288
+ // watermark.
289
+ static inline void FlipMeaningOfInvalidatedWatermarkFlag();
290
+
291
+ // Returns true if the page allocation watermark was not altered during
292
+ // scavenge.
293
+ inline bool IsWatermarkValid();
294
+
295
+ inline void InvalidateWatermark(bool value);
296
+
297
+ inline bool GetPageFlag(PageFlag flag);
298
+ inline void SetPageFlag(PageFlag flag, bool value);
299
+ inline void ClearPageFlags();
300
+
301
+ inline void ClearGCFields();
302
+
303
+ static const int kAllocationWatermarkOffsetShift = WATERMARK_INVALIDATED + 1;
304
+ static const int kAllocationWatermarkOffsetBits = kPageSizeBits + 1;
305
+ static const uint32_t kAllocationWatermarkOffsetMask =
306
+ ((1 << kAllocationWatermarkOffsetBits) - 1) <<
307
+ kAllocationWatermarkOffsetShift;
308
+
309
+ static const uint32_t kFlagsMask =
310
+ ((1 << kAllocationWatermarkOffsetShift) - 1);
311
+
312
+ STATIC_CHECK(kBitsPerInt - kAllocationWatermarkOffsetShift >=
313
+ kAllocationWatermarkOffsetBits);
314
+
315
+ // This field contains the meaning of the WATERMARK_INVALIDATED flag.
316
+ // Instead of clearing this flag from all pages we just flip
317
+ // its meaning at the beginning of a scavenge.
318
+ static intptr_t watermark_invalidated_mark_;
319
+
320
+ //---------------------------------------------------------------------------
321
+ // Page header description.
322
+ //
323
+ // If a page is not in the large object space, the first word,
324
+ // opaque_header, encodes the next page address (aligned to kPageSize 8K)
325
+ // and the chunk number (0 ~ 8K-1). Only MemoryAllocator should use
326
+ // opaque_header. The value range of the opaque_header is [0..kPageSize[,
327
+ // or [next_page_start, next_page_end[. It cannot point to a valid address
328
+ // in the current page. If a page is in the large object space, the first
329
+ // word *may* (if the page start and large object chunk start are the
330
+ // same) contain the address of the next large object chunk.
331
+ intptr_t opaque_header;
332
+
333
+ // If the page is not in the large object space, the low-order bit of the
334
+ // second word is set. If the page is in the large object space, the
335
+ // second word *may* (if the page start and large object chunk start are
336
+ // the same) contain the large object chunk size. In either case, the
337
+ // low-order bit for large object pages will be cleared.
338
+ // For normal pages this word is used to store page flags and
339
+ // offset of allocation top.
340
+ intptr_t flags_;
341
+
342
+ // This field contains dirty marks for regions covering the page. Only dirty
343
+ // regions might contain intergenerational references.
344
+ // Only 32 dirty marks are supported so for large object pages several regions
345
+ // might be mapped to a single dirty mark.
346
+ uint32_t dirty_regions_;
347
+
348
+ // The index of the page in its owner space.
349
+ int mc_page_index;
350
+
351
+ // During mark-compact collections this field contains the forwarding address
352
+ // of the first live object in this page.
353
+ // During scavenge collection this field is used to store allocation watermark
354
+ // if it is altered during scavenge.
355
+ Address mc_first_forwarded;
356
+ };
357
+
358
+
359
+ // ----------------------------------------------------------------------------
360
+ // Space is the abstract superclass for all allocation spaces.
361
+ class Space : public Malloced {
362
+ public:
363
+ Space(AllocationSpace id, Executability executable)
364
+ : id_(id), executable_(executable) {}
365
+
366
+ virtual ~Space() {}
367
+
368
+ // Does the space need executable memory?
369
+ Executability executable() { return executable_; }
370
+
371
+ // Identity used in error reporting.
372
+ AllocationSpace identity() { return id_; }
373
+
374
+ // Returns allocated size.
375
+ virtual intptr_t Size() = 0;
376
+
377
+ // Returns size of objects. Can differ from the allocated size
378
+ // (e.g. see LargeObjectSpace).
379
+ virtual intptr_t SizeOfObjects() { return Size(); }
380
+
381
+ #ifdef ENABLE_HEAP_PROTECTION
382
+ // Protect/unprotect the space by marking it read-only/writable.
383
+ virtual void Protect() = 0;
384
+ virtual void Unprotect() = 0;
385
+ #endif
386
+
387
+ #ifdef DEBUG
388
+ virtual void Print() = 0;
389
+ #endif
390
+
391
+ // After calling this we can allocate a certain number of bytes using only
392
+ // linear allocation (with a LinearAllocationScope and an AlwaysAllocateScope)
393
+ // without using freelists or causing a GC. This is used by partial
394
+ // snapshots. It returns true of space was reserved or false if a GC is
395
+ // needed. For paged spaces the space requested must include the space wasted
396
+ // at the end of each when allocating linearly.
397
+ virtual bool ReserveSpace(int bytes) = 0;
398
+
399
+ private:
400
+ AllocationSpace id_;
401
+ Executability executable_;
402
+ };
403
+
404
+
405
+ // ----------------------------------------------------------------------------
406
+ // All heap objects containing executable code (code objects) must be allocated
407
+ // from a 2 GB range of memory, so that they can call each other using 32-bit
408
+ // displacements. This happens automatically on 32-bit platforms, where 32-bit
409
+ // displacements cover the entire 4GB virtual address space. On 64-bit
410
+ // platforms, we support this using the CodeRange object, which reserves and
411
+ // manages a range of virtual memory.
412
+ class CodeRange : public AllStatic {
413
+ public:
414
+ // Reserves a range of virtual memory, but does not commit any of it.
415
+ // Can only be called once, at heap initialization time.
416
+ // Returns false on failure.
417
+ static bool Setup(const size_t requested_size);
418
+
419
+ // Frees the range of virtual memory, and frees the data structures used to
420
+ // manage it.
421
+ static void TearDown();
422
+
423
+ static bool exists() { return code_range_ != NULL; }
424
+ static bool contains(Address address) {
425
+ if (code_range_ == NULL) return false;
426
+ Address start = static_cast<Address>(code_range_->address());
427
+ return start <= address && address < start + code_range_->size();
428
+ }
429
+
430
+ // Allocates a chunk of memory from the large-object portion of
431
+ // the code range. On platforms with no separate code range, should
432
+ // not be called.
433
+ MUST_USE_RESULT static void* AllocateRawMemory(const size_t requested,
434
+ size_t* allocated);
435
+ static void FreeRawMemory(void* buf, size_t length);
436
+
437
+ private:
438
+ // The reserved range of virtual memory that all code objects are put in.
439
+ static VirtualMemory* code_range_;
440
+ // Plain old data class, just a struct plus a constructor.
441
+ class FreeBlock {
442
+ public:
443
+ FreeBlock(Address start_arg, size_t size_arg)
444
+ : start(start_arg), size(size_arg) {}
445
+ FreeBlock(void* start_arg, size_t size_arg)
446
+ : start(static_cast<Address>(start_arg)), size(size_arg) {}
447
+
448
+ Address start;
449
+ size_t size;
450
+ };
451
+
452
+ // Freed blocks of memory are added to the free list. When the allocation
453
+ // list is exhausted, the free list is sorted and merged to make the new
454
+ // allocation list.
455
+ static List<FreeBlock> free_list_;
456
+ // Memory is allocated from the free blocks on the allocation list.
457
+ // The block at current_allocation_block_index_ is the current block.
458
+ static List<FreeBlock> allocation_list_;
459
+ static int current_allocation_block_index_;
460
+
461
+ // Finds a block on the allocation list that contains at least the
462
+ // requested amount of memory. If none is found, sorts and merges
463
+ // the existing free memory blocks, and searches again.
464
+ // If none can be found, terminates V8 with FatalProcessOutOfMemory.
465
+ static void GetNextAllocationBlock(size_t requested);
466
+ // Compares the start addresses of two free blocks.
467
+ static int CompareFreeBlockAddress(const FreeBlock* left,
468
+ const FreeBlock* right);
469
+ };
470
+
471
+
472
+ // ----------------------------------------------------------------------------
473
+ // A space acquires chunks of memory from the operating system. The memory
474
+ // allocator manages chunks for the paged heap spaces (old space and map
475
+ // space). A paged chunk consists of pages. Pages in a chunk have contiguous
476
+ // addresses and are linked as a list.
477
+ //
478
+ // The allocator keeps an initial chunk which is used for the new space. The
479
+ // leftover regions of the initial chunk are used for the initial chunks of
480
+ // old space and map space if they are big enough to hold at least one page.
481
+ // The allocator assumes that there is one old space and one map space, each
482
+ // expands the space by allocating kPagesPerChunk pages except the last
483
+ // expansion (before running out of space). The first chunk may contain fewer
484
+ // than kPagesPerChunk pages as well.
485
+ //
486
+ // The memory allocator also allocates chunks for the large object space, but
487
+ // they are managed by the space itself. The new space does not expand.
488
+ //
489
+ // The fact that pages for paged spaces are allocated and deallocated in chunks
490
+ // induces a constraint on the order of pages in a linked lists. We say that
491
+ // pages are linked in the chunk-order if and only if every two consecutive
492
+ // pages from the same chunk are consecutive in the linked list.
493
+ //
494
+
495
+
496
+ class MemoryAllocator : public AllStatic {
497
+ public:
498
+ // Initializes its internal bookkeeping structures.
499
+ // Max capacity of the total space and executable memory limit.
500
+ static bool Setup(intptr_t max_capacity, intptr_t capacity_executable);
501
+
502
+ // Deletes valid chunks.
503
+ static void TearDown();
504
+
505
+ // Reserves an initial address range of virtual memory to be split between
506
+ // the two new space semispaces, the old space, and the map space. The
507
+ // memory is not yet committed or assigned to spaces and split into pages.
508
+ // The initial chunk is unmapped when the memory allocator is torn down.
509
+ // This function should only be called when there is not already a reserved
510
+ // initial chunk (initial_chunk_ should be NULL). It returns the start
511
+ // address of the initial chunk if successful, with the side effect of
512
+ // setting the initial chunk, or else NULL if unsuccessful and leaves the
513
+ // initial chunk NULL.
514
+ static void* ReserveInitialChunk(const size_t requested);
515
+
516
+ // Commits pages from an as-yet-unmanaged block of virtual memory into a
517
+ // paged space. The block should be part of the initial chunk reserved via
518
+ // a call to ReserveInitialChunk. The number of pages is always returned in
519
+ // the output parameter num_pages. This function assumes that the start
520
+ // address is non-null and that it is big enough to hold at least one
521
+ // page-aligned page. The call always succeeds, and num_pages is always
522
+ // greater than zero.
523
+ static Page* CommitPages(Address start, size_t size, PagedSpace* owner,
524
+ int* num_pages);
525
+
526
+ // Commit a contiguous block of memory from the initial chunk. Assumes that
527
+ // the address is not NULL, the size is greater than zero, and that the
528
+ // block is contained in the initial chunk. Returns true if it succeeded
529
+ // and false otherwise.
530
+ static bool CommitBlock(Address start, size_t size, Executability executable);
531
+
532
+ // Uncommit a contiguous block of memory [start..(start+size)[.
533
+ // start is not NULL, the size is greater than zero, and the
534
+ // block is contained in the initial chunk. Returns true if it succeeded
535
+ // and false otherwise.
536
+ static bool UncommitBlock(Address start, size_t size);
537
+
538
+ // Zaps a contiguous block of memory [start..(start+size)[ thus
539
+ // filling it up with a recognizable non-NULL bit pattern.
540
+ static void ZapBlock(Address start, size_t size);
541
+
542
+ // Attempts to allocate the requested (non-zero) number of pages from the
543
+ // OS. Fewer pages might be allocated than requested. If it fails to
544
+ // allocate memory for the OS or cannot allocate a single page, this
545
+ // function returns an invalid page pointer (NULL). The caller must check
546
+ // whether the returned page is valid (by calling Page::is_valid()). It is
547
+ // guaranteed that allocated pages have contiguous addresses. The actual
548
+ // number of allocated pages is returned in the output parameter
549
+ // allocated_pages. If the PagedSpace owner is executable and there is
550
+ // a code range, the pages are allocated from the code range.
551
+ static Page* AllocatePages(int requested_pages, int* allocated_pages,
552
+ PagedSpace* owner);
553
+
554
+ // Frees pages from a given page and after. Requires pages to be
555
+ // linked in chunk-order (see comment for class).
556
+ // If 'p' is the first page of a chunk, pages from 'p' are freed
557
+ // and this function returns an invalid page pointer.
558
+ // Otherwise, the function searches a page after 'p' that is
559
+ // the first page of a chunk. Pages after the found page
560
+ // are freed and the function returns 'p'.
561
+ static Page* FreePages(Page* p);
562
+
563
+ // Frees all pages owned by given space.
564
+ static void FreeAllPages(PagedSpace* space);
565
+
566
+ // Allocates and frees raw memory of certain size.
567
+ // These are just thin wrappers around OS::Allocate and OS::Free,
568
+ // but keep track of allocated bytes as part of heap.
569
+ // If the flag is EXECUTABLE and a code range exists, the requested
570
+ // memory is allocated from the code range. If a code range exists
571
+ // and the freed memory is in it, the code range manages the freed memory.
572
+ MUST_USE_RESULT static void* AllocateRawMemory(const size_t requested,
573
+ size_t* allocated,
574
+ Executability executable);
575
+ static void FreeRawMemory(void* buf,
576
+ size_t length,
577
+ Executability executable);
578
+ static void PerformAllocationCallback(ObjectSpace space,
579
+ AllocationAction action,
580
+ size_t size);
581
+
582
+ static void AddMemoryAllocationCallback(MemoryAllocationCallback callback,
583
+ ObjectSpace space,
584
+ AllocationAction action);
585
+ static void RemoveMemoryAllocationCallback(
586
+ MemoryAllocationCallback callback);
587
+ static bool MemoryAllocationCallbackRegistered(
588
+ MemoryAllocationCallback callback);
589
+
590
+ // Returns the maximum available bytes of heaps.
591
+ static intptr_t Available() {
592
+ return capacity_ < size_ ? 0 : capacity_ - size_;
593
+ }
594
+
595
+ // Returns allocated spaces in bytes.
596
+ static intptr_t Size() { return size_; }
597
+
598
+ // Returns the maximum available executable bytes of heaps.
599
+ static intptr_t AvailableExecutable() {
600
+ if (capacity_executable_ < size_executable_) return 0;
601
+ return capacity_executable_ - size_executable_;
602
+ }
603
+
604
+ // Returns allocated executable spaces in bytes.
605
+ static intptr_t SizeExecutable() { return size_executable_; }
606
+
607
+ // Returns maximum available bytes that the old space can have.
608
+ static intptr_t MaxAvailable() {
609
+ return (Available() / Page::kPageSize) * Page::kObjectAreaSize;
610
+ }
611
+
612
+ // Sanity check on a pointer.
613
+ static bool SafeIsInAPageChunk(Address addr);
614
+
615
+ // Links two pages.
616
+ static inline void SetNextPage(Page* prev, Page* next);
617
+
618
+ // Returns the next page of a given page.
619
+ static inline Page* GetNextPage(Page* p);
620
+
621
+ // Checks whether a page belongs to a space.
622
+ static inline bool IsPageInSpace(Page* p, PagedSpace* space);
623
+
624
+ // Returns the space that owns the given page.
625
+ static inline PagedSpace* PageOwner(Page* page);
626
+
627
+ // Finds the first/last page in the same chunk as a given page.
628
+ static Page* FindFirstPageInSameChunk(Page* p);
629
+ static Page* FindLastPageInSameChunk(Page* p);
630
+
631
+ // Relinks list of pages owned by space to make it chunk-ordered.
632
+ // Returns new first and last pages of space.
633
+ // Also returns last page in relinked list which has WasInUsedBeforeMC
634
+ // flag set.
635
+ static void RelinkPageListInChunkOrder(PagedSpace* space,
636
+ Page** first_page,
637
+ Page** last_page,
638
+ Page** last_page_in_use);
639
+
640
+ #ifdef ENABLE_HEAP_PROTECTION
641
+ // Protect/unprotect a block of memory by marking it read-only/writable.
642
+ static inline void Protect(Address start, size_t size);
643
+ static inline void Unprotect(Address start, size_t size,
644
+ Executability executable);
645
+
646
+ // Protect/unprotect a chunk given a page in the chunk.
647
+ static inline void ProtectChunkFromPage(Page* page);
648
+ static inline void UnprotectChunkFromPage(Page* page);
649
+ #endif
650
+
651
+ #ifdef DEBUG
652
+ // Reports statistic info of the space.
653
+ static void ReportStatistics();
654
+ #endif
655
+
656
+ static void AddToAllocatedChunks(Address addr, intptr_t size);
657
+ static void RemoveFromAllocatedChunks(Address addr, intptr_t size);
658
+ // Note: This only checks the regular chunks, not the odd-sized initial
659
+ // chunk.
660
+ static bool InAllocatedChunks(Address addr);
661
+
662
+ // Due to encoding limitation, we can only have 8K chunks.
663
+ static const int kMaxNofChunks = 1 << kPageSizeBits;
664
+ // If a chunk has at least 16 pages, the maximum heap size is about
665
+ // 8K * 8K * 16 = 1G bytes.
666
+ #ifdef V8_TARGET_ARCH_X64
667
+ static const int kPagesPerChunk = 32;
668
+ // On 64 bit the chunk table consists of 4 levels of 4096-entry tables.
669
+ static const int kPagesPerChunkLog2 = 5;
670
+ static const int kChunkTableLevels = 4;
671
+ static const int kChunkTableBitsPerLevel = 12;
672
+ #else
673
+ static const int kPagesPerChunk = 16;
674
+ // On 32 bit the chunk table consists of 2 levels of 256-entry tables.
675
+ static const int kPagesPerChunkLog2 = 4;
676
+ static const int kChunkTableLevels = 2;
677
+ static const int kChunkTableBitsPerLevel = 8;
678
+ #endif
679
+
680
+ private:
681
+ static const int kChunkSize = kPagesPerChunk * Page::kPageSize;
682
+ static const int kChunkSizeLog2 = kPagesPerChunkLog2 + kPageSizeBits;
683
+ static const int kChunkTableTopLevelEntries =
684
+ 1 << (sizeof(intptr_t) * kBitsPerByte - kChunkSizeLog2 -
685
+ (kChunkTableLevels - 1) * kChunkTableBitsPerLevel);
686
+
687
+ // The chunks are not chunk-size aligned so for a given chunk-sized area of
688
+ // memory there can be two chunks that cover it.
689
+ static const int kChunkTableFineGrainedWordsPerEntry = 2;
690
+ static const uintptr_t kUnusedChunkTableEntry = 0;
691
+
692
+ // Maximum space size in bytes.
693
+ static intptr_t capacity_;
694
+ // Maximum subset of capacity_ that can be executable
695
+ static intptr_t capacity_executable_;
696
+
697
+ // Top level table to track whether memory is part of a chunk or not.
698
+ static uintptr_t chunk_table_[kChunkTableTopLevelEntries];
699
+
700
+ // Allocated space size in bytes.
701
+ static intptr_t size_;
702
+ // Allocated executable space size in bytes.
703
+ static intptr_t size_executable_;
704
+
705
+ struct MemoryAllocationCallbackRegistration {
706
+ MemoryAllocationCallbackRegistration(MemoryAllocationCallback callback,
707
+ ObjectSpace space,
708
+ AllocationAction action)
709
+ : callback(callback), space(space), action(action) {
710
+ }
711
+ MemoryAllocationCallback callback;
712
+ ObjectSpace space;
713
+ AllocationAction action;
714
+ };
715
+ // A List of callback that are triggered when memory is allocated or free'd
716
+ static List<MemoryAllocationCallbackRegistration>
717
+ memory_allocation_callbacks_;
718
+
719
+ // The initial chunk of virtual memory.
720
+ static VirtualMemory* initial_chunk_;
721
+
722
+ // Allocated chunk info: chunk start address, chunk size, and owning space.
723
+ class ChunkInfo BASE_EMBEDDED {
724
+ public:
725
+ ChunkInfo() : address_(NULL),
726
+ size_(0),
727
+ owner_(NULL),
728
+ executable_(NOT_EXECUTABLE) {}
729
+ inline void init(Address a, size_t s, PagedSpace* o);
730
+ Address address() { return address_; }
731
+ size_t size() { return size_; }
732
+ PagedSpace* owner() { return owner_; }
733
+ // We save executability of the owner to allow using it
734
+ // when collecting stats after the owner has been destroyed.
735
+ Executability executable() const { return executable_; }
736
+
737
+ private:
738
+ Address address_;
739
+ size_t size_;
740
+ PagedSpace* owner_;
741
+ Executability executable_;
742
+ };
743
+
744
+ // Chunks_, free_chunk_ids_ and top_ act as a stack of free chunk ids.
745
+ static List<ChunkInfo> chunks_;
746
+ static List<int> free_chunk_ids_;
747
+ static int max_nof_chunks_;
748
+ static int top_;
749
+
750
+ // Push/pop a free chunk id onto/from the stack.
751
+ static void Push(int free_chunk_id);
752
+ static int Pop();
753
+ static bool OutOfChunkIds() { return top_ == 0; }
754
+
755
+ // Frees a chunk.
756
+ static void DeleteChunk(int chunk_id);
757
+
758
+ // Helpers to maintain and query the chunk tables.
759
+ static void AddChunkUsingAddress(
760
+ uintptr_t chunk_start, // Where the chunk starts.
761
+ uintptr_t chunk_index_base); // Used to place the chunk in the tables.
762
+ static void RemoveChunkFoundUsingAddress(
763
+ uintptr_t chunk_start, // Where the chunk starts.
764
+ uintptr_t chunk_index_base); // Used to locate the entry in the tables.
765
+ // Controls whether the lookup creates intermediate levels of tables as
766
+ // needed.
767
+ enum CreateTables { kDontCreateTables, kCreateTablesAsNeeded };
768
+ static uintptr_t* AllocatedChunksFinder(uintptr_t* table,
769
+ uintptr_t address,
770
+ int bit_position,
771
+ CreateTables create_as_needed);
772
+ static void FreeChunkTables(uintptr_t* array, int length, int level);
773
+ static int FineGrainedIndexForAddress(uintptr_t address) {
774
+ int index = ((address >> kChunkSizeLog2) &
775
+ ((1 << kChunkTableBitsPerLevel) - 1));
776
+ return index * kChunkTableFineGrainedWordsPerEntry;
777
+ }
778
+
779
+
780
+ // Basic check whether a chunk id is in the valid range.
781
+ static inline bool IsValidChunkId(int chunk_id);
782
+
783
+ // Checks whether a chunk id identifies an allocated chunk.
784
+ static inline bool IsValidChunk(int chunk_id);
785
+
786
+ // Returns the chunk id that a page belongs to.
787
+ static inline int GetChunkId(Page* p);
788
+
789
+ // True if the address lies in the initial chunk.
790
+ static inline bool InInitialChunk(Address address);
791
+
792
+ // Initializes pages in a chunk. Returns the first page address.
793
+ // This function and GetChunkId() are provided for the mark-compact
794
+ // collector to rebuild page headers in the from space, which is
795
+ // used as a marking stack and its page headers are destroyed.
796
+ static Page* InitializePagesInChunk(int chunk_id, int pages_in_chunk,
797
+ PagedSpace* owner);
798
+
799
+ static Page* RelinkPagesInChunk(int chunk_id,
800
+ Address chunk_start,
801
+ size_t chunk_size,
802
+ Page* prev,
803
+ Page** last_page_in_use);
804
+ };
805
+
806
+
807
+ // -----------------------------------------------------------------------------
808
+ // Interface for heap object iterator to be implemented by all object space
809
+ // object iterators.
810
+ //
811
+ // NOTE: The space specific object iterators also implements the own next()
812
+ // method which is used to avoid using virtual functions
813
+ // iterating a specific space.
814
+
815
+ class ObjectIterator : public Malloced {
816
+ public:
817
+ virtual ~ObjectIterator() { }
818
+
819
+ virtual HeapObject* next_object() = 0;
820
+ };
821
+
822
+
823
+ // -----------------------------------------------------------------------------
824
+ // Heap object iterator in new/old/map spaces.
825
+ //
826
+ // A HeapObjectIterator iterates objects from a given address to the
827
+ // top of a space. The given address must be below the current
828
+ // allocation pointer (space top). There are some caveats.
829
+ //
830
+ // (1) If the space top changes upward during iteration (because of
831
+ // allocating new objects), the iterator does not iterate objects
832
+ // above the original space top. The caller must create a new
833
+ // iterator starting from the old top in order to visit these new
834
+ // objects.
835
+ //
836
+ // (2) If new objects are allocated below the original allocation top
837
+ // (e.g., free-list allocation in paged spaces), the new objects
838
+ // may or may not be iterated depending on their position with
839
+ // respect to the current point of iteration.
840
+ //
841
+ // (3) The space top should not change downward during iteration,
842
+ // otherwise the iterator will return not-necessarily-valid
843
+ // objects.
844
+
845
+ class HeapObjectIterator: public ObjectIterator {
846
+ public:
847
+ // Creates a new object iterator in a given space. If a start
848
+ // address is not given, the iterator starts from the space bottom.
849
+ // If the size function is not given, the iterator calls the default
850
+ // Object::Size().
851
+ explicit HeapObjectIterator(PagedSpace* space);
852
+ HeapObjectIterator(PagedSpace* space, HeapObjectCallback size_func);
853
+ HeapObjectIterator(PagedSpace* space, Address start);
854
+ HeapObjectIterator(PagedSpace* space,
855
+ Address start,
856
+ HeapObjectCallback size_func);
857
+ HeapObjectIterator(Page* page, HeapObjectCallback size_func);
858
+
859
+ inline HeapObject* next() {
860
+ return (cur_addr_ < cur_limit_) ? FromCurrentPage() : FromNextPage();
861
+ }
862
+
863
+ // implementation of ObjectIterator.
864
+ virtual HeapObject* next_object() { return next(); }
865
+
866
+ private:
867
+ Address cur_addr_; // current iteration point
868
+ Address end_addr_; // end iteration point
869
+ Address cur_limit_; // current page limit
870
+ HeapObjectCallback size_func_; // size function
871
+ Page* end_page_; // caches the page of the end address
872
+
873
+ HeapObject* FromCurrentPage() {
874
+ ASSERT(cur_addr_ < cur_limit_);
875
+
876
+ HeapObject* obj = HeapObject::FromAddress(cur_addr_);
877
+ int obj_size = (size_func_ == NULL) ? obj->Size() : size_func_(obj);
878
+ ASSERT_OBJECT_SIZE(obj_size);
879
+
880
+ cur_addr_ += obj_size;
881
+ ASSERT(cur_addr_ <= cur_limit_);
882
+
883
+ return obj;
884
+ }
885
+
886
+ // Slow path of next, goes into the next page.
887
+ HeapObject* FromNextPage();
888
+
889
+ // Initializes fields.
890
+ void Initialize(Address start, Address end, HeapObjectCallback size_func);
891
+
892
+ #ifdef DEBUG
893
+ // Verifies whether fields have valid values.
894
+ void Verify();
895
+ #endif
896
+ };
897
+
898
+
899
+ // -----------------------------------------------------------------------------
900
+ // A PageIterator iterates the pages in a paged space.
901
+ //
902
+ // The PageIterator class provides three modes for iterating pages in a space:
903
+ // PAGES_IN_USE iterates pages containing allocated objects.
904
+ // PAGES_USED_BY_MC iterates pages that hold relocated objects during a
905
+ // mark-compact collection.
906
+ // ALL_PAGES iterates all pages in the space.
907
+ //
908
+ // There are some caveats.
909
+ //
910
+ // (1) If the space expands during iteration, new pages will not be
911
+ // returned by the iterator in any mode.
912
+ //
913
+ // (2) If new objects are allocated during iteration, they will appear
914
+ // in pages returned by the iterator. Allocation may cause the
915
+ // allocation pointer or MC allocation pointer in the last page to
916
+ // change between constructing the iterator and iterating the last
917
+ // page.
918
+ //
919
+ // (3) The space should not shrink during iteration, otherwise the
920
+ // iterator will return deallocated pages.
921
+
922
+ class PageIterator BASE_EMBEDDED {
923
+ public:
924
+ enum Mode {
925
+ PAGES_IN_USE,
926
+ PAGES_USED_BY_MC,
927
+ ALL_PAGES
928
+ };
929
+
930
+ PageIterator(PagedSpace* space, Mode mode);
931
+
932
+ inline bool has_next();
933
+ inline Page* next();
934
+
935
+ private:
936
+ PagedSpace* space_;
937
+ Page* prev_page_; // Previous page returned.
938
+ Page* stop_page_; // Page to stop at (last page returned by the iterator).
939
+ };
940
+
941
+
942
+ // -----------------------------------------------------------------------------
943
+ // A space has a list of pages. The next page can be accessed via
944
+ // Page::next_page() call. The next page of the last page is an
945
+ // invalid page pointer. A space can expand and shrink dynamically.
946
+
947
+ // An abstraction of allocation and relocation pointers in a page-structured
948
+ // space.
949
+ class AllocationInfo {
950
+ public:
951
+ Address top; // current allocation top
952
+ Address limit; // current allocation limit
953
+
954
+ #ifdef DEBUG
955
+ bool VerifyPagedAllocation() {
956
+ return (Page::FromAllocationTop(top) == Page::FromAllocationTop(limit))
957
+ && (top <= limit);
958
+ }
959
+ #endif
960
+ };
961
+
962
+
963
+ // An abstraction of the accounting statistics of a page-structured space.
964
+ // The 'capacity' of a space is the number of object-area bytes (ie, not
965
+ // including page bookkeeping structures) currently in the space. The 'size'
966
+ // of a space is the number of allocated bytes, the 'waste' in the space is
967
+ // the number of bytes that are not allocated and not available to
968
+ // allocation without reorganizing the space via a GC (eg, small blocks due
969
+ // to internal fragmentation, top of page areas in map space), and the bytes
970
+ // 'available' is the number of unallocated bytes that are not waste. The
971
+ // capacity is the sum of size, waste, and available.
972
+ //
973
+ // The stats are only set by functions that ensure they stay balanced. These
974
+ // functions increase or decrease one of the non-capacity stats in
975
+ // conjunction with capacity, or else they always balance increases and
976
+ // decreases to the non-capacity stats.
977
+ class AllocationStats BASE_EMBEDDED {
978
+ public:
979
+ AllocationStats() { Clear(); }
980
+
981
+ // Zero out all the allocation statistics (ie, no capacity).
982
+ void Clear() {
983
+ capacity_ = 0;
984
+ available_ = 0;
985
+ size_ = 0;
986
+ waste_ = 0;
987
+ }
988
+
989
+ // Reset the allocation statistics (ie, available = capacity with no
990
+ // wasted or allocated bytes).
991
+ void Reset() {
992
+ available_ = capacity_;
993
+ size_ = 0;
994
+ waste_ = 0;
995
+ }
996
+
997
+ // Accessors for the allocation statistics.
998
+ intptr_t Capacity() { return capacity_; }
999
+ intptr_t Available() { return available_; }
1000
+ intptr_t Size() { return size_; }
1001
+ intptr_t Waste() { return waste_; }
1002
+
1003
+ // Grow the space by adding available bytes.
1004
+ void ExpandSpace(int size_in_bytes) {
1005
+ capacity_ += size_in_bytes;
1006
+ available_ += size_in_bytes;
1007
+ }
1008
+
1009
+ // Shrink the space by removing available bytes.
1010
+ void ShrinkSpace(int size_in_bytes) {
1011
+ capacity_ -= size_in_bytes;
1012
+ available_ -= size_in_bytes;
1013
+ }
1014
+
1015
+ // Allocate from available bytes (available -> size).
1016
+ void AllocateBytes(intptr_t size_in_bytes) {
1017
+ available_ -= size_in_bytes;
1018
+ size_ += size_in_bytes;
1019
+ }
1020
+
1021
+ // Free allocated bytes, making them available (size -> available).
1022
+ void DeallocateBytes(intptr_t size_in_bytes) {
1023
+ size_ -= size_in_bytes;
1024
+ available_ += size_in_bytes;
1025
+ }
1026
+
1027
+ // Waste free bytes (available -> waste).
1028
+ void WasteBytes(int size_in_bytes) {
1029
+ available_ -= size_in_bytes;
1030
+ waste_ += size_in_bytes;
1031
+ }
1032
+
1033
+ // Consider the wasted bytes to be allocated, as they contain filler
1034
+ // objects (waste -> size).
1035
+ void FillWastedBytes(intptr_t size_in_bytes) {
1036
+ waste_ -= size_in_bytes;
1037
+ size_ += size_in_bytes;
1038
+ }
1039
+
1040
+ private:
1041
+ intptr_t capacity_;
1042
+ intptr_t available_;
1043
+ intptr_t size_;
1044
+ intptr_t waste_;
1045
+ };
1046
+
1047
+
1048
+ class PagedSpace : public Space {
1049
+ public:
1050
+ // Creates a space with a maximum capacity, and an id.
1051
+ PagedSpace(intptr_t max_capacity,
1052
+ AllocationSpace id,
1053
+ Executability executable);
1054
+
1055
+ virtual ~PagedSpace() {}
1056
+
1057
+ // Set up the space using the given address range of virtual memory (from
1058
+ // the memory allocator's initial chunk) if possible. If the block of
1059
+ // addresses is not big enough to contain a single page-aligned page, a
1060
+ // fresh chunk will be allocated.
1061
+ bool Setup(Address start, size_t size);
1062
+
1063
+ // Returns true if the space has been successfully set up and not
1064
+ // subsequently torn down.
1065
+ bool HasBeenSetup();
1066
+
1067
+ // Cleans up the space, frees all pages in this space except those belonging
1068
+ // to the initial chunk, uncommits addresses in the initial chunk.
1069
+ void TearDown();
1070
+
1071
+ // Checks whether an object/address is in this space.
1072
+ inline bool Contains(Address a);
1073
+ bool Contains(HeapObject* o) { return Contains(o->address()); }
1074
+ // Never crashes even if a is not a valid pointer.
1075
+ inline bool SafeContains(Address a);
1076
+
1077
+ // Given an address occupied by a live object, return that object if it is
1078
+ // in this space, or Failure::Exception() if it is not. The implementation
1079
+ // iterates over objects in the page containing the address, the cost is
1080
+ // linear in the number of objects in the page. It may be slow.
1081
+ MUST_USE_RESULT MaybeObject* FindObject(Address addr);
1082
+
1083
+ // Checks whether page is currently in use by this space.
1084
+ bool IsUsed(Page* page);
1085
+
1086
+ void MarkAllPagesClean();
1087
+
1088
+ // Prepares for a mark-compact GC.
1089
+ virtual void PrepareForMarkCompact(bool will_compact);
1090
+
1091
+ // The top of allocation in a page in this space. Undefined if page is unused.
1092
+ Address PageAllocationTop(Page* page) {
1093
+ return page == TopPageOf(allocation_info_) ? top()
1094
+ : PageAllocationLimit(page);
1095
+ }
1096
+
1097
+ // The limit of allocation for a page in this space.
1098
+ virtual Address PageAllocationLimit(Page* page) = 0;
1099
+
1100
+ void FlushTopPageWatermark() {
1101
+ AllocationTopPage()->SetCachedAllocationWatermark(top());
1102
+ AllocationTopPage()->InvalidateWatermark(true);
1103
+ }
1104
+
1105
+ // Current capacity without growing (Size() + Available() + Waste()).
1106
+ intptr_t Capacity() { return accounting_stats_.Capacity(); }
1107
+
1108
+ // Total amount of memory committed for this space. For paged
1109
+ // spaces this equals the capacity.
1110
+ intptr_t CommittedMemory() { return Capacity(); }
1111
+
1112
+ // Available bytes without growing.
1113
+ intptr_t Available() { return accounting_stats_.Available(); }
1114
+
1115
+ // Allocated bytes in this space.
1116
+ virtual intptr_t Size() { return accounting_stats_.Size(); }
1117
+
1118
+ // Wasted bytes due to fragmentation and not recoverable until the
1119
+ // next GC of this space.
1120
+ intptr_t Waste() { return accounting_stats_.Waste(); }
1121
+
1122
+ // Returns the address of the first object in this space.
1123
+ Address bottom() { return first_page_->ObjectAreaStart(); }
1124
+
1125
+ // Returns the allocation pointer in this space.
1126
+ Address top() { return allocation_info_.top; }
1127
+
1128
+ // Allocate the requested number of bytes in the space if possible, return a
1129
+ // failure object if not.
1130
+ MUST_USE_RESULT inline MaybeObject* AllocateRaw(int size_in_bytes);
1131
+
1132
+ // Allocate the requested number of bytes for relocation during mark-compact
1133
+ // collection.
1134
+ MUST_USE_RESULT inline MaybeObject* MCAllocateRaw(int size_in_bytes);
1135
+
1136
+ virtual bool ReserveSpace(int bytes);
1137
+
1138
+ // Used by ReserveSpace.
1139
+ virtual void PutRestOfCurrentPageOnFreeList(Page* current_page) = 0;
1140
+
1141
+ // Free all pages in range from prev (exclusive) to last (inclusive).
1142
+ // Freed pages are moved to the end of page list.
1143
+ void FreePages(Page* prev, Page* last);
1144
+
1145
+ // Deallocates a block.
1146
+ virtual void DeallocateBlock(Address start,
1147
+ int size_in_bytes,
1148
+ bool add_to_freelist) = 0;
1149
+
1150
+ // Set space allocation info.
1151
+ void SetTop(Address top) {
1152
+ allocation_info_.top = top;
1153
+ allocation_info_.limit = PageAllocationLimit(Page::FromAllocationTop(top));
1154
+ }
1155
+
1156
+ // ---------------------------------------------------------------------------
1157
+ // Mark-compact collection support functions
1158
+
1159
+ // Set the relocation point to the beginning of the space.
1160
+ void MCResetRelocationInfo();
1161
+
1162
+ // Writes relocation info to the top page.
1163
+ void MCWriteRelocationInfoToPage() {
1164
+ TopPageOf(mc_forwarding_info_)->
1165
+ SetAllocationWatermark(mc_forwarding_info_.top);
1166
+ }
1167
+
1168
+ // Computes the offset of a given address in this space to the beginning
1169
+ // of the space.
1170
+ int MCSpaceOffsetForAddress(Address addr);
1171
+
1172
+ // Updates the allocation pointer to the relocation top after a mark-compact
1173
+ // collection.
1174
+ virtual void MCCommitRelocationInfo() = 0;
1175
+
1176
+ // Releases half of unused pages.
1177
+ void Shrink();
1178
+
1179
+ // Ensures that the capacity is at least 'capacity'. Returns false on failure.
1180
+ bool EnsureCapacity(int capacity);
1181
+
1182
+ #ifdef ENABLE_HEAP_PROTECTION
1183
+ // Protect/unprotect the space by marking it read-only/writable.
1184
+ void Protect();
1185
+ void Unprotect();
1186
+ #endif
1187
+
1188
+ #ifdef DEBUG
1189
+ // Print meta info and objects in this space.
1190
+ virtual void Print();
1191
+
1192
+ // Verify integrity of this space.
1193
+ virtual void Verify(ObjectVisitor* visitor);
1194
+
1195
+ // Overridden by subclasses to verify space-specific object
1196
+ // properties (e.g., only maps or free-list nodes are in map space).
1197
+ virtual void VerifyObject(HeapObject* obj) {}
1198
+
1199
+ // Report code object related statistics
1200
+ void CollectCodeStatistics();
1201
+ static void ReportCodeStatistics();
1202
+ static void ResetCodeStatistics();
1203
+ #endif
1204
+
1205
+ // Returns the page of the allocation pointer.
1206
+ Page* AllocationTopPage() { return TopPageOf(allocation_info_); }
1207
+
1208
+ void RelinkPageListInChunkOrder(bool deallocate_blocks);
1209
+
1210
+ protected:
1211
+ // Maximum capacity of this space.
1212
+ intptr_t max_capacity_;
1213
+
1214
+ // Accounting information for this space.
1215
+ AllocationStats accounting_stats_;
1216
+
1217
+ // The first page in this space.
1218
+ Page* first_page_;
1219
+
1220
+ // The last page in this space. Initially set in Setup, updated in
1221
+ // Expand and Shrink.
1222
+ Page* last_page_;
1223
+
1224
+ // True if pages owned by this space are linked in chunk-order.
1225
+ // See comment for class MemoryAllocator for definition of chunk-order.
1226
+ bool page_list_is_chunk_ordered_;
1227
+
1228
+ // Normal allocation information.
1229
+ AllocationInfo allocation_info_;
1230
+
1231
+ // Relocation information during mark-compact collections.
1232
+ AllocationInfo mc_forwarding_info_;
1233
+
1234
+ // Bytes of each page that cannot be allocated. Possibly non-zero
1235
+ // for pages in spaces with only fixed-size objects. Always zero
1236
+ // for pages in spaces with variable sized objects (those pages are
1237
+ // padded with free-list nodes).
1238
+ int page_extra_;
1239
+
1240
+ // Sets allocation pointer to a page bottom.
1241
+ static void SetAllocationInfo(AllocationInfo* alloc_info, Page* p);
1242
+
1243
+ // Returns the top page specified by an allocation info structure.
1244
+ static Page* TopPageOf(AllocationInfo alloc_info) {
1245
+ return Page::FromAllocationTop(alloc_info.limit);
1246
+ }
1247
+
1248
+ int CountPagesToTop() {
1249
+ Page* p = Page::FromAllocationTop(allocation_info_.top);
1250
+ PageIterator it(this, PageIterator::ALL_PAGES);
1251
+ int counter = 1;
1252
+ while (it.has_next()) {
1253
+ if (it.next() == p) return counter;
1254
+ counter++;
1255
+ }
1256
+ UNREACHABLE();
1257
+ return -1;
1258
+ }
1259
+
1260
+ // Expands the space by allocating a fixed number of pages. Returns false if
1261
+ // it cannot allocate requested number of pages from OS. Newly allocated
1262
+ // pages are append to the last_page;
1263
+ bool Expand(Page* last_page);
1264
+
1265
+ // Generic fast case allocation function that tries linear allocation in
1266
+ // the top page of 'alloc_info'. Returns NULL on failure.
1267
+ inline HeapObject* AllocateLinearly(AllocationInfo* alloc_info,
1268
+ int size_in_bytes);
1269
+
1270
+ // During normal allocation or deserialization, roll to the next page in
1271
+ // the space (there is assumed to be one) and allocate there. This
1272
+ // function is space-dependent.
1273
+ virtual HeapObject* AllocateInNextPage(Page* current_page,
1274
+ int size_in_bytes) = 0;
1275
+
1276
+ // Slow path of AllocateRaw. This function is space-dependent.
1277
+ MUST_USE_RESULT virtual HeapObject* SlowAllocateRaw(int size_in_bytes) = 0;
1278
+
1279
+ // Slow path of MCAllocateRaw.
1280
+ MUST_USE_RESULT HeapObject* SlowMCAllocateRaw(int size_in_bytes);
1281
+
1282
+ #ifdef DEBUG
1283
+ // Returns the number of total pages in this space.
1284
+ int CountTotalPages();
1285
+ #endif
1286
+ private:
1287
+
1288
+ // Returns a pointer to the page of the relocation pointer.
1289
+ Page* MCRelocationTopPage() { return TopPageOf(mc_forwarding_info_); }
1290
+
1291
+ friend class PageIterator;
1292
+ };
1293
+
1294
+
1295
+ #if defined(DEBUG) || defined(ENABLE_LOGGING_AND_PROFILING)
1296
+ class NumberAndSizeInfo BASE_EMBEDDED {
1297
+ public:
1298
+ NumberAndSizeInfo() : number_(0), bytes_(0) {}
1299
+
1300
+ int number() const { return number_; }
1301
+ void increment_number(int num) { number_ += num; }
1302
+
1303
+ int bytes() const { return bytes_; }
1304
+ void increment_bytes(int size) { bytes_ += size; }
1305
+
1306
+ void clear() {
1307
+ number_ = 0;
1308
+ bytes_ = 0;
1309
+ }
1310
+
1311
+ private:
1312
+ int number_;
1313
+ int bytes_;
1314
+ };
1315
+
1316
+
1317
+ // HistogramInfo class for recording a single "bar" of a histogram. This
1318
+ // class is used for collecting statistics to print to stdout (when compiled
1319
+ // with DEBUG) or to the log file (when compiled with
1320
+ // ENABLE_LOGGING_AND_PROFILING).
1321
+ class HistogramInfo: public NumberAndSizeInfo {
1322
+ public:
1323
+ HistogramInfo() : NumberAndSizeInfo() {}
1324
+
1325
+ const char* name() { return name_; }
1326
+ void set_name(const char* name) { name_ = name; }
1327
+
1328
+ private:
1329
+ const char* name_;
1330
+ };
1331
+ #endif
1332
+
1333
+
1334
+ // -----------------------------------------------------------------------------
1335
+ // SemiSpace in young generation
1336
+ //
1337
+ // A semispace is a contiguous chunk of memory. The mark-compact collector
1338
+ // uses the memory in the from space as a marking stack when tracing live
1339
+ // objects.
1340
+
1341
+ class SemiSpace : public Space {
1342
+ public:
1343
+ // Constructor.
1344
+ SemiSpace() :Space(NEW_SPACE, NOT_EXECUTABLE) {
1345
+ start_ = NULL;
1346
+ age_mark_ = NULL;
1347
+ }
1348
+
1349
+ // Sets up the semispace using the given chunk.
1350
+ bool Setup(Address start, int initial_capacity, int maximum_capacity);
1351
+
1352
+ // Tear down the space. Heap memory was not allocated by the space, so it
1353
+ // is not deallocated here.
1354
+ void TearDown();
1355
+
1356
+ // True if the space has been set up but not torn down.
1357
+ bool HasBeenSetup() { return start_ != NULL; }
1358
+
1359
+ // Grow the size of the semispace by committing extra virtual memory.
1360
+ // Assumes that the caller has checked that the semispace has not reached
1361
+ // its maximum capacity (and thus there is space available in the reserved
1362
+ // address range to grow).
1363
+ bool Grow();
1364
+
1365
+ // Grow the semispace to the new capacity. The new capacity
1366
+ // requested must be larger than the current capacity.
1367
+ bool GrowTo(int new_capacity);
1368
+
1369
+ // Shrinks the semispace to the new capacity. The new capacity
1370
+ // requested must be more than the amount of used memory in the
1371
+ // semispace and less than the current capacity.
1372
+ bool ShrinkTo(int new_capacity);
1373
+
1374
+ // Returns the start address of the space.
1375
+ Address low() { return start_; }
1376
+ // Returns one past the end address of the space.
1377
+ Address high() { return low() + capacity_; }
1378
+
1379
+ // Age mark accessors.
1380
+ Address age_mark() { return age_mark_; }
1381
+ void set_age_mark(Address mark) { age_mark_ = mark; }
1382
+
1383
+ // True if the address is in the address range of this semispace (not
1384
+ // necessarily below the allocation pointer).
1385
+ bool Contains(Address a) {
1386
+ return (reinterpret_cast<uintptr_t>(a) & address_mask_)
1387
+ == reinterpret_cast<uintptr_t>(start_);
1388
+ }
1389
+
1390
+ // True if the object is a heap object in the address range of this
1391
+ // semispace (not necessarily below the allocation pointer).
1392
+ bool Contains(Object* o) {
1393
+ return (reinterpret_cast<uintptr_t>(o) & object_mask_) == object_expected_;
1394
+ }
1395
+
1396
+ // The offset of an address from the beginning of the space.
1397
+ int SpaceOffsetForAddress(Address addr) {
1398
+ return static_cast<int>(addr - low());
1399
+ }
1400
+
1401
+ // If we don't have these here then SemiSpace will be abstract. However
1402
+ // they should never be called.
1403
+ virtual intptr_t Size() {
1404
+ UNREACHABLE();
1405
+ return 0;
1406
+ }
1407
+
1408
+ virtual bool ReserveSpace(int bytes) {
1409
+ UNREACHABLE();
1410
+ return false;
1411
+ }
1412
+
1413
+ bool is_committed() { return committed_; }
1414
+ bool Commit();
1415
+ bool Uncommit();
1416
+
1417
+ #ifdef ENABLE_HEAP_PROTECTION
1418
+ // Protect/unprotect the space by marking it read-only/writable.
1419
+ virtual void Protect() {}
1420
+ virtual void Unprotect() {}
1421
+ #endif
1422
+
1423
+ #ifdef DEBUG
1424
+ virtual void Print();
1425
+ virtual void Verify();
1426
+ #endif
1427
+
1428
+ // Returns the current capacity of the semi space.
1429
+ int Capacity() { return capacity_; }
1430
+
1431
+ // Returns the maximum capacity of the semi space.
1432
+ int MaximumCapacity() { return maximum_capacity_; }
1433
+
1434
+ // Returns the initial capacity of the semi space.
1435
+ int InitialCapacity() { return initial_capacity_; }
1436
+
1437
+ private:
1438
+ // The current and maximum capacity of the space.
1439
+ int capacity_;
1440
+ int maximum_capacity_;
1441
+ int initial_capacity_;
1442
+
1443
+ // The start address of the space.
1444
+ Address start_;
1445
+ // Used to govern object promotion during mark-compact collection.
1446
+ Address age_mark_;
1447
+
1448
+ // Masks and comparison values to test for containment in this semispace.
1449
+ uintptr_t address_mask_;
1450
+ uintptr_t object_mask_;
1451
+ uintptr_t object_expected_;
1452
+
1453
+ bool committed_;
1454
+
1455
+ public:
1456
+ TRACK_MEMORY("SemiSpace")
1457
+ };
1458
+
1459
+
1460
+ // A SemiSpaceIterator is an ObjectIterator that iterates over the active
1461
+ // semispace of the heap's new space. It iterates over the objects in the
1462
+ // semispace from a given start address (defaulting to the bottom of the
1463
+ // semispace) to the top of the semispace. New objects allocated after the
1464
+ // iterator is created are not iterated.
1465
+ class SemiSpaceIterator : public ObjectIterator {
1466
+ public:
1467
+ // Create an iterator over the objects in the given space. If no start
1468
+ // address is given, the iterator starts from the bottom of the space. If
1469
+ // no size function is given, the iterator calls Object::Size().
1470
+ explicit SemiSpaceIterator(NewSpace* space);
1471
+ SemiSpaceIterator(NewSpace* space, HeapObjectCallback size_func);
1472
+ SemiSpaceIterator(NewSpace* space, Address start);
1473
+
1474
+ HeapObject* next() {
1475
+ if (current_ == limit_) return NULL;
1476
+
1477
+ HeapObject* object = HeapObject::FromAddress(current_);
1478
+ int size = (size_func_ == NULL) ? object->Size() : size_func_(object);
1479
+
1480
+ current_ += size;
1481
+ return object;
1482
+ }
1483
+
1484
+ // Implementation of the ObjectIterator functions.
1485
+ virtual HeapObject* next_object() { return next(); }
1486
+
1487
+ private:
1488
+ void Initialize(NewSpace* space, Address start, Address end,
1489
+ HeapObjectCallback size_func);
1490
+
1491
+ // The semispace.
1492
+ SemiSpace* space_;
1493
+ // The current iteration point.
1494
+ Address current_;
1495
+ // The end of iteration.
1496
+ Address limit_;
1497
+ // The callback function.
1498
+ HeapObjectCallback size_func_;
1499
+ };
1500
+
1501
+
1502
+ // -----------------------------------------------------------------------------
1503
+ // The young generation space.
1504
+ //
1505
+ // The new space consists of a contiguous pair of semispaces. It simply
1506
+ // forwards most functions to the appropriate semispace.
1507
+
1508
+ class NewSpace : public Space {
1509
+ public:
1510
+ // Constructor.
1511
+ NewSpace() : Space(NEW_SPACE, NOT_EXECUTABLE) {}
1512
+
1513
+ // Sets up the new space using the given chunk.
1514
+ bool Setup(Address start, int size);
1515
+
1516
+ // Tears down the space. Heap memory was not allocated by the space, so it
1517
+ // is not deallocated here.
1518
+ void TearDown();
1519
+
1520
+ // True if the space has been set up but not torn down.
1521
+ bool HasBeenSetup() {
1522
+ return to_space_.HasBeenSetup() && from_space_.HasBeenSetup();
1523
+ }
1524
+
1525
+ // Flip the pair of spaces.
1526
+ void Flip();
1527
+
1528
+ // Grow the capacity of the semispaces. Assumes that they are not at
1529
+ // their maximum capacity.
1530
+ void Grow();
1531
+
1532
+ // Shrink the capacity of the semispaces.
1533
+ void Shrink();
1534
+
1535
+ // True if the address or object lies in the address range of either
1536
+ // semispace (not necessarily below the allocation pointer).
1537
+ bool Contains(Address a) {
1538
+ return (reinterpret_cast<uintptr_t>(a) & address_mask_)
1539
+ == reinterpret_cast<uintptr_t>(start_);
1540
+ }
1541
+ bool Contains(Object* o) {
1542
+ return (reinterpret_cast<uintptr_t>(o) & object_mask_) == object_expected_;
1543
+ }
1544
+
1545
+ // Return the allocated bytes in the active semispace.
1546
+ virtual intptr_t Size() { return static_cast<int>(top() - bottom()); }
1547
+ // The same, but returning an int. We have to have the one that returns
1548
+ // intptr_t because it is inherited, but if we know we are dealing with the
1549
+ // new space, which can't get as big as the other spaces then this is useful:
1550
+ int SizeAsInt() { return static_cast<int>(Size()); }
1551
+
1552
+ // Return the current capacity of a semispace.
1553
+ intptr_t Capacity() {
1554
+ ASSERT(to_space_.Capacity() == from_space_.Capacity());
1555
+ return to_space_.Capacity();
1556
+ }
1557
+
1558
+ // Return the total amount of memory committed for new space.
1559
+ intptr_t CommittedMemory() {
1560
+ if (from_space_.is_committed()) return 2 * Capacity();
1561
+ return Capacity();
1562
+ }
1563
+
1564
+ // Return the available bytes without growing in the active semispace.
1565
+ intptr_t Available() { return Capacity() - Size(); }
1566
+
1567
+ // Return the maximum capacity of a semispace.
1568
+ int MaximumCapacity() {
1569
+ ASSERT(to_space_.MaximumCapacity() == from_space_.MaximumCapacity());
1570
+ return to_space_.MaximumCapacity();
1571
+ }
1572
+
1573
+ // Returns the initial capacity of a semispace.
1574
+ int InitialCapacity() {
1575
+ ASSERT(to_space_.InitialCapacity() == from_space_.InitialCapacity());
1576
+ return to_space_.InitialCapacity();
1577
+ }
1578
+
1579
+ // Return the address of the allocation pointer in the active semispace.
1580
+ Address top() { return allocation_info_.top; }
1581
+ // Return the address of the first object in the active semispace.
1582
+ Address bottom() { return to_space_.low(); }
1583
+
1584
+ // Get the age mark of the inactive semispace.
1585
+ Address age_mark() { return from_space_.age_mark(); }
1586
+ // Set the age mark in the active semispace.
1587
+ void set_age_mark(Address mark) { to_space_.set_age_mark(mark); }
1588
+
1589
+ // The start address of the space and a bit mask. Anding an address in the
1590
+ // new space with the mask will result in the start address.
1591
+ Address start() { return start_; }
1592
+ uintptr_t mask() { return address_mask_; }
1593
+
1594
+ // The allocation top and limit addresses.
1595
+ Address* allocation_top_address() { return &allocation_info_.top; }
1596
+ Address* allocation_limit_address() { return &allocation_info_.limit; }
1597
+
1598
+ MUST_USE_RESULT MaybeObject* AllocateRaw(int size_in_bytes) {
1599
+ return AllocateRawInternal(size_in_bytes, &allocation_info_);
1600
+ }
1601
+
1602
+ // Allocate the requested number of bytes for relocation during mark-compact
1603
+ // collection.
1604
+ MUST_USE_RESULT MaybeObject* MCAllocateRaw(int size_in_bytes) {
1605
+ return AllocateRawInternal(size_in_bytes, &mc_forwarding_info_);
1606
+ }
1607
+
1608
+ // Reset the allocation pointer to the beginning of the active semispace.
1609
+ void ResetAllocationInfo();
1610
+ // Reset the reloction pointer to the bottom of the inactive semispace in
1611
+ // preparation for mark-compact collection.
1612
+ void MCResetRelocationInfo();
1613
+ // Update the allocation pointer in the active semispace after a
1614
+ // mark-compact collection.
1615
+ void MCCommitRelocationInfo();
1616
+
1617
+ // Get the extent of the inactive semispace (for use as a marking stack).
1618
+ Address FromSpaceLow() { return from_space_.low(); }
1619
+ Address FromSpaceHigh() { return from_space_.high(); }
1620
+
1621
+ // Get the extent of the active semispace (to sweep newly copied objects
1622
+ // during a scavenge collection).
1623
+ Address ToSpaceLow() { return to_space_.low(); }
1624
+ Address ToSpaceHigh() { return to_space_.high(); }
1625
+
1626
+ // Offsets from the beginning of the semispaces.
1627
+ int ToSpaceOffsetForAddress(Address a) {
1628
+ return to_space_.SpaceOffsetForAddress(a);
1629
+ }
1630
+ int FromSpaceOffsetForAddress(Address a) {
1631
+ return from_space_.SpaceOffsetForAddress(a);
1632
+ }
1633
+
1634
+ // True if the object is a heap object in the address range of the
1635
+ // respective semispace (not necessarily below the allocation pointer of the
1636
+ // semispace).
1637
+ bool ToSpaceContains(Object* o) { return to_space_.Contains(o); }
1638
+ bool FromSpaceContains(Object* o) { return from_space_.Contains(o); }
1639
+
1640
+ bool ToSpaceContains(Address a) { return to_space_.Contains(a); }
1641
+ bool FromSpaceContains(Address a) { return from_space_.Contains(a); }
1642
+
1643
+ virtual bool ReserveSpace(int bytes);
1644
+
1645
+ // Resizes a sequential string which must be the most recent thing that was
1646
+ // allocated in new space.
1647
+ template <typename StringType>
1648
+ inline void ShrinkStringAtAllocationBoundary(String* string, int len);
1649
+
1650
+ #ifdef ENABLE_HEAP_PROTECTION
1651
+ // Protect/unprotect the space by marking it read-only/writable.
1652
+ virtual void Protect();
1653
+ virtual void Unprotect();
1654
+ #endif
1655
+
1656
+ #ifdef DEBUG
1657
+ // Verify the active semispace.
1658
+ virtual void Verify();
1659
+ // Print the active semispace.
1660
+ virtual void Print() { to_space_.Print(); }
1661
+ #endif
1662
+
1663
+ #if defined(DEBUG) || defined(ENABLE_LOGGING_AND_PROFILING)
1664
+ // Iterates the active semispace to collect statistics.
1665
+ void CollectStatistics();
1666
+ // Reports previously collected statistics of the active semispace.
1667
+ void ReportStatistics();
1668
+ // Clears previously collected statistics.
1669
+ void ClearHistograms();
1670
+
1671
+ // Record the allocation or promotion of a heap object. Note that we don't
1672
+ // record every single allocation, but only those that happen in the
1673
+ // to space during a scavenge GC.
1674
+ void RecordAllocation(HeapObject* obj);
1675
+ void RecordPromotion(HeapObject* obj);
1676
+ #endif
1677
+
1678
+ // Return whether the operation succeded.
1679
+ bool CommitFromSpaceIfNeeded() {
1680
+ if (from_space_.is_committed()) return true;
1681
+ return from_space_.Commit();
1682
+ }
1683
+
1684
+ bool UncommitFromSpace() {
1685
+ if (!from_space_.is_committed()) return true;
1686
+ return from_space_.Uncommit();
1687
+ }
1688
+
1689
+ private:
1690
+ // The semispaces.
1691
+ SemiSpace to_space_;
1692
+ SemiSpace from_space_;
1693
+
1694
+ // Start address and bit mask for containment testing.
1695
+ Address start_;
1696
+ uintptr_t address_mask_;
1697
+ uintptr_t object_mask_;
1698
+ uintptr_t object_expected_;
1699
+
1700
+ // Allocation pointer and limit for normal allocation and allocation during
1701
+ // mark-compact collection.
1702
+ AllocationInfo allocation_info_;
1703
+ AllocationInfo mc_forwarding_info_;
1704
+
1705
+ #if defined(DEBUG) || defined(ENABLE_LOGGING_AND_PROFILING)
1706
+ HistogramInfo* allocated_histogram_;
1707
+ HistogramInfo* promoted_histogram_;
1708
+ #endif
1709
+
1710
+ // Implementation of AllocateRaw and MCAllocateRaw.
1711
+ MUST_USE_RESULT inline MaybeObject* AllocateRawInternal(
1712
+ int size_in_bytes,
1713
+ AllocationInfo* alloc_info);
1714
+
1715
+ friend class SemiSpaceIterator;
1716
+
1717
+ public:
1718
+ TRACK_MEMORY("NewSpace")
1719
+ };
1720
+
1721
+
1722
+ // -----------------------------------------------------------------------------
1723
+ // Free lists for old object spaces
1724
+ //
1725
+ // Free-list nodes are free blocks in the heap. They look like heap objects
1726
+ // (free-list node pointers have the heap object tag, and they have a map like
1727
+ // a heap object). They have a size and a next pointer. The next pointer is
1728
+ // the raw address of the next free list node (or NULL).
1729
+ class FreeListNode: public HeapObject {
1730
+ public:
1731
+ // Obtain a free-list node from a raw address. This is not a cast because
1732
+ // it does not check nor require that the first word at the address is a map
1733
+ // pointer.
1734
+ static FreeListNode* FromAddress(Address address) {
1735
+ return reinterpret_cast<FreeListNode*>(HeapObject::FromAddress(address));
1736
+ }
1737
+
1738
+ static inline bool IsFreeListNode(HeapObject* object);
1739
+
1740
+ // Set the size in bytes, which can be read with HeapObject::Size(). This
1741
+ // function also writes a map to the first word of the block so that it
1742
+ // looks like a heap object to the garbage collector and heap iteration
1743
+ // functions.
1744
+ void set_size(int size_in_bytes);
1745
+
1746
+ // Accessors for the next field.
1747
+ inline Address next();
1748
+ inline void set_next(Address next);
1749
+
1750
+ private:
1751
+ static const int kNextOffset = POINTER_SIZE_ALIGN(ByteArray::kHeaderSize);
1752
+
1753
+ DISALLOW_IMPLICIT_CONSTRUCTORS(FreeListNode);
1754
+ };
1755
+
1756
+
1757
+ // The free list for the old space.
1758
+ class OldSpaceFreeList BASE_EMBEDDED {
1759
+ public:
1760
+ explicit OldSpaceFreeList(AllocationSpace owner);
1761
+
1762
+ // Clear the free list.
1763
+ void Reset();
1764
+
1765
+ // Return the number of bytes available on the free list.
1766
+ intptr_t available() { return available_; }
1767
+
1768
+ // Place a node on the free list. The block of size 'size_in_bytes'
1769
+ // starting at 'start' is placed on the free list. The return value is the
1770
+ // number of bytes that have been lost due to internal fragmentation by
1771
+ // freeing the block. Bookkeeping information will be written to the block,
1772
+ // ie, its contents will be destroyed. The start address should be word
1773
+ // aligned, and the size should be a non-zero multiple of the word size.
1774
+ int Free(Address start, int size_in_bytes);
1775
+
1776
+ // Allocate a block of size 'size_in_bytes' from the free list. The block
1777
+ // is unitialized. A failure is returned if no block is available. The
1778
+ // number of bytes lost to fragmentation is returned in the output parameter
1779
+ // 'wasted_bytes'. The size should be a non-zero multiple of the word size.
1780
+ MUST_USE_RESULT MaybeObject* Allocate(int size_in_bytes, int* wasted_bytes);
1781
+
1782
+ void MarkNodes();
1783
+
1784
+ private:
1785
+ // The size range of blocks, in bytes. (Smaller allocations are allowed, but
1786
+ // will always result in waste.)
1787
+ static const int kMinBlockSize = 2 * kPointerSize;
1788
+ static const int kMaxBlockSize = Page::kMaxHeapObjectSize;
1789
+
1790
+ // The identity of the owning space, for building allocation Failure
1791
+ // objects.
1792
+ AllocationSpace owner_;
1793
+
1794
+ // Total available bytes in all blocks on this free list.
1795
+ int available_;
1796
+
1797
+ // Blocks are put on exact free lists in an array, indexed by size in words.
1798
+ // The available sizes are kept in an increasingly ordered list. Entries
1799
+ // corresponding to sizes < kMinBlockSize always have an empty free list
1800
+ // (but index kHead is used for the head of the size list).
1801
+ struct SizeNode {
1802
+ // Address of the head FreeListNode of the implied block size or NULL.
1803
+ Address head_node_;
1804
+ // Size (words) of the next larger available size if head_node_ != NULL.
1805
+ int next_size_;
1806
+ };
1807
+ static const int kFreeListsLength = kMaxBlockSize / kPointerSize + 1;
1808
+ SizeNode free_[kFreeListsLength];
1809
+
1810
+ // Sentinel elements for the size list. Real elements are in ]kHead..kEnd[.
1811
+ static const int kHead = kMinBlockSize / kPointerSize - 1;
1812
+ static const int kEnd = kMaxInt;
1813
+
1814
+ // We keep a "finger" in the size list to speed up a common pattern:
1815
+ // repeated requests for the same or increasing sizes.
1816
+ int finger_;
1817
+
1818
+ // Starting from *prev, find and return the smallest size >= index (words),
1819
+ // or kEnd. Update *prev to be the largest size < index, or kHead.
1820
+ int FindSize(int index, int* prev) {
1821
+ int cur = free_[*prev].next_size_;
1822
+ while (cur < index) {
1823
+ *prev = cur;
1824
+ cur = free_[cur].next_size_;
1825
+ }
1826
+ return cur;
1827
+ }
1828
+
1829
+ // Remove an existing element from the size list.
1830
+ void RemoveSize(int index) {
1831
+ int prev = kHead;
1832
+ int cur = FindSize(index, &prev);
1833
+ ASSERT(cur == index);
1834
+ free_[prev].next_size_ = free_[cur].next_size_;
1835
+ finger_ = prev;
1836
+ }
1837
+
1838
+ // Insert a new element into the size list.
1839
+ void InsertSize(int index) {
1840
+ int prev = kHead;
1841
+ int cur = FindSize(index, &prev);
1842
+ ASSERT(cur != index);
1843
+ free_[prev].next_size_ = index;
1844
+ free_[index].next_size_ = cur;
1845
+ }
1846
+
1847
+ // The size list is not updated during a sequence of calls to Free, but is
1848
+ // rebuilt before the next allocation.
1849
+ void RebuildSizeList();
1850
+ bool needs_rebuild_;
1851
+
1852
+ #ifdef DEBUG
1853
+ // Does this free list contain a free block located at the address of 'node'?
1854
+ bool Contains(FreeListNode* node);
1855
+ #endif
1856
+
1857
+ DISALLOW_COPY_AND_ASSIGN(OldSpaceFreeList);
1858
+ };
1859
+
1860
+
1861
+ // The free list for the map space.
1862
+ class FixedSizeFreeList BASE_EMBEDDED {
1863
+ public:
1864
+ FixedSizeFreeList(AllocationSpace owner, int object_size);
1865
+
1866
+ // Clear the free list.
1867
+ void Reset();
1868
+
1869
+ // Return the number of bytes available on the free list.
1870
+ intptr_t available() { return available_; }
1871
+
1872
+ // Place a node on the free list. The block starting at 'start' (assumed to
1873
+ // have size object_size_) is placed on the free list. Bookkeeping
1874
+ // information will be written to the block, ie, its contents will be
1875
+ // destroyed. The start address should be word aligned.
1876
+ void Free(Address start);
1877
+
1878
+ // Allocate a fixed sized block from the free list. The block is unitialized.
1879
+ // A failure is returned if no block is available.
1880
+ MUST_USE_RESULT MaybeObject* Allocate();
1881
+
1882
+ void MarkNodes();
1883
+
1884
+ private:
1885
+ // Available bytes on the free list.
1886
+ intptr_t available_;
1887
+
1888
+ // The head of the free list.
1889
+ Address head_;
1890
+
1891
+ // The tail of the free list.
1892
+ Address tail_;
1893
+
1894
+ // The identity of the owning space, for building allocation Failure
1895
+ // objects.
1896
+ AllocationSpace owner_;
1897
+
1898
+ // The size of the objects in this space.
1899
+ int object_size_;
1900
+
1901
+ DISALLOW_COPY_AND_ASSIGN(FixedSizeFreeList);
1902
+ };
1903
+
1904
+
1905
+ // -----------------------------------------------------------------------------
1906
+ // Old object space (excluding map objects)
1907
+
1908
+ class OldSpace : public PagedSpace {
1909
+ public:
1910
+ // Creates an old space object with a given maximum capacity.
1911
+ // The constructor does not allocate pages from OS.
1912
+ explicit OldSpace(intptr_t max_capacity,
1913
+ AllocationSpace id,
1914
+ Executability executable)
1915
+ : PagedSpace(max_capacity, id, executable), free_list_(id) {
1916
+ page_extra_ = 0;
1917
+ }
1918
+
1919
+ // The bytes available on the free list (ie, not above the linear allocation
1920
+ // pointer).
1921
+ intptr_t AvailableFree() { return free_list_.available(); }
1922
+
1923
+ // The limit of allocation for a page in this space.
1924
+ virtual Address PageAllocationLimit(Page* page) {
1925
+ return page->ObjectAreaEnd();
1926
+ }
1927
+
1928
+ // Give a block of memory to the space's free list. It might be added to
1929
+ // the free list or accounted as waste.
1930
+ // If add_to_freelist is false then just accounting stats are updated and
1931
+ // no attempt to add area to free list is made.
1932
+ void Free(Address start, int size_in_bytes, bool add_to_freelist) {
1933
+ accounting_stats_.DeallocateBytes(size_in_bytes);
1934
+
1935
+ if (add_to_freelist) {
1936
+ int wasted_bytes = free_list_.Free(start, size_in_bytes);
1937
+ accounting_stats_.WasteBytes(wasted_bytes);
1938
+ }
1939
+ }
1940
+
1941
+ virtual void DeallocateBlock(Address start,
1942
+ int size_in_bytes,
1943
+ bool add_to_freelist);
1944
+
1945
+ // Prepare for full garbage collection. Resets the relocation pointer and
1946
+ // clears the free list.
1947
+ virtual void PrepareForMarkCompact(bool will_compact);
1948
+
1949
+ // Updates the allocation pointer to the relocation top after a mark-compact
1950
+ // collection.
1951
+ virtual void MCCommitRelocationInfo();
1952
+
1953
+ virtual void PutRestOfCurrentPageOnFreeList(Page* current_page);
1954
+
1955
+ void MarkFreeListNodes() { free_list_.MarkNodes(); }
1956
+
1957
+ #ifdef DEBUG
1958
+ // Reports statistics for the space
1959
+ void ReportStatistics();
1960
+ #endif
1961
+
1962
+ protected:
1963
+ // Virtual function in the superclass. Slow path of AllocateRaw.
1964
+ MUST_USE_RESULT HeapObject* SlowAllocateRaw(int size_in_bytes);
1965
+
1966
+ // Virtual function in the superclass. Allocate linearly at the start of
1967
+ // the page after current_page (there is assumed to be one).
1968
+ HeapObject* AllocateInNextPage(Page* current_page, int size_in_bytes);
1969
+
1970
+ private:
1971
+ // The space's free list.
1972
+ OldSpaceFreeList free_list_;
1973
+
1974
+ public:
1975
+ TRACK_MEMORY("OldSpace")
1976
+ };
1977
+
1978
+
1979
+ // -----------------------------------------------------------------------------
1980
+ // Old space for objects of a fixed size
1981
+
1982
+ class FixedSpace : public PagedSpace {
1983
+ public:
1984
+ FixedSpace(intptr_t max_capacity,
1985
+ AllocationSpace id,
1986
+ int object_size_in_bytes,
1987
+ const char* name)
1988
+ : PagedSpace(max_capacity, id, NOT_EXECUTABLE),
1989
+ object_size_in_bytes_(object_size_in_bytes),
1990
+ name_(name),
1991
+ free_list_(id, object_size_in_bytes) {
1992
+ page_extra_ = Page::kObjectAreaSize % object_size_in_bytes;
1993
+ }
1994
+
1995
+ // The limit of allocation for a page in this space.
1996
+ virtual Address PageAllocationLimit(Page* page) {
1997
+ return page->ObjectAreaEnd() - page_extra_;
1998
+ }
1999
+
2000
+ int object_size_in_bytes() { return object_size_in_bytes_; }
2001
+
2002
+ // Give a fixed sized block of memory to the space's free list.
2003
+ // If add_to_freelist is false then just accounting stats are updated and
2004
+ // no attempt to add area to free list is made.
2005
+ void Free(Address start, bool add_to_freelist) {
2006
+ if (add_to_freelist) {
2007
+ free_list_.Free(start);
2008
+ }
2009
+ accounting_stats_.DeallocateBytes(object_size_in_bytes_);
2010
+ }
2011
+
2012
+ // Prepares for a mark-compact GC.
2013
+ virtual void PrepareForMarkCompact(bool will_compact);
2014
+
2015
+ // Updates the allocation pointer to the relocation top after a mark-compact
2016
+ // collection.
2017
+ virtual void MCCommitRelocationInfo();
2018
+
2019
+ virtual void PutRestOfCurrentPageOnFreeList(Page* current_page);
2020
+
2021
+ virtual void DeallocateBlock(Address start,
2022
+ int size_in_bytes,
2023
+ bool add_to_freelist);
2024
+
2025
+ void MarkFreeListNodes() { free_list_.MarkNodes(); }
2026
+
2027
+ #ifdef DEBUG
2028
+ // Reports statistic info of the space
2029
+ void ReportStatistics();
2030
+ #endif
2031
+
2032
+ protected:
2033
+ // Virtual function in the superclass. Slow path of AllocateRaw.
2034
+ MUST_USE_RESULT HeapObject* SlowAllocateRaw(int size_in_bytes);
2035
+
2036
+ // Virtual function in the superclass. Allocate linearly at the start of
2037
+ // the page after current_page (there is assumed to be one).
2038
+ HeapObject* AllocateInNextPage(Page* current_page, int size_in_bytes);
2039
+
2040
+ void ResetFreeList() {
2041
+ free_list_.Reset();
2042
+ }
2043
+
2044
+ private:
2045
+ // The size of objects in this space.
2046
+ int object_size_in_bytes_;
2047
+
2048
+ // The name of this space.
2049
+ const char* name_;
2050
+
2051
+ // The space's free list.
2052
+ FixedSizeFreeList free_list_;
2053
+ };
2054
+
2055
+
2056
+ // -----------------------------------------------------------------------------
2057
+ // Old space for all map objects
2058
+
2059
+ class MapSpace : public FixedSpace {
2060
+ public:
2061
+ // Creates a map space object with a maximum capacity.
2062
+ MapSpace(intptr_t max_capacity, int max_map_space_pages, AllocationSpace id)
2063
+ : FixedSpace(max_capacity, id, Map::kSize, "map"),
2064
+ max_map_space_pages_(max_map_space_pages) {
2065
+ ASSERT(max_map_space_pages < kMaxMapPageIndex);
2066
+ }
2067
+
2068
+ // Prepares for a mark-compact GC.
2069
+ virtual void PrepareForMarkCompact(bool will_compact);
2070
+
2071
+ // Given an index, returns the page address.
2072
+ Address PageAddress(int page_index) { return page_addresses_[page_index]; }
2073
+
2074
+ static const int kMaxMapPageIndex = 1 << MapWord::kMapPageIndexBits;
2075
+
2076
+ // Are map pointers encodable into map word?
2077
+ bool MapPointersEncodable() {
2078
+ if (!FLAG_use_big_map_space) {
2079
+ ASSERT(CountPagesToTop() <= kMaxMapPageIndex);
2080
+ return true;
2081
+ }
2082
+ return CountPagesToTop() <= max_map_space_pages_;
2083
+ }
2084
+
2085
+ // Should be called after forced sweep to find out if map space needs
2086
+ // compaction.
2087
+ bool NeedsCompaction(int live_maps) {
2088
+ return !MapPointersEncodable() && live_maps <= CompactionThreshold();
2089
+ }
2090
+
2091
+ Address TopAfterCompaction(int live_maps) {
2092
+ ASSERT(NeedsCompaction(live_maps));
2093
+
2094
+ int pages_left = live_maps / kMapsPerPage;
2095
+ PageIterator it(this, PageIterator::ALL_PAGES);
2096
+ while (pages_left-- > 0) {
2097
+ ASSERT(it.has_next());
2098
+ it.next()->SetRegionMarks(Page::kAllRegionsCleanMarks);
2099
+ }
2100
+ ASSERT(it.has_next());
2101
+ Page* top_page = it.next();
2102
+ top_page->SetRegionMarks(Page::kAllRegionsCleanMarks);
2103
+ ASSERT(top_page->is_valid());
2104
+
2105
+ int offset = live_maps % kMapsPerPage * Map::kSize;
2106
+ Address top = top_page->ObjectAreaStart() + offset;
2107
+ ASSERT(top < top_page->ObjectAreaEnd());
2108
+ ASSERT(Contains(top));
2109
+
2110
+ return top;
2111
+ }
2112
+
2113
+ void FinishCompaction(Address new_top, int live_maps) {
2114
+ Page* top_page = Page::FromAddress(new_top);
2115
+ ASSERT(top_page->is_valid());
2116
+
2117
+ SetAllocationInfo(&allocation_info_, top_page);
2118
+ allocation_info_.top = new_top;
2119
+
2120
+ int new_size = live_maps * Map::kSize;
2121
+ accounting_stats_.DeallocateBytes(accounting_stats_.Size());
2122
+ accounting_stats_.AllocateBytes(new_size);
2123
+
2124
+ // Flush allocation watermarks.
2125
+ for (Page* p = first_page_; p != top_page; p = p->next_page()) {
2126
+ p->SetAllocationWatermark(p->AllocationTop());
2127
+ }
2128
+ top_page->SetAllocationWatermark(new_top);
2129
+
2130
+ #ifdef DEBUG
2131
+ if (FLAG_enable_slow_asserts) {
2132
+ intptr_t actual_size = 0;
2133
+ for (Page* p = first_page_; p != top_page; p = p->next_page())
2134
+ actual_size += kMapsPerPage * Map::kSize;
2135
+ actual_size += (new_top - top_page->ObjectAreaStart());
2136
+ ASSERT(accounting_stats_.Size() == actual_size);
2137
+ }
2138
+ #endif
2139
+
2140
+ Shrink();
2141
+ ResetFreeList();
2142
+ }
2143
+
2144
+ protected:
2145
+ #ifdef DEBUG
2146
+ virtual void VerifyObject(HeapObject* obj);
2147
+ #endif
2148
+
2149
+ private:
2150
+ static const int kMapsPerPage = Page::kObjectAreaSize / Map::kSize;
2151
+
2152
+ // Do map space compaction if there is a page gap.
2153
+ int CompactionThreshold() {
2154
+ return kMapsPerPage * (max_map_space_pages_ - 1);
2155
+ }
2156
+
2157
+ const int max_map_space_pages_;
2158
+
2159
+ // An array of page start address in a map space.
2160
+ Address page_addresses_[kMaxMapPageIndex];
2161
+
2162
+ public:
2163
+ TRACK_MEMORY("MapSpace")
2164
+ };
2165
+
2166
+
2167
+ // -----------------------------------------------------------------------------
2168
+ // Old space for all global object property cell objects
2169
+
2170
+ class CellSpace : public FixedSpace {
2171
+ public:
2172
+ // Creates a property cell space object with a maximum capacity.
2173
+ CellSpace(intptr_t max_capacity, AllocationSpace id)
2174
+ : FixedSpace(max_capacity, id, JSGlobalPropertyCell::kSize, "cell") {}
2175
+
2176
+ protected:
2177
+ #ifdef DEBUG
2178
+ virtual void VerifyObject(HeapObject* obj);
2179
+ #endif
2180
+
2181
+ public:
2182
+ TRACK_MEMORY("CellSpace")
2183
+ };
2184
+
2185
+
2186
+ // -----------------------------------------------------------------------------
2187
+ // Large objects ( > Page::kMaxHeapObjectSize ) are allocated and managed by
2188
+ // the large object space. A large object is allocated from OS heap with
2189
+ // extra padding bytes (Page::kPageSize + Page::kObjectStartOffset).
2190
+ // A large object always starts at Page::kObjectStartOffset to a page.
2191
+ // Large objects do not move during garbage collections.
2192
+
2193
+ // A LargeObjectChunk holds exactly one large object page with exactly one
2194
+ // large object.
2195
+ class LargeObjectChunk {
2196
+ public:
2197
+ // Allocates a new LargeObjectChunk that contains a large object page
2198
+ // (Page::kPageSize aligned) that has at least size_in_bytes (for a large
2199
+ // object) bytes after the object area start of that page.
2200
+ static LargeObjectChunk* New(int size_in_bytes, Executability executable);
2201
+
2202
+ // Free the memory associated with the chunk.
2203
+ inline void Free(Executability executable);
2204
+
2205
+ // Interpret a raw address as a large object chunk.
2206
+ static LargeObjectChunk* FromAddress(Address address) {
2207
+ return reinterpret_cast<LargeObjectChunk*>(address);
2208
+ }
2209
+
2210
+ // Returns the address of this chunk.
2211
+ Address address() { return reinterpret_cast<Address>(this); }
2212
+
2213
+ // Accessors for the fields of the chunk.
2214
+ LargeObjectChunk* next() { return next_; }
2215
+ void set_next(LargeObjectChunk* chunk) { next_ = chunk; }
2216
+ size_t size() { return size_ & ~Page::kPageFlagMask; }
2217
+
2218
+ // Compute the start address in the chunk.
2219
+ inline Address GetStartAddress();
2220
+
2221
+ // Returns the object in this chunk.
2222
+ HeapObject* GetObject() { return HeapObject::FromAddress(GetStartAddress()); }
2223
+
2224
+ // Given a requested size returns the physical size of a chunk to be
2225
+ // allocated.
2226
+ static int ChunkSizeFor(int size_in_bytes);
2227
+
2228
+ // Given a chunk size, returns the object size it can accommodate. Used by
2229
+ // LargeObjectSpace::Available.
2230
+ static intptr_t ObjectSizeFor(intptr_t chunk_size) {
2231
+ if (chunk_size <= (Page::kPageSize + Page::kObjectStartOffset)) return 0;
2232
+ return chunk_size - Page::kPageSize - Page::kObjectStartOffset;
2233
+ }
2234
+
2235
+ private:
2236
+ // A pointer to the next large object chunk in the space or NULL.
2237
+ LargeObjectChunk* next_;
2238
+
2239
+ // The total size of this chunk.
2240
+ size_t size_;
2241
+
2242
+ public:
2243
+ TRACK_MEMORY("LargeObjectChunk")
2244
+ };
2245
+
2246
+
2247
+ class LargeObjectSpace : public Space {
2248
+ public:
2249
+ explicit LargeObjectSpace(AllocationSpace id);
2250
+ virtual ~LargeObjectSpace() {}
2251
+
2252
+ // Initializes internal data structures.
2253
+ bool Setup();
2254
+
2255
+ // Releases internal resources, frees objects in this space.
2256
+ void TearDown();
2257
+
2258
+ // Allocates a (non-FixedArray, non-Code) large object.
2259
+ MUST_USE_RESULT MaybeObject* AllocateRaw(int size_in_bytes);
2260
+ // Allocates a large Code object.
2261
+ MUST_USE_RESULT MaybeObject* AllocateRawCode(int size_in_bytes);
2262
+ // Allocates a large FixedArray.
2263
+ MUST_USE_RESULT MaybeObject* AllocateRawFixedArray(int size_in_bytes);
2264
+
2265
+ // Available bytes for objects in this space.
2266
+ intptr_t Available() {
2267
+ return LargeObjectChunk::ObjectSizeFor(MemoryAllocator::Available());
2268
+ }
2269
+
2270
+ virtual intptr_t Size() {
2271
+ return size_;
2272
+ }
2273
+
2274
+ virtual intptr_t SizeOfObjects() {
2275
+ return objects_size_;
2276
+ }
2277
+
2278
+ int PageCount() {
2279
+ return page_count_;
2280
+ }
2281
+
2282
+ // Finds an object for a given address, returns Failure::Exception()
2283
+ // if it is not found. The function iterates through all objects in this
2284
+ // space, may be slow.
2285
+ MaybeObject* FindObject(Address a);
2286
+
2287
+ // Finds a large object page containing the given pc, returns NULL
2288
+ // if such a page doesn't exist.
2289
+ LargeObjectChunk* FindChunkContainingPc(Address pc);
2290
+
2291
+ // Iterates objects covered by dirty regions.
2292
+ void IterateDirtyRegions(ObjectSlotCallback func);
2293
+
2294
+ // Frees unmarked objects.
2295
+ void FreeUnmarkedObjects();
2296
+
2297
+ // Checks whether a heap object is in this space; O(1).
2298
+ bool Contains(HeapObject* obj);
2299
+
2300
+ // Checks whether the space is empty.
2301
+ bool IsEmpty() { return first_chunk_ == NULL; }
2302
+
2303
+ // See the comments for ReserveSpace in the Space class. This has to be
2304
+ // called after ReserveSpace has been called on the paged spaces, since they
2305
+ // may use some memory, leaving less for large objects.
2306
+ virtual bool ReserveSpace(int bytes);
2307
+
2308
+ #ifdef ENABLE_HEAP_PROTECTION
2309
+ // Protect/unprotect the space by marking it read-only/writable.
2310
+ void Protect();
2311
+ void Unprotect();
2312
+ #endif
2313
+
2314
+ #ifdef DEBUG
2315
+ virtual void Verify();
2316
+ virtual void Print();
2317
+ void ReportStatistics();
2318
+ void CollectCodeStatistics();
2319
+ #endif
2320
+ // Checks whether an address is in the object area in this space. It
2321
+ // iterates all objects in the space. May be slow.
2322
+ bool SlowContains(Address addr) { return !FindObject(addr)->IsFailure(); }
2323
+
2324
+ private:
2325
+ // The head of the linked list of large object chunks.
2326
+ LargeObjectChunk* first_chunk_;
2327
+ intptr_t size_; // allocated bytes
2328
+ int page_count_; // number of chunks
2329
+ intptr_t objects_size_; // size of objects
2330
+
2331
+ // Shared implementation of AllocateRaw, AllocateRawCode and
2332
+ // AllocateRawFixedArray.
2333
+ MUST_USE_RESULT MaybeObject* AllocateRawInternal(int requested_size,
2334
+ int object_size,
2335
+ Executability executable);
2336
+
2337
+ friend class LargeObjectIterator;
2338
+
2339
+ public:
2340
+ TRACK_MEMORY("LargeObjectSpace")
2341
+ };
2342
+
2343
+
2344
+ class LargeObjectIterator: public ObjectIterator {
2345
+ public:
2346
+ explicit LargeObjectIterator(LargeObjectSpace* space);
2347
+ LargeObjectIterator(LargeObjectSpace* space, HeapObjectCallback size_func);
2348
+
2349
+ HeapObject* next();
2350
+
2351
+ // implementation of ObjectIterator.
2352
+ virtual HeapObject* next_object() { return next(); }
2353
+
2354
+ private:
2355
+ LargeObjectChunk* current_;
2356
+ HeapObjectCallback size_func_;
2357
+ };
2358
+
2359
+
2360
+ } } // namespace v8::internal
2361
+
2362
+ #endif // V8_SPACES_H_