libv8-node 20.12.1.0-aarch64-linux → 22.5.1.0-aarch64-linux
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/lib/libv8/node/version.rb +3 -3
- data/vendor/v8/aarch64-linux/libv8/obj/libv8_monolith.a +0 -0
- data/vendor/v8/include/cppgc/internal/api-constants.h +24 -5
- data/vendor/v8/include/cppgc/internal/caged-heap-local-data.h +16 -6
- data/vendor/v8/include/cppgc/internal/caged-heap.h +12 -5
- data/vendor/v8/include/cppgc/internal/gc-info.h +82 -91
- data/vendor/v8/include/cppgc/internal/member-storage.h +16 -8
- data/vendor/v8/include/cppgc/member.h +25 -0
- data/vendor/v8/include/cppgc/persistent.h +4 -0
- data/vendor/v8/include/cppgc/platform.h +6 -1
- data/vendor/v8/include/cppgc/sentinel-pointer.h +7 -0
- data/vendor/v8/include/cppgc/source-location.h +2 -78
- data/vendor/v8/include/cppgc/trace-trait.h +8 -0
- data/vendor/v8/include/cppgc/type-traits.h +25 -4
- data/vendor/v8/include/cppgc/visitor.h +82 -4
- data/vendor/v8/include/libplatform/libplatform.h +7 -1
- data/vendor/v8/include/v8-array-buffer.h +6 -0
- data/vendor/v8/include/v8-callbacks.h +57 -19
- data/vendor/v8/include/v8-container.h +54 -0
- data/vendor/v8/include/v8-context.h +58 -32
- data/vendor/v8/include/v8-embedder-heap.h +31 -3
- data/vendor/v8/include/v8-embedder-state-scope.h +2 -1
- data/vendor/v8/include/v8-exception.h +15 -9
- data/vendor/v8/include/v8-fast-api-calls.h +58 -31
- data/vendor/v8/include/v8-forward.h +1 -0
- data/vendor/v8/include/v8-function-callback.h +135 -30
- data/vendor/v8/include/v8-function.h +6 -0
- data/vendor/v8/include/v8-handle-base.h +137 -0
- data/vendor/v8/include/v8-inspector.h +35 -13
- data/vendor/v8/include/v8-internal.h +510 -71
- data/vendor/v8/include/v8-isolate.h +176 -100
- data/vendor/v8/include/v8-local-handle.h +383 -112
- data/vendor/v8/include/v8-memory-span.h +157 -2
- data/vendor/v8/include/v8-message.h +22 -3
- data/vendor/v8/include/v8-metrics.h +1 -0
- data/vendor/v8/include/v8-object.h +98 -77
- data/vendor/v8/include/v8-persistent-handle.h +68 -90
- data/vendor/v8/include/v8-platform.h +191 -23
- data/vendor/v8/include/v8-primitive.h +12 -8
- data/vendor/v8/include/v8-profiler.h +16 -2
- data/vendor/v8/include/v8-script.h +88 -14
- data/vendor/v8/include/v8-snapshot.h +96 -22
- data/vendor/v8/include/v8-source-location.h +92 -0
- data/vendor/v8/include/v8-statistics.h +31 -10
- data/vendor/v8/include/v8-template.h +410 -131
- data/vendor/v8/include/v8-traced-handle.h +108 -90
- data/vendor/v8/include/v8-typed-array.h +115 -7
- data/vendor/v8/include/v8-unwinder.h +1 -1
- data/vendor/v8/include/v8-util.h +23 -20
- data/vendor/v8/include/v8-value-serializer.h +14 -0
- data/vendor/v8/include/v8-value.h +105 -3
- data/vendor/v8/include/v8-version.h +4 -4
- data/vendor/v8/include/v8config.h +54 -20
- metadata +4 -2
@@ -10,10 +10,11 @@
|
|
10
10
|
#include <string.h>
|
11
11
|
|
12
12
|
#include <atomic>
|
13
|
+
#include <iterator>
|
14
|
+
#include <memory>
|
13
15
|
#include <type_traits>
|
14
16
|
|
15
|
-
#include "
|
16
|
-
#include "v8config.h" // NOLINT(build/include_directory)
|
17
|
+
#include "v8config.h" // NOLINT(build/include_directory)
|
17
18
|
|
18
19
|
namespace v8 {
|
19
20
|
|
@@ -24,6 +25,7 @@ class Isolate;
|
|
24
25
|
|
25
26
|
namespace internal {
|
26
27
|
|
28
|
+
class Heap;
|
27
29
|
class Isolate;
|
28
30
|
|
29
31
|
typedef uintptr_t Address;
|
@@ -80,7 +82,7 @@ struct SmiTagging<4> {
|
|
80
82
|
static_cast<intptr_t>(kUintptrAllBitsSet << (kSmiValueSize - 1));
|
81
83
|
static constexpr intptr_t kSmiMaxValue = -(kSmiMinValue + 1);
|
82
84
|
|
83
|
-
V8_INLINE static int SmiToInt(Address value) {
|
85
|
+
V8_INLINE static constexpr int SmiToInt(Address value) {
|
84
86
|
int shift_bits = kSmiTagSize + kSmiShiftSize;
|
85
87
|
// Truncate and shift down (requires >> to be sign extending).
|
86
88
|
return static_cast<int32_t>(static_cast<uint32_t>(value)) >> shift_bits;
|
@@ -105,7 +107,7 @@ struct SmiTagging<8> {
|
|
105
107
|
static_cast<intptr_t>(kUintptrAllBitsSet << (kSmiValueSize - 1));
|
106
108
|
static constexpr intptr_t kSmiMaxValue = -(kSmiMinValue + 1);
|
107
109
|
|
108
|
-
V8_INLINE static int SmiToInt(Address value) {
|
110
|
+
V8_INLINE static constexpr int SmiToInt(Address value) {
|
109
111
|
int shift_bits = kSmiTagSize + kSmiShiftSize;
|
110
112
|
// Shift down and throw away top 32 bits.
|
111
113
|
return static_cast<int>(static_cast<intptr_t>(value) >> shift_bits);
|
@@ -173,11 +175,15 @@ using SandboxedPointer_t = Address;
|
|
173
175
|
#ifdef V8_ENABLE_SANDBOX
|
174
176
|
|
175
177
|
// Size of the sandbox, excluding the guard regions surrounding it.
|
176
|
-
#
|
178
|
+
#if defined(V8_TARGET_OS_ANDROID)
|
177
179
|
// On Android, most 64-bit devices seem to be configured with only 39 bits of
|
178
180
|
// virtual address space for userspace. As such, limit the sandbox to 128GB (a
|
179
181
|
// quarter of the total available address space).
|
180
182
|
constexpr size_t kSandboxSizeLog2 = 37; // 128 GB
|
183
|
+
#elif defined(V8_TARGET_ARCH_LOONG64)
|
184
|
+
// Some Linux distros on LoongArch64 configured with only 40 bits of virtual
|
185
|
+
// address space for userspace. Limit the sandbox to 256GB here.
|
186
|
+
constexpr size_t kSandboxSizeLog2 = 38; // 256 GB
|
181
187
|
#else
|
182
188
|
// Everywhere else use a 1TB sandbox.
|
183
189
|
constexpr size_t kSandboxSizeLog2 = 40; // 1 TB
|
@@ -247,20 +253,22 @@ static_assert(1ULL << (64 - kBoundedSizeShift) ==
|
|
247
253
|
// size allows omitting bounds checks on table accesses if the indices are
|
248
254
|
// guaranteed (e.g. through shifting) to be below the maximum index. This
|
249
255
|
// value must be a power of two.
|
250
|
-
|
256
|
+
constexpr size_t kExternalPointerTableReservationSize = 512 * MB;
|
251
257
|
|
252
258
|
// The external pointer table indices stored in HeapObjects as external
|
253
259
|
// pointers are shifted to the left by this amount to guarantee that they are
|
254
260
|
// smaller than the maximum table size.
|
255
|
-
|
261
|
+
constexpr uint32_t kExternalPointerIndexShift = 6;
|
256
262
|
#else
|
257
|
-
|
258
|
-
|
263
|
+
constexpr size_t kExternalPointerTableReservationSize = 1024 * MB;
|
264
|
+
constexpr uint32_t kExternalPointerIndexShift = 5;
|
259
265
|
#endif // V8_TARGET_OS_ANDROID
|
260
266
|
|
261
267
|
// The maximum number of entries in an external pointer table.
|
262
|
-
|
263
|
-
|
268
|
+
constexpr int kExternalPointerTableEntrySize = 8;
|
269
|
+
constexpr int kExternalPointerTableEntrySizeLog2 = 3;
|
270
|
+
constexpr size_t kMaxExternalPointers =
|
271
|
+
kExternalPointerTableReservationSize / kExternalPointerTableEntrySize;
|
264
272
|
static_assert((1 << (32 - kExternalPointerIndexShift)) == kMaxExternalPointers,
|
265
273
|
"kExternalPointerTableReservationSize and "
|
266
274
|
"kExternalPointerIndexShift don't match");
|
@@ -268,7 +276,7 @@ static_assert((1 << (32 - kExternalPointerIndexShift)) == kMaxExternalPointers,
|
|
268
276
|
#else // !V8_COMPRESS_POINTERS
|
269
277
|
|
270
278
|
// Needed for the V8.SandboxedExternalPointersCount histogram.
|
271
|
-
|
279
|
+
constexpr size_t kMaxExternalPointers = 0;
|
272
280
|
|
273
281
|
#endif // V8_COMPRESS_POINTERS
|
274
282
|
|
@@ -281,15 +289,21 @@ static const size_t kMaxExternalPointers = 0;
|
|
281
289
|
// that it is smaller than the size of the table.
|
282
290
|
using ExternalPointerHandle = uint32_t;
|
283
291
|
|
284
|
-
// ExternalPointers point to objects located outside the sandbox. When
|
285
|
-
//
|
286
|
-
//
|
292
|
+
// ExternalPointers point to objects located outside the sandbox. When the V8
|
293
|
+
// sandbox is enabled, these are stored on heap as ExternalPointerHandles,
|
294
|
+
// otherwise they are simply raw pointers.
|
287
295
|
#ifdef V8_ENABLE_SANDBOX
|
288
296
|
using ExternalPointer_t = ExternalPointerHandle;
|
289
297
|
#else
|
290
298
|
using ExternalPointer_t = Address;
|
291
299
|
#endif
|
292
300
|
|
301
|
+
constexpr ExternalPointer_t kNullExternalPointer = 0;
|
302
|
+
constexpr ExternalPointerHandle kNullExternalPointerHandle = 0;
|
303
|
+
|
304
|
+
//
|
305
|
+
// External Pointers.
|
306
|
+
//
|
293
307
|
// When the sandbox is enabled, external pointers are stored in an external
|
294
308
|
// pointer table and are referenced from HeapObjects through an index (a
|
295
309
|
// "handle"). When stored in the table, the pointers are tagged with per-type
|
@@ -359,6 +373,7 @@ using ExternalPointer_t = Address;
|
|
359
373
|
// ExternalPointerTable.
|
360
374
|
constexpr uint64_t kExternalPointerMarkBit = 1ULL << 62;
|
361
375
|
constexpr uint64_t kExternalPointerTagMask = 0x40ff000000000000;
|
376
|
+
constexpr uint64_t kExternalPointerTagMaskWithoutMarkBit = 0xff000000000000;
|
362
377
|
constexpr uint64_t kExternalPointerTagShift = 48;
|
363
378
|
|
364
379
|
// All possible 8-bit type tags.
|
@@ -410,14 +425,15 @@ constexpr uint64_t kAllExternalPointerTypeTags[] = {
|
|
410
425
|
/* it is the Embedder's responsibility to ensure type safety (against */ \
|
411
426
|
/* substitution) and lifetime validity of these objects. */ \
|
412
427
|
V(kExternalObjectValueTag, TAG(13)) \
|
413
|
-
V(
|
428
|
+
V(kFunctionTemplateInfoCallbackTag, TAG(14)) \
|
414
429
|
V(kAccessorInfoGetterTag, TAG(15)) \
|
415
430
|
V(kAccessorInfoSetterTag, TAG(16)) \
|
416
431
|
V(kWasmInternalFunctionCallTargetTag, TAG(17)) \
|
417
432
|
V(kWasmTypeInfoNativeTypeTag, TAG(18)) \
|
418
433
|
V(kWasmExportedFunctionDataSignatureTag, TAG(19)) \
|
419
434
|
V(kWasmContinuationJmpbufTag, TAG(20)) \
|
420
|
-
V(
|
435
|
+
V(kWasmIndirectFunctionTargetTag, TAG(21)) \
|
436
|
+
V(kArrayBufferExtensionTag, TAG(22))
|
421
437
|
|
422
438
|
// All external pointer tags.
|
423
439
|
#define ALL_EXTERNAL_POINTER_TAGS(V) \
|
@@ -430,7 +446,7 @@ constexpr uint64_t kAllExternalPointerTypeTags[] = {
|
|
430
446
|
(HasMarkBit ? kExternalPointerMarkBit : 0))
|
431
447
|
enum ExternalPointerTag : uint64_t {
|
432
448
|
// Empty tag value. Mostly used as placeholder.
|
433
|
-
kExternalPointerNullTag = MAKE_TAG(
|
449
|
+
kExternalPointerNullTag = MAKE_TAG(1, 0b00000000),
|
434
450
|
// External pointer tag that will match any external pointer. Use with care!
|
435
451
|
kAnyExternalPointerTag = MAKE_TAG(1, 0b11111111),
|
436
452
|
// The free entry tag has all type bits set so every type check with a
|
@@ -456,6 +472,15 @@ V8_INLINE static constexpr bool IsSharedExternalPointerType(
|
|
456
472
|
return tag >= kFirstSharedTag && tag <= kLastSharedTag;
|
457
473
|
}
|
458
474
|
|
475
|
+
// True if the external pointer may live in a read-only object, in which case
|
476
|
+
// the table entry will be in the shared read-only segment of the external
|
477
|
+
// pointer table.
|
478
|
+
V8_INLINE static constexpr bool IsMaybeReadOnlyExternalPointerType(
|
479
|
+
ExternalPointerTag tag) {
|
480
|
+
return tag == kAccessorInfoGetterTag || tag == kAccessorInfoSetterTag ||
|
481
|
+
tag == kFunctionTemplateInfoCallbackTag;
|
482
|
+
}
|
483
|
+
|
459
484
|
// Sanity checks.
|
460
485
|
#define CHECK_SHARED_EXTERNAL_POINTER_TAGS(Tag, ...) \
|
461
486
|
static_assert(IsSharedExternalPointerType(Tag));
|
@@ -471,6 +496,126 @@ PER_ISOLATE_EXTERNAL_POINTER_TAGS(CHECK_NON_SHARED_EXTERNAL_POINTER_TAGS)
|
|
471
496
|
#undef SHARED_EXTERNAL_POINTER_TAGS
|
472
497
|
#undef EXTERNAL_POINTER_TAGS
|
473
498
|
|
499
|
+
//
|
500
|
+
// Indirect Pointers.
|
501
|
+
//
|
502
|
+
// When the sandbox is enabled, indirect pointers are used to reference
|
503
|
+
// HeapObjects that live outside of the sandbox (but are still managed by V8's
|
504
|
+
// garbage collector). When object A references an object B through an indirect
|
505
|
+
// pointer, object A will contain a IndirectPointerHandle, i.e. a shifted
|
506
|
+
// 32-bit index, which identifies an entry in a pointer table (either the
|
507
|
+
// trusted pointer table for TrustedObjects, or the code pointer table if it is
|
508
|
+
// a Code object). This table entry then contains the actual pointer to object
|
509
|
+
// B. Further, object B owns this pointer table entry, and it is responsible
|
510
|
+
// for updating the "self-pointer" in the entry when it is relocated in memory.
|
511
|
+
// This way, in contrast to "normal" pointers, indirect pointers never need to
|
512
|
+
// be tracked by the GC (i.e. there is no remembered set for them).
|
513
|
+
// These pointers do not exist when the sandbox is disabled.
|
514
|
+
|
515
|
+
// An IndirectPointerHandle represents a 32-bit index into a pointer table.
|
516
|
+
using IndirectPointerHandle = uint32_t;
|
517
|
+
|
518
|
+
// A null handle always references an entry that contains nullptr.
|
519
|
+
constexpr IndirectPointerHandle kNullIndirectPointerHandle = 0;
|
520
|
+
|
521
|
+
// When the sandbox is enabled, indirect pointers are used to implement:
|
522
|
+
// - TrustedPointers: an indirect pointer using the trusted pointer table (TPT)
|
523
|
+
// and referencing a TrustedObject in one of the trusted heap spaces.
|
524
|
+
// - CodePointers, an indirect pointer using the code pointer table (CPT) and
|
525
|
+
// referencing a Code object together with its instruction stream.
|
526
|
+
|
527
|
+
//
|
528
|
+
// Trusted Pointers.
|
529
|
+
//
|
530
|
+
// A pointer to a TrustedObject.
|
531
|
+
// When the sandbox is enabled, these are indirect pointers using the trusted
|
532
|
+
// pointer table (TPT). They are used to reference trusted objects (located in
|
533
|
+
// one of V8's trusted heap spaces, outside of the sandbox) from inside the
|
534
|
+
// sandbox in a memory-safe way. When the sandbox is disabled, these are
|
535
|
+
// regular tagged pointers.
|
536
|
+
using TrustedPointerHandle = IndirectPointerHandle;
|
537
|
+
|
538
|
+
// The size of the virtual memory reservation for the trusted pointer table.
|
539
|
+
// As with the external pointer table, a maximum table size in combination with
|
540
|
+
// shifted indices allows omitting bounds checks.
|
541
|
+
constexpr size_t kTrustedPointerTableReservationSize = 64 * MB;
|
542
|
+
|
543
|
+
// The trusted pointer handles are stores shifted to the left by this amount
|
544
|
+
// to guarantee that they are smaller than the maximum table size.
|
545
|
+
constexpr uint32_t kTrustedPointerHandleShift = 9;
|
546
|
+
|
547
|
+
// A null handle always references an entry that contains nullptr.
|
548
|
+
constexpr TrustedPointerHandle kNullTrustedPointerHandle =
|
549
|
+
kNullIndirectPointerHandle;
|
550
|
+
|
551
|
+
// The maximum number of entries in an trusted pointer table.
|
552
|
+
constexpr int kTrustedPointerTableEntrySize = 8;
|
553
|
+
constexpr int kTrustedPointerTableEntrySizeLog2 = 3;
|
554
|
+
constexpr size_t kMaxTrustedPointers =
|
555
|
+
kTrustedPointerTableReservationSize / kTrustedPointerTableEntrySize;
|
556
|
+
static_assert((1 << (32 - kTrustedPointerHandleShift)) == kMaxTrustedPointers,
|
557
|
+
"kTrustedPointerTableReservationSize and "
|
558
|
+
"kTrustedPointerHandleShift don't match");
|
559
|
+
|
560
|
+
//
|
561
|
+
// Code Pointers.
|
562
|
+
//
|
563
|
+
// A pointer to a Code object.
|
564
|
+
// Essentially a specialized version of a trusted pointer that (when the
|
565
|
+
// sandbox is enabled) uses the code pointer table (CPT) instead of the TPT.
|
566
|
+
// Each entry in the CPT contains both a pointer to a Code object as well as a
|
567
|
+
// pointer to the Code's entrypoint. This allows calling/jumping into Code with
|
568
|
+
// one fewer memory access (compared to the case where the entrypoint pointer
|
569
|
+
// first needs to be loaded from the Code object). As such, a CodePointerHandle
|
570
|
+
// can be used both to obtain the referenced Code object and to directly load
|
571
|
+
// its entrypoint.
|
572
|
+
//
|
573
|
+
// When the sandbox is disabled, these are regular tagged pointers.
|
574
|
+
using CodePointerHandle = IndirectPointerHandle;
|
575
|
+
|
576
|
+
// The size of the virtual memory reservation for the code pointer table.
|
577
|
+
// As with the other tables, a maximum table size in combination with shifted
|
578
|
+
// indices allows omitting bounds checks.
|
579
|
+
constexpr size_t kCodePointerTableReservationSize = 16 * MB;
|
580
|
+
|
581
|
+
// Code pointer handles are shifted by a different amount than indirect pointer
|
582
|
+
// handles as the tables have a different maximum size.
|
583
|
+
constexpr uint32_t kCodePointerHandleShift = 12;
|
584
|
+
|
585
|
+
// A null handle always references an entry that contains nullptr.
|
586
|
+
constexpr CodePointerHandle kNullCodePointerHandle = kNullIndirectPointerHandle;
|
587
|
+
|
588
|
+
// It can sometimes be necessary to distinguish a code pointer handle from a
|
589
|
+
// trusted pointer handle. A typical example would be a union trusted pointer
|
590
|
+
// field that can refer to both Code objects and other trusted objects. To
|
591
|
+
// support these use-cases, we use a simple marking scheme where some of the
|
592
|
+
// low bits of a code pointer handle are set, while they will be unset on a
|
593
|
+
// trusted pointer handle. This way, the correct table to resolve the handle
|
594
|
+
// can be determined even in the absence of a type tag.
|
595
|
+
constexpr uint32_t kCodePointerHandleMarker = 0x1;
|
596
|
+
static_assert(kCodePointerHandleShift > 0);
|
597
|
+
static_assert(kTrustedPointerHandleShift > 0);
|
598
|
+
|
599
|
+
// The maximum number of entries in a code pointer table.
|
600
|
+
constexpr int kCodePointerTableEntrySize = 16;
|
601
|
+
constexpr int kCodePointerTableEntrySizeLog2 = 4;
|
602
|
+
constexpr size_t kMaxCodePointers =
|
603
|
+
kCodePointerTableReservationSize / kCodePointerTableEntrySize;
|
604
|
+
static_assert(
|
605
|
+
(1 << (32 - kCodePointerHandleShift)) == kMaxCodePointers,
|
606
|
+
"kCodePointerTableReservationSize and kCodePointerHandleShift don't match");
|
607
|
+
|
608
|
+
constexpr int kCodePointerTableEntryEntrypointOffset = 0;
|
609
|
+
constexpr int kCodePointerTableEntryCodeObjectOffset = 8;
|
610
|
+
|
611
|
+
// Constants that can be used to mark places that should be modified once
|
612
|
+
// certain types of objects are moved out of the sandbox and into trusted space.
|
613
|
+
constexpr bool kRuntimeGeneratedCodeObjectsLiveInTrustedSpace = true;
|
614
|
+
constexpr bool kBuiltinCodeObjectsLiveInTrustedSpace = false;
|
615
|
+
constexpr bool kAllCodeObjectsLiveInTrustedSpace =
|
616
|
+
kRuntimeGeneratedCodeObjectsLiveInTrustedSpace &&
|
617
|
+
kBuiltinCodeObjectsLiveInTrustedSpace;
|
618
|
+
|
474
619
|
// {obj} must be the raw tagged pointer representation of a HeapObject
|
475
620
|
// that's guaranteed to never be in ReadOnlySpace.
|
476
621
|
V8_EXPORT internal::Isolate* IsolateFromNeverReadOnlySpaceObject(Address obj);
|
@@ -517,15 +662,22 @@ class Internals {
|
|
517
662
|
static const int kExternalOneByteRepresentationTag = 0x0a;
|
518
663
|
|
519
664
|
static const uint32_t kNumIsolateDataSlots = 4;
|
520
|
-
static const int kStackGuardSize =
|
665
|
+
static const int kStackGuardSize = 8 * kApiSystemPointerSize;
|
666
|
+
static const int kNumberOfBooleanFlags = 6;
|
667
|
+
static const int kErrorMessageParamSize = 1;
|
668
|
+
static const int kTablesAlignmentPaddingSize = 1;
|
521
669
|
static const int kBuiltinTier0EntryTableSize = 7 * kApiSystemPointerSize;
|
522
670
|
static const int kBuiltinTier0TableSize = 7 * kApiSystemPointerSize;
|
523
671
|
static const int kLinearAllocationAreaSize = 3 * kApiSystemPointerSize;
|
524
|
-
static const int kThreadLocalTopSize =
|
672
|
+
static const int kThreadLocalTopSize = 30 * kApiSystemPointerSize;
|
673
|
+
static const int kHandleScopeDataSize =
|
674
|
+
2 * kApiSystemPointerSize + 2 * kApiInt32Size;
|
525
675
|
|
526
|
-
// ExternalPointerTable layout guarantees.
|
527
|
-
static const int
|
528
|
-
static const int kExternalPointerTableSize =
|
676
|
+
// ExternalPointerTable and TrustedPointerTable layout guarantees.
|
677
|
+
static const int kExternalPointerTableBasePointerOffset = 0;
|
678
|
+
static const int kExternalPointerTableSize = 2 * kApiSystemPointerSize;
|
679
|
+
static const int kTrustedPointerTableSize = 2 * kApiSystemPointerSize;
|
680
|
+
static const int kTrustedPointerTableBasePointerOffset = 0;
|
529
681
|
|
530
682
|
// IsolateData layout guarantees.
|
531
683
|
static const int kIsolateCageBaseOffset = 0;
|
@@ -533,16 +685,23 @@ class Internals {
|
|
533
685
|
kIsolateCageBaseOffset + kApiSystemPointerSize;
|
534
686
|
static const int kVariousBooleanFlagsOffset =
|
535
687
|
kIsolateStackGuardOffset + kStackGuardSize;
|
536
|
-
static const int
|
537
|
-
kVariousBooleanFlagsOffset +
|
688
|
+
static const int kErrorMessageParamOffset =
|
689
|
+
kVariousBooleanFlagsOffset + kNumberOfBooleanFlags;
|
690
|
+
static const int kBuiltinTier0EntryTableOffset = kErrorMessageParamOffset +
|
691
|
+
kErrorMessageParamSize +
|
692
|
+
kTablesAlignmentPaddingSize;
|
538
693
|
static const int kBuiltinTier0TableOffset =
|
539
694
|
kBuiltinTier0EntryTableOffset + kBuiltinTier0EntryTableSize;
|
540
695
|
static const int kNewAllocationInfoOffset =
|
541
696
|
kBuiltinTier0TableOffset + kBuiltinTier0TableSize;
|
542
697
|
static const int kOldAllocationInfoOffset =
|
543
698
|
kNewAllocationInfoOffset + kLinearAllocationAreaSize;
|
699
|
+
|
700
|
+
static const int kFastCCallAlignmentPaddingSize =
|
701
|
+
kApiSystemPointerSize == 8 ? 0 : kApiSystemPointerSize;
|
544
702
|
static const int kIsolateFastCCallCallerFpOffset =
|
545
|
-
kOldAllocationInfoOffset + kLinearAllocationAreaSize
|
703
|
+
kOldAllocationInfoOffset + kLinearAllocationAreaSize +
|
704
|
+
kFastCCallAlignmentPaddingSize;
|
546
705
|
static const int kIsolateFastCCallCallerPcOffset =
|
547
706
|
kIsolateFastCCallCallerFpOffset + kApiSystemPointerSize;
|
548
707
|
static const int kIsolateFastApiCallTargetOffset =
|
@@ -551,39 +710,64 @@ class Internals {
|
|
551
710
|
kIsolateFastApiCallTargetOffset + kApiSystemPointerSize;
|
552
711
|
static const int kIsolateThreadLocalTopOffset =
|
553
712
|
kIsolateLongTaskStatsCounterOffset + kApiSizetSize;
|
554
|
-
static const int
|
713
|
+
static const int kIsolateHandleScopeDataOffset =
|
555
714
|
kIsolateThreadLocalTopOffset + kThreadLocalTopSize;
|
715
|
+
static const int kIsolateEmbedderDataOffset =
|
716
|
+
kIsolateHandleScopeDataOffset + kHandleScopeDataSize;
|
556
717
|
#ifdef V8_COMPRESS_POINTERS
|
557
718
|
static const int kIsolateExternalPointerTableOffset =
|
558
719
|
kIsolateEmbedderDataOffset + kNumIsolateDataSlots * kApiSystemPointerSize;
|
559
720
|
static const int kIsolateSharedExternalPointerTableAddressOffset =
|
560
721
|
kIsolateExternalPointerTableOffset + kExternalPointerTableSize;
|
561
|
-
|
722
|
+
#ifdef V8_ENABLE_SANDBOX
|
723
|
+
static const int kIsolateTrustedCageBaseOffset =
|
562
724
|
kIsolateSharedExternalPointerTableAddressOffset + kApiSystemPointerSize;
|
725
|
+
static const int kIsolateTrustedPointerTableOffset =
|
726
|
+
kIsolateTrustedCageBaseOffset + kApiSystemPointerSize;
|
727
|
+
static const int kIsolateApiCallbackThunkArgumentOffset =
|
728
|
+
kIsolateTrustedPointerTableOffset + kTrustedPointerTableSize;
|
563
729
|
#else
|
564
|
-
static const int
|
730
|
+
static const int kIsolateApiCallbackThunkArgumentOffset =
|
731
|
+
kIsolateSharedExternalPointerTableAddressOffset + kApiSystemPointerSize;
|
732
|
+
#endif // V8_ENABLE_SANDBOX
|
733
|
+
#else
|
734
|
+
static const int kIsolateApiCallbackThunkArgumentOffset =
|
565
735
|
kIsolateEmbedderDataOffset + kNumIsolateDataSlots * kApiSystemPointerSize;
|
566
|
-
#endif
|
736
|
+
#endif // V8_COMPRESS_POINTERS
|
737
|
+
static const int kContinuationPreservedEmbedderDataOffset =
|
738
|
+
kIsolateApiCallbackThunkArgumentOffset + kApiSystemPointerSize;
|
739
|
+
|
740
|
+
static const int kWasm64OOBOffsetAlignmentPaddingSize = 0;
|
741
|
+
static const int kWasm64OOBOffsetOffset =
|
742
|
+
kContinuationPreservedEmbedderDataOffset + kApiSystemPointerSize +
|
743
|
+
kWasm64OOBOffsetAlignmentPaddingSize;
|
744
|
+
static const int kIsolateRootsOffset =
|
745
|
+
kWasm64OOBOffsetOffset + sizeof(int64_t);
|
567
746
|
|
568
747
|
#if V8_STATIC_ROOTS_BOOL
|
569
748
|
|
570
|
-
// These constants
|
749
|
+
// These constants are copied from static-roots.h and guarded by static asserts.
|
571
750
|
#define EXPORTED_STATIC_ROOTS_PTR_LIST(V) \
|
572
|
-
V(UndefinedValue)
|
573
|
-
V(NullValue)
|
574
|
-
V(TrueValue)
|
575
|
-
V(FalseValue)
|
576
|
-
V(EmptyString)
|
577
|
-
V(TheHoleValue)
|
751
|
+
V(UndefinedValue, 0x69) \
|
752
|
+
V(NullValue, 0x85) \
|
753
|
+
V(TrueValue, 0xc9) \
|
754
|
+
V(FalseValue, 0xad) \
|
755
|
+
V(EmptyString, 0xa1) \
|
756
|
+
V(TheHoleValue, 0x719)
|
578
757
|
|
579
758
|
using Tagged_t = uint32_t;
|
580
759
|
struct StaticReadOnlyRoot {
|
581
|
-
#define DEF_ROOT(name)
|
760
|
+
#define DEF_ROOT(name, value) static constexpr Tagged_t k##name = value;
|
582
761
|
EXPORTED_STATIC_ROOTS_PTR_LIST(DEF_ROOT)
|
583
762
|
#undef DEF_ROOT
|
584
763
|
|
585
|
-
|
586
|
-
|
764
|
+
static constexpr Tagged_t kFirstStringMap = 0xe5;
|
765
|
+
static constexpr Tagged_t kLastStringMap = 0x47d;
|
766
|
+
|
767
|
+
#define PLUSONE(...) +1
|
768
|
+
static constexpr size_t kNumberOfExportedStaticRoots =
|
769
|
+
2 + EXPORTED_STATIC_ROOTS_PTR_LIST(PLUSONE);
|
770
|
+
#undef PLUSONE
|
587
771
|
};
|
588
772
|
|
589
773
|
#endif // V8_STATIC_ROOTS_BOOL
|
@@ -600,8 +784,6 @@ class Internals {
|
|
600
784
|
static const int kNodeStateMask = 0x3;
|
601
785
|
static const int kNodeStateIsWeakValue = 2;
|
602
786
|
|
603
|
-
static const int kTracedNodeClassIdOffset = kApiSystemPointerSize;
|
604
|
-
|
605
787
|
static const int kFirstNonstringType = 0x80;
|
606
788
|
static const int kOddballType = 0x83;
|
607
789
|
static const int kForeignType = 0xcc;
|
@@ -609,8 +791,13 @@ class Internals {
|
|
609
791
|
static const int kJSObjectType = 0x421;
|
610
792
|
static const int kFirstJSApiObjectType = 0x422;
|
611
793
|
static const int kLastJSApiObjectType = 0x80A;
|
794
|
+
// Defines a range [kFirstEmbedderJSApiObjectType, kJSApiObjectTypesCount]
|
795
|
+
// of JSApiObject instance type values that an embedder can use.
|
796
|
+
static const int kFirstEmbedderJSApiObjectType = 0;
|
797
|
+
static const int kLastEmbedderJSApiObjectType =
|
798
|
+
kLastJSApiObjectType - kFirstJSApiObjectType;
|
612
799
|
|
613
|
-
static const int kUndefinedOddballKind =
|
800
|
+
static const int kUndefinedOddballKind = 4;
|
614
801
|
static const int kNullOddballKind = 3;
|
615
802
|
|
616
803
|
// Constants used by PropertyCallbackInfo to check if we should throw when an
|
@@ -641,11 +828,11 @@ class Internals {
|
|
641
828
|
#endif
|
642
829
|
}
|
643
830
|
|
644
|
-
V8_INLINE static bool HasHeapObjectTag(Address value) {
|
831
|
+
V8_INLINE static constexpr bool HasHeapObjectTag(Address value) {
|
645
832
|
return (value & kHeapObjectTagMask) == static_cast<Address>(kHeapObjectTag);
|
646
833
|
}
|
647
834
|
|
648
|
-
V8_INLINE static int SmiValue(Address value) {
|
835
|
+
V8_INLINE static constexpr int SmiValue(Address value) {
|
649
836
|
return PlatformSmiTagging::SmiToInt(value);
|
650
837
|
}
|
651
838
|
|
@@ -680,6 +867,15 @@ class Internals {
|
|
680
867
|
return ReadRawField<uint16_t>(map, kMapInstanceTypeOffset);
|
681
868
|
}
|
682
869
|
|
870
|
+
V8_INLINE static Address LoadMap(Address obj) {
|
871
|
+
if (!HasHeapObjectTag(obj)) return kNullAddress;
|
872
|
+
Address map = ReadTaggedPointerField(obj, kHeapObjectMapOffset);
|
873
|
+
#ifdef V8_MAP_PACKING
|
874
|
+
map = UnpackMapWord(map);
|
875
|
+
#endif
|
876
|
+
return map;
|
877
|
+
}
|
878
|
+
|
683
879
|
V8_INLINE static int GetOddballKind(Address obj) {
|
684
880
|
return SmiValue(ReadTaggedSignedField(obj, kOddballKindOffset));
|
685
881
|
}
|
@@ -753,15 +949,15 @@ class Internals {
|
|
753
949
|
Address base = *reinterpret_cast<Address*>(
|
754
950
|
reinterpret_cast<uintptr_t>(isolate) + kIsolateCageBaseOffset);
|
755
951
|
switch (index) {
|
756
|
-
#define DECOMPRESS_ROOT(name) \
|
757
|
-
case k##name##RootIndex:
|
952
|
+
#define DECOMPRESS_ROOT(name, ...) \
|
953
|
+
case k##name##RootIndex: \
|
758
954
|
return base + StaticReadOnlyRoot::k##name;
|
759
955
|
EXPORTED_STATIC_ROOTS_PTR_LIST(DECOMPRESS_ROOT)
|
760
956
|
#undef DECOMPRESS_ROOT
|
957
|
+
#undef EXPORTED_STATIC_ROOTS_PTR_LIST
|
761
958
|
default:
|
762
959
|
break;
|
763
960
|
}
|
764
|
-
#undef EXPORTED_STATIC_ROOTS_PTR_LIST
|
765
961
|
#endif // V8_STATIC_ROOTS_BOOL
|
766
962
|
return *GetRootSlot(isolate, index);
|
767
963
|
}
|
@@ -770,7 +966,7 @@ class Internals {
|
|
770
966
|
V8_INLINE static Address* GetExternalPointerTableBase(v8::Isolate* isolate) {
|
771
967
|
Address addr = reinterpret_cast<Address>(isolate) +
|
772
968
|
kIsolateExternalPointerTableOffset +
|
773
|
-
|
969
|
+
kExternalPointerTableBasePointerOffset;
|
774
970
|
return *reinterpret_cast<Address**>(addr);
|
775
971
|
}
|
776
972
|
|
@@ -779,7 +975,7 @@ class Internals {
|
|
779
975
|
Address addr = reinterpret_cast<Address>(isolate) +
|
780
976
|
kIsolateSharedExternalPointerTableAddressOffset;
|
781
977
|
addr = *reinterpret_cast<Address*>(addr);
|
782
|
-
addr +=
|
978
|
+
addr += kExternalPointerTableBasePointerOffset;
|
783
979
|
return *reinterpret_cast<Address**>(addr);
|
784
980
|
}
|
785
981
|
#endif
|
@@ -860,6 +1056,10 @@ class Internals {
|
|
860
1056
|
return addr & -static_cast<intptr_t>(kPtrComprCageBaseAlignment);
|
861
1057
|
}
|
862
1058
|
|
1059
|
+
V8_INLINE static uint32_t CompressTagged(Address value) {
|
1060
|
+
return static_cast<uint32_t>(value);
|
1061
|
+
}
|
1062
|
+
|
863
1063
|
V8_INLINE static Address DecompressTaggedField(Address heap_object_ptr,
|
864
1064
|
uint32_t value) {
|
865
1065
|
Address base = GetPtrComprCageBaseFromOnHeapAddress(heap_object_ptr);
|
@@ -901,57 +1101,296 @@ class BackingStoreBase {};
|
|
901
1101
|
// This is needed for histograms sampling garbage collection reasons.
|
902
1102
|
constexpr int kGarbageCollectionReasonMaxValue = 27;
|
903
1103
|
|
1104
|
+
// Base class for the address block allocator compatible with standard
|
1105
|
+
// containers, which registers its allocated range as strong roots.
|
1106
|
+
class V8_EXPORT StrongRootAllocatorBase {
|
1107
|
+
public:
|
1108
|
+
Heap* heap() const { return heap_; }
|
1109
|
+
|
1110
|
+
bool operator==(const StrongRootAllocatorBase& other) const {
|
1111
|
+
return heap_ == other.heap_;
|
1112
|
+
}
|
1113
|
+
bool operator!=(const StrongRootAllocatorBase& other) const {
|
1114
|
+
return heap_ != other.heap_;
|
1115
|
+
}
|
1116
|
+
|
1117
|
+
protected:
|
1118
|
+
explicit StrongRootAllocatorBase(Heap* heap) : heap_(heap) {}
|
1119
|
+
explicit StrongRootAllocatorBase(v8::Isolate* isolate);
|
1120
|
+
|
1121
|
+
// Allocate/deallocate a range of n elements of type internal::Address.
|
1122
|
+
Address* allocate_impl(size_t n);
|
1123
|
+
void deallocate_impl(Address* p, size_t n) noexcept;
|
1124
|
+
|
1125
|
+
private:
|
1126
|
+
Heap* heap_;
|
1127
|
+
};
|
1128
|
+
|
1129
|
+
// The general version of this template behaves just as std::allocator, with
|
1130
|
+
// the exception that the constructor takes the isolate as parameter. Only
|
1131
|
+
// specialized versions, e.g., internal::StrongRootAllocator<internal::Address>
|
1132
|
+
// and internal::StrongRootAllocator<v8::Local<T>> register the allocated range
|
1133
|
+
// as strong roots.
|
1134
|
+
template <typename T>
|
1135
|
+
class StrongRootAllocator : public StrongRootAllocatorBase,
|
1136
|
+
private std::allocator<T> {
|
1137
|
+
public:
|
1138
|
+
using value_type = T;
|
1139
|
+
|
1140
|
+
explicit StrongRootAllocator(Heap* heap) : StrongRootAllocatorBase(heap) {}
|
1141
|
+
explicit StrongRootAllocator(v8::Isolate* isolate)
|
1142
|
+
: StrongRootAllocatorBase(isolate) {}
|
1143
|
+
template <typename U>
|
1144
|
+
StrongRootAllocator(const StrongRootAllocator<U>& other) noexcept
|
1145
|
+
: StrongRootAllocatorBase(other) {}
|
1146
|
+
|
1147
|
+
using std::allocator<T>::allocate;
|
1148
|
+
using std::allocator<T>::deallocate;
|
1149
|
+
};
|
1150
|
+
|
1151
|
+
// A class of iterators that wrap some different iterator type.
|
1152
|
+
// If specified, ElementType is the type of element accessed by the wrapper
|
1153
|
+
// iterator; in this case, the actual reference and pointer types of Iterator
|
1154
|
+
// must be convertible to ElementType& and ElementType*, respectively.
|
1155
|
+
template <typename Iterator, typename ElementType = void>
|
1156
|
+
class WrappedIterator {
|
1157
|
+
public:
|
1158
|
+
static_assert(
|
1159
|
+
!std::is_void_v<ElementType> ||
|
1160
|
+
(std::is_convertible_v<typename std::iterator_traits<Iterator>::pointer,
|
1161
|
+
ElementType*> &&
|
1162
|
+
std::is_convertible_v<typename std::iterator_traits<Iterator>::reference,
|
1163
|
+
ElementType&>));
|
1164
|
+
|
1165
|
+
using iterator_category =
|
1166
|
+
typename std::iterator_traits<Iterator>::iterator_category;
|
1167
|
+
using difference_type =
|
1168
|
+
typename std::iterator_traits<Iterator>::difference_type;
|
1169
|
+
using value_type =
|
1170
|
+
std::conditional_t<std::is_void_v<ElementType>,
|
1171
|
+
typename std::iterator_traits<Iterator>::value_type,
|
1172
|
+
ElementType>;
|
1173
|
+
using pointer =
|
1174
|
+
std::conditional_t<std::is_void_v<ElementType>,
|
1175
|
+
typename std::iterator_traits<Iterator>::pointer,
|
1176
|
+
ElementType*>;
|
1177
|
+
using reference =
|
1178
|
+
std::conditional_t<std::is_void_v<ElementType>,
|
1179
|
+
typename std::iterator_traits<Iterator>::reference,
|
1180
|
+
ElementType&>;
|
1181
|
+
|
1182
|
+
constexpr WrappedIterator() noexcept : it_() {}
|
1183
|
+
constexpr explicit WrappedIterator(Iterator it) noexcept : it_(it) {}
|
1184
|
+
|
1185
|
+
template <typename OtherIterator, typename OtherElementType,
|
1186
|
+
std::enable_if_t<std::is_convertible_v<OtherIterator, Iterator>,
|
1187
|
+
bool> = true>
|
1188
|
+
constexpr WrappedIterator(
|
1189
|
+
const WrappedIterator<OtherIterator, OtherElementType>& it) noexcept
|
1190
|
+
: it_(it.base()) {}
|
1191
|
+
|
1192
|
+
constexpr reference operator*() const noexcept { return *it_; }
|
1193
|
+
constexpr pointer operator->() const noexcept { return it_.operator->(); }
|
1194
|
+
|
1195
|
+
constexpr WrappedIterator& operator++() noexcept {
|
1196
|
+
++it_;
|
1197
|
+
return *this;
|
1198
|
+
}
|
1199
|
+
constexpr WrappedIterator operator++(int) noexcept {
|
1200
|
+
WrappedIterator result(*this);
|
1201
|
+
++(*this);
|
1202
|
+
return result;
|
1203
|
+
}
|
1204
|
+
|
1205
|
+
constexpr WrappedIterator& operator--() noexcept {
|
1206
|
+
--it_;
|
1207
|
+
return *this;
|
1208
|
+
}
|
1209
|
+
constexpr WrappedIterator operator--(int) noexcept {
|
1210
|
+
WrappedIterator result(*this);
|
1211
|
+
--(*this);
|
1212
|
+
return result;
|
1213
|
+
}
|
1214
|
+
constexpr WrappedIterator operator+(difference_type n) const noexcept {
|
1215
|
+
WrappedIterator result(*this);
|
1216
|
+
result += n;
|
1217
|
+
return result;
|
1218
|
+
}
|
1219
|
+
constexpr WrappedIterator& operator+=(difference_type n) noexcept {
|
1220
|
+
it_ += n;
|
1221
|
+
return *this;
|
1222
|
+
}
|
1223
|
+
constexpr WrappedIterator operator-(difference_type n) const noexcept {
|
1224
|
+
return *this + (-n);
|
1225
|
+
}
|
1226
|
+
constexpr WrappedIterator& operator-=(difference_type n) noexcept {
|
1227
|
+
*this += -n;
|
1228
|
+
return *this;
|
1229
|
+
}
|
1230
|
+
constexpr reference operator[](difference_type n) const noexcept {
|
1231
|
+
return it_[n];
|
1232
|
+
}
|
1233
|
+
|
1234
|
+
constexpr Iterator base() const noexcept { return it_; }
|
1235
|
+
|
1236
|
+
private:
|
1237
|
+
template <typename OtherIterator, typename OtherElementType>
|
1238
|
+
friend class WrappedIterator;
|
1239
|
+
|
1240
|
+
private:
|
1241
|
+
Iterator it_;
|
1242
|
+
};
|
1243
|
+
|
1244
|
+
template <typename Iterator, typename ElementType, typename OtherIterator,
|
1245
|
+
typename OtherElementType>
|
1246
|
+
constexpr bool operator==(
|
1247
|
+
const WrappedIterator<Iterator, ElementType>& x,
|
1248
|
+
const WrappedIterator<OtherIterator, OtherElementType>& y) noexcept {
|
1249
|
+
return x.base() == y.base();
|
1250
|
+
}
|
1251
|
+
|
1252
|
+
template <typename Iterator, typename ElementType, typename OtherIterator,
|
1253
|
+
typename OtherElementType>
|
1254
|
+
constexpr bool operator<(
|
1255
|
+
const WrappedIterator<Iterator, ElementType>& x,
|
1256
|
+
const WrappedIterator<OtherIterator, OtherElementType>& y) noexcept {
|
1257
|
+
return x.base() < y.base();
|
1258
|
+
}
|
1259
|
+
|
1260
|
+
template <typename Iterator, typename ElementType, typename OtherIterator,
|
1261
|
+
typename OtherElementType>
|
1262
|
+
constexpr bool operator!=(
|
1263
|
+
const WrappedIterator<Iterator, ElementType>& x,
|
1264
|
+
const WrappedIterator<OtherIterator, OtherElementType>& y) noexcept {
|
1265
|
+
return !(x == y);
|
1266
|
+
}
|
1267
|
+
|
1268
|
+
template <typename Iterator, typename ElementType, typename OtherIterator,
|
1269
|
+
typename OtherElementType>
|
1270
|
+
constexpr bool operator>(
|
1271
|
+
const WrappedIterator<Iterator, ElementType>& x,
|
1272
|
+
const WrappedIterator<OtherIterator, OtherElementType>& y) noexcept {
|
1273
|
+
return y < x;
|
1274
|
+
}
|
1275
|
+
|
1276
|
+
template <typename Iterator, typename ElementType, typename OtherIterator,
|
1277
|
+
typename OtherElementType>
|
1278
|
+
constexpr bool operator>=(
|
1279
|
+
const WrappedIterator<Iterator, ElementType>& x,
|
1280
|
+
const WrappedIterator<OtherIterator, OtherElementType>& y) noexcept {
|
1281
|
+
return !(x < y);
|
1282
|
+
}
|
1283
|
+
|
1284
|
+
template <typename Iterator, typename ElementType, typename OtherIterator,
|
1285
|
+
typename OtherElementType>
|
1286
|
+
constexpr bool operator<=(
|
1287
|
+
const WrappedIterator<Iterator, ElementType>& x,
|
1288
|
+
const WrappedIterator<OtherIterator, OtherElementType>& y) noexcept {
|
1289
|
+
return !(y < x);
|
1290
|
+
}
|
1291
|
+
|
1292
|
+
template <typename Iterator, typename ElementType, typename OtherIterator,
|
1293
|
+
typename OtherElementType>
|
1294
|
+
constexpr auto operator-(
|
1295
|
+
const WrappedIterator<Iterator, ElementType>& x,
|
1296
|
+
const WrappedIterator<OtherIterator, OtherElementType>& y) noexcept
|
1297
|
+
-> decltype(x.base() - y.base()) {
|
1298
|
+
return x.base() - y.base();
|
1299
|
+
}
|
1300
|
+
|
1301
|
+
template <typename Iterator, typename ElementType>
|
1302
|
+
constexpr WrappedIterator<Iterator> operator+(
|
1303
|
+
typename WrappedIterator<Iterator, ElementType>::difference_type n,
|
1304
|
+
const WrappedIterator<Iterator, ElementType>& x) noexcept {
|
1305
|
+
x += n;
|
1306
|
+
return x;
|
1307
|
+
}
|
1308
|
+
|
904
1309
|
// Helper functions about values contained in handles.
|
1310
|
+
// A value is either an indirect pointer or a direct pointer, depending on
|
1311
|
+
// whether direct local support is enabled.
|
905
1312
|
class ValueHelper final {
|
906
1313
|
public:
|
907
|
-
#ifdef
|
908
|
-
static constexpr Address
|
1314
|
+
#ifdef V8_ENABLE_DIRECT_LOCAL
|
1315
|
+
static constexpr Address kTaggedNullAddress = 1;
|
1316
|
+
static constexpr Address kEmpty = kTaggedNullAddress;
|
1317
|
+
#else
|
1318
|
+
static constexpr Address kEmpty = kNullAddress;
|
1319
|
+
#endif // V8_ENABLE_DIRECT_LOCAL
|
909
1320
|
|
910
1321
|
template <typename T>
|
911
|
-
static
|
912
|
-
return reinterpret_cast<
|
1322
|
+
V8_INLINE static bool IsEmpty(T* value) {
|
1323
|
+
return reinterpret_cast<Address>(value) == kEmpty;
|
1324
|
+
}
|
1325
|
+
|
1326
|
+
// Returns a handle's "value" for all kinds of abstract handles. For Local,
|
1327
|
+
// it is equivalent to `*handle`. The variadic parameters support handle
|
1328
|
+
// types with extra type parameters, like `Persistent<T, M>`.
|
1329
|
+
template <template <typename T, typename... Ms> typename H, typename T,
|
1330
|
+
typename... Ms>
|
1331
|
+
V8_INLINE static T* HandleAsValue(const H<T, Ms...>& handle) {
|
1332
|
+
return handle.template value<T>();
|
913
1333
|
}
|
914
1334
|
|
1335
|
+
#ifdef V8_ENABLE_DIRECT_LOCAL
|
1336
|
+
|
915
1337
|
template <typename T>
|
916
1338
|
V8_INLINE static Address ValueAsAddress(const T* value) {
|
917
1339
|
return reinterpret_cast<Address>(value);
|
918
1340
|
}
|
919
1341
|
|
920
|
-
template <typename T, typename S>
|
1342
|
+
template <typename T, bool check_null = true, typename S>
|
921
1343
|
V8_INLINE static T* SlotAsValue(S* slot) {
|
1344
|
+
if (check_null && slot == nullptr) {
|
1345
|
+
return reinterpret_cast<T*>(kTaggedNullAddress);
|
1346
|
+
}
|
922
1347
|
return *reinterpret_cast<T**>(slot);
|
923
1348
|
}
|
924
1349
|
|
925
|
-
|
926
|
-
V8_INLINE static T* ValueAsSlot(T* const& value) {
|
927
|
-
return reinterpret_cast<T*>(const_cast<T**>(&value));
|
928
|
-
}
|
929
|
-
|
930
|
-
#else // !V8_ENABLE_CONSERVATIVE_STACK_SCANNING
|
931
|
-
|
932
|
-
template <typename T>
|
933
|
-
static constexpr T* EmptyValue() {
|
934
|
-
return nullptr;
|
935
|
-
}
|
1350
|
+
#else // !V8_ENABLE_DIRECT_LOCAL
|
936
1351
|
|
937
1352
|
template <typename T>
|
938
1353
|
V8_INLINE static Address ValueAsAddress(const T* value) {
|
939
1354
|
return *reinterpret_cast<const Address*>(value);
|
940
1355
|
}
|
941
1356
|
|
942
|
-
template <typename T, typename S>
|
1357
|
+
template <typename T, bool check_null = true, typename S>
|
943
1358
|
V8_INLINE static T* SlotAsValue(S* slot) {
|
944
1359
|
return reinterpret_cast<T*>(slot);
|
945
1360
|
}
|
946
1361
|
|
947
|
-
|
948
|
-
|
949
|
-
|
1362
|
+
#endif // V8_ENABLE_DIRECT_LOCAL
|
1363
|
+
};
|
1364
|
+
|
1365
|
+
/**
|
1366
|
+
* Helper functions about handles.
|
1367
|
+
*/
|
1368
|
+
class HandleHelper final {
|
1369
|
+
public:
|
1370
|
+
/**
|
1371
|
+
* Checks whether two handles are equal.
|
1372
|
+
* They are equal iff they are both empty or they are both non-empty and the
|
1373
|
+
* objects to which they refer are physically equal.
|
1374
|
+
*
|
1375
|
+
* If both handles refer to JS objects, this is the same as strict equality.
|
1376
|
+
* For primitives, such as numbers or strings, a `false` return value does not
|
1377
|
+
* indicate that the values aren't equal in the JavaScript sense.
|
1378
|
+
* Use `Value::StrictEquals()` to check primitives for equality.
|
1379
|
+
*/
|
1380
|
+
template <typename T1, typename T2>
|
1381
|
+
V8_INLINE static bool EqualHandles(const T1& lhs, const T2& rhs) {
|
1382
|
+
if (lhs.IsEmpty()) return rhs.IsEmpty();
|
1383
|
+
if (rhs.IsEmpty()) return false;
|
1384
|
+
return lhs.ptr() == rhs.ptr();
|
950
1385
|
}
|
951
1386
|
|
952
|
-
|
1387
|
+
static V8_EXPORT bool IsOnStack(const void* ptr);
|
1388
|
+
static V8_EXPORT void VerifyOnStack(const void* ptr);
|
1389
|
+
static V8_EXPORT void VerifyOnMainThread();
|
953
1390
|
};
|
954
1391
|
|
1392
|
+
V8_EXPORT void VerifyHandleIsNonEmpty(bool is_empty);
|
1393
|
+
|
955
1394
|
} // namespace internal
|
956
1395
|
} // namespace v8
|
957
1396
|
|