2022-06-27 20:50:08 +00:00
|
|
|
From 0000000000000000000000000000000000000000 Mon Sep 17 00:00:00 2001
|
|
|
|
From: Jeremy Rose <japthorp@slack-corp.com>
|
|
|
|
Date: Tue, 21 Jun 2022 10:04:21 -0700
|
|
|
|
Subject: support V8 sandboxed pointers
|
|
|
|
|
|
|
|
This refactors several allocators to allocate within the V8 memory cage,
|
|
|
|
allowing them to be compatible with the V8_SANDBOXED_POINTERS feature.
|
|
|
|
|
|
|
|
diff --git a/src/api/environment.cc b/src/api/environment.cc
|
2023-08-08 22:52:51 +00:00
|
|
|
index 7ef6d04794c31064c70dbbb0bfc1dd7bf4d1b8fc..d5a03d5e10faaa204b3f9f290fed79be824c78b1 100644
|
2022-06-27 20:50:08 +00:00
|
|
|
--- a/src/api/environment.cc
|
|
|
|
+++ b/src/api/environment.cc
|
2023-08-08 22:52:51 +00:00
|
|
|
@@ -87,6 +87,14 @@ MaybeLocal<Value> PrepareStackTraceCallback(Local<Context> context,
|
2022-07-05 15:28:22 +00:00
|
|
|
return result;
|
|
|
|
}
|
|
|
|
|
|
|
|
+NodeArrayBufferAllocator::NodeArrayBufferAllocator() {
|
|
|
|
+ zero_fill_field_ = static_cast<uint32_t*>(allocator_->Allocate(sizeof(*zero_fill_field_)));
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+NodeArrayBufferAllocator::~NodeArrayBufferAllocator() {
|
|
|
|
+ allocator_->Free(zero_fill_field_, sizeof(*zero_fill_field_));
|
|
|
|
+}
|
|
|
|
+
|
2022-06-27 20:50:08 +00:00
|
|
|
void* NodeArrayBufferAllocator::Allocate(size_t size) {
|
|
|
|
void* ret;
|
2022-08-29 13:55:36 +00:00
|
|
|
if (zero_fill_field_ || per_process::cli_options->zero_fill_all_buffers)
|
2022-06-27 20:50:08 +00:00
|
|
|
diff --git a/src/crypto/crypto_util.cc b/src/crypto/crypto_util.cc
|
2023-08-08 22:52:51 +00:00
|
|
|
index 4d79fbb4a30a7e57a7456413685706d9af36b1b3..13e4d8b532cdb4c39fab30a8f4118ff8d2bca33c 100644
|
2022-06-27 20:50:08 +00:00
|
|
|
--- a/src/crypto/crypto_util.cc
|
|
|
|
+++ b/src/crypto/crypto_util.cc
|
2023-08-08 22:52:51 +00:00
|
|
|
@@ -349,10 +349,35 @@ ByteSource& ByteSource::operator=(ByteSource&& other) noexcept {
|
2022-06-27 20:50:08 +00:00
|
|
|
return *this;
|
|
|
|
}
|
|
|
|
|
|
|
|
-std::unique_ptr<BackingStore> ByteSource::ReleaseToBackingStore() {
|
|
|
|
+std::unique_ptr<BackingStore> ByteSource::ReleaseToBackingStore(Environment* env) {
|
|
|
|
// It's ok for allocated_data_ to be nullptr but
|
|
|
|
// only if size_ is zero.
|
|
|
|
CHECK_IMPLIES(size_ > 0, allocated_data_ != nullptr);
|
2022-09-07 07:46:37 +00:00
|
|
|
+#if defined(V8_ENABLE_SANDBOX)
|
2022-06-27 20:50:08 +00:00
|
|
|
+ // When V8 sandboxed pointers are enabled, we have to copy into the memory
|
|
|
|
+ // cage. We still want to ensure we erase the data on free though, so
|
|
|
|
+ // provide a custom deleter that calls OPENSSL_cleanse.
|
|
|
|
+ if (!size())
|
|
|
|
+ return ArrayBuffer::NewBackingStore(env->isolate(), 0);
|
|
|
|
+ std::unique_ptr<ArrayBuffer::Allocator> allocator(ArrayBuffer::Allocator::NewDefaultAllocator());
|
|
|
|
+ void* v8_data = allocator->Allocate(size());
|
|
|
|
+ CHECK(v8_data);
|
|
|
|
+ memcpy(v8_data, allocated_data_, size());
|
|
|
|
+ OPENSSL_clear_free(allocated_data_, size());
|
|
|
|
+ std::unique_ptr<BackingStore> ptr = ArrayBuffer::NewBackingStore(
|
|
|
|
+ v8_data,
|
|
|
|
+ size(),
|
|
|
|
+ [](void* data, size_t length, void*) {
|
|
|
|
+ OPENSSL_cleanse(data, length);
|
|
|
|
+ std::unique_ptr<ArrayBuffer::Allocator> allocator(ArrayBuffer::Allocator::NewDefaultAllocator());
|
|
|
|
+ allocator->Free(data, length);
|
|
|
|
+ }, nullptr);
|
|
|
|
+ CHECK(ptr);
|
|
|
|
+ allocated_data_ = nullptr;
|
|
|
|
+ data_ = nullptr;
|
|
|
|
+ size_ = 0;
|
|
|
|
+ return ptr;
|
|
|
|
+#else
|
|
|
|
std::unique_ptr<BackingStore> ptr = ArrayBuffer::NewBackingStore(
|
|
|
|
allocated_data_,
|
|
|
|
size(),
|
2023-08-08 22:52:51 +00:00
|
|
|
@@ -364,10 +389,11 @@ std::unique_ptr<BackingStore> ByteSource::ReleaseToBackingStore() {
|
2022-06-27 20:50:08 +00:00
|
|
|
data_ = nullptr;
|
|
|
|
size_ = 0;
|
|
|
|
return ptr;
|
2022-09-07 07:46:37 +00:00
|
|
|
+#endif // defined(V8_ENABLE_SANDBOX)
|
2022-06-27 20:50:08 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
Local<ArrayBuffer> ByteSource::ToArrayBuffer(Environment* env) {
|
|
|
|
- std::unique_ptr<BackingStore> store = ReleaseToBackingStore();
|
|
|
|
+ std::unique_ptr<BackingStore> store = ReleaseToBackingStore(env);
|
|
|
|
return ArrayBuffer::New(env->isolate(), std::move(store));
|
|
|
|
}
|
|
|
|
|
2023-08-08 22:52:51 +00:00
|
|
|
@@ -696,6 +722,16 @@ namespace {
|
2022-11-10 21:31:20 +00:00
|
|
|
// in which case this has the same semantics as
|
|
|
|
// using OPENSSL_malloc. However, if the secure heap is
|
|
|
|
// initialized, SecureBuffer will automatically use it.
|
2022-09-07 07:46:37 +00:00
|
|
|
+#if defined(V8_ENABLE_SANDBOX)
|
2022-06-27 20:50:08 +00:00
|
|
|
+// When V8 sandboxed pointers are enabled, the secure heap cannot be used as
|
|
|
|
+// all ArrayBuffers must be allocated inside the V8 memory cage.
|
|
|
|
+void SecureBuffer(const FunctionCallbackInfo<Value>& args) {
|
|
|
|
+ CHECK(args[0]->IsUint32());
|
|
|
|
+ uint32_t len = args[0].As<Uint32>()->Value();
|
|
|
|
+ Local<ArrayBuffer> buffer = ArrayBuffer::New(args.GetIsolate(), len);
|
|
|
|
+ args.GetReturnValue().Set(Uint8Array::New(buffer, 0, len));
|
|
|
|
+}
|
|
|
|
+#else
|
2022-11-10 21:31:20 +00:00
|
|
|
void SecureBuffer(const FunctionCallbackInfo<Value>& args) {
|
|
|
|
CHECK(args[0]->IsUint32());
|
|
|
|
Environment* env = Environment::GetCurrent(args);
|
2023-08-08 22:52:51 +00:00
|
|
|
@@ -717,6 +753,7 @@ void SecureBuffer(const FunctionCallbackInfo<Value>& args) {
|
2022-06-27 20:50:08 +00:00
|
|
|
Local<ArrayBuffer> buffer = ArrayBuffer::New(env->isolate(), store);
|
|
|
|
args.GetReturnValue().Set(Uint8Array::New(buffer, 0, len));
|
|
|
|
}
|
2022-09-07 07:46:37 +00:00
|
|
|
+#endif // defined(V8_ENABLE_SANDBOX)
|
2022-06-27 20:50:08 +00:00
|
|
|
|
|
|
|
void SecureHeapUsed(const FunctionCallbackInfo<Value>& args) {
|
|
|
|
#ifndef OPENSSL_IS_BORINGSSL
|
|
|
|
diff --git a/src/crypto/crypto_util.h b/src/crypto/crypto_util.h
|
2023-03-13 13:51:03 +00:00
|
|
|
index bf19334cf61fa497c9325c1d2e996a16545f1b7f..333039b3b7cdf29e911a9c09932b2588d4cccf1a 100644
|
2022-06-27 20:50:08 +00:00
|
|
|
--- a/src/crypto/crypto_util.h
|
|
|
|
+++ b/src/crypto/crypto_util.h
|
2023-02-03 11:43:42 +00:00
|
|
|
@@ -280,7 +280,7 @@ class ByteSource {
|
2022-06-27 20:50:08 +00:00
|
|
|
// Creates a v8::BackingStore that takes over responsibility for
|
|
|
|
// any allocated data. The ByteSource will be reset with size = 0
|
|
|
|
// after being called.
|
|
|
|
- std::unique_ptr<v8::BackingStore> ReleaseToBackingStore();
|
|
|
|
+ std::unique_ptr<v8::BackingStore> ReleaseToBackingStore(Environment* env);
|
|
|
|
|
|
|
|
v8::Local<v8::ArrayBuffer> ToArrayBuffer(Environment* env);
|
|
|
|
|
|
|
|
diff --git a/src/node_i18n.cc b/src/node_i18n.cc
|
2023-03-13 13:51:03 +00:00
|
|
|
index bb810632ee6617759d9cbd24c84a5d1a3a6081aa..3faf9840eddf2db993baef0866bc8854b49c0700 100644
|
2022-06-27 20:50:08 +00:00
|
|
|
--- a/src/node_i18n.cc
|
|
|
|
+++ b/src/node_i18n.cc
|
|
|
|
@@ -104,7 +104,7 @@ namespace {
|
|
|
|
|
|
|
|
template <typename T>
|
|
|
|
MaybeLocal<Object> ToBufferEndian(Environment* env, MaybeStackBuffer<T>* buf) {
|
|
|
|
- MaybeLocal<Object> ret = Buffer::New(env, buf);
|
|
|
|
+ MaybeLocal<Object> ret = Buffer::Copy(env, reinterpret_cast<char*>(buf->out()), buf->length() * sizeof(T));
|
|
|
|
if (ret.IsEmpty())
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
diff --git a/src/node_internals.h b/src/node_internals.h
|
2023-04-18 20:23:11 +00:00
|
|
|
index 427cfab4eebcab0aed33e42915f4a0e5a9db7cdf..d174b2720b0b2561ebe30437df609e0366388527 100644
|
2022-06-27 20:50:08 +00:00
|
|
|
--- a/src/node_internals.h
|
|
|
|
+++ b/src/node_internals.h
|
2023-04-18 20:23:11 +00:00
|
|
|
@@ -102,7 +102,9 @@ v8::Maybe<bool> InitializePrimordials(v8::Local<v8::Context> context);
|
2022-07-05 15:28:22 +00:00
|
|
|
|
|
|
|
class NodeArrayBufferAllocator : public ArrayBufferAllocator {
|
|
|
|
public:
|
|
|
|
- inline uint32_t* zero_fill_field() { return &zero_fill_field_; }
|
|
|
|
+ NodeArrayBufferAllocator();
|
|
|
|
+ ~NodeArrayBufferAllocator() override;
|
|
|
|
+ inline uint32_t* zero_fill_field() { return zero_fill_field_; }
|
|
|
|
|
|
|
|
void* Allocate(size_t size) override; // Defined in src/node.cc
|
|
|
|
void* AllocateUninitialized(size_t size) override;
|
2023-04-18 20:23:11 +00:00
|
|
|
@@ -121,7 +123,7 @@ class NodeArrayBufferAllocator : public ArrayBufferAllocator {
|
2022-07-05 15:28:22 +00:00
|
|
|
}
|
|
|
|
|
2022-06-27 20:50:08 +00:00
|
|
|
private:
|
2022-07-05 15:28:22 +00:00
|
|
|
- uint32_t zero_fill_field_ = 1; // Boolean but exposed as uint32 to JS land.
|
|
|
|
+ uint32_t* zero_fill_field_ = nullptr; // Boolean but exposed as uint32 to JS land.
|
2022-06-27 20:50:08 +00:00
|
|
|
std::atomic<size_t> total_mem_usage_ {0};
|
|
|
|
|
2022-08-29 13:55:36 +00:00
|
|
|
// Delegate to V8's allocator for compatibility with the V8 memory cage.
|
2022-06-27 20:50:08 +00:00
|
|
|
diff --git a/src/node_serdes.cc b/src/node_serdes.cc
|
2023-02-16 14:18:47 +00:00
|
|
|
index 6864f2d88b34abfa4090780d6993684cd0b366a3..0249574c4431fb5b98852699f1368f71b49691c1 100644
|
2022-06-27 20:50:08 +00:00
|
|
|
--- a/src/node_serdes.cc
|
|
|
|
+++ b/src/node_serdes.cc
|
|
|
|
@@ -29,6 +29,11 @@ using v8::ValueSerializer;
|
|
|
|
|
|
|
|
namespace serdes {
|
|
|
|
|
|
|
|
+v8::ArrayBuffer::Allocator* GetAllocator() {
|
|
|
|
+ static v8::ArrayBuffer::Allocator* allocator = v8::ArrayBuffer::Allocator::NewDefaultAllocator();
|
|
|
|
+ return allocator;
|
2023-02-16 14:18:47 +00:00
|
|
|
+}
|
2022-06-27 20:50:08 +00:00
|
|
|
+
|
|
|
|
class SerializerContext : public BaseObject,
|
|
|
|
public ValueSerializer::Delegate {
|
|
|
|
public:
|
|
|
|
@@ -37,10 +42,15 @@ class SerializerContext : public BaseObject,
|
|
|
|
|
|
|
|
~SerializerContext() override = default;
|
|
|
|
|
|
|
|
+ // v8::ValueSerializer::Delegate
|
|
|
|
void ThrowDataCloneError(Local<String> message) override;
|
|
|
|
Maybe<bool> WriteHostObject(Isolate* isolate, Local<Object> object) override;
|
|
|
|
Maybe<uint32_t> GetSharedArrayBufferId(
|
|
|
|
Isolate* isolate, Local<SharedArrayBuffer> shared_array_buffer) override;
|
|
|
|
+ void* ReallocateBufferMemory(void* old_buffer,
|
|
|
|
+ size_t old_length,
|
|
|
|
+ size_t* new_length) override;
|
|
|
|
+ void FreeBufferMemory(void* buffer) override;
|
|
|
|
|
|
|
|
static void SetTreatArrayBufferViewsAsHostObjects(
|
|
|
|
const FunctionCallbackInfo<Value>& args);
|
|
|
|
@@ -61,6 +71,7 @@ class SerializerContext : public BaseObject,
|
|
|
|
|
|
|
|
private:
|
|
|
|
ValueSerializer serializer_;
|
|
|
|
+ size_t last_length_ = 0;
|
|
|
|
};
|
|
|
|
|
|
|
|
class DeserializerContext : public BaseObject,
|
|
|
|
@@ -144,6 +155,24 @@ Maybe<uint32_t> SerializerContext::GetSharedArrayBufferId(
|
|
|
|
return id.ToLocalChecked()->Uint32Value(env()->context());
|
|
|
|
}
|
|
|
|
|
|
|
|
+void* SerializerContext::ReallocateBufferMemory(void* old_buffer,
|
|
|
|
+ size_t requested_size,
|
|
|
|
+ size_t* new_length) {
|
|
|
|
+ *new_length = std::max(static_cast<size_t>(4096), requested_size);
|
|
|
|
+ if (old_buffer) {
|
|
|
|
+ void* ret = GetAllocator()->Reallocate(old_buffer, last_length_, *new_length);
|
|
|
|
+ last_length_ = *new_length;
|
|
|
|
+ return ret;
|
|
|
|
+ } else {
|
|
|
|
+ last_length_ = *new_length;
|
|
|
|
+ return GetAllocator()->Allocate(*new_length);
|
|
|
|
+ }
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+void SerializerContext::FreeBufferMemory(void* buffer) {
|
|
|
|
+ GetAllocator()->Free(buffer, last_length_);
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
Maybe<bool> SerializerContext::WriteHostObject(Isolate* isolate,
|
|
|
|
Local<Object> input) {
|
|
|
|
MaybeLocal<Value> ret;
|
2023-01-26 06:43:57 +00:00
|
|
|
@@ -209,9 +238,14 @@ void SerializerContext::ReleaseBuffer(const FunctionCallbackInfo<Value>& args) {
|
|
|
|
// Note: Both ValueSerializer and this Buffer::New() variant use malloc()
|
|
|
|
// as the underlying allocator.
|
2022-06-27 20:50:08 +00:00
|
|
|
std::pair<uint8_t*, size_t> ret = ctx->serializer_.Release();
|
2023-01-26 06:43:57 +00:00
|
|
|
- auto buf = Buffer::New(ctx->env(),
|
|
|
|
- reinterpret_cast<char*>(ret.first),
|
2022-06-27 20:50:08 +00:00
|
|
|
- ret.second);
|
2023-01-26 06:43:57 +00:00
|
|
|
+ std::unique_ptr<v8::BackingStore> bs =
|
|
|
|
+ v8::ArrayBuffer::NewBackingStore(reinterpret_cast<char*>(ret.first), ret.second,
|
|
|
|
+ [](void* data, size_t length, void* deleter_data) {
|
|
|
|
+ if (data) GetAllocator()->Free(reinterpret_cast<char*>(data), length);
|
|
|
|
+ }, nullptr);
|
|
|
|
+ Local<ArrayBuffer> ab = v8::ArrayBuffer::New(ctx->env()->isolate(), std::move(bs));
|
|
|
|
+
|
|
|
|
+ auto buf = Buffer::New(ctx->env(), ab, 0, ret.second);
|
2022-06-27 20:50:08 +00:00
|
|
|
|
|
|
|
if (!buf.IsEmpty()) {
|
|
|
|
args.GetReturnValue().Set(buf.ToLocalChecked());
|
2023-01-26 06:43:57 +00:00
|
|
|
diff --git a/test/parallel/test-v8-serialize-leak.js b/test/parallel/test-v8-serialize-leak.js
|
2023-02-07 20:51:07 +00:00
|
|
|
index 696dbfea65ba95b1157cb6f469762d4a6e196199..b342ec59100809187689d0770a462b0b99e75e58 100644
|
2023-01-26 06:43:57 +00:00
|
|
|
--- a/test/parallel/test-v8-serialize-leak.js
|
|
|
|
+++ b/test/parallel/test-v8-serialize-leak.js
|
2023-02-07 20:51:07 +00:00
|
|
|
@@ -25,5 +25,5 @@ if (process.config.variables.asan) {
|
|
|
|
} else if (process.config.variables.node_builtin_modules_path) {
|
|
|
|
assert(after < before * 4, `node_builtin_modules_path: before=${before} after=${after}`);
|
2023-01-26 06:43:57 +00:00
|
|
|
} else {
|
|
|
|
- assert(after < before * 2, `before=${before} after=${after}`);
|
|
|
|
+ assert(after < before * 3, `before=${before} after=${after}`);
|
|
|
|
}
|