eaf2c61bef
* refactor: use gin in Promise * refactor: separate Promise impl that returns nothing * refactor: use Promise<void> for promise that returns nothing * fix: methods should be able to run on both browser and renderer process * fix: should not pass base::StringPiece across threads * refactor: no more need to use different ResolvePromise for empty Promise * refactor: move Promise to gin_helper
237 lines
9.8 KiB
C++
237 lines
9.8 KiB
C++
// Copyright (c) 2013 GitHub, Inc.
|
|
// Use of this source code is governed by the MIT license that can be
|
|
// found in the LICENSE file.
|
|
|
|
#include <string>
|
|
|
|
#include "base/task/post_task.h"
|
|
#include "base/values.h"
|
|
#include "content/public/renderer/render_frame.h"
|
|
#include "gin/dictionary.h"
|
|
#include "gin/handle.h"
|
|
#include "gin/object_template_builder.h"
|
|
#include "gin/wrappable.h"
|
|
#include "services/service_manager/public/cpp/interface_provider.h"
|
|
#include "shell/common/api/api.mojom.h"
|
|
#include "shell/common/gin_converters/blink_converter.h"
|
|
#include "shell/common/gin_converters/value_converter.h"
|
|
#include "shell/common/gin_helper/promise.h"
|
|
#include "shell/common/node_bindings.h"
|
|
#include "shell/common/node_includes.h"
|
|
#include "third_party/blink/public/web/web_local_frame.h"
|
|
|
|
using blink::WebLocalFrame;
|
|
using content::RenderFrame;
|
|
|
|
namespace {
|
|
|
|
RenderFrame* GetCurrentRenderFrame() {
|
|
WebLocalFrame* frame = WebLocalFrame::FrameForCurrentContext();
|
|
if (!frame)
|
|
return nullptr;
|
|
|
|
return RenderFrame::FromWebFrame(frame);
|
|
}
|
|
|
|
class IPCRenderer : public gin::Wrappable<IPCRenderer> {
|
|
public:
|
|
static gin::WrapperInfo kWrapperInfo;
|
|
|
|
static gin::Handle<IPCRenderer> Create(v8::Isolate* isolate) {
|
|
return gin::CreateHandle(isolate, new IPCRenderer(isolate));
|
|
}
|
|
|
|
explicit IPCRenderer(v8::Isolate* isolate)
|
|
: task_runner_(base::CreateSingleThreadTaskRunner({base::ThreadPool()})) {
|
|
RenderFrame* render_frame = GetCurrentRenderFrame();
|
|
DCHECK(render_frame);
|
|
|
|
// Bind the interface on the background runner. All accesses will be via
|
|
// the thread-safe pointer. This is to support our "fake-sync"
|
|
// MessageSync() hack; see the comment in IPCRenderer::SendSync.
|
|
electron::mojom::ElectronBrowserPtrInfo info;
|
|
render_frame->GetRemoteInterfaces()->GetInterface(mojo::MakeRequest(&info));
|
|
electron_browser_ptr_ =
|
|
electron::mojom::ThreadSafeElectronBrowserPtr::Create(std::move(info),
|
|
task_runner_);
|
|
}
|
|
|
|
// gin::Wrappable:
|
|
gin::ObjectTemplateBuilder GetObjectTemplateBuilder(
|
|
v8::Isolate* isolate) override {
|
|
return gin::Wrappable<IPCRenderer>::GetObjectTemplateBuilder(isolate)
|
|
.SetMethod("send", &IPCRenderer::Send)
|
|
.SetMethod("sendSync", &IPCRenderer::SendSync)
|
|
.SetMethod("sendTo", &IPCRenderer::SendTo)
|
|
.SetMethod("sendToHost", &IPCRenderer::SendToHost)
|
|
.SetMethod("invoke", &IPCRenderer::Invoke);
|
|
}
|
|
|
|
const char* GetTypeName() override { return "IPCRenderer"; }
|
|
|
|
private:
|
|
void Send(v8::Isolate* isolate,
|
|
bool internal,
|
|
const std::string& channel,
|
|
v8::Local<v8::Value> arguments) {
|
|
blink::CloneableMessage message;
|
|
if (!gin::ConvertFromV8(isolate, arguments, &message)) {
|
|
return;
|
|
}
|
|
electron_browser_ptr_->get()->Message(internal, channel,
|
|
std::move(message));
|
|
}
|
|
|
|
v8::Local<v8::Promise> Invoke(v8::Isolate* isolate,
|
|
bool internal,
|
|
const std::string& channel,
|
|
v8::Local<v8::Value> arguments) {
|
|
blink::CloneableMessage message;
|
|
if (!gin::ConvertFromV8(isolate, arguments, &message)) {
|
|
return v8::Local<v8::Promise>();
|
|
}
|
|
gin_helper::Promise<blink::CloneableMessage> p(isolate);
|
|
auto handle = p.GetHandle();
|
|
|
|
electron_browser_ptr_->get()->Invoke(
|
|
internal, channel, std::move(message),
|
|
base::BindOnce(
|
|
[](gin_helper::Promise<blink::CloneableMessage> p,
|
|
blink::CloneableMessage result) { p.Resolve(result); },
|
|
std::move(p)));
|
|
|
|
return handle;
|
|
}
|
|
|
|
void SendTo(v8::Isolate* isolate,
|
|
bool internal,
|
|
bool send_to_all,
|
|
int32_t web_contents_id,
|
|
const std::string& channel,
|
|
v8::Local<v8::Value> arguments) {
|
|
blink::CloneableMessage message;
|
|
if (!gin::ConvertFromV8(isolate, arguments, &message)) {
|
|
return;
|
|
}
|
|
electron_browser_ptr_->get()->MessageTo(
|
|
internal, send_to_all, web_contents_id, channel, std::move(message));
|
|
}
|
|
|
|
void SendToHost(v8::Isolate* isolate,
|
|
const std::string& channel,
|
|
v8::Local<v8::Value> arguments) {
|
|
blink::CloneableMessage message;
|
|
if (!gin::ConvertFromV8(isolate, arguments, &message)) {
|
|
return;
|
|
}
|
|
electron_browser_ptr_->get()->MessageHost(channel, std::move(message));
|
|
}
|
|
|
|
blink::CloneableMessage SendSync(v8::Isolate* isolate,
|
|
bool internal,
|
|
const std::string& channel,
|
|
v8::Local<v8::Value> arguments) {
|
|
blink::CloneableMessage message;
|
|
if (!gin::ConvertFromV8(isolate, arguments, &message)) {
|
|
return blink::CloneableMessage();
|
|
}
|
|
// We aren't using a true synchronous mojo call here. We're calling an
|
|
// asynchronous method and blocking on the result. The reason we're doing
|
|
// this is a little complicated, so buckle up.
|
|
//
|
|
// Mojo has a concept of synchronous calls. However, synchronous calls are
|
|
// dangerous. In particular, it's quite possible for two processes to call
|
|
// synchronous methods on each other and cause a deadlock. Mojo has a
|
|
// mechanism to avoid this kind of deadlock: if a process is waiting on the
|
|
// result of a synchronous call, and it receives an incoming call for a
|
|
// synchronous method, it will process that request immediately, even
|
|
// though it's currently blocking. However, if it receives an incoming
|
|
// request for an _asynchronous_ method, that can't cause a deadlock, so it
|
|
// stashes the request on a queue to be processed once the synchronous
|
|
// thing it's waiting on returns.
|
|
//
|
|
// This behavior is useful for preventing deadlocks, but it is inconvenient
|
|
// here because it can result in messages being reordered. If the main
|
|
// process is awaiting the result of a synchronous call (which it does only
|
|
// very rarely, since it's bad to block the main process), and we send
|
|
// first an asynchronous message to the main process, followed by a
|
|
// synchronous message, then the main process will process the synchronous
|
|
// one first.
|
|
//
|
|
// It turns out, Electron has some dependency on message ordering,
|
|
// especially during window shutdown, and getting messages out of order can
|
|
// result in, for example, remote objects disappearing unexpectedly. To
|
|
// avoid these issues and guarantee consistent message ordering, we send
|
|
// all messages to the main process as asynchronous messages. This causes
|
|
// them to always be queued and processed in the same order they were
|
|
// received, even if they were received while the main process was waiting
|
|
// on a synchronous call.
|
|
//
|
|
// However, in the calling process, we still need to block on the result,
|
|
// because the caller is expecting a result synchronously. So we do a bit
|
|
// of a trick: we pass the Mojo handle over to a worker thread, send the
|
|
// asynchronous message from that thread, and then block on the result.
|
|
// It's important that we pass the handle over to the worker thread,
|
|
// because that allows Mojo to process incoming messages (most importantly,
|
|
// the response to our request) on that thread. If we didn't pass it to a
|
|
// worker thread, and instead sent the call from the main thread, we would
|
|
// never receive a response because Mojo wouldn't be able to run its
|
|
// message handling code, because the main thread would be tied up blocking
|
|
// on the WaitableEvent.
|
|
//
|
|
// Phew. If you got this far, here's a gold star: ⭐️
|
|
|
|
blink::CloneableMessage result;
|
|
|
|
// A task is posted to a worker thread to execute the request so that
|
|
// this thread may block on a waitable event. It is safe to pass raw
|
|
// pointers to |result| and |response_received_event| as this stack frame
|
|
// will survive until the request is complete.
|
|
|
|
base::WaitableEvent response_received_event;
|
|
task_runner_->PostTask(
|
|
FROM_HERE, base::BindOnce(&IPCRenderer::SendMessageSyncOnWorkerThread,
|
|
base::Unretained(this),
|
|
base::Unretained(&response_received_event),
|
|
base::Unretained(&result), internal, channel,
|
|
std::move(message)));
|
|
response_received_event.Wait();
|
|
return result;
|
|
}
|
|
|
|
void SendMessageSyncOnWorkerThread(base::WaitableEvent* event,
|
|
blink::CloneableMessage* result,
|
|
bool internal,
|
|
const std::string& channel,
|
|
blink::CloneableMessage arguments) {
|
|
electron_browser_ptr_->get()->MessageSync(
|
|
internal, channel, std::move(arguments),
|
|
base::BindOnce(&IPCRenderer::ReturnSyncResponseToMainThread,
|
|
base::Unretained(event), base::Unretained(result)));
|
|
}
|
|
|
|
static void ReturnSyncResponseToMainThread(base::WaitableEvent* event,
|
|
blink::CloneableMessage* result,
|
|
blink::CloneableMessage response) {
|
|
*result = std::move(response);
|
|
event->Signal();
|
|
}
|
|
|
|
scoped_refptr<base::SequencedTaskRunner> task_runner_;
|
|
scoped_refptr<electron::mojom::ThreadSafeElectronBrowserPtr>
|
|
electron_browser_ptr_;
|
|
};
|
|
|
|
gin::WrapperInfo IPCRenderer::kWrapperInfo = {gin::kEmbedderNativeGin};
|
|
|
|
void Initialize(v8::Local<v8::Object> exports,
|
|
v8::Local<v8::Value> unused,
|
|
v8::Local<v8::Context> context,
|
|
void* priv) {
|
|
gin::Dictionary dict(context->GetIsolate(), exports);
|
|
dict.Set("ipc", IPCRenderer::Create(context->GetIsolate()));
|
|
}
|
|
|
|
} // namespace
|
|
|
|
NODE_LINKED_MODULE_CONTEXT_AWARE(atom_renderer_ipc, Initialize)
|