Merge pull request #5769 from electron/fetch-job-cleanup
Refactor the URLRequestFetchJob code
This commit is contained in:
commit
0c93b21a5f
3 changed files with 61 additions and 59 deletions
|
@ -77,7 +77,8 @@ class ResponsePiper : public net::URLFetcherResponseWriter {
|
||||||
URLRequestFetchJob::URLRequestFetchJob(
|
URLRequestFetchJob::URLRequestFetchJob(
|
||||||
net::URLRequest* request, net::NetworkDelegate* network_delegate)
|
net::URLRequest* request, net::NetworkDelegate* network_delegate)
|
||||||
: JsAsker<net::URLRequestJob>(request, network_delegate),
|
: JsAsker<net::URLRequestJob>(request, network_delegate),
|
||||||
pending_buffer_size_(0) {
|
pending_buffer_size_(0),
|
||||||
|
write_num_bytes_(0) {
|
||||||
}
|
}
|
||||||
|
|
||||||
void URLRequestFetchJob::BeforeStartInUI(
|
void URLRequestFetchJob::BeforeStartInUI(
|
||||||
|
@ -169,22 +170,21 @@ void URLRequestFetchJob::HeadersCompleted() {
|
||||||
int URLRequestFetchJob::DataAvailable(net::IOBuffer* buffer,
|
int URLRequestFetchJob::DataAvailable(net::IOBuffer* buffer,
|
||||||
int num_bytes,
|
int num_bytes,
|
||||||
const net::CompletionCallback& callback) {
|
const net::CompletionCallback& callback) {
|
||||||
// Store the data in a drainable IO buffer if pending_buffer_ is empty, i.e.
|
// When pending_buffer_ is empty, there's no ReadRawData() operation waiting
|
||||||
// there's no ReadRawData() operation waiting for IO completion.
|
// for IO completion, we have to save the parameters until the request is
|
||||||
|
// ready to read data.
|
||||||
if (!pending_buffer_.get()) {
|
if (!pending_buffer_.get()) {
|
||||||
response_piper_callback_ = callback;
|
write_buffer_ = buffer;
|
||||||
drainable_buffer_ = new net::DrainableIOBuffer(buffer, num_bytes);
|
write_num_bytes_ = num_bytes;
|
||||||
|
write_callback_ = callback;
|
||||||
return net::ERR_IO_PENDING;
|
return net::ERR_IO_PENDING;
|
||||||
}
|
}
|
||||||
|
|
||||||
// pending_buffer_ is set to the IOBuffer instance provided to ReadRawData()
|
// Write data to the pending buffer and clear them after the writing.
|
||||||
// by URLRequestJob.
|
int bytes_read = BufferCopy(buffer, num_bytes,
|
||||||
int bytes_read = std::min(num_bytes, pending_buffer_size_);
|
pending_buffer_.get(), pending_buffer_size_);
|
||||||
memcpy(pending_buffer_->data(), buffer->data(), bytes_read);
|
|
||||||
|
|
||||||
ClearPendingBuffer();
|
|
||||||
|
|
||||||
ReadRawDataComplete(bytes_read);
|
ReadRawDataComplete(bytes_read);
|
||||||
|
ClearPendingBuffer();
|
||||||
return bytes_read;
|
return bytes_read;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -193,46 +193,26 @@ void URLRequestFetchJob::Kill() {
|
||||||
fetcher_.reset();
|
fetcher_.reset();
|
||||||
}
|
}
|
||||||
|
|
||||||
void URLRequestFetchJob::StartReading() {
|
|
||||||
if (drainable_buffer_.get()) {
|
|
||||||
auto num_bytes = drainable_buffer_->BytesRemaining();
|
|
||||||
if (num_bytes <= 0 && !response_piper_callback_.is_null()) {
|
|
||||||
int size = drainable_buffer_->BytesConsumed();
|
|
||||||
drainable_buffer_ = nullptr;
|
|
||||||
response_piper_callback_.Run(size);
|
|
||||||
response_piper_callback_.Reset();
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
int bytes_read = std::min(num_bytes, pending_buffer_size_);
|
|
||||||
memcpy(pending_buffer_->data(), drainable_buffer_->data(), bytes_read);
|
|
||||||
|
|
||||||
drainable_buffer_->DidConsume(bytes_read);
|
|
||||||
|
|
||||||
ClearPendingBuffer();
|
|
||||||
|
|
||||||
ReadRawDataComplete(bytes_read);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
void URLRequestFetchJob::ClearPendingBuffer() {
|
|
||||||
// Clear the buffers before notifying the read is complete, so that it is
|
|
||||||
// safe for the observer to read.
|
|
||||||
pending_buffer_ = nullptr;
|
|
||||||
pending_buffer_size_ = 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
int URLRequestFetchJob::ReadRawData(net::IOBuffer* dest, int dest_size) {
|
int URLRequestFetchJob::ReadRawData(net::IOBuffer* dest, int dest_size) {
|
||||||
if (GetResponseCode() == 204) {
|
if (GetResponseCode() == 204) {
|
||||||
request()->set_received_response_content_length(prefilter_bytes_read());
|
request()->set_received_response_content_length(prefilter_bytes_read());
|
||||||
return net::OK;
|
return net::OK;
|
||||||
}
|
}
|
||||||
pending_buffer_ = dest;
|
|
||||||
pending_buffer_size_ = dest_size;
|
// When write_buffer_ is empty, there is no data valable yet, we have to save
|
||||||
// Read from drainable_buffer_ if available, i.e. response data became
|
// the dest buffer util DataAvailable.
|
||||||
// available before ReadRawData was called.
|
if (!write_buffer_.get()) {
|
||||||
StartReading();
|
pending_buffer_ = dest;
|
||||||
return net::ERR_IO_PENDING;
|
pending_buffer_size_ = dest_size;
|
||||||
|
return net::ERR_IO_PENDING;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Read from the write buffer and clear them after reading.
|
||||||
|
int bytes_read = BufferCopy(write_buffer_.get(), write_num_bytes_,
|
||||||
|
dest, dest_size);
|
||||||
|
write_callback_.Run(bytes_read);
|
||||||
|
ClearWriteBuffer();
|
||||||
|
return bytes_read;
|
||||||
}
|
}
|
||||||
|
|
||||||
bool URLRequestFetchJob::GetMimeType(std::string* mime_type) const {
|
bool URLRequestFetchJob::GetMimeType(std::string* mime_type) const {
|
||||||
|
@ -264,6 +244,7 @@ void URLRequestFetchJob::OnURLFetchComplete(const net::URLFetcher* source) {
|
||||||
}
|
}
|
||||||
|
|
||||||
ClearPendingBuffer();
|
ClearPendingBuffer();
|
||||||
|
ClearWriteBuffer();
|
||||||
|
|
||||||
if (fetcher_->GetStatus().is_success())
|
if (fetcher_->GetStatus().is_success())
|
||||||
ReadRawDataComplete(0);
|
ReadRawDataComplete(0);
|
||||||
|
@ -271,4 +252,22 @@ void URLRequestFetchJob::OnURLFetchComplete(const net::URLFetcher* source) {
|
||||||
NotifyStartError(fetcher_->GetStatus());
|
NotifyStartError(fetcher_->GetStatus());
|
||||||
}
|
}
|
||||||
|
|
||||||
|
int URLRequestFetchJob::BufferCopy(net::IOBuffer* source, int num_bytes,
|
||||||
|
net::IOBuffer* target, int target_size) {
|
||||||
|
int bytes_written = std::min(num_bytes, target_size);
|
||||||
|
memcpy(target->data(), source->data(), bytes_written);
|
||||||
|
return bytes_written;
|
||||||
|
}
|
||||||
|
|
||||||
|
void URLRequestFetchJob::ClearPendingBuffer() {
|
||||||
|
pending_buffer_ = nullptr;
|
||||||
|
pending_buffer_size_ = 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
void URLRequestFetchJob::ClearWriteBuffer() {
|
||||||
|
write_buffer_ = nullptr;
|
||||||
|
write_num_bytes_ = 0;
|
||||||
|
write_callback_.Reset();
|
||||||
|
}
|
||||||
|
|
||||||
} // namespace atom
|
} // namespace atom
|
||||||
|
|
|
@ -9,14 +9,10 @@
|
||||||
|
|
||||||
#include "atom/browser/net/js_asker.h"
|
#include "atom/browser/net/js_asker.h"
|
||||||
#include "browser/url_request_context_getter.h"
|
#include "browser/url_request_context_getter.h"
|
||||||
#include "net/url_request/url_request_context_getter.h"
|
|
||||||
#include "net/url_request/url_fetcher_delegate.h"
|
#include "net/url_request/url_fetcher_delegate.h"
|
||||||
#include "net/url_request/url_request_job.h"
|
|
||||||
|
|
||||||
namespace atom {
|
namespace atom {
|
||||||
|
|
||||||
class AtomBrowserContext;
|
|
||||||
|
|
||||||
class URLRequestFetchJob : public JsAsker<net::URLRequestJob>,
|
class URLRequestFetchJob : public JsAsker<net::URLRequestJob>,
|
||||||
public net::URLFetcherDelegate,
|
public net::URLFetcherDelegate,
|
||||||
public brightray::URLRequestContextGetter::Delegate {
|
public brightray::URLRequestContextGetter::Delegate {
|
||||||
|
@ -30,9 +26,6 @@ class URLRequestFetchJob : public JsAsker<net::URLRequestJob>,
|
||||||
const net::CompletionCallback& callback);
|
const net::CompletionCallback& callback);
|
||||||
|
|
||||||
protected:
|
protected:
|
||||||
void StartReading();
|
|
||||||
void ClearPendingBuffer();
|
|
||||||
|
|
||||||
// JsAsker:
|
// JsAsker:
|
||||||
void BeforeStartInUI(v8::Isolate*, v8::Local<v8::Value>) override;
|
void BeforeStartInUI(v8::Isolate*, v8::Local<v8::Value>) override;
|
||||||
void StartAsync(std::unique_ptr<base::Value> options) override;
|
void StartAsync(std::unique_ptr<base::Value> options) override;
|
||||||
|
@ -48,13 +41,23 @@ class URLRequestFetchJob : public JsAsker<net::URLRequestJob>,
|
||||||
void OnURLFetchComplete(const net::URLFetcher* source) override;
|
void OnURLFetchComplete(const net::URLFetcher* source) override;
|
||||||
|
|
||||||
private:
|
private:
|
||||||
|
int BufferCopy(net::IOBuffer* source, int num_bytes,
|
||||||
|
net::IOBuffer* target, int target_size);
|
||||||
|
void ClearPendingBuffer();
|
||||||
|
void ClearWriteBuffer();
|
||||||
|
|
||||||
scoped_refptr<net::URLRequestContextGetter> url_request_context_getter_;
|
scoped_refptr<net::URLRequestContextGetter> url_request_context_getter_;
|
||||||
std::unique_ptr<net::URLFetcher> fetcher_;
|
std::unique_ptr<net::URLFetcher> fetcher_;
|
||||||
scoped_refptr<net::IOBuffer> pending_buffer_;
|
|
||||||
scoped_refptr<net::DrainableIOBuffer> drainable_buffer_;
|
|
||||||
int pending_buffer_size_;
|
|
||||||
std::unique_ptr<net::HttpResponseInfo> response_info_;
|
std::unique_ptr<net::HttpResponseInfo> response_info_;
|
||||||
net::CompletionCallback response_piper_callback_;
|
|
||||||
|
// Saved arguments passed to ReadRawData.
|
||||||
|
scoped_refptr<net::IOBuffer> pending_buffer_;
|
||||||
|
int pending_buffer_size_;
|
||||||
|
|
||||||
|
// Saved arguments passed to DataAvailable.
|
||||||
|
scoped_refptr<net::IOBuffer> write_buffer_;
|
||||||
|
int write_num_bytes_;
|
||||||
|
net::CompletionCallback write_callback_;
|
||||||
|
|
||||||
DISALLOW_COPY_AND_ASSIGN(URLRequestFetchJob);
|
DISALLOW_COPY_AND_ASSIGN(URLRequestFetchJob);
|
||||||
};
|
};
|
||||||
|
|
|
@ -511,7 +511,7 @@ describe('protocol module', function () {
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
||||||
it('loads resource for webContents', function (done) {
|
it('works when target URL redirects', function (done) {
|
||||||
var contents = null
|
var contents = null
|
||||||
var server = http.createServer(function (req, res) {
|
var server = http.createServer(function (req, res) {
|
||||||
if (req.url == '/serverRedirect') {
|
if (req.url == '/serverRedirect') {
|
||||||
|
|
Loading…
Reference in a new issue