Don't need buffer for piping data
This commit is contained in:
parent
af05f26a5f
commit
81db8e098e
3 changed files with 26 additions and 47 deletions
|
@ -5,7 +5,6 @@
|
||||||
#include "atom/browser/net/url_request_fetch_job.h"
|
#include "atom/browser/net/url_request_fetch_job.h"
|
||||||
|
|
||||||
#include <algorithm>
|
#include <algorithm>
|
||||||
#include <vector>
|
|
||||||
|
|
||||||
#include "atom/browser/atom_browser_context.h"
|
#include "atom/browser/atom_browser_context.h"
|
||||||
#include "net/base/io_buffer.h"
|
#include "net/base/io_buffer.h"
|
||||||
|
@ -32,14 +31,13 @@ class ResponsePiper : public net::URLFetcherResponseWriter {
|
||||||
int Write(net::IOBuffer* buffer,
|
int Write(net::IOBuffer* buffer,
|
||||||
int num_bytes,
|
int num_bytes,
|
||||||
const net::CompletionCallback& callback) override {
|
const net::CompletionCallback& callback) override {
|
||||||
job_->DataAvailable(buffer, num_bytes);
|
|
||||||
if (first_write_) {
|
if (first_write_) {
|
||||||
// The URLFetcherResponseWriter doesn't have an event when headers have
|
// The URLFetcherResponseWriter doesn't have an event when headers have
|
||||||
// been read, so we have to emulate by hooking to first write event.
|
// been read, so we have to emulate by hooking to first write event.
|
||||||
job_->HeadersCompleted();
|
job_->HeadersCompleted();
|
||||||
first_write_ = false;
|
first_write_ = false;
|
||||||
}
|
}
|
||||||
return num_bytes;
|
return job_->DataAvailable(buffer, num_bytes);
|
||||||
}
|
}
|
||||||
int Finish(const net::CompletionCallback& callback) override {
|
int Finish(const net::CompletionCallback& callback) override {
|
||||||
return net::OK;
|
return net::OK;
|
||||||
|
@ -60,8 +58,7 @@ URLRequestFetchJob::URLRequestFetchJob(
|
||||||
const GURL& url)
|
const GURL& url)
|
||||||
: net::URLRequestJob(request, network_delegate),
|
: net::URLRequestJob(request, network_delegate),
|
||||||
url_(url),
|
url_(url),
|
||||||
finished_(false),
|
pending_buffer_size_(0) {}
|
||||||
weak_factory_(this) {}
|
|
||||||
|
|
||||||
void URLRequestFetchJob::HeadersCompleted() {
|
void URLRequestFetchJob::HeadersCompleted() {
|
||||||
response_info_.reset(new net::HttpResponseInfo);
|
response_info_.reset(new net::HttpResponseInfo);
|
||||||
|
@ -69,12 +66,21 @@ void URLRequestFetchJob::HeadersCompleted() {
|
||||||
NotifyHeadersComplete();
|
NotifyHeadersComplete();
|
||||||
}
|
}
|
||||||
|
|
||||||
void URLRequestFetchJob::DataAvailable(net::IOBuffer* buffer, int num_bytes) {
|
int URLRequestFetchJob::DataAvailable(net::IOBuffer* buffer, int num_bytes) {
|
||||||
buffer_.resize(buffer_.size() + num_bytes);
|
// Clear the IO_PENDING status.
|
||||||
memcpy(buffer_.data() + buffer_.size() - num_bytes,
|
|
||||||
buffer->data(),
|
|
||||||
num_bytes);
|
|
||||||
SetStatus(net::URLRequestStatus());
|
SetStatus(net::URLRequestStatus());
|
||||||
|
// Do nothing if pending_buffer_ is empty, i.e. there's no ReadRawData()
|
||||||
|
// operation waiting for IO completion.
|
||||||
|
if (!pending_buffer_.get())
|
||||||
|
return net::ERR_IO_PENDING;;
|
||||||
|
|
||||||
|
// pending_buffer_ is set to the IOBuffer instance provided to ReadRawData()
|
||||||
|
// by URLRequestJob.
|
||||||
|
|
||||||
|
int bytes_read = std::min(num_bytes, pending_buffer_size_);
|
||||||
|
memcpy(pending_buffer_->data(), buffer->data(), bytes_read);
|
||||||
|
NotifyReadComplete(bytes_read);
|
||||||
|
return bytes_read;
|
||||||
}
|
}
|
||||||
|
|
||||||
void URLRequestFetchJob::Start() {
|
void URLRequestFetchJob::Start() {
|
||||||
|
@ -93,29 +99,10 @@ void URLRequestFetchJob::Kill() {
|
||||||
bool URLRequestFetchJob::ReadRawData(net::IOBuffer* dest,
|
bool URLRequestFetchJob::ReadRawData(net::IOBuffer* dest,
|
||||||
int dest_size,
|
int dest_size,
|
||||||
int* bytes_read) {
|
int* bytes_read) {
|
||||||
if (finished_) {
|
pending_buffer_ = dest;
|
||||||
*bytes_read = 0;
|
pending_buffer_size_ = dest_size;
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (buffer_.size() == 0) {
|
|
||||||
SetStatus(net::URLRequestStatus(net::URLRequestStatus::IO_PENDING, 0));
|
SetStatus(net::URLRequestStatus(net::URLRequestStatus::IO_PENDING, 0));
|
||||||
return false;
|
return false;
|
||||||
}
|
|
||||||
|
|
||||||
if (static_cast<size_t>(dest_size) >= buffer_.size()) {
|
|
||||||
// Copy all data at once (quick).
|
|
||||||
memcpy(dest->data(), buffer_.data(), buffer_.size());
|
|
||||||
*bytes_read = buffer_.size();
|
|
||||||
buffer_.clear();
|
|
||||||
} else {
|
|
||||||
// Can not fit all data, strip them (slow).
|
|
||||||
memcpy(dest->data(), buffer_.data(), dest_size);
|
|
||||||
*bytes_read = dest_size;
|
|
||||||
std::rotate(buffer_.begin(), buffer_.begin() + dest_size, buffer_.end());
|
|
||||||
buffer_.resize(buffer_.size() - dest_size);
|
|
||||||
}
|
|
||||||
return true;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
bool URLRequestFetchJob::GetMimeType(std::string* mime_type) const {
|
bool URLRequestFetchJob::GetMimeType(std::string* mime_type) const {
|
||||||
|
@ -138,7 +125,8 @@ int URLRequestFetchJob::GetResponseCode() const {
|
||||||
}
|
}
|
||||||
|
|
||||||
void URLRequestFetchJob::OnURLFetchComplete(const net::URLFetcher* source) {
|
void URLRequestFetchJob::OnURLFetchComplete(const net::URLFetcher* source) {
|
||||||
finished_ = true;
|
NotifyDone(fetcher_->GetStatus());
|
||||||
|
NotifyReadComplete(0);
|
||||||
}
|
}
|
||||||
|
|
||||||
} // namespace atom
|
} // namespace atom
|
||||||
|
|
|
@ -5,9 +5,6 @@
|
||||||
#ifndef ATOM_BROWSER_NET_URL_REQUEST_FETCH_JOB_H_
|
#ifndef ATOM_BROWSER_NET_URL_REQUEST_FETCH_JOB_H_
|
||||||
#define ATOM_BROWSER_NET_URL_REQUEST_FETCH_JOB_H_
|
#define ATOM_BROWSER_NET_URL_REQUEST_FETCH_JOB_H_
|
||||||
|
|
||||||
#include <vector>
|
|
||||||
|
|
||||||
#include "base/memory/weak_ptr.h"
|
|
||||||
#include "net/url_request/url_fetcher_delegate.h"
|
#include "net/url_request/url_fetcher_delegate.h"
|
||||||
#include "net/url_request/url_request_job.h"
|
#include "net/url_request/url_request_job.h"
|
||||||
|
|
||||||
|
@ -20,12 +17,8 @@ class URLRequestFetchJob : public net::URLRequestJob,
|
||||||
net::NetworkDelegate* network_delegate,
|
net::NetworkDelegate* network_delegate,
|
||||||
const GURL& url);
|
const GURL& url);
|
||||||
|
|
||||||
base::WeakPtr<URLRequestFetchJob> GetWeakPtr() {
|
|
||||||
return weak_factory_.GetWeakPtr();
|
|
||||||
};
|
|
||||||
|
|
||||||
void HeadersCompleted();
|
void HeadersCompleted();
|
||||||
void DataAvailable(net::IOBuffer* buffer, int num_bytes);
|
int DataAvailable(net::IOBuffer* buffer, int num_bytes);
|
||||||
|
|
||||||
// net::URLRequestJob:
|
// net::URLRequestJob:
|
||||||
void Start() override;
|
void Start() override;
|
||||||
|
@ -43,11 +36,9 @@ class URLRequestFetchJob : public net::URLRequestJob,
|
||||||
private:
|
private:
|
||||||
GURL url_;
|
GURL url_;
|
||||||
scoped_ptr<net::URLFetcher> fetcher_;
|
scoped_ptr<net::URLFetcher> fetcher_;
|
||||||
std::vector<char> buffer_;
|
scoped_refptr<net::IOBuffer> pending_buffer_;
|
||||||
|
int pending_buffer_size_;
|
||||||
scoped_ptr<net::HttpResponseInfo> response_info_;
|
scoped_ptr<net::HttpResponseInfo> response_info_;
|
||||||
bool finished_;
|
|
||||||
|
|
||||||
base::WeakPtrFactory<URLRequestFetchJob> weak_factory_;
|
|
||||||
|
|
||||||
DISALLOW_COPY_AND_ASSIGN(URLRequestFetchJob);
|
DISALLOW_COPY_AND_ASSIGN(URLRequestFetchJob);
|
||||||
};
|
};
|
||||||
|
|
|
@ -82,7 +82,7 @@ describe 'protocol module', ->
|
||||||
server.listen 0, '127.0.0.1', ->
|
server.listen 0, '127.0.0.1', ->
|
||||||
{port} = server.address()
|
{port} = server.address()
|
||||||
url = "http://127.0.0.1:#{port}"
|
url = "http://127.0.0.1:#{port}"
|
||||||
job = new protocol.RequestHttpJob(url)
|
job = new protocol.RequestHttpJob({url})
|
||||||
handler = remote.createFunctionWithReturnValue job
|
handler = remote.createFunctionWithReturnValue job
|
||||||
protocol.registerProtocol 'atom-http-job', handler
|
protocol.registerProtocol 'atom-http-job', handler
|
||||||
|
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue