629 lines
24 KiB
C++
629 lines
24 KiB
C++
#ifndef CLIENT_HTTP_HPP
|
|
#define CLIENT_HTTP_HPP
|
|
|
|
#include "utility.hpp"
|
|
#include <mutex>
|
|
#include <random>
|
|
#include <unordered_set>
|
|
#include <vector>
|
|
|
|
#ifdef USE_STANDALONE_ASIO
|
|
#include <asio.hpp>
|
|
namespace SimpleWeb {
|
|
using error_code = std::error_code;
|
|
using errc = std::errc;
|
|
using system_error = std::system_error;
|
|
namespace make_error_code = std;
|
|
using string_view = const std::string &; // TODO c++17: use std::string_view
|
|
} // namespace SimpleWeb
|
|
#else
|
|
#include <boost/asio.hpp>
|
|
#include <boost/utility/string_ref.hpp>
|
|
namespace SimpleWeb {
|
|
namespace asio = boost::asio;
|
|
using error_code = boost::system::error_code;
|
|
namespace errc = boost::system::errc;
|
|
using system_error = boost::system::system_error;
|
|
namespace make_error_code = boost::system::errc;
|
|
using string_view = boost::string_ref;
|
|
} // namespace SimpleWeb
|
|
#endif
|
|
|
|
namespace SimpleWeb {
|
|
template <class socket_type>
|
|
class Client;
|
|
|
|
template <class socket_type>
|
|
class ClientBase {
|
|
public:
|
|
class Content : public std::istream {
|
|
friend class ClientBase<socket_type>;
|
|
|
|
public:
|
|
size_t size() {
|
|
return streambuf.size();
|
|
}
|
|
/// Convenience function to return std::string. Note that the stream buffer is emptied when this functions is used.
|
|
std::string string() {
|
|
std::stringstream ss;
|
|
ss << rdbuf();
|
|
return ss.str();
|
|
}
|
|
|
|
private:
|
|
asio::streambuf &streambuf;
|
|
Content(asio::streambuf &streambuf) : std::istream(&streambuf), streambuf(streambuf) {}
|
|
};
|
|
|
|
class Response {
|
|
friend class ClientBase<socket_type>;
|
|
friend class Client<socket_type>;
|
|
|
|
public:
|
|
std::string http_version, status_code;
|
|
|
|
Content content;
|
|
|
|
CaseInsensitiveMultimap header;
|
|
|
|
private:
|
|
asio::streambuf content_buffer;
|
|
|
|
Response() : content(content_buffer) {}
|
|
};
|
|
|
|
class Config {
|
|
friend class ClientBase<socket_type>;
|
|
|
|
private:
|
|
Config() {}
|
|
|
|
public:
|
|
/// Set timeout on requests in seconds. Default value: 0 (no timeout).
|
|
long timeout = 0;
|
|
/// Set connect timeout in seconds. Default value: 0 (Config::timeout is then used instead).
|
|
long timeout_connect = 0;
|
|
/// Set proxy server (server:port)
|
|
std::string proxy_server;
|
|
};
|
|
|
|
protected:
|
|
class Connection : public std::enable_shared_from_this<Connection> {
|
|
public:
|
|
template <typename... Args>
|
|
Connection(std::shared_ptr<ScopeRunner> handler_runner, long timeout, Args &&... args)
|
|
: handler_runner(std::move(handler_runner)), timeout(timeout), socket(new socket_type(std::forward<Args>(args)...)) {}
|
|
|
|
std::shared_ptr<ScopeRunner> handler_runner;
|
|
long timeout;
|
|
|
|
std::unique_ptr<socket_type> socket; // Socket must be unique_ptr since asio::ssl::stream<asio::ip::tcp::socket> is not movable
|
|
bool in_use = false;
|
|
bool attempt_reconnect = true;
|
|
|
|
std::unique_ptr<asio::deadline_timer> timer;
|
|
|
|
void set_timeout(long seconds = 0) {
|
|
if(seconds == 0)
|
|
seconds = timeout;
|
|
if(seconds == 0) {
|
|
timer = nullptr;
|
|
return;
|
|
}
|
|
timer = std::unique_ptr<asio::deadline_timer>(new asio::deadline_timer(socket->get_io_service()));
|
|
timer->expires_from_now(boost::posix_time::seconds(seconds));
|
|
auto self = this->shared_from_this();
|
|
timer->async_wait([self](const error_code &ec) {
|
|
if(!ec) {
|
|
error_code ec;
|
|
self->socket->lowest_layer().cancel(ec);
|
|
}
|
|
});
|
|
}
|
|
|
|
void cancel_timeout() {
|
|
if(timer)
|
|
timer->cancel();
|
|
}
|
|
};
|
|
|
|
class Session {
|
|
public:
|
|
Session(std::shared_ptr<Connection> connection, std::unique_ptr<asio::streambuf> request_buffer)
|
|
: connection(std::move(connection)), request_buffer(std::move(request_buffer)), response(new Response()) {}
|
|
|
|
std::shared_ptr<Connection> connection;
|
|
std::unique_ptr<asio::streambuf> request_buffer;
|
|
std::shared_ptr<Response> response;
|
|
std::function<void(const std::shared_ptr<Connection> &, const error_code &)> callback;
|
|
};
|
|
|
|
public:
|
|
/// Set before calling request
|
|
Config config;
|
|
|
|
/// If you have your own asio::io_service, store its pointer here before calling request().
|
|
/// When using asynchronous requests, running the io_service is up to the programmer.
|
|
std::shared_ptr<asio::io_service> io_service;
|
|
|
|
/// Convenience function to perform synchronous request. The io_service is run within this function.
|
|
/// If reusing the io_service for other tasks, use the asynchronous request functions instead.
|
|
/// Do not use concurrently with the asynchronous request functions.
|
|
std::shared_ptr<Response> request(const std::string &method, const std::string &path = std::string("/"),
|
|
string_view content = "", const CaseInsensitiveMultimap &header = CaseInsensitiveMultimap()) {
|
|
std::shared_ptr<Response> response;
|
|
error_code ec;
|
|
request(method, path, content, header, [&response, &ec](std::shared_ptr<Response> response_, const error_code &ec_) {
|
|
response = response_;
|
|
ec = ec_;
|
|
});
|
|
|
|
{
|
|
std::unique_lock<std::mutex> lock(concurrent_synchronous_requests_mutex);
|
|
++concurrent_synchronous_requests;
|
|
}
|
|
io_service->run();
|
|
{
|
|
std::unique_lock<std::mutex> lock(concurrent_synchronous_requests_mutex);
|
|
--concurrent_synchronous_requests;
|
|
if(!concurrent_synchronous_requests)
|
|
io_service->reset();
|
|
}
|
|
|
|
if(ec)
|
|
throw system_error(ec);
|
|
|
|
return response;
|
|
}
|
|
|
|
/// Convenience function to perform synchronous request. The io_service is run within this function.
|
|
/// If reusing the io_service for other tasks, use the asynchronous request functions instead.
|
|
/// Do not use concurrently with the asynchronous request functions.
|
|
std::shared_ptr<Response> request(const std::string &method, const std::string &path, std::istream &content,
|
|
const CaseInsensitiveMultimap &header = CaseInsensitiveMultimap()) {
|
|
std::shared_ptr<Response> response;
|
|
error_code ec;
|
|
request(method, path, content, header, [&response, &ec](std::shared_ptr<Response> response_, const error_code &ec_) {
|
|
response = response_;
|
|
ec = ec_;
|
|
});
|
|
|
|
{
|
|
std::unique_lock<std::mutex> lock(concurrent_synchronous_requests_mutex);
|
|
++concurrent_synchronous_requests;
|
|
}
|
|
io_service->run();
|
|
{
|
|
std::unique_lock<std::mutex> lock(concurrent_synchronous_requests_mutex);
|
|
--concurrent_synchronous_requests;
|
|
if(!concurrent_synchronous_requests)
|
|
io_service->reset();
|
|
}
|
|
|
|
if(ec)
|
|
throw system_error(ec);
|
|
|
|
return response;
|
|
}
|
|
|
|
/// Asynchronous request where setting and/or running Client's io_service is required.
|
|
/// Do not use concurrently with the synchronous request functions.
|
|
void request(const std::string &method, const std::string &path, string_view content, const CaseInsensitiveMultimap &header,
|
|
std::function<void(std::shared_ptr<Response>, const error_code &)> &&request_callback_) {
|
|
auto session = std::make_shared<Session>(get_connection(), create_request_header(method, path, header));
|
|
auto response = session->response;
|
|
auto request_callback = std::make_shared<std::function<void(std::shared_ptr<Response>, const error_code &)>>(std::move(request_callback_));
|
|
session->callback = [this, response, request_callback](const std::shared_ptr<Connection> &connection, const error_code &ec) {
|
|
{
|
|
std::unique_lock<std::mutex> lock(this->connections_mutex);
|
|
connection->in_use = false;
|
|
|
|
// Remove unused connections, but keep one open for HTTP persistent connection:
|
|
size_t unused_connections = 0;
|
|
for(auto it = this->connections.begin(); it != this->connections.end();) {
|
|
if(ec && connection == *it)
|
|
it = this->connections.erase(it);
|
|
else if((*it)->in_use)
|
|
++it;
|
|
else {
|
|
++unused_connections;
|
|
if(unused_connections > 1)
|
|
it = this->connections.erase(it);
|
|
else
|
|
++it;
|
|
}
|
|
}
|
|
}
|
|
|
|
if(*request_callback)
|
|
(*request_callback)(response, ec);
|
|
};
|
|
|
|
std::ostream write_stream(session->request_buffer.get());
|
|
if(content.size() > 0)
|
|
write_stream << "Content-Length: " << content.size() << "\r\n";
|
|
write_stream << "\r\n"
|
|
<< content;
|
|
|
|
connect(session);
|
|
}
|
|
|
|
/// Asynchronous request where setting and/or running Client's io_service is required.
|
|
/// Do not use concurrently with the synchronous request functions.
|
|
void request(const std::string &method, const std::string &path, string_view content,
|
|
std::function<void(std::shared_ptr<Response>, const error_code &)> &&request_callback) {
|
|
request(method, path, content, CaseInsensitiveMultimap(), std::move(request_callback));
|
|
}
|
|
|
|
/// Asynchronous request where setting and/or running Client's io_service is required.
|
|
void request(const std::string &method, const std::string &path,
|
|
std::function<void(std::shared_ptr<Response>, const error_code &)> &&request_callback) {
|
|
request(method, path, std::string(), CaseInsensitiveMultimap(), std::move(request_callback));
|
|
}
|
|
|
|
/// Asynchronous request where setting and/or running Client's io_service is required.
|
|
void request(const std::string &method, std::function<void(std::shared_ptr<Response>, const error_code &)> &&request_callback) {
|
|
request(method, std::string("/"), std::string(), CaseInsensitiveMultimap(), std::move(request_callback));
|
|
}
|
|
|
|
/// Asynchronous request where setting and/or running Client's io_service is required.
|
|
void request(const std::string &method, const std::string &path, std::istream &content, const CaseInsensitiveMultimap &header,
|
|
std::function<void(std::shared_ptr<Response>, const error_code &)> &&request_callback_) {
|
|
auto session = std::make_shared<Session>(get_connection(), create_request_header(method, path, header));
|
|
auto response = session->response;
|
|
auto request_callback = std::make_shared<std::function<void(std::shared_ptr<Response>, const error_code &)>>(std::move(request_callback_));
|
|
session->callback = [this, response, request_callback](const std::shared_ptr<Connection> &connection, const error_code &ec) {
|
|
{
|
|
std::unique_lock<std::mutex> lock(this->connections_mutex);
|
|
connection->in_use = false;
|
|
|
|
// Remove unused connections, but keep one open for HTTP persistent connection:
|
|
size_t unused_connections = 0;
|
|
for(auto it = this->connections.begin(); it != this->connections.end();) {
|
|
if(ec && connection == *it)
|
|
it = this->connections.erase(it);
|
|
else if((*it)->in_use)
|
|
++it;
|
|
else {
|
|
++unused_connections;
|
|
if(unused_connections > 1)
|
|
it = this->connections.erase(it);
|
|
else
|
|
++it;
|
|
}
|
|
}
|
|
}
|
|
|
|
if(*request_callback)
|
|
(*request_callback)(response, ec);
|
|
};
|
|
|
|
content.seekg(0, std::ios::end);
|
|
auto content_length = content.tellg();
|
|
content.seekg(0, std::ios::beg);
|
|
std::ostream write_stream(session->request_buffer.get());
|
|
if(content_length > 0)
|
|
write_stream << "Content-Length: " << content_length << "\r\n";
|
|
write_stream << "\r\n";
|
|
if(content_length > 0)
|
|
write_stream << content.rdbuf();
|
|
|
|
connect(session);
|
|
}
|
|
|
|
/// Asynchronous request where setting and/or running Client's io_service is required.
|
|
void request(const std::string &method, const std::string &path, std::istream &content,
|
|
std::function<void(std::shared_ptr<Response>, const error_code &)> &&request_callback) {
|
|
request(method, path, content, CaseInsensitiveMultimap(), std::move(request_callback));
|
|
}
|
|
|
|
/// Close connections
|
|
void stop() {
|
|
std::unique_lock<std::mutex> lock(connections_mutex);
|
|
for(auto it = connections.begin(); it != connections.end();) {
|
|
error_code ec;
|
|
(*it)->socket->lowest_layer().cancel(ec);
|
|
it = connections.erase(it);
|
|
}
|
|
}
|
|
|
|
virtual ~ClientBase() {
|
|
handler_runner->stop();
|
|
stop();
|
|
}
|
|
|
|
protected:
|
|
bool internal_io_service = false;
|
|
|
|
std::string host;
|
|
unsigned short port;
|
|
|
|
std::unique_ptr<asio::ip::tcp::resolver::query> query;
|
|
|
|
std::unordered_set<std::shared_ptr<Connection>> connections;
|
|
std::mutex connections_mutex;
|
|
|
|
std::shared_ptr<ScopeRunner> handler_runner;
|
|
|
|
size_t concurrent_synchronous_requests = 0;
|
|
std::mutex concurrent_synchronous_requests_mutex;
|
|
|
|
ClientBase(const std::string &host_port, unsigned short default_port) : handler_runner(new ScopeRunner()) {
|
|
auto parsed_host_port = parse_host_port(host_port, default_port);
|
|
host = parsed_host_port.first;
|
|
port = parsed_host_port.second;
|
|
}
|
|
|
|
std::shared_ptr<Connection> get_connection() {
|
|
std::shared_ptr<Connection> connection;
|
|
std::unique_lock<std::mutex> lock(connections_mutex);
|
|
|
|
if(!io_service) {
|
|
io_service = std::make_shared<asio::io_service>();
|
|
internal_io_service = true;
|
|
}
|
|
|
|
for(auto it = connections.begin(); it != connections.end(); ++it) {
|
|
if(!(*it)->in_use && !connection) {
|
|
connection = *it;
|
|
break;
|
|
}
|
|
}
|
|
if(!connection) {
|
|
connection = create_connection();
|
|
connections.emplace(connection);
|
|
}
|
|
connection->attempt_reconnect = true;
|
|
connection->in_use = true;
|
|
|
|
if(!query) {
|
|
if(config.proxy_server.empty())
|
|
query = std::unique_ptr<asio::ip::tcp::resolver::query>(new asio::ip::tcp::resolver::query(host, std::to_string(port)));
|
|
else {
|
|
auto proxy_host_port = parse_host_port(config.proxy_server, 8080);
|
|
query = std::unique_ptr<asio::ip::tcp::resolver::query>(new asio::ip::tcp::resolver::query(proxy_host_port.first, std::to_string(proxy_host_port.second)));
|
|
}
|
|
}
|
|
|
|
return connection;
|
|
}
|
|
|
|
virtual std::shared_ptr<Connection> create_connection() = 0;
|
|
virtual void connect(const std::shared_ptr<Session> &) = 0;
|
|
|
|
std::unique_ptr<asio::streambuf> create_request_header(const std::string &method, const std::string &path, const CaseInsensitiveMultimap &header) const {
|
|
auto corrected_path = path;
|
|
if(corrected_path == "")
|
|
corrected_path = "/";
|
|
if(!config.proxy_server.empty() && std::is_same<socket_type, asio::ip::tcp::socket>::value)
|
|
corrected_path = "http://" + host + ':' + std::to_string(port) + corrected_path;
|
|
|
|
std::unique_ptr<asio::streambuf> request_buffer(new asio::streambuf());
|
|
std::ostream write_stream(request_buffer.get());
|
|
write_stream << method << " " << corrected_path << " HTTP/1.1\r\n";
|
|
write_stream << "Host: " << host << "\r\n";
|
|
for(auto &h : header)
|
|
write_stream << h.first << ": " << h.second << "\r\n";
|
|
return request_buffer;
|
|
}
|
|
|
|
std::pair<std::string, unsigned short> parse_host_port(const std::string &host_port, unsigned short default_port) const {
|
|
std::pair<std::string, unsigned short> parsed_host_port;
|
|
size_t host_end = host_port.find(':');
|
|
if(host_end == std::string::npos) {
|
|
parsed_host_port.first = host_port;
|
|
parsed_host_port.second = default_port;
|
|
}
|
|
else {
|
|
parsed_host_port.first = host_port.substr(0, host_end);
|
|
parsed_host_port.second = static_cast<unsigned short>(stoul(host_port.substr(host_end + 1)));
|
|
}
|
|
return parsed_host_port;
|
|
}
|
|
|
|
void write(const std::shared_ptr<Session> &session) {
|
|
session->connection->set_timeout();
|
|
asio::async_write(*session->connection->socket, session->request_buffer->data(), [this, session](const error_code &ec, size_t /*bytes_transferred*/) {
|
|
session->connection->cancel_timeout();
|
|
auto lock = session->connection->handler_runner->continue_lock();
|
|
if(!lock)
|
|
return;
|
|
if(!ec)
|
|
this->read(session);
|
|
else
|
|
session->callback(session->connection, ec);
|
|
});
|
|
}
|
|
|
|
void read(const std::shared_ptr<Session> &session) {
|
|
session->connection->set_timeout();
|
|
asio::async_read_until(*session->connection->socket, session->response->content_buffer, "\r\n\r\n", [this, session](const error_code &ec, size_t bytes_transferred) {
|
|
session->connection->cancel_timeout();
|
|
auto lock = session->connection->handler_runner->continue_lock();
|
|
if(!lock)
|
|
return;
|
|
if(!ec) {
|
|
session->connection->attempt_reconnect = true;
|
|
|
|
size_t num_additional_bytes = session->response->content_buffer.size() - bytes_transferred;
|
|
|
|
if(!ResponseMessage::parse(session->response->content, session->response->http_version, session->response->status_code, session->response->header)) {
|
|
session->callback(session->connection, make_error_code::make_error_code(errc::protocol_error));
|
|
return;
|
|
}
|
|
|
|
auto header_it = session->response->header.find("Content-Length");
|
|
if(header_it != session->response->header.end()) {
|
|
auto content_length = stoull(header_it->second);
|
|
if(content_length > num_additional_bytes) {
|
|
session->connection->set_timeout();
|
|
asio::async_read(*session->connection->socket, session->response->content_buffer, asio::transfer_exactly(content_length - num_additional_bytes), [this, session](const error_code &ec, size_t /*bytes_transferred*/) {
|
|
session->connection->cancel_timeout();
|
|
auto lock = session->connection->handler_runner->continue_lock();
|
|
if(!lock)
|
|
return;
|
|
if(!ec)
|
|
session->callback(session->connection, ec);
|
|
else
|
|
session->callback(session->connection, ec);
|
|
});
|
|
}
|
|
else
|
|
session->callback(session->connection, ec);
|
|
}
|
|
else if((header_it = session->response->header.find("Transfer-Encoding")) != session->response->header.end() && header_it->second == "chunked") {
|
|
auto tmp_streambuf = std::make_shared<asio::streambuf>();
|
|
this->read_chunked(session, tmp_streambuf);
|
|
}
|
|
else if(session->response->http_version < "1.1" || ((header_it = session->response->header.find("Session")) != session->response->header.end() && header_it->second == "close")) {
|
|
session->connection->set_timeout();
|
|
asio::async_read(*session->connection->socket, session->response->content_buffer, [this, session](const error_code &ec, size_t /*bytes_transferred*/) {
|
|
session->connection->cancel_timeout();
|
|
auto lock = session->connection->handler_runner->continue_lock();
|
|
if(!lock)
|
|
return;
|
|
if(!ec)
|
|
session->callback(session->connection, ec);
|
|
else
|
|
session->callback(session->connection, ec == asio::error::eof ? error_code() : ec);
|
|
});
|
|
}
|
|
else
|
|
session->callback(session->connection, ec);
|
|
}
|
|
else {
|
|
if(session->connection->attempt_reconnect && ec != asio::error::operation_aborted) {
|
|
std::unique_lock<std::mutex> lock(connections_mutex);
|
|
auto it = connections.find(session->connection);
|
|
if(it != connections.end()) {
|
|
connections.erase(it);
|
|
session->connection = create_connection();
|
|
session->connection->attempt_reconnect = false;
|
|
session->connection->in_use = true;
|
|
connections.emplace(session->connection);
|
|
lock.unlock();
|
|
this->connect(session);
|
|
}
|
|
else {
|
|
lock.unlock();
|
|
session->callback(session->connection, ec);
|
|
}
|
|
}
|
|
else
|
|
session->callback(session->connection, ec);
|
|
}
|
|
});
|
|
}
|
|
|
|
void read_chunked(const std::shared_ptr<Session> &session, const std::shared_ptr<asio::streambuf> &tmp_streambuf) {
|
|
session->connection->set_timeout();
|
|
asio::async_read_until(*session->connection->socket, session->response->content_buffer, "\r\n", [this, session, tmp_streambuf](const error_code &ec, size_t bytes_transferred) {
|
|
session->connection->cancel_timeout();
|
|
auto lock = session->connection->handler_runner->continue_lock();
|
|
if(!lock)
|
|
return;
|
|
if(!ec) {
|
|
std::string line;
|
|
getline(session->response->content, line);
|
|
bytes_transferred -= line.size() + 1;
|
|
line.pop_back();
|
|
std::streamsize length = stol(line, 0, 16);
|
|
|
|
auto num_additional_bytes = static_cast<std::streamsize>(session->response->content_buffer.size() - bytes_transferred);
|
|
|
|
auto post_process = [this, session, tmp_streambuf, length]() {
|
|
std::ostream tmp_stream(tmp_streambuf.get());
|
|
if(length > 0) {
|
|
std::vector<char> buffer(static_cast<size_t>(length));
|
|
session->response->content.read(&buffer[0], length);
|
|
tmp_stream.write(&buffer[0], length);
|
|
}
|
|
|
|
// Remove "\r\n"
|
|
session->response->content.get();
|
|
session->response->content.get();
|
|
|
|
if(length > 0)
|
|
this->read_chunked(session, tmp_streambuf);
|
|
else {
|
|
std::ostream response_stream(&session->response->content_buffer);
|
|
response_stream << tmp_stream.rdbuf();
|
|
error_code ec;
|
|
session->callback(session->connection, ec);
|
|
}
|
|
};
|
|
|
|
if((2 + length) > num_additional_bytes) {
|
|
session->connection->set_timeout();
|
|
asio::async_read(*session->connection->socket, session->response->content_buffer, asio::transfer_exactly(2 + length - num_additional_bytes), [this, session, post_process](const error_code &ec, size_t /*bytes_transferred*/) {
|
|
session->connection->cancel_timeout();
|
|
auto lock = session->connection->handler_runner->continue_lock();
|
|
if(!lock)
|
|
return;
|
|
if(!ec)
|
|
post_process();
|
|
else
|
|
session->callback(session->connection, ec);
|
|
});
|
|
}
|
|
else
|
|
post_process();
|
|
}
|
|
else
|
|
session->callback(session->connection, ec);
|
|
});
|
|
}
|
|
};
|
|
|
|
template <class socket_type>
|
|
class Client : public ClientBase<socket_type> {};
|
|
|
|
typedef asio::ip::tcp::socket HTTP;
|
|
|
|
template <>
|
|
class Client<HTTP> : public ClientBase<HTTP> {
|
|
public:
|
|
Client(const std::string &server_port_path) : ClientBase<HTTP>::ClientBase(server_port_path, 80) {}
|
|
|
|
protected:
|
|
std::shared_ptr<Connection> create_connection() override {
|
|
return std::make_shared<Connection>(handler_runner, config.timeout, *io_service);
|
|
}
|
|
|
|
void connect(const std::shared_ptr<Session> &session) override {
|
|
if(!session->connection->socket->lowest_layer().is_open()) {
|
|
auto resolver = std::make_shared<asio::ip::tcp::resolver>(*io_service);
|
|
session->connection->set_timeout(config.timeout_connect);
|
|
resolver->async_resolve(*query, [this, session, resolver](const error_code &ec, asio::ip::tcp::resolver::iterator it) {
|
|
session->connection->cancel_timeout();
|
|
auto lock = session->connection->handler_runner->continue_lock();
|
|
if(!lock)
|
|
return;
|
|
if(!ec) {
|
|
session->connection->set_timeout(config.timeout_connect);
|
|
asio::async_connect(*session->connection->socket, it, [this, session, resolver](const error_code &ec, asio::ip::tcp::resolver::iterator /*it*/) {
|
|
session->connection->cancel_timeout();
|
|
auto lock = session->connection->handler_runner->continue_lock();
|
|
if(!lock)
|
|
return;
|
|
if(!ec) {
|
|
asio::ip::tcp::no_delay option(true);
|
|
error_code ec;
|
|
session->connection->socket->set_option(option, ec);
|
|
this->write(session);
|
|
}
|
|
else
|
|
session->callback(session->connection, ec);
|
|
});
|
|
}
|
|
else
|
|
session->callback(session->connection, ec);
|
|
});
|
|
}
|
|
else
|
|
write(session);
|
|
}
|
|
};
|
|
} // namespace SimpleWeb
|
|
|
|
#endif /* CLIENT_HTTP_HPP */
|