0
0
Fork 0
mirror of https://github.com/matrix-construct/construct synced 2024-11-25 16:22:35 +01:00
construct/ircd/client.cc
2017-10-11 20:54:02 -07:00

513 lines
12 KiB
C++

/*
* charybdis: an advanced ircd.
* client.c: Controls clients.
*
* Copyright (C) 1990 Jarkko Oikarinen and University of Oulu, Co Center
* Copyright (C) 1996-2002 Hybrid Development Team
* Copyright (C) 2002-2005 ircd-ratbox development team
* Copyright (C) 2007 William Pitcock
* Copyright (C) 2016 Charybdis Development Team
* Copyright (C) 2016 Jason Volk <jason@zemos.net>
*
* This program is free software; you can redistribute it and/or modify
* it under the terms of the GNU General Public License as published by
* the Free Software Foundation; either version 2 of the License, or
* (at your option) any later version.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License
* along with this program; if not, write to the Free Software
* Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307
* USA
*/
#include <ircd/asio.h>
namespace ircd {
// Default time limit for how long a client connection can be in "async mode"
// (or idle mode) after which it is disconnected.
const auto async_timeout
{
300s
};
// Time limit for how long a connected client can be in "request mode." This
// should never be hit unless there's an error in the handling code.
const auto request_timeout
{
60s
};
// The pool of request contexts. When a client makes a request it does so by acquiring
// a stack from this pool. The request handling and response logic can then be written
// in a synchronous manner as if each connection had its own thread.
ctx::pool request
{
"request", 256_KiB
};
// Container for all active clients (connections) for iteration purposes.
client::list client::clients;
bool handle_ec_timeout(client &);
bool handle_ec_eof(client &);
bool handle_ec_success(client &);
bool handle_ec(client &, const net::error_code &);
void async_recv_next(std::shared_ptr<client>, const milliseconds &timeout);
void async_recv_next(std::shared_ptr<client>);
void disconnect(client &, const socket::dc & = socket::dc::RST);
void disconnect_all();
template<class... args> std::shared_ptr<client> make_client(args&&...);
} // namespace ircd
ircd::client::init::init()
{
request.add(2);
}
ircd::client::init::~init()
noexcept
{
log::debug("Interrupting %zu requests; dropping %zu requests; disconnecting %zu clients.",
request.active(),
request.pending(),
client::clients.size());
request.interrupt();
disconnect_all();
request.join();
}
ircd::hostport
ircd::local(const client &client)
{
if(!client.sock)
return { "0.0.0.0"s, 0 };
return net::local_hostport(*client.sock);
}
ircd::hostport
ircd::remote(const client &client)
{
if(!client.sock)
return { "0.0.0.0"s, 0 };
return net::remote_hostport(*client.sock);
}
ircd::http::response::write_closure
ircd::write_closure(client &client)
{
// returns a function that can be called to send an iovector of data to a client
return [&client](const ilist<const_buffer> &iov)
{
//std::cout << "<<<<" << std::endl;
//std::cout << iov << std::endl;
//std::cout << "----" << std::endl;
const auto written
{
write(*client.sock, iov)
};
};
}
ircd::parse::read_closure
ircd::read_closure(client &client)
{
static const auto handle_error([]
(const boost::system::system_error &e)
{
using namespace boost::system::errc;
switch(e.code().value())
{
case operation_canceled: throw http::error(http::REQUEST_TIMEOUT);
default: throw boost::system::system_error(e);
}
});
// Returns a function the parser can call when it wants more data
return [&client](char *&start, char *const &stop)
{
try
{
const char *const got(start);
read(client, start, stop);
//std::cout << ">>>>" << std::endl;
//std::cout << string_view{got, start} << std::endl;
//std::cout << "----" << std::endl;
}
catch(const boost::system::system_error &e)
{
handle_error(e);
}
};
}
char *
ircd::read(client &client,
char *&start,
char *const &stop)
{
auto &sock(*client.sock);
const std::array<mutable_buffer, 1> bufs
{{
{ start, stop }
}};
char *const base(start);
start += sock.read_some(bufs);
return base;
}
const char *
ircd::write(client &client,
const char *&start,
const char *const &stop)
{
auto &sock(*client.sock);
const std::array<const_buffer, 1> bufs
{{
{ start, stop }
}};
const char *const base(start);
start += sock.write(bufs);
return base;
}
//
// client
//
ircd::client::client()
:client{std::shared_ptr<socket>{}}
{
}
ircd::client::client(const hostport &host_port,
const seconds &timeout)
:client
{
std::make_shared<socket>(host_port.first, host_port.second, timeout)
}
{
}
ircd::client::client(std::shared_ptr<socket> sock)
:clit{clients, clients.emplace(end(clients), this)}
,sock{std::move(sock)}
,request_timer{ircd::timer::nostart}
{
}
ircd::client::~client()
noexcept
{
}
namespace ircd
{
void handle_request(client &client, parse::capstan &pc, const http::request::head &head);
bool handle_request(client &client, parse::capstan &pc);
}
bool
ircd::client::main()
noexcept try
{
const auto header_max{8192};
const auto content_max{65536};
unique_buffer<mutable_buffer> buffer
{
header_max + content_max
};
parse::buffer pb{buffer};
parse::capstan pc{pb, read_closure(*this)}; do
{
if(!handle_request(*this, pc))
return false;
pb.remove();
}
while(pc.unparsed());
return true;
}
catch(const boost::system::system_error &e)
{
using boost::asio::error::eof;
using boost::asio::error::broken_pipe;
using boost::asio::error::connection_reset;
using namespace boost::system::errc;
switch(e.code().value())
{
case success:
assert(0);
return true;
case eof:
case broken_pipe:
case connection_reset:
case not_connected:
case operation_canceled:
return false;
default:
break;
}
log::critical("(unexpected) system_error: %s", e.what());
if(ircd::debugmode)
std::terminate();
return false;
}
catch(const std::exception &e)
{
log::error("client[%s] [500 Internal Error]: %s",
string(remote(*this)),
e.what());
if(ircd::debugmode)
std::terminate();
return false;
}
bool
ircd::handle_request(client &client,
parse::capstan &pc)
try
{
client.request_timer = ircd::timer{};
client.sock->set_timeout(request_timeout, [&client]
(const net::error_code &ec)
{
if(!ec)
client.sock->cancel();
});
http::request
{
pc, nullptr, write_closure(client), [&client, &pc]
(const auto &head)
{
client.sock->timer.cancel();
handle_request(client, pc, head);
}
};
return true;
}
catch(const http::error &e)
{
log::debug("client[%s] HTTP %s in %ld$us %s",
string(remote(client)),
e.what(),
client.request_timer.at<microseconds>().count(),
e.content);
switch(e.code)
{
case http::BAD_REQUEST: return false;
case http::INTERNAL_SERVER_ERROR: return false;
case http::REQUEST_TIMEOUT: return false;
default: return true;
}
}
void
ircd::handle_request(client &client,
parse::capstan &pc,
const http::request::head &head)
{
log::debug("client[%s] HTTP %s `%s' (content-length: %zu)",
string(remote(client)),
head.method,
head.path,
head.content_length);
auto &resource
{
ircd::resource::find(head.path)
};
resource(client, pc, head);
}
std::shared_ptr<ircd::client>
ircd::add_client(std::shared_ptr<socket> s)
{
const auto client
{
make_client(std::move(s))
};
log::debug("client[%s] CONNECTED local[%s]",
string(remote(*client)),
string(local(*client)));
async_recv_next(client, async_timeout);
return client;
}
template<class... args>
std::shared_ptr<ircd::client>
ircd::make_client(args&&... a)
{
return std::make_shared<client>(std::forward<args>(a)...);
}
void
ircd::disconnect_all()
{
for(auto &client : client::clients) try
{
disconnect(*client, socket::dc::RST);
}
catch(const std::exception &e)
{
log::warning("Error disconnecting client @%p: %s", &client, e.what());
}
}
void
ircd::disconnect(client &client,
const socket::dc &type)
{
auto &sock(*client.sock);
sock.disconnect(type);
}
void
ircd::async_recv_next(std::shared_ptr<client> client)
{
async_recv_next(std::move(client), milliseconds(-1));
}
///
/// This function is the basis for the client's request loop. We still use
/// an asynchronous pattern until there is activity on the socket (a request)
/// in which case the switch to synchronous mode is made by jumping into an
/// ircd::context drawn from the request pool. When the request is finished,
/// the client exits back into asynchronous mode until the next request is
/// received and rinse and repeat.
//
/// This sequence exists to avoid any possible c10k-style limitation imposed by
/// dedicating a context and its stack space to the lifetime of a connection.
/// This is similar to the thread-per-request pattern before async was in vogue.
//
/// Developers: Pay close attention to the comments to know exactly where you
/// are and what you can do at any given point in this sequence.
///
void
ircd::async_recv_next(std::shared_ptr<client> client,
const milliseconds &timeout)
{
auto &sock(*client->sock);
// This call returns immediately so we no longer block the current context and
// its stack while waiting for activity on idle connections between requests.
sock(timeout, [client(std::move(client)), timeout](const net::error_code &ec)
noexcept
{
// Right here this handler is executing on the main stack (not in any
// ircd::context). We handle any socket errors now, and if this function
// returns here the client's shared_ptr may expire and that will be the
// end of this client, socket, and connection...
if(!handle_ec(*client, ec))
return;
// This call returns immediately because we can never block the main stack outside
// of the ircd::context system. The context the closure ends up getting is the next
// available from the request pool, which may not be available immediately so this
// handler might be queued for some time after this call returns.
request([client(std::move(client)), timeout]
{
// Right here this handler is executing on an ircd::context with its own
// stack dedicated to the lifetime of this request. If client::main()
// returns true, we bring the client back into async mode to wait for
// the next request.
if(client->main())
async_recv_next(client, timeout);
});
});
}
bool
ircd::handle_ec(client &client,
const net::error_code &ec)
{
using namespace boost::system::errc;
using boost::asio::error::eof;
switch(ec.value())
{
case success: return handle_ec_success(client);
case eof: return handle_ec_eof(client);
case operation_canceled: return handle_ec_timeout(client);
default:
{
log::debug("client(%p): %s", &client, ec.message());
return false;
}
}
}
bool
ircd::handle_ec_success(client &client)
{
return true;
}
bool
ircd::handle_ec_eof(client &client)
try
{
log::debug("client[%s]: EOF",
string(remote(client)));
client.sock->disconnect(socket::FIN_RECV);
return false;
}
catch(const std::exception &e)
{
log::warning("client(%p): EOF: %s",
&client,
e.what());
return false;
}
bool
ircd::handle_ec_timeout(client &client)
try
{
auto &sock(*client.sock);
log::debug("client[%s]: disconnecting after inactivity timeout",
string(remote(client)));
sock.disconnect();
return false;
}
catch(const std::exception &e)
{
log::warning("client(%p): timeout: %s",
&client,
e.what());
return false;
}