mirror of
https://github.com/matrix-construct/construct
synced 2024-11-30 02:32:43 +01:00
516 lines
12 KiB
C++
516 lines
12 KiB
C++
// Copyright (C) Matrix Construct Developers, Authors & Contributors
|
|
// Copyright (C) 2016-2018 Jason Volk
|
|
//
|
|
// Permission to use, copy, modify, and/or distribute this software for any
|
|
// purpose with or without fee is hereby granted, provided that the above
|
|
// copyright notice and this permission notice is present in all copies.
|
|
//
|
|
// THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
|
|
// IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
|
|
// WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
|
// DISCLAIMED. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT,
|
|
// INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
|
|
// (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
|
|
// SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
// HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,
|
|
// STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING
|
|
// IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
|
// POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
#pragma once
|
|
#define HAVE_AIO_H
|
|
|
|
#include <RB_INC_SYS_SYSCALL_H
|
|
#include <RB_INC_SYS_EVENTFD_H
|
|
#include <linux/aio_abi.h>
|
|
|
|
/// AIO context instance. Right now this is a singleton with an extern
|
|
/// instance at fs::aioctx.
|
|
struct ircd::fs::aio
|
|
{
|
|
struct request;
|
|
|
|
/// Maximum number of events we can submit to kernel
|
|
static constexpr const size_t &MAX_EVENTS
|
|
{
|
|
64
|
|
};
|
|
|
|
/// The semaphore value for the eventfd which we keep here.
|
|
uint64_t semval{0};
|
|
|
|
/// An eventfd which will be notified by the kernel; we integrate this with
|
|
/// the ircd io_service core epoll() event loop. The EFD_SEMAPHORE flag is
|
|
/// not used to reduce the number of triggers. We can collect multiple AIO
|
|
/// completions after a single trigger to this fd. With EFD_SEMAPHORE, we
|
|
/// would collect all the completions on the first trigger and then
|
|
/// continue to get polled. Because EFD_SEMAPHORE is not set, the semval
|
|
/// which is kept above will reflect a hint for how many AIO's are done.
|
|
asio::posix::stream_descriptor resfd
|
|
{
|
|
*ircd::ios, int(syscall(::eventfd, semval, EFD_NONBLOCK))
|
|
};
|
|
|
|
/// Handler to the io context we submit requests to the kernel with
|
|
aio_context_t idp
|
|
{
|
|
0
|
|
};
|
|
|
|
// Callback stack invoked when the sigfd is notified of completed events.
|
|
void handle_event(const io_event &) noexcept;
|
|
void handle_events() noexcept;
|
|
void handle(const error_code &, const size_t) noexcept;
|
|
|
|
void set_handle();
|
|
|
|
aio()
|
|
{
|
|
syscall<SYS_io_setup>(MAX_EVENTS, &idp);
|
|
set_handle();
|
|
}
|
|
|
|
~aio() noexcept
|
|
{
|
|
resfd.cancel();
|
|
syscall<SYS_io_destroy>(idp);
|
|
}
|
|
};
|
|
|
|
/// Generic request control block.
|
|
struct ircd::fs::aio::request
|
|
:iocb
|
|
{
|
|
struct read;
|
|
struct write;
|
|
|
|
ssize_t retval {0};
|
|
ssize_t errcode {0};
|
|
ctx::ctx *waiter {nullptr};
|
|
bool close_fd {false};
|
|
bool free_req {false};
|
|
|
|
/// called if close_fd is true
|
|
void close_fildes() noexcept;
|
|
|
|
/// Overriden by types of requests
|
|
virtual void handle() = 0;
|
|
|
|
public:
|
|
// Submit a request. ctx overload will properly yield
|
|
size_t operator()(ctx::ctx &waiter);
|
|
void operator()();
|
|
|
|
void cancel();
|
|
|
|
request(const int &fd);
|
|
virtual ~request() noexcept;
|
|
};
|
|
|
|
ircd::fs::aio::request::request(const int &fd)
|
|
:iocb{0}
|
|
{
|
|
assert(aioctx);
|
|
aio_flags = IOCB_FLAG_RESFD;
|
|
aio_resfd = aioctx->resfd.native_handle();
|
|
aio_fildes = fd;
|
|
}
|
|
|
|
/// vtable base
|
|
ircd::fs::aio::request::~request()
|
|
noexcept
|
|
{
|
|
}
|
|
|
|
/// Cancel a request. The handler callstack is invoked directly from here
|
|
/// which means any callback will be invoked or ctx will be notified if
|
|
/// appropriate.
|
|
void
|
|
ircd::fs::aio::request::cancel()
|
|
{
|
|
io_event result {0};
|
|
const auto &cb{static_cast<iocb *>(this)};
|
|
|
|
assert(aioctx);
|
|
syscall_nointr<SYS_io_cancel>(aioctx->idp, cb, &result);
|
|
aioctx->handle_event(result);
|
|
}
|
|
|
|
/// Submit a request and properly yield the ircd::ctx. The ctx argument
|
|
/// must be the currently running ctx (or ctx::current) for now; there is
|
|
/// no other usage. When this returns the result will be available or
|
|
/// exception will be thrown.
|
|
size_t
|
|
ircd::fs::aio::request::operator()(ctx::ctx &waiter)
|
|
try
|
|
{
|
|
// Cooperation strategy is to set this->waiter to the ctx and wait for
|
|
// notification with this->waiter set to null to ignore spurious notes.
|
|
assert(&waiter == ctx::current);
|
|
this->waiter = &waiter;
|
|
operator()(); do
|
|
{
|
|
ctx::wait();
|
|
}
|
|
while(this->waiter);
|
|
|
|
if(retval == -1)
|
|
throw_system_error(errcode);
|
|
|
|
return size_t(retval);
|
|
}
|
|
catch(const ctx::interrupted &e)
|
|
{
|
|
// When the ctx is interrupted we're obligated to cancel the request.
|
|
// The handler callstack is invoked directly from here by cancel() for
|
|
// what it's worth but we rethrow the interrupt anyway.
|
|
this->waiter = nullptr;
|
|
cancel();
|
|
throw;
|
|
}
|
|
|
|
/// Submit a request.
|
|
void
|
|
ircd::fs::aio::request::operator()()
|
|
{
|
|
struct iocb *const cbs[]
|
|
{
|
|
static_cast<iocb *>(this)
|
|
};
|
|
|
|
assert(aioctx);
|
|
syscall<SYS_io_submit>(aioctx->idp, 1, &cbs);
|
|
/*
|
|
log::debug("AIO request(%p) fd:%u op:%d bytes:%lu off:%ld prio:%d ctx:%p submit",
|
|
this,
|
|
this->aio_fildes,
|
|
this->aio_lio_opcode,
|
|
this->aio_nbytes,
|
|
this->aio_offset,
|
|
this->aio_reqprio,
|
|
this->waiter);
|
|
*/
|
|
}
|
|
|
|
void
|
|
ircd::fs::aio::set_handle()
|
|
{
|
|
semval = 0;
|
|
const asio::mutable_buffers_1 bufs(&semval, sizeof(semval));
|
|
auto handler{std::bind(&aio::handle, this, ph::_1, ph::_2)};
|
|
asio::async_read(resfd, bufs, std::move(handler));
|
|
}
|
|
|
|
/// Handle notifications that requests are complete.
|
|
void
|
|
ircd::fs::aio::handle(const error_code &ec,
|
|
const size_t bytes)
|
|
noexcept
|
|
{
|
|
assert((bytes == 8 && !ec && semval >= 1) || (bytes == 0 && ec));
|
|
assert(!ec || ec.category() == asio::error::get_system_category());
|
|
|
|
switch(ec.value())
|
|
{
|
|
case boost::system::errc::success:
|
|
handle_events();
|
|
break;
|
|
|
|
case boost::system::errc::operation_canceled:
|
|
return;
|
|
|
|
default:
|
|
throw boost::system::system_error(ec);
|
|
}
|
|
|
|
set_handle();
|
|
}
|
|
|
|
void
|
|
ircd::fs::aio::handle_events()
|
|
noexcept try
|
|
{
|
|
std::array<io_event, MAX_EVENTS> event;
|
|
|
|
// The number of completed requests available in events[]. This syscall
|
|
// is restarted on EINTR. After restart, it may or may not find any ready
|
|
// events but it never blocks to do so.
|
|
const auto count
|
|
{
|
|
syscall_nointr<SYS_io_getevents>(idp, 0, event.size(), event.data(), nullptr)
|
|
};
|
|
|
|
// The count should be at least 1 event. The only reason to return 0 might
|
|
// be related to an INTR; this assert will find out and may be commented.
|
|
assert(count > 0);
|
|
|
|
for(ssize_t i(0); i < count; ++i)
|
|
handle_event(event[i]);
|
|
}
|
|
catch(const std::exception &e)
|
|
{
|
|
log::error("AIO(%p) handle_events: %s",
|
|
this,
|
|
e.what());
|
|
}
|
|
|
|
void
|
|
ircd::fs::aio::handle_event(const io_event &event)
|
|
noexcept try
|
|
{
|
|
// Our extended control block is passed in event.data
|
|
auto *const request
|
|
{
|
|
reinterpret_cast<aio::request *>(event.data)
|
|
};
|
|
|
|
// The relevant iocb is repeated back to us in the result; we assert
|
|
// some basic sanity here about the layout of the request conglomerate.
|
|
assert(reinterpret_cast<iocb *>(event.obj) == static_cast<iocb *>(request));
|
|
|
|
// error conventions are like so
|
|
assert(event.res >= -1); // unix syscall return value semantic
|
|
assert(event.res2 >= 0); // errno code semantic
|
|
assert(event.res == -1 || event.res2 == 0);
|
|
|
|
// Set result indicators
|
|
request->retval = event.res;
|
|
request->errcode = event.res2;
|
|
/*
|
|
log::debug("AIO request(%p) fd:%d op:%d bytes:%lu off:%ld prio:%d ctx:%p result: bytes:%ld errno:%ld",
|
|
request,
|
|
request->aio_fildes,
|
|
request->aio_lio_opcode,
|
|
request->aio_nbytes,
|
|
request->aio_offset,
|
|
request->aio_reqprio,
|
|
request->waiter,
|
|
request->retval,
|
|
request->errcode);
|
|
*/
|
|
const unwind cleanup{[&request]
|
|
{
|
|
// The user might want us to cleanup their fd after this event
|
|
if(request->close_fd)
|
|
request->close_fildes();
|
|
|
|
// The user might want us to free a dynamically allocated request struct
|
|
// to simplify their callback sequence. The noexcept guarantees h
|
|
if(request->free_req)
|
|
delete request;
|
|
}};
|
|
|
|
// virtual dispatch based on the request type. Alternatively, we could
|
|
// switch() on the iocb lio_opcode and downcast... but that's what vtable
|
|
// does for us right here.
|
|
request->handle();
|
|
}
|
|
catch(const std::exception &e)
|
|
{
|
|
log::critical("Unhandled request(%lu) event(%p) error: %s",
|
|
event.data,
|
|
&event,
|
|
e.what());
|
|
}
|
|
|
|
/// If requested, close the file descriptor. Errors here can be logged but
|
|
/// must be otherwise ignored.
|
|
void
|
|
ircd::fs::aio::request::close_fildes()
|
|
noexcept try
|
|
{
|
|
syscall(::close, aio_fildes);
|
|
}
|
|
catch(const std::exception &e)
|
|
{
|
|
log::error("Failed to close request(%p) fd:%d: %s",
|
|
this,
|
|
aio_fildes);
|
|
}
|
|
|
|
///////////////////////////////////////////////////////////////////////////////
|
|
//
|
|
// fs/read.h
|
|
//
|
|
|
|
namespace ircd::fs
|
|
{
|
|
void read__aio(const string_view &path, const mutable_raw_buffer &, const read_opts &, read_callback);
|
|
string_view read__aio(const string_view &path, const mutable_raw_buffer &, const read_opts &);
|
|
std::string read__aio(const string_view &path, const read_opts &);
|
|
}
|
|
|
|
/// Read request control block
|
|
struct ircd::fs::aio::request::read
|
|
:request
|
|
{
|
|
read_callback callback;
|
|
|
|
virtual void handle() final override;
|
|
|
|
read(const int &fd, const mutable_raw_buffer &buf, const read_opts &opts, read_callback callback);
|
|
};
|
|
|
|
ircd::fs::aio::request::read::read(const int &fd,
|
|
const mutable_raw_buffer &buf,
|
|
const read_opts &opts,
|
|
read_callback callback)
|
|
:request{fd}
|
|
,callback{std::move(callback)}
|
|
{
|
|
aio_data = uintptr_t(this);
|
|
aio_reqprio = opts.priority;
|
|
aio_lio_opcode = IOCB_CMD_PREAD;
|
|
|
|
aio_buf = uintptr_t(buffer::data(buf));
|
|
aio_nbytes = buffer::size(buf);
|
|
aio_offset = opts.offset;
|
|
}
|
|
|
|
void
|
|
ircd::fs::aio::request::read::handle()
|
|
{
|
|
if(waiter)
|
|
{
|
|
ircd::ctx::notify(*waiter);
|
|
waiter = nullptr;
|
|
}
|
|
|
|
if(callback)
|
|
{
|
|
const size_t bytes
|
|
{
|
|
retval >= 0? size_t(retval) : 0
|
|
};
|
|
|
|
const string_view view
|
|
{
|
|
reinterpret_cast<const char *>(aio_buf), bytes
|
|
};
|
|
|
|
std::exception_ptr eptr
|
|
{
|
|
errcode != 0? make_system_error(errcode) : std::exception_ptr{}
|
|
};
|
|
|
|
callback(std::move(eptr), view);
|
|
}
|
|
}
|
|
|
|
//
|
|
// Interface
|
|
//
|
|
|
|
std::string
|
|
ircd::fs::read__aio(const string_view &path,
|
|
const read_opts &opts)
|
|
{
|
|
// Path to open(2) must be null terminated;
|
|
static thread_local char pathstr[2048];
|
|
strlcpy(pathstr, path, sizeof(pathstr));
|
|
|
|
const auto fd
|
|
{
|
|
syscall(::open, pathstr, O_CLOEXEC, O_RDONLY)
|
|
};
|
|
|
|
const unwind cfd{[&fd]
|
|
{
|
|
syscall(::close, fd);
|
|
}};
|
|
|
|
// This fstat may be defeating; to be sure, don't use this overload
|
|
// and progressively buffer chunks into your application.
|
|
struct stat stat;
|
|
syscall(::fstat, fd, &stat);
|
|
const auto &size
|
|
{
|
|
stat.st_size
|
|
};
|
|
|
|
std::string ret(size, char{});
|
|
const mutable_buffer buf
|
|
{
|
|
const_cast<char *>(ret.data()), ret.size()
|
|
};
|
|
|
|
aio::request::read request
|
|
{
|
|
int(fd), buf, opts, nullptr
|
|
};
|
|
|
|
const size_t bytes
|
|
{
|
|
request(ctx::cur())
|
|
};
|
|
|
|
ret.resize(bytes);
|
|
return ret;
|
|
}
|
|
|
|
ircd::string_view
|
|
ircd::fs::read__aio(const string_view &path,
|
|
const mutable_raw_buffer &buf,
|
|
const read_opts &opts)
|
|
{
|
|
// Path to open(2) must be null terminated;
|
|
static thread_local char pathstr[2048];
|
|
strlcpy(pathstr, path, sizeof(pathstr));
|
|
|
|
const auto fd
|
|
{
|
|
syscall(::open, pathstr, O_CLOEXEC, O_RDONLY)
|
|
};
|
|
|
|
const unwind cfd{[&fd]
|
|
{
|
|
syscall(::close, fd);
|
|
}};
|
|
|
|
aio::request::read request
|
|
{
|
|
int(fd), buf, opts, nullptr
|
|
};
|
|
|
|
const size_t bytes
|
|
{
|
|
request(ctx::cur())
|
|
};
|
|
|
|
const string_view view
|
|
{
|
|
reinterpret_cast<const char *>(data(buf)), bytes
|
|
};
|
|
|
|
return view;
|
|
}
|
|
|
|
void
|
|
ircd::fs::read__aio(const string_view &path,
|
|
const mutable_raw_buffer &buf,
|
|
const read_opts &opts,
|
|
read_callback callback)
|
|
{
|
|
static thread_local char pathstr[2048];
|
|
strlcpy(pathstr, path, sizeof(pathstr));
|
|
|
|
const auto fd
|
|
{
|
|
syscall(::open, pathstr, O_CLOEXEC, O_RDONLY)
|
|
};
|
|
|
|
const unwind::exceptional cfd{[&fd]
|
|
{
|
|
syscall(::close, fd);
|
|
}};
|
|
|
|
auto request
|
|
{
|
|
std::make_unique<aio::request::read>(int(fd), buf, opts, std::move(callback))
|
|
};
|
|
|
|
request->close_fd = true;
|
|
request->free_req = true;
|
|
(*request)();
|
|
request.release();
|
|
}
|