2020-09-12 06:11:07 +03:00
|
|
|
/*
|
|
|
|
* Copyright (c) 2020, the SerenityOS developers.
|
2021-09-05 00:48:46 +03:00
|
|
|
* Copyright (c) 2021, Andreas Kling <kling@serenityos.org>
|
2020-09-12 06:11:07 +03:00
|
|
|
*
|
2021-04-22 11:24:48 +03:00
|
|
|
* SPDX-License-Identifier: BSD-2-Clause
|
2020-09-12 06:11:07 +03:00
|
|
|
*/
|
|
|
|
|
|
|
|
#pragma once
|
|
|
|
|
|
|
|
#include <AK/Types.h>
|
|
|
|
#include <AK/Userspace.h>
|
2021-08-06 11:45:34 +03:00
|
|
|
#include <Kernel/Memory/MemoryManager.h>
|
2020-09-12 06:11:07 +03:00
|
|
|
#include <Kernel/StdLib.h>
|
|
|
|
#include <Kernel/UnixTypes.h>
|
|
|
|
#include <LibC/errno_numbers.h>
|
|
|
|
|
|
|
|
namespace Kernel {
|
|
|
|
|
2021-02-15 02:22:28 +03:00
|
|
|
class [[nodiscard]] UserOrKernelBuffer {
|
2020-09-12 06:11:07 +03:00
|
|
|
public:
|
|
|
|
UserOrKernelBuffer() = delete;
|
|
|
|
|
|
|
|
static UserOrKernelBuffer for_kernel_buffer(u8* kernel_buffer)
|
|
|
|
{
|
2021-08-06 14:49:36 +03:00
|
|
|
VERIFY(!kernel_buffer || !Memory::is_user_address(VirtualAddress(kernel_buffer)));
|
2020-09-12 06:11:07 +03:00
|
|
|
return UserOrKernelBuffer(kernel_buffer);
|
|
|
|
}
|
|
|
|
|
|
|
|
static Optional<UserOrKernelBuffer> for_user_buffer(u8* user_buffer, size_t size)
|
|
|
|
{
|
2021-08-06 14:49:36 +03:00
|
|
|
if (user_buffer && !Memory::is_user_range(VirtualAddress(user_buffer), size))
|
2020-09-12 06:11:07 +03:00
|
|
|
return {};
|
|
|
|
return UserOrKernelBuffer(user_buffer);
|
|
|
|
}
|
|
|
|
|
|
|
|
template<typename UserspaceType>
|
|
|
|
static Optional<UserOrKernelBuffer> for_user_buffer(UserspaceType userspace, size_t size)
|
|
|
|
{
|
2021-08-06 14:49:36 +03:00
|
|
|
if (!Memory::is_user_range(VirtualAddress(userspace.unsafe_userspace_ptr()), size))
|
2020-09-12 06:11:07 +03:00
|
|
|
return {};
|
|
|
|
return UserOrKernelBuffer(const_cast<u8*>((const u8*)userspace.unsafe_userspace_ptr()));
|
|
|
|
}
|
|
|
|
|
2021-02-15 02:22:28 +03:00
|
|
|
[[nodiscard]] bool is_kernel_buffer() const;
|
|
|
|
[[nodiscard]] const void* user_or_kernel_ptr() const { return m_buffer; }
|
2020-09-12 06:11:07 +03:00
|
|
|
|
2021-06-16 17:44:15 +03:00
|
|
|
[[nodiscard]] UserOrKernelBuffer offset(size_t offset) const
|
2020-09-12 06:11:07 +03:00
|
|
|
{
|
|
|
|
if (!m_buffer)
|
|
|
|
return *this;
|
|
|
|
UserOrKernelBuffer offset_buffer = *this;
|
|
|
|
offset_buffer.m_buffer += offset;
|
2021-02-23 22:42:32 +03:00
|
|
|
VERIFY(offset_buffer.is_kernel_buffer() == is_kernel_buffer());
|
2020-09-12 06:11:07 +03:00
|
|
|
return offset_buffer;
|
|
|
|
}
|
|
|
|
|
2021-09-06 23:25:02 +03:00
|
|
|
KResultOr<NonnullOwnPtr<KString>> try_copy_into_kstring(size_t) const;
|
2021-09-07 13:09:52 +03:00
|
|
|
KResult write(const void* src, size_t offset, size_t len);
|
|
|
|
KResult write(const void* src, size_t len)
|
2020-09-12 06:11:07 +03:00
|
|
|
{
|
|
|
|
return write(src, 0, len);
|
|
|
|
}
|
2021-09-07 13:09:52 +03:00
|
|
|
KResult write(ReadonlyBytes bytes)
|
2020-09-15 13:24:14 +03:00
|
|
|
{
|
|
|
|
return write(bytes.data(), bytes.size());
|
|
|
|
}
|
|
|
|
|
2021-09-07 13:09:52 +03:00
|
|
|
KResult read(void* dest, size_t offset, size_t len) const;
|
|
|
|
KResult read(void* dest, size_t len) const
|
2020-09-12 06:11:07 +03:00
|
|
|
{
|
|
|
|
return read(dest, 0, len);
|
|
|
|
}
|
2021-09-07 13:09:52 +03:00
|
|
|
|
|
|
|
KResult read(Bytes bytes) const
|
2020-09-15 13:24:14 +03:00
|
|
|
{
|
|
|
|
return read(bytes.data(), bytes.size());
|
|
|
|
}
|
2020-09-12 06:11:07 +03:00
|
|
|
|
2021-09-07 13:09:52 +03:00
|
|
|
KResult memset(int value, size_t offset, size_t len);
|
|
|
|
KResult memset(int value, size_t len)
|
2020-09-12 06:11:07 +03:00
|
|
|
{
|
|
|
|
return memset(value, 0, len);
|
|
|
|
}
|
|
|
|
|
|
|
|
template<size_t BUFFER_BYTES, typename F>
|
2021-09-06 23:25:02 +03:00
|
|
|
KResultOr<size_t> write_buffered(size_t offset, size_t len, F f)
|
2020-09-12 06:11:07 +03:00
|
|
|
{
|
|
|
|
if (!m_buffer)
|
2021-05-13 10:08:44 +03:00
|
|
|
return EFAULT;
|
2020-09-12 06:11:07 +03:00
|
|
|
if (is_kernel_buffer()) {
|
|
|
|
// We're transferring directly to a kernel buffer, bypass
|
2021-09-01 09:44:55 +03:00
|
|
|
Bytes bytes { m_buffer + offset, len };
|
|
|
|
return f(bytes);
|
2020-09-12 06:11:07 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// The purpose of using a buffer on the stack is that we can
|
|
|
|
// avoid a bunch of small (e.g. 1-byte) copy_to_user calls
|
|
|
|
u8 buffer[BUFFER_BYTES];
|
|
|
|
size_t nwritten = 0;
|
|
|
|
while (nwritten < len) {
|
|
|
|
auto to_copy = min(sizeof(buffer), len - nwritten);
|
2021-09-01 09:44:55 +03:00
|
|
|
Bytes bytes { buffer, to_copy };
|
|
|
|
KResultOr<size_t> copied_or_error = f(bytes);
|
2021-06-16 17:44:15 +03:00
|
|
|
if (copied_or_error.is_error())
|
|
|
|
return copied_or_error.error();
|
|
|
|
auto copied = copied_or_error.value();
|
2021-06-17 12:15:55 +03:00
|
|
|
VERIFY(copied <= to_copy);
|
2021-09-07 13:09:52 +03:00
|
|
|
TRY(write(buffer, nwritten, copied));
|
2021-06-17 12:15:55 +03:00
|
|
|
nwritten += copied;
|
|
|
|
if (copied < to_copy)
|
2020-09-12 06:11:07 +03:00
|
|
|
break;
|
|
|
|
}
|
2021-05-13 10:08:44 +03:00
|
|
|
return nwritten;
|
2020-09-12 06:11:07 +03:00
|
|
|
}
|
|
|
|
template<size_t BUFFER_BYTES, typename F>
|
2021-09-06 23:25:02 +03:00
|
|
|
KResultOr<size_t> write_buffered(size_t len, F f)
|
2020-09-12 06:11:07 +03:00
|
|
|
{
|
|
|
|
return write_buffered<BUFFER_BYTES, F>(0, len, f);
|
|
|
|
}
|
|
|
|
|
|
|
|
template<size_t BUFFER_BYTES, typename F>
|
2021-09-06 23:25:02 +03:00
|
|
|
KResultOr<size_t> read_buffered(size_t offset, size_t len, F f) const
|
2020-09-12 06:11:07 +03:00
|
|
|
{
|
|
|
|
if (!m_buffer)
|
2021-05-13 10:08:44 +03:00
|
|
|
return EFAULT;
|
2020-09-12 06:11:07 +03:00
|
|
|
if (is_kernel_buffer()) {
|
|
|
|
// We're transferring directly from a kernel buffer, bypass
|
2021-09-01 09:44:55 +03:00
|
|
|
return f({ m_buffer + offset, len });
|
2020-09-12 06:11:07 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// The purpose of using a buffer on the stack is that we can
|
|
|
|
// avoid a bunch of small (e.g. 1-byte) copy_from_user calls
|
|
|
|
u8 buffer[BUFFER_BYTES];
|
|
|
|
size_t nread = 0;
|
|
|
|
while (nread < len) {
|
|
|
|
auto to_copy = min(sizeof(buffer), len - nread);
|
2021-09-07 13:09:52 +03:00
|
|
|
TRY(read(buffer, nread, to_copy));
|
2021-09-01 09:44:55 +03:00
|
|
|
ReadonlyBytes read_only_bytes { buffer, to_copy };
|
|
|
|
KResultOr<size_t> copied_or_error = f(read_only_bytes);
|
2021-06-16 16:20:35 +03:00
|
|
|
if (copied_or_error.is_error())
|
|
|
|
return copied_or_error.error();
|
|
|
|
auto copied = copied_or_error.value();
|
|
|
|
VERIFY(copied <= to_copy);
|
|
|
|
nread += copied;
|
|
|
|
if (copied < to_copy)
|
2020-09-12 06:11:07 +03:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
return nread;
|
|
|
|
}
|
|
|
|
template<size_t BUFFER_BYTES, typename F>
|
2021-09-06 23:25:02 +03:00
|
|
|
KResultOr<size_t> read_buffered(size_t len, F f) const
|
2020-09-12 06:11:07 +03:00
|
|
|
{
|
|
|
|
return read_buffered<BUFFER_BYTES, F>(0, len, f);
|
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
|
|
|
explicit UserOrKernelBuffer(u8* buffer)
|
|
|
|
: m_buffer(buffer)
|
|
|
|
{
|
|
|
|
}
|
|
|
|
|
|
|
|
u8* m_buffer;
|
|
|
|
};
|
|
|
|
|
|
|
|
}
|