ladybird/Kernel/VM/PurgeablePageRanges.cpp
Andreas Kling 5d180d1f99 Everywhere: Rename ASSERT => VERIFY
(...and ASSERT_NOT_REACHED => VERIFY_NOT_REACHED)

Since all of these checks are done in release builds as well,
let's rename them to VERIFY to prevent confusion, as everyone is
used to assertions being compiled out in release.

We can introduce a new ASSERT macro that is specifically for debug
checks, but I'm doing this wholesale conversion first since we've
accumulated thousands of these already, and it's not immediately
obvious which ones are suitable for ASSERT.
2021-02-23 20:56:54 +01:00

328 lines
12 KiB
C++

/*
* Copyright (c) 2018-2020, Andreas Kling <kling@serenityos.org>
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions are met:
*
* 1. Redistributions of source code must retain the above copyright notice, this
* list of conditions and the following disclaimer.
*
* 2. Redistributions in binary form must reproduce the above copyright notice,
* this list of conditions and the following disclaimer in the documentation
* and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
* DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
* SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
* CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
* OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#include <AK/BinarySearch.h>
#include <AK/ScopeGuard.h>
#include <Kernel/Debug.h>
#include <Kernel/Process.h>
#include <Kernel/VM/AnonymousVMObject.h>
#include <Kernel/VM/MemoryManager.h>
#include <Kernel/VM/PhysicalPage.h>
#include <Kernel/VM/PurgeablePageRanges.h>
namespace Kernel {
#if VOLATILE_PAGE_RANGES_DEBUG
inline LogStream& operator<<(const LogStream& stream, const VolatilePageRange& range)
{
stream << "{" << range.base << " (" << range.count << ") purged: " << range.was_purged << "}";
return const_cast<LogStream&>(stream);
}
static void dump_volatile_page_ranges(const Vector<VolatilePageRange>& ranges)
{
for (size_t i = 0; i < ranges.size(); i++) {
const auto& range = ranges[i];
klog() << " [" << i << "] " << range;
}
}
#endif
void VolatilePageRanges::add_unchecked(const VolatilePageRange& range)
{
auto add_range = m_total_range.intersected(range);
if (add_range.is_empty())
return;
m_ranges.append(range);
}
bool VolatilePageRanges::add(const VolatilePageRange& range)
{
auto add_range = m_total_range.intersected(range);
if (add_range.is_empty())
return false;
add_range.was_purged = range.was_purged;
#if VOLATILE_PAGE_RANGES_DEBUG
klog() << "ADD " << range << " (total range: " << m_total_range << ") -->";
dump_volatile_page_ranges(m_ranges);
ScopeGuard debug_guard([&]() {
klog() << "After adding " << range << " (total range: " << m_total_range << ")";
dump_volatile_page_ranges(m_ranges);
klog() << "<-- ADD " << range << " (total range: " << m_total_range << ")";
});
#endif
size_t nearby_index = 0;
auto* existing_range = binary_search(
m_ranges.span(), add_range, &nearby_index, [](auto& a, auto& b) {
if (a.intersects_or_adjacent(b))
return 0;
return (signed)(a.base - (b.base + b.count - 1));
});
size_t inserted_index = 0;
if (existing_range) {
if (*existing_range == add_range)
return false;
if (existing_range->was_purged != add_range.was_purged) {
// Found an intersecting or adjacent range, but the purge flag
// doesn't match. Subtract what we're adding from it, and
existing_range->subtract_intersecting(add_range);
if (existing_range->is_empty()) {
*existing_range = add_range;
} else {
m_ranges.insert(++nearby_index, add_range);
existing_range = &m_ranges[nearby_index];
}
} else {
// Found an intersecting or adjacent range that can be merged
existing_range->combine_intersecting_or_adjacent(add_range);
}
inserted_index = nearby_index;
} else {
// Insert into the sorted list
m_ranges.insert_before_matching(
VolatilePageRange(add_range), [&](auto& entry) {
return entry.base >= add_range.base + add_range.count;
},
nearby_index, &inserted_index);
existing_range = &m_ranges[inserted_index];
}
// See if we can merge any of the following ranges
inserted_index++;
while (inserted_index < m_ranges.size()) {
auto& next_range = m_ranges[inserted_index];
if (!next_range.intersects_or_adjacent(*existing_range))
break;
if (next_range.was_purged != existing_range->was_purged) {
// The purged flag of following range is not the same.
// Subtract the added/combined range from it
next_range.subtract_intersecting(*existing_range);
if (next_range.is_empty())
m_ranges.remove(inserted_index);
} else {
existing_range->combine_intersecting_or_adjacent(next_range);
m_ranges.remove(inserted_index);
}
}
return true;
}
bool VolatilePageRanges::remove(const VolatilePageRange& range, bool& was_purged)
{
auto remove_range = m_total_range.intersected(range);
if (remove_range.is_empty())
return false;
#if VOLATILE_PAGE_RANGES_DEBUG
klog() << "REMOVE " << range << " (total range: " << m_total_range << ") -->";
dump_volatile_page_ranges(m_ranges);
ScopeGuard debug_guard([&]() {
klog() << "After removing " << range << " (total range: " << m_total_range << ")";
dump_volatile_page_ranges(m_ranges);
klog() << "<-- REMOVE " << range << " (total range: " << m_total_range << ") was_purged: " << was_purged;
});
#endif
size_t nearby_index = 0;
auto* existing_range = binary_search(
m_ranges.span(), remove_range, &nearby_index, [](auto& a, auto& b) {
if (a.intersects(b))
return 0;
return (signed)(a.base - (b.base + b.count - 1));
});
if (!existing_range)
return false;
was_purged = existing_range->was_purged;
if (existing_range->range_equals(remove_range)) {
m_ranges.remove(nearby_index);
} else {
// See if we need to remove any of the following ranges
VERIFY(existing_range == &m_ranges[nearby_index]); // sanity check
while (nearby_index < m_ranges.size()) {
existing_range = &m_ranges[nearby_index];
if (!existing_range->intersects(range))
break;
was_purged |= existing_range->was_purged;
existing_range->subtract_intersecting(remove_range);
if (existing_range->is_empty()) {
m_ranges.remove(nearby_index);
break;
}
}
}
return true;
}
bool VolatilePageRanges::intersects(const VolatilePageRange& range) const
{
auto* existing_range = binary_search(
m_ranges.span(), range, nullptr, [](auto& a, auto& b) {
if (a.intersects(b))
return 0;
return (signed)(a.base - (b.base + b.count - 1));
});
return existing_range != nullptr;
}
PurgeablePageRanges::PurgeablePageRanges(const VMObject& vmobject)
: m_volatile_ranges({ 0, vmobject.is_anonymous() ? vmobject.page_count() : 0 })
{
}
bool PurgeablePageRanges::add_volatile_range(const VolatilePageRange& range)
{
if (range.is_empty())
return false;
// Since we may need to call into AnonymousVMObject we need to acquire
// its lock as well, and acquire it first. This is important so that
// we don't deadlock when a page fault (e.g. on another processor)
// happens that is meant to lazy-allocate a committed page. It would
// call into AnonymousVMObject::range_made_volatile, which then would
// also call into this object and need to acquire m_lock. By acquiring
// the vmobject lock first in both cases, we avoid deadlocking.
// We can access m_vmobject without any locks for that purpose because
// add_volatile_range and remove_volatile_range can only be called
// by same object that calls set_vmobject.
ScopedSpinLock vmobject_lock(m_vmobject->m_lock);
ScopedSpinLock lock(m_volatile_ranges_lock);
bool added = m_volatile_ranges.add(range);
if (added)
m_vmobject->range_made_volatile(range);
return added;
}
auto PurgeablePageRanges::remove_volatile_range(const VolatilePageRange& range, bool& was_purged) -> RemoveVolatileError
{
if (range.is_empty()) {
was_purged = false;
return RemoveVolatileError::Success;
}
ScopedSpinLock vmobject_lock(m_vmobject->m_lock); // see comment in add_volatile_range
ScopedSpinLock lock(m_volatile_ranges_lock);
VERIFY(m_vmobject);
// Before we actually remove this range, we need to check if we need
// to commit any pages, which may fail. If it fails, we don't actually
// want to make any modifications. COW pages are already accounted for
// in m_shared_committed_cow_pages
size_t need_commit_pages = 0;
m_volatile_ranges.for_each_intersecting_range(range, [&](const VolatilePageRange& intersected_range) {
need_commit_pages += m_vmobject->count_needed_commit_pages_for_nonvolatile_range(intersected_range);
return IterationDecision::Continue;
});
if (need_commit_pages > 0) {
// See if we can grab enough pages for what we're marking non-volatile
if (!MM.commit_user_physical_pages(need_commit_pages))
return RemoveVolatileError::OutOfMemory;
// Now that we are committed to these pages, mark them for lazy-commit allocation
auto pages_to_mark = need_commit_pages;
m_volatile_ranges.for_each_intersecting_range(range, [&](const VolatilePageRange& intersected_range) {
auto pages_marked = m_vmobject->mark_committed_pages_for_nonvolatile_range(intersected_range, pages_to_mark);
pages_to_mark -= pages_marked;
return IterationDecision::Continue;
});
}
// Now actually remove the range
if (m_volatile_ranges.remove(range, was_purged)) {
m_vmobject->range_made_nonvolatile(range);
return RemoveVolatileError::Success;
}
VERIFY(need_commit_pages == 0); // We should have not touched anything
return RemoveVolatileError::SuccessNoChange;
}
bool PurgeablePageRanges::is_volatile_range(const VolatilePageRange& range) const
{
if (range.is_empty())
return false;
ScopedSpinLock lock(m_volatile_ranges_lock);
return m_volatile_ranges.intersects(range);
}
bool PurgeablePageRanges::is_volatile(size_t index) const
{
ScopedSpinLock lock(m_volatile_ranges_lock);
return m_volatile_ranges.contains(index);
}
void PurgeablePageRanges::set_was_purged(const VolatilePageRange& range)
{
ScopedSpinLock lock(m_volatile_ranges_lock);
m_volatile_ranges.add({ range.base, range.count, true });
}
void PurgeablePageRanges::set_vmobject(AnonymousVMObject* vmobject)
{
// No lock needed here
if (vmobject) {
VERIFY(!m_vmobject);
m_vmobject = vmobject;
} else {
VERIFY(m_vmobject);
m_vmobject = nullptr;
}
}
CommittedCowPages::CommittedCowPages(size_t committed_pages)
: m_committed_pages(committed_pages)
{
}
CommittedCowPages::~CommittedCowPages()
{
// Return unused committed pages
if (m_committed_pages > 0)
MM.uncommit_user_physical_pages(m_committed_pages);
}
NonnullRefPtr<PhysicalPage> CommittedCowPages::allocate_one()
{
VERIFY(m_committed_pages > 0);
m_committed_pages--;
return MM.allocate_committed_user_physical_page(MemoryManager::ShouldZeroFill::Yes);
}
bool CommittedCowPages::return_one()
{
VERIFY(m_committed_pages > 0);
m_committed_pages--;
MM.uncommit_user_physical_pages(1);
return m_committed_pages == 0;
}
}