Kernel: Writes to disk are not synchronous anymore
Implement "proper" DiskCache syncing
This commit is contained in:
parent
82b049204d
commit
b0c22b61ec
|
@ -7,16 +7,18 @@
|
||||||
namespace Kernel
|
namespace Kernel
|
||||||
{
|
{
|
||||||
|
|
||||||
|
class StorageDevice;
|
||||||
|
|
||||||
class DiskCache
|
class DiskCache
|
||||||
{
|
{
|
||||||
public:
|
public:
|
||||||
DiskCache(size_t sector_size);
|
DiskCache(size_t sector_size, StorageDevice&);
|
||||||
~DiskCache();
|
~DiskCache();
|
||||||
|
|
||||||
bool read_from_cache(uint64_t sector, uint8_t* buffer);
|
bool read_from_cache(uint64_t sector, uint8_t* buffer);
|
||||||
BAN::ErrorOr<void> write_to_cache(uint64_t sector, const uint8_t* buffer, bool dirty);
|
BAN::ErrorOr<void> write_to_cache(uint64_t sector, const uint8_t* buffer, bool dirty);
|
||||||
|
|
||||||
void sync();
|
BAN::ErrorOr<void> sync();
|
||||||
size_t release_clean_pages(size_t);
|
size_t release_clean_pages(size_t);
|
||||||
size_t release_pages(size_t);
|
size_t release_pages(size_t);
|
||||||
void release_all_pages();
|
void release_all_pages();
|
||||||
|
@ -32,6 +34,7 @@ namespace Kernel
|
||||||
|
|
||||||
private:
|
private:
|
||||||
const size_t m_sector_size;
|
const size_t m_sector_size;
|
||||||
|
StorageDevice& m_device;
|
||||||
BAN::Vector<PageCache> m_cache;
|
BAN::Vector<PageCache> m_cache;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
|
@ -76,12 +76,15 @@ namespace Kernel
|
||||||
BAN::Vector<Partition*>& partitions() { return m_partitions; }
|
BAN::Vector<Partition*>& partitions() { return m_partitions; }
|
||||||
const BAN::Vector<Partition*>& partitions() const { return m_partitions; }
|
const BAN::Vector<Partition*>& partitions() const { return m_partitions; }
|
||||||
|
|
||||||
|
BAN::ErrorOr<void> sync_disk_cache();
|
||||||
|
|
||||||
protected:
|
protected:
|
||||||
virtual BAN::ErrorOr<void> read_sectors_impl(uint64_t lba, uint8_t sector_count, uint8_t* buffer) = 0;
|
virtual BAN::ErrorOr<void> read_sectors_impl(uint64_t lba, uint8_t sector_count, uint8_t* buffer) = 0;
|
||||||
virtual BAN::ErrorOr<void> write_sectors_impl(uint64_t lba, uint8_t sector_count, const uint8_t* buffer) = 0;
|
virtual BAN::ErrorOr<void> write_sectors_impl(uint64_t lba, uint8_t sector_count, const uint8_t* buffer) = 0;
|
||||||
void add_disk_cache();
|
void add_disk_cache();
|
||||||
|
|
||||||
private:
|
private:
|
||||||
|
SpinLock m_lock;
|
||||||
BAN::Optional<DiskCache> m_disk_cache;
|
BAN::Optional<DiskCache> m_disk_cache;
|
||||||
BAN::Vector<Partition*> m_partitions;
|
BAN::Vector<Partition*> m_partitions;
|
||||||
|
|
||||||
|
|
|
@ -8,8 +8,9 @@
|
||||||
namespace Kernel
|
namespace Kernel
|
||||||
{
|
{
|
||||||
|
|
||||||
DiskCache::DiskCache(size_t sector_size)
|
DiskCache::DiskCache(size_t sector_size, StorageDevice& device)
|
||||||
: m_sector_size(sector_size)
|
: m_sector_size(sector_size)
|
||||||
|
, m_device(device)
|
||||||
{
|
{
|
||||||
ASSERT(PAGE_SIZE % m_sector_size == 0);
|
ASSERT(PAGE_SIZE % m_sector_size == 0);
|
||||||
ASSERT(PAGE_SIZE / m_sector_size <= sizeof(PageCache::sector_mask) * 8);
|
ASSERT(PAGE_SIZE / m_sector_size <= sizeof(PageCache::sector_mask) * 8);
|
||||||
|
@ -31,7 +32,6 @@ namespace Kernel
|
||||||
LockGuard page_table_locker(page_table);
|
LockGuard page_table_locker(page_table);
|
||||||
ASSERT(page_table.is_page_free(0));
|
ASSERT(page_table.is_page_free(0));
|
||||||
|
|
||||||
CriticalScope _;
|
|
||||||
for (auto& cache : m_cache)
|
for (auto& cache : m_cache)
|
||||||
{
|
{
|
||||||
if (cache.first_sector < page_cache_start)
|
if (cache.first_sector < page_cache_start)
|
||||||
|
@ -42,6 +42,7 @@ namespace Kernel
|
||||||
if (!(cache.sector_mask & (1 << page_cache_offset)))
|
if (!(cache.sector_mask & (1 << page_cache_offset)))
|
||||||
continue;
|
continue;
|
||||||
|
|
||||||
|
CriticalScope _;
|
||||||
page_table.map_page_at(cache.paddr, 0, PageTable::Flags::Present);
|
page_table.map_page_at(cache.paddr, 0, PageTable::Flags::Present);
|
||||||
memcpy(buffer, (void*)(page_cache_offset * m_sector_size), m_sector_size);
|
memcpy(buffer, (void*)(page_cache_offset * m_sector_size), m_sector_size);
|
||||||
page_table.unmap_page(0);
|
page_table.unmap_page(0);
|
||||||
|
@ -62,8 +63,6 @@ namespace Kernel
|
||||||
LockGuard page_table_locker(page_table);
|
LockGuard page_table_locker(page_table);
|
||||||
ASSERT(page_table.is_page_free(0));
|
ASSERT(page_table.is_page_free(0));
|
||||||
|
|
||||||
CriticalScope _;
|
|
||||||
|
|
||||||
size_t index = 0;
|
size_t index = 0;
|
||||||
|
|
||||||
// Search the cache if the have this sector in memory
|
// Search the cache if the have this sector in memory
|
||||||
|
@ -76,9 +75,12 @@ namespace Kernel
|
||||||
if (cache.first_sector > page_cache_start)
|
if (cache.first_sector > page_cache_start)
|
||||||
break;
|
break;
|
||||||
|
|
||||||
page_table.map_page_at(cache.paddr, 0, PageTable::Flags::ReadWrite | PageTable::Flags::Present);
|
{
|
||||||
memcpy((void*)(page_cache_offset * m_sector_size), buffer, m_sector_size);
|
CriticalScope _;
|
||||||
page_table.unmap_page(0);
|
page_table.map_page_at(cache.paddr, 0, PageTable::Flags::ReadWrite | PageTable::Flags::Present);
|
||||||
|
memcpy((void*)(page_cache_offset * m_sector_size), buffer, m_sector_size);
|
||||||
|
page_table.unmap_page(0);
|
||||||
|
}
|
||||||
|
|
||||||
cache.sector_mask |= 1 << page_cache_offset;
|
cache.sector_mask |= 1 << page_cache_offset;
|
||||||
if (dirty)
|
if (dirty)
|
||||||
|
@ -104,25 +106,51 @@ namespace Kernel
|
||||||
return ret.error();
|
return ret.error();
|
||||||
}
|
}
|
||||||
|
|
||||||
page_table.map_page_at(cache.paddr, 0, PageTable::Flags::Present);
|
{
|
||||||
memcpy((void*)(page_cache_offset * m_sector_size), buffer, m_sector_size);
|
CriticalScope _;
|
||||||
page_table.unmap_page(0);
|
page_table.map_page_at(cache.paddr, 0, PageTable::Flags::Present);
|
||||||
|
memcpy((void*)(page_cache_offset * m_sector_size), buffer, m_sector_size);
|
||||||
|
page_table.unmap_page(0);
|
||||||
|
}
|
||||||
|
|
||||||
return {};
|
return {};
|
||||||
}
|
}
|
||||||
|
|
||||||
void DiskCache::sync()
|
BAN::ErrorOr<void> DiskCache::sync()
|
||||||
{
|
{
|
||||||
CriticalScope _;
|
BAN::Vector<uint8_t> sector_buffer;
|
||||||
|
TRY(sector_buffer.resize(m_sector_size));
|
||||||
|
|
||||||
|
PageTable& page_table = PageTable::current();
|
||||||
|
LockGuard page_table_locker(page_table);
|
||||||
|
ASSERT(page_table.is_page_free(0));
|
||||||
|
|
||||||
for (auto& cache : m_cache)
|
for (auto& cache : m_cache)
|
||||||
ASSERT(cache.dirty_mask == 0);
|
{
|
||||||
|
for (int i = 0; cache.dirty_mask; i++)
|
||||||
|
{
|
||||||
|
if (!(cache.dirty_mask & (1 << i)))
|
||||||
|
continue;
|
||||||
|
|
||||||
|
{
|
||||||
|
CriticalScope _;
|
||||||
|
page_table.map_page_at(cache.paddr, 0, PageTable::Flags::Present);
|
||||||
|
memcpy(sector_buffer.data(), (void*)(i * m_sector_size), m_sector_size);
|
||||||
|
page_table.unmap_page(0);
|
||||||
|
}
|
||||||
|
|
||||||
|
TRY(m_device.write_sectors_impl(cache.first_sector + i, 1, sector_buffer.data()));
|
||||||
|
cache.dirty_mask &= ~(1 << i);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return {};
|
||||||
}
|
}
|
||||||
|
|
||||||
size_t DiskCache::release_clean_pages(size_t page_count)
|
size_t DiskCache::release_clean_pages(size_t page_count)
|
||||||
{
|
{
|
||||||
// NOTE: There might not actually be page_count pages after this
|
// NOTE: There might not actually be page_count pages after this
|
||||||
// function returns. The synchronization must be done elsewhere.
|
// function returns. The synchronization must be done elsewhere.
|
||||||
CriticalScope _;
|
|
||||||
|
|
||||||
size_t released = 0;
|
size_t released = 0;
|
||||||
for (size_t i = 0; i < m_cache.size() && released < page_count;)
|
for (size_t i = 0; i < m_cache.size() && released < page_count;)
|
||||||
|
@ -147,8 +175,9 @@ namespace Kernel
|
||||||
size_t released = release_clean_pages(page_count);
|
size_t released = release_clean_pages(page_count);
|
||||||
if (released >= page_count)
|
if (released >= page_count)
|
||||||
return released;
|
return released;
|
||||||
|
if (!sync().is_error())
|
||||||
ASSERT_NOT_REACHED();
|
released += release_clean_pages(page_count - released);
|
||||||
|
return released;
|
||||||
}
|
}
|
||||||
|
|
||||||
void DiskCache::release_all_pages()
|
void DiskCache::release_all_pages()
|
||||||
|
|
|
@ -3,6 +3,7 @@
|
||||||
#include <BAN/StringView.h>
|
#include <BAN/StringView.h>
|
||||||
#include <BAN/UTF8.h>
|
#include <BAN/UTF8.h>
|
||||||
#include <kernel/FS/VirtualFileSystem.h>
|
#include <kernel/FS/VirtualFileSystem.h>
|
||||||
|
#include <kernel/LockGuard.h>
|
||||||
#include <kernel/PCI.h>
|
#include <kernel/PCI.h>
|
||||||
#include <kernel/Storage/StorageDevice.h>
|
#include <kernel/Storage/StorageDevice.h>
|
||||||
#include <kernel/Thread.h>
|
#include <kernel/Thread.h>
|
||||||
|
@ -257,15 +258,26 @@ namespace Kernel
|
||||||
|
|
||||||
void StorageDevice::add_disk_cache()
|
void StorageDevice::add_disk_cache()
|
||||||
{
|
{
|
||||||
|
LockGuard _(m_lock);
|
||||||
ASSERT(!m_disk_cache.has_value());
|
ASSERT(!m_disk_cache.has_value());
|
||||||
m_disk_cache = DiskCache(sector_size());
|
m_disk_cache = DiskCache(sector_size(), *this);
|
||||||
|
}
|
||||||
|
|
||||||
|
BAN::ErrorOr<void> StorageDevice::sync_disk_cache()
|
||||||
|
{
|
||||||
|
LockGuard _(m_lock);
|
||||||
|
if (m_disk_cache.has_value())
|
||||||
|
TRY(m_disk_cache->sync());
|
||||||
|
return {};
|
||||||
}
|
}
|
||||||
|
|
||||||
BAN::ErrorOr<void> StorageDevice::read_sectors(uint64_t lba, uint8_t sector_count, uint8_t* buffer)
|
BAN::ErrorOr<void> StorageDevice::read_sectors(uint64_t lba, uint8_t sector_count, uint8_t* buffer)
|
||||||
{
|
{
|
||||||
for (uint8_t offset = 0; offset < sector_count; offset++)
|
for (uint8_t offset = 0; offset < sector_count; offset++)
|
||||||
{
|
{
|
||||||
Thread::TerminateBlocker _(Thread::current());
|
LockGuard _(m_lock);
|
||||||
|
Thread::TerminateBlocker blocker(Thread::current());
|
||||||
|
|
||||||
uint8_t* buffer_ptr = buffer + offset * sector_size();
|
uint8_t* buffer_ptr = buffer + offset * sector_size();
|
||||||
if (m_disk_cache.has_value())
|
if (m_disk_cache.has_value())
|
||||||
if (m_disk_cache->read_from_cache(lba + offset, buffer_ptr))
|
if (m_disk_cache->read_from_cache(lba + offset, buffer_ptr))
|
||||||
|
@ -283,11 +295,12 @@ namespace Kernel
|
||||||
// TODO: use disk cache for dirty pages. I don't wanna think about how to do it safely now
|
// TODO: use disk cache for dirty pages. I don't wanna think about how to do it safely now
|
||||||
for (uint8_t offset = 0; offset < sector_count; offset++)
|
for (uint8_t offset = 0; offset < sector_count; offset++)
|
||||||
{
|
{
|
||||||
Thread::TerminateBlocker _(Thread::current());
|
LockGuard _(m_lock);
|
||||||
|
Thread::TerminateBlocker blocker(Thread::current());
|
||||||
|
|
||||||
const uint8_t* buffer_ptr = buffer + offset * sector_size();
|
const uint8_t* buffer_ptr = buffer + offset * sector_size();
|
||||||
TRY(write_sectors_impl(lba + offset, 1, buffer_ptr));
|
if (!m_disk_cache.has_value() || m_disk_cache->write_to_cache(lba + offset, buffer_ptr, true).is_error())
|
||||||
if (m_disk_cache.has_value())
|
TRY(write_sectors_impl(lba + offset, 1, buffer_ptr));
|
||||||
(void)m_disk_cache->write_to_cache(lba + offset, buffer_ptr, false);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return {};
|
return {};
|
||||||
|
|
Loading…
Reference in New Issue