forked from Bananymous/banan-os
Kernel: Rewrite DiskCache
We now cache only clean pages since I don't want to think about syncing the disk later.
This commit is contained in:
parent
6c0f864a6e
commit
a9cf9bceef
|
@ -7,16 +7,14 @@
|
||||||
namespace Kernel
|
namespace Kernel
|
||||||
{
|
{
|
||||||
|
|
||||||
class StorageDevice;
|
|
||||||
|
|
||||||
class DiskCache
|
class DiskCache
|
||||||
{
|
{
|
||||||
public:
|
public:
|
||||||
DiskCache(StorageDevice&);
|
DiskCache(size_t sector_size);
|
||||||
~DiskCache();
|
~DiskCache();
|
||||||
|
|
||||||
BAN::ErrorOr<void> read_sector(uint64_t sector, uint8_t* buffer);
|
bool read_from_cache(uint64_t sector, uint8_t* buffer);
|
||||||
BAN::ErrorOr<void> write_sector(uint64_t sector, const uint8_t* buffer);
|
BAN::ErrorOr<void> write_to_cache(uint64_t sector, const uint8_t* buffer, bool dirty);
|
||||||
|
|
||||||
void sync();
|
void sync();
|
||||||
size_t release_clean_pages(size_t);
|
size_t release_clean_pages(size_t);
|
||||||
|
@ -30,16 +28,11 @@ namespace Kernel
|
||||||
uint64_t first_sector { 0 };
|
uint64_t first_sector { 0 };
|
||||||
uint8_t sector_mask { 0 };
|
uint8_t sector_mask { 0 };
|
||||||
uint8_t dirty_mask { 0 };
|
uint8_t dirty_mask { 0 };
|
||||||
|
|
||||||
void sync(StorageDevice&);
|
|
||||||
BAN::ErrorOr<void> read_sector(StorageDevice&, uint64_t sector, uint8_t* buffer);
|
|
||||||
BAN::ErrorOr<void> write_sector(StorageDevice&, uint64_t sector, const uint8_t* buffer);
|
|
||||||
};
|
};
|
||||||
|
|
||||||
private:
|
private:
|
||||||
SpinLock m_lock;
|
const size_t m_sector_size;
|
||||||
StorageDevice& m_device;
|
BAN::Vector<PageCache> m_cache;
|
||||||
BAN::Vector<PageCache> m_cache;
|
|
||||||
};
|
};
|
||||||
|
|
||||||
}
|
}
|
|
@ -81,8 +81,8 @@ namespace Kernel
|
||||||
void add_disk_cache();
|
void add_disk_cache();
|
||||||
|
|
||||||
private:
|
private:
|
||||||
DiskCache* m_disk_cache { nullptr };
|
BAN::Optional<DiskCache> m_disk_cache;
|
||||||
BAN::Vector<Partition*> m_partitions;
|
BAN::Vector<Partition*> m_partitions;
|
||||||
|
|
||||||
friend class DiskCache;
|
friend class DiskCache;
|
||||||
};
|
};
|
||||||
|
|
|
@ -8,105 +8,121 @@
|
||||||
namespace Kernel
|
namespace Kernel
|
||||||
{
|
{
|
||||||
|
|
||||||
DiskCache::DiskCache(StorageDevice& device)
|
DiskCache::DiskCache(size_t sector_size)
|
||||||
: m_device(device)
|
: m_sector_size(sector_size)
|
||||||
{ }
|
{
|
||||||
|
ASSERT(PAGE_SIZE % m_sector_size == 0);
|
||||||
|
ASSERT(PAGE_SIZE / m_sector_size <= sizeof(PageCache::sector_mask) * 8);
|
||||||
|
ASSERT(PAGE_SIZE / m_sector_size <= sizeof(PageCache::dirty_mask) * 8);
|
||||||
|
}
|
||||||
|
|
||||||
DiskCache::~DiskCache()
|
DiskCache::~DiskCache()
|
||||||
{
|
{
|
||||||
release_all_pages();
|
release_all_pages();
|
||||||
}
|
}
|
||||||
|
|
||||||
BAN::ErrorOr<void> DiskCache::read_sector(uint64_t sector, uint8_t* buffer)
|
bool DiskCache::read_from_cache(uint64_t sector, uint8_t* buffer)
|
||||||
{
|
{
|
||||||
ASSERT(m_device.sector_size() <= PAGE_SIZE);
|
uint64_t sectors_per_page = PAGE_SIZE / m_sector_size;
|
||||||
|
uint64_t page_cache_offset = sector % sectors_per_page;
|
||||||
|
uint64_t page_cache_start = sector - page_cache_offset;
|
||||||
|
|
||||||
LockGuard _(m_lock);
|
PageTable& page_table = PageTable::current();
|
||||||
|
LockGuard page_table_locker(page_table);
|
||||||
|
ASSERT(page_table.is_page_free(0));
|
||||||
|
|
||||||
uint64_t sectors_per_page = PAGE_SIZE / m_device.sector_size();
|
CriticalScope _;
|
||||||
ASSERT(sectors_per_page <= sizeof(PageCache::sector_mask) * 8);
|
for (auto& cache : m_cache)
|
||||||
|
{
|
||||||
|
if (cache.first_sector < page_cache_start)
|
||||||
|
continue;
|
||||||
|
if (cache.first_sector > page_cache_start)
|
||||||
|
break;
|
||||||
|
|
||||||
uint64_t page_cache_start = sector / sectors_per_page * sectors_per_page;
|
if (!(cache.sector_mask & (1 << page_cache_offset)))
|
||||||
|
continue;
|
||||||
|
|
||||||
|
page_table.map_page_at(cache.paddr, 0, PageTable::Flags::Present);
|
||||||
|
memcpy(buffer, (void*)(page_cache_offset * m_sector_size), m_sector_size);
|
||||||
|
page_table.unmap_page(0);
|
||||||
|
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
return false;
|
||||||
|
};
|
||||||
|
|
||||||
|
BAN::ErrorOr<void> DiskCache::write_to_cache(uint64_t sector, const uint8_t* buffer, bool dirty)
|
||||||
|
{
|
||||||
|
uint64_t sectors_per_page = PAGE_SIZE / m_sector_size;
|
||||||
|
uint64_t page_cache_offset = sector % sectors_per_page;
|
||||||
|
uint64_t page_cache_start = sector - page_cache_offset;
|
||||||
|
|
||||||
|
PageTable& page_table = PageTable::current();
|
||||||
|
LockGuard page_table_locker(page_table);
|
||||||
|
ASSERT(page_table.is_page_free(0));
|
||||||
|
|
||||||
|
CriticalScope _;
|
||||||
|
|
||||||
// Check if we already have a cache for this page
|
|
||||||
// FIXME: binary search
|
|
||||||
size_t index = 0;
|
size_t index = 0;
|
||||||
|
|
||||||
|
// Search the cache if the have this sector in memory
|
||||||
for (; index < m_cache.size(); index++)
|
for (; index < m_cache.size(); index++)
|
||||||
{
|
{
|
||||||
if (m_cache[index].first_sector < page_cache_start)
|
auto& cache = m_cache[index];
|
||||||
|
|
||||||
|
if (cache.first_sector < page_cache_start)
|
||||||
continue;
|
continue;
|
||||||
if (m_cache[index].first_sector > page_cache_start)
|
if (cache.first_sector > page_cache_start)
|
||||||
break;
|
break;
|
||||||
TRY(m_cache[index].read_sector(m_device, sector, buffer));
|
|
||||||
|
page_table.map_page_at(cache.paddr, 0, PageTable::Flags::ReadWrite | PageTable::Flags::Present);
|
||||||
|
memcpy((void*)(page_cache_offset * m_sector_size), buffer, m_sector_size);
|
||||||
|
page_table.unmap_page(0);
|
||||||
|
|
||||||
|
cache.sector_mask |= 1 << page_cache_offset;
|
||||||
|
if (dirty)
|
||||||
|
cache.dirty_mask |= 1 << page_cache_offset;
|
||||||
|
|
||||||
return {};
|
return {};
|
||||||
}
|
}
|
||||||
|
|
||||||
// Try to allocate new cache
|
// Try to add new page to the cache
|
||||||
if (paddr_t paddr = Heap::get().take_free_page())
|
paddr_t paddr = Heap::get().take_free_page();
|
||||||
|
if (paddr == 0)
|
||||||
|
return BAN::Error::from_errno(ENOMEM);
|
||||||
|
|
||||||
|
PageCache cache;
|
||||||
|
cache.paddr = paddr;
|
||||||
|
cache.first_sector = page_cache_start;
|
||||||
|
cache.sector_mask = 1 << page_cache_offset;
|
||||||
|
cache.dirty_mask = dirty ? cache.sector_mask : 0;
|
||||||
|
|
||||||
|
if (auto ret = m_cache.insert(index, cache); ret.is_error())
|
||||||
{
|
{
|
||||||
MUST(m_cache.insert(index, { .paddr = paddr, .first_sector = page_cache_start }));
|
Heap::get().release_page(paddr);
|
||||||
TRY(m_cache[index].read_sector(m_device, sector, buffer));
|
return ret.error();
|
||||||
return {};
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Could not allocate new cache, read from disk
|
page_table.map_page_at(cache.paddr, 0, PageTable::Flags::Present);
|
||||||
TRY(m_device.read_sectors_impl(sector, 1, buffer));
|
memcpy((void*)(page_cache_offset * m_sector_size), buffer, m_sector_size);
|
||||||
return {};
|
page_table.unmap_page(0);
|
||||||
}
|
|
||||||
|
|
||||||
BAN::ErrorOr<void> DiskCache::write_sector(uint64_t sector, const uint8_t* buffer)
|
|
||||||
{
|
|
||||||
ASSERT(m_device.sector_size() <= PAGE_SIZE);
|
|
||||||
|
|
||||||
LockGuard _(m_lock);
|
|
||||||
|
|
||||||
uint64_t sectors_per_page = PAGE_SIZE / m_device.sector_size();
|
|
||||||
ASSERT(sectors_per_page <= sizeof(PageCache::sector_mask) * 8);
|
|
||||||
|
|
||||||
uint64_t page_cache_start = sector / sectors_per_page * sectors_per_page;
|
|
||||||
|
|
||||||
// Check if we already have a cache for this page
|
|
||||||
// FIXME: binary search
|
|
||||||
size_t index = 0;
|
|
||||||
for (; index < m_cache.size(); index++)
|
|
||||||
{
|
|
||||||
if (m_cache[index].first_sector < page_cache_start)
|
|
||||||
continue;
|
|
||||||
if (m_cache[index].first_sector > page_cache_start)
|
|
||||||
break;
|
|
||||||
TRY(m_cache[index].write_sector(m_device, sector, buffer));
|
|
||||||
return {};
|
|
||||||
}
|
|
||||||
|
|
||||||
// Try to allocate new cache
|
|
||||||
if (paddr_t paddr = Heap::get().take_free_page())
|
|
||||||
{
|
|
||||||
MUST(m_cache.insert(index, { .paddr = paddr, .first_sector = page_cache_start }));
|
|
||||||
TRY(m_cache[index].write_sector(m_device, sector, buffer));
|
|
||||||
return {};
|
|
||||||
}
|
|
||||||
|
|
||||||
// Could not allocate new cache, write to disk
|
|
||||||
TRY(m_device.write_sectors_impl(sector, 1, buffer));
|
|
||||||
return {};
|
return {};
|
||||||
}
|
}
|
||||||
|
|
||||||
void DiskCache::sync()
|
void DiskCache::sync()
|
||||||
{
|
{
|
||||||
ASSERT(m_device.sector_size() <= PAGE_SIZE);
|
CriticalScope _;
|
||||||
|
for (auto& cache : m_cache)
|
||||||
LockGuard _(m_lock);
|
ASSERT(cache.dirty_mask == 0);
|
||||||
for (auto& cache_block : m_cache)
|
|
||||||
cache_block.sync(m_device);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
size_t DiskCache::release_clean_pages(size_t page_count)
|
size_t DiskCache::release_clean_pages(size_t page_count)
|
||||||
{
|
{
|
||||||
ASSERT(m_device.sector_size() <= PAGE_SIZE);
|
|
||||||
|
|
||||||
// NOTE: There might not actually be page_count pages after this
|
// NOTE: There might not actually be page_count pages after this
|
||||||
// function returns. The synchronization must be done elsewhere.
|
// function returns. The synchronization must be done elsewhere.
|
||||||
LockGuard _(m_lock);
|
CriticalScope _;
|
||||||
|
|
||||||
size_t released = 0;
|
size_t released = 0;
|
||||||
for (size_t i = 0; i < m_cache.size() && released < page_count;)
|
for (size_t i = 0; i < m_cache.size() && released < page_count;)
|
||||||
|
@ -128,138 +144,16 @@ namespace Kernel
|
||||||
|
|
||||||
size_t DiskCache::release_pages(size_t page_count)
|
size_t DiskCache::release_pages(size_t page_count)
|
||||||
{
|
{
|
||||||
ASSERT(m_device.sector_size() <= PAGE_SIZE);
|
|
||||||
|
|
||||||
size_t released = release_clean_pages(page_count);
|
size_t released = release_clean_pages(page_count);
|
||||||
if (released >= page_count)
|
if (released >= page_count)
|
||||||
return released;
|
return released;
|
||||||
|
|
||||||
// NOTE: There might not actually be page_count pages after this
|
ASSERT_NOT_REACHED();
|
||||||
// function returns. The synchronization must be done elsewhere.
|
|
||||||
LockGuard _(m_lock);
|
|
||||||
|
|
||||||
while (!m_cache.empty() && released < page_count)
|
|
||||||
{
|
|
||||||
m_cache.back().sync(m_device);
|
|
||||||
Heap::get().release_page(m_cache.back().paddr);
|
|
||||||
m_cache.pop_back();
|
|
||||||
released++;
|
|
||||||
}
|
|
||||||
|
|
||||||
(void)m_cache.shrink_to_fit();
|
|
||||||
|
|
||||||
return released;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
void DiskCache::release_all_pages()
|
void DiskCache::release_all_pages()
|
||||||
{
|
{
|
||||||
ASSERT(m_device.sector_size() <= PAGE_SIZE);
|
release_pages(m_cache.size());
|
||||||
|
|
||||||
LockGuard _(m_lock);
|
|
||||||
|
|
||||||
for (auto& cache_block : m_cache)
|
|
||||||
{
|
|
||||||
cache_block.sync(m_device);
|
|
||||||
Heap::get().release_page(cache_block.paddr);
|
|
||||||
}
|
|
||||||
|
|
||||||
m_cache.clear();
|
|
||||||
}
|
|
||||||
|
|
||||||
void DiskCache::PageCache::sync(StorageDevice& device)
|
|
||||||
{
|
|
||||||
if (this->dirty_mask == 0)
|
|
||||||
return;
|
|
||||||
|
|
||||||
ASSERT(device.sector_size() <= PAGE_SIZE);
|
|
||||||
|
|
||||||
PageTable& page_table = PageTable::current();
|
|
||||||
|
|
||||||
page_table.lock();
|
|
||||||
ASSERT(page_table.is_page_free(0));
|
|
||||||
|
|
||||||
page_table.map_page_at(this->paddr, 0, PageTable::Flags::ReadWrite | PageTable::Flags::Present);
|
|
||||||
|
|
||||||
for (size_t i = 0; i < PAGE_SIZE / device.sector_size(); i++)
|
|
||||||
{
|
|
||||||
if (!(this->dirty_mask & (1 << i)))
|
|
||||||
continue;
|
|
||||||
MUST(device.write_sectors_impl(this->first_sector + i, 1, (const uint8_t*)(i * device.sector_size())));
|
|
||||||
// FIXME: race condition between here :)
|
|
||||||
this->dirty_mask &= ~(1 << i);
|
|
||||||
}
|
|
||||||
|
|
||||||
page_table.unmap_page(0);
|
|
||||||
|
|
||||||
page_table.unlock();
|
|
||||||
}
|
|
||||||
|
|
||||||
BAN::ErrorOr<void> DiskCache::PageCache::read_sector(StorageDevice& device, uint64_t sector, uint8_t* buffer)
|
|
||||||
{
|
|
||||||
ASSERT(device.sector_size() <= PAGE_SIZE);
|
|
||||||
|
|
||||||
uint64_t sectors_per_page = PAGE_SIZE / device.sector_size();
|
|
||||||
uint64_t sector_offset = sector - this->first_sector;
|
|
||||||
|
|
||||||
ASSERT(sector_offset < sectors_per_page);
|
|
||||||
|
|
||||||
PageTable& page_table = PageTable::current();
|
|
||||||
|
|
||||||
page_table.lock();
|
|
||||||
ASSERT(page_table.is_page_free(0));
|
|
||||||
|
|
||||||
page_table.map_page_at(this->paddr, 0, PageTable::Flags::ReadWrite | PageTable::Flags::Present);
|
|
||||||
|
|
||||||
// Sector not yet cached
|
|
||||||
if (!(this->sector_mask & (1 << sector_offset)))
|
|
||||||
{
|
|
||||||
TRY(device.read_sectors_impl(sector, 1, buffer));
|
|
||||||
|
|
||||||
CriticalScope _;
|
|
||||||
memcpy((void*)(sector_offset * device.sector_size()), buffer, device.sector_size());
|
|
||||||
this->sector_mask |= 1 << sector_offset;
|
|
||||||
}
|
|
||||||
else
|
|
||||||
{
|
|
||||||
CriticalScope _;
|
|
||||||
memcpy(buffer, (const void*)(sector_offset * device.sector_size()), device.sector_size());
|
|
||||||
}
|
|
||||||
|
|
||||||
page_table.unmap_page(0);
|
|
||||||
|
|
||||||
page_table.unlock();
|
|
||||||
|
|
||||||
return {};
|
|
||||||
}
|
|
||||||
|
|
||||||
BAN::ErrorOr<void> DiskCache::PageCache::write_sector(StorageDevice& device, uint64_t sector, const uint8_t* buffer)
|
|
||||||
{
|
|
||||||
ASSERT(device.sector_size() <= PAGE_SIZE);
|
|
||||||
|
|
||||||
uint64_t sectors_per_page = PAGE_SIZE / device.sector_size();
|
|
||||||
uint64_t sector_offset = sector - this->first_sector;
|
|
||||||
|
|
||||||
ASSERT(sector_offset < sectors_per_page);
|
|
||||||
|
|
||||||
PageTable& page_table = PageTable::current();
|
|
||||||
|
|
||||||
page_table.lock();
|
|
||||||
ASSERT(page_table.is_page_free(0));
|
|
||||||
|
|
||||||
page_table.map_page_at(this->paddr, 0, PageTable::Flags::ReadWrite | PageTable::Flags::Present);
|
|
||||||
|
|
||||||
{
|
|
||||||
CriticalScope _;
|
|
||||||
memcpy((void*)(sector_offset * device.sector_size()), buffer, device.sector_size());
|
|
||||||
this->sector_mask |= 1 << sector_offset;
|
|
||||||
this->dirty_mask |= 1 << sector_offset;
|
|
||||||
}
|
|
||||||
|
|
||||||
page_table.unmap_page(0);
|
|
||||||
|
|
||||||
page_table.unlock();
|
|
||||||
|
|
||||||
return {};
|
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
|
@ -251,35 +251,41 @@ namespace Kernel
|
||||||
|
|
||||||
StorageDevice::~StorageDevice()
|
StorageDevice::~StorageDevice()
|
||||||
{
|
{
|
||||||
if (m_disk_cache)
|
|
||||||
delete m_disk_cache;
|
|
||||||
m_disk_cache = nullptr;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
void StorageDevice::add_disk_cache()
|
void StorageDevice::add_disk_cache()
|
||||||
{
|
{
|
||||||
ASSERT(m_disk_cache == nullptr);
|
ASSERT(!m_disk_cache.has_value());
|
||||||
m_disk_cache = new DiskCache(*this);
|
m_disk_cache = DiskCache(sector_size());
|
||||||
ASSERT(m_disk_cache);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
BAN::ErrorOr<void> StorageDevice::read_sectors(uint64_t lba, uint8_t sector_count, uint8_t* buffer)
|
BAN::ErrorOr<void> StorageDevice::read_sectors(uint64_t lba, uint8_t sector_count, uint8_t* buffer)
|
||||||
{
|
{
|
||||||
if (!m_disk_cache)
|
for (uint8_t offset = 0; offset < sector_count; offset++)
|
||||||
return read_sectors_impl(lba, sector_count, buffer);
|
{
|
||||||
for (uint8_t sector = 0; sector < sector_count; sector++)
|
uint8_t* buffer_ptr = buffer + offset * sector_size();
|
||||||
TRY(m_disk_cache->read_sector(lba + sector, buffer + sector * sector_size()));
|
if (m_disk_cache.has_value())
|
||||||
|
if (m_disk_cache->read_from_cache(lba + offset, buffer_ptr))
|
||||||
|
continue;
|
||||||
|
TRY(read_sectors_impl(lba + offset, 1, buffer_ptr));
|
||||||
|
if (m_disk_cache.has_value())
|
||||||
|
(void)m_disk_cache->write_to_cache(lba + offset, buffer_ptr, false);
|
||||||
|
}
|
||||||
|
|
||||||
return {};
|
return {};
|
||||||
}
|
}
|
||||||
|
|
||||||
BAN::ErrorOr<void> StorageDevice::write_sectors(uint64_t lba, uint8_t sector_count, const uint8_t* buffer)
|
BAN::ErrorOr<void> StorageDevice::write_sectors(uint64_t lba, uint8_t sector_count, const uint8_t* buffer)
|
||||||
{
|
{
|
||||||
if (!m_disk_cache)
|
// TODO: use disk cache for dirty pages. I don't wanna think about how to do it safely now
|
||||||
return write_sectors_impl(lba, sector_count, buffer);
|
|
||||||
for (uint8_t sector = 0; sector < sector_count; sector++)
|
for (uint8_t sector = 0; sector < sector_count; sector++)
|
||||||
TRY(m_disk_cache->write_sector(lba + sector, buffer + sector * sector_size()));
|
{
|
||||||
|
TRY(write_sectors_impl(lba + sector, 1, buffer));
|
||||||
|
if (m_disk_cache.has_value())
|
||||||
|
(void)m_disk_cache->write_to_cache(lba + sector, buffer + sector * sector_size(), false);
|
||||||
|
}
|
||||||
|
|
||||||
return {};
|
return {};
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
}
|
}
|
Loading…
Reference in New Issue