mirror of
https://github.com/RGBCube/serenity
synced 2025-05-16 12:04:58 +00:00

In most cases it's safe to abort the requested operation and go forward, however, in some places it's not clear yet how to handle these failures, therefore, we use the MUST() wrapper to force a kernel panic for now.
241 lines
9.5 KiB
C++
241 lines
9.5 KiB
C++
/*
|
|
* Copyright (c) 2021, Liav A. <liavalb@hotmail.co.il>
|
|
*
|
|
* SPDX-License-Identifier: BSD-2-Clause
|
|
*/
|
|
|
|
#include <Kernel/Bus/PCI/API.h>
|
|
#include <Kernel/Sections.h>
|
|
#include <Kernel/Storage/ATA/ATA.h>
|
|
#include <Kernel/Storage/ATA/BMIDEChannel.h>
|
|
#include <Kernel/Storage/ATA/IDEController.h>
|
|
#include <Kernel/WorkQueue.h>
|
|
|
|
namespace Kernel {
|
|
|
|
UNMAP_AFTER_INIT NonnullRefPtr<BMIDEChannel> BMIDEChannel::create(IDEController const& ide_controller, IDEChannel::IOAddressGroup io_group, IDEChannel::ChannelType type)
|
|
{
|
|
return adopt_ref(*new BMIDEChannel(ide_controller, io_group, type));
|
|
}
|
|
|
|
UNMAP_AFTER_INIT NonnullRefPtr<BMIDEChannel> BMIDEChannel::create(IDEController const& ide_controller, u8 irq, IDEChannel::IOAddressGroup io_group, IDEChannel::ChannelType type)
|
|
{
|
|
return adopt_ref(*new BMIDEChannel(ide_controller, irq, io_group, type));
|
|
}
|
|
|
|
UNMAP_AFTER_INIT BMIDEChannel::BMIDEChannel(IDEController const& controller, IDEChannel::IOAddressGroup io_group, IDEChannel::ChannelType type)
|
|
: IDEChannel(controller, io_group, type)
|
|
{
|
|
initialize();
|
|
}
|
|
|
|
UNMAP_AFTER_INIT BMIDEChannel::BMIDEChannel(IDEController const& controller, u8 irq, IDEChannel::IOAddressGroup io_group, IDEChannel::ChannelType type)
|
|
: IDEChannel(controller, irq, io_group, type)
|
|
{
|
|
initialize();
|
|
}
|
|
|
|
UNMAP_AFTER_INIT void BMIDEChannel::initialize()
|
|
{
|
|
VERIFY(m_io_group.bus_master_base().has_value());
|
|
// Let's try to set up DMA transfers.
|
|
{
|
|
auto region_or_error = MM.allocate_dma_buffer_page("IDE PRDT", Memory::Region::Access::ReadWrite, m_prdt_page);
|
|
if (region_or_error.is_error())
|
|
TODO();
|
|
m_prdt_region = region_or_error.release_value();
|
|
}
|
|
{
|
|
auto region_or_error = MM.allocate_dma_buffer_page("IDE DMA region", Memory::Region::Access::ReadWrite, m_dma_buffer_page);
|
|
if (region_or_error.is_error())
|
|
TODO();
|
|
m_dma_buffer_region = region_or_error.release_value();
|
|
}
|
|
|
|
prdt().end_of_table = 0x8000;
|
|
|
|
// clear bus master interrupt status
|
|
m_io_group.bus_master_base().value().offset(2).out<u8>(m_io_group.bus_master_base().value().offset(2).in<u8>() | 4);
|
|
}
|
|
|
|
static void print_ide_status(u8 status)
|
|
{
|
|
dbgln("BMIDEChannel: print_ide_status: DRQ={} BSY={}, DRDY={}, DSC={}, DF={}, CORR={}, IDX={}, ERR={}",
|
|
(status & ATA_SR_DRQ) != 0,
|
|
(status & ATA_SR_BSY) != 0,
|
|
(status & ATA_SR_DRDY) != 0,
|
|
(status & ATA_SR_DSC) != 0,
|
|
(status & ATA_SR_DF) != 0,
|
|
(status & ATA_SR_CORR) != 0,
|
|
(status & ATA_SR_IDX) != 0,
|
|
(status & ATA_SR_ERR) != 0);
|
|
}
|
|
|
|
bool BMIDEChannel::handle_irq(RegisterState const&)
|
|
{
|
|
u8 status = m_io_group.io_base().offset(ATA_REG_STATUS).in<u8>();
|
|
|
|
m_entropy_source.add_random_event(status);
|
|
|
|
VERIFY(m_io_group.bus_master_base().has_value());
|
|
u8 bstatus = m_io_group.bus_master_base().value().offset(2).in<u8>();
|
|
if (!(bstatus & 0x4)) {
|
|
// interrupt not from this device, ignore
|
|
dbgln_if(PATA_DEBUG, "BMIDEChannel: ignore interrupt");
|
|
return false;
|
|
}
|
|
// clear bus master interrupt status
|
|
m_io_group.bus_master_base().value().offset(2).out<u8>(m_io_group.bus_master_base().value().offset(2).in<u8>() | 4);
|
|
|
|
SpinlockLocker lock(m_request_lock);
|
|
dbgln_if(PATA_DEBUG, "BMIDEChannel: interrupt: DRQ={}, BSY={}, DRDY={}",
|
|
(status & ATA_SR_DRQ) != 0,
|
|
(status & ATA_SR_BSY) != 0,
|
|
(status & ATA_SR_DRDY) != 0);
|
|
|
|
if (!m_current_request) {
|
|
dbgln("BMIDEChannel: IRQ but no pending request!");
|
|
return false;
|
|
}
|
|
|
|
if (status & ATA_SR_ERR) {
|
|
print_ide_status(status);
|
|
m_device_error = m_io_group.io_base().offset(ATA_REG_ERROR).in<u8>();
|
|
dbgln("BMIDEChannel: Error {:#02x}!", (u8)m_device_error);
|
|
try_disambiguate_error();
|
|
complete_current_request(AsyncDeviceRequest::Failure);
|
|
return true;
|
|
}
|
|
m_device_error = 0;
|
|
complete_current_request(AsyncDeviceRequest::Success);
|
|
return true;
|
|
}
|
|
|
|
void BMIDEChannel::complete_current_request(AsyncDeviceRequest::RequestResult result)
|
|
{
|
|
// NOTE: this may be called from the interrupt handler!
|
|
VERIFY(m_current_request);
|
|
VERIFY(m_request_lock.is_locked());
|
|
|
|
// Now schedule reading back the buffer as soon as we leave the irq handler.
|
|
// This is important so that we can safely write the buffer back,
|
|
// which could cause page faults. Note that this may be called immediately
|
|
// before Processor::deferred_call_queue returns!
|
|
auto work_item_creation_result = g_io_work->try_queue([this, result]() {
|
|
dbgln_if(PATA_DEBUG, "BMIDEChannel::complete_current_request result: {}", (int)result);
|
|
SpinlockLocker lock(m_request_lock);
|
|
VERIFY(m_current_request);
|
|
auto current_request = m_current_request;
|
|
m_current_request.clear();
|
|
|
|
if (result == AsyncDeviceRequest::Success) {
|
|
if (current_request->request_type() == AsyncBlockDeviceRequest::Read) {
|
|
if (auto result = current_request->write_to_buffer(current_request->buffer(), m_dma_buffer_region->vaddr().as_ptr(), current_request->buffer_size()); result.is_error()) {
|
|
lock.unlock();
|
|
current_request->complete(AsyncDeviceRequest::MemoryFault);
|
|
return;
|
|
}
|
|
}
|
|
|
|
// I read somewhere that this may trigger a cache flush so let's do it.
|
|
VERIFY(m_io_group.bus_master_base().has_value());
|
|
m_io_group.bus_master_base().value().offset(2).out<u8>(m_io_group.bus_master_base().value().offset(2).in<u8>() | 0x6);
|
|
}
|
|
|
|
lock.unlock();
|
|
current_request->complete(result);
|
|
});
|
|
if (work_item_creation_result.is_error()) {
|
|
auto current_request = m_current_request;
|
|
m_current_request.clear();
|
|
current_request->complete(AsyncDeviceRequest::OutOfMemory);
|
|
}
|
|
}
|
|
|
|
void BMIDEChannel::ata_write_sectors(bool slave_request, u16 capabilities)
|
|
{
|
|
VERIFY(m_lock.is_locked());
|
|
VERIFY(!m_current_request.is_null());
|
|
VERIFY(m_current_request->block_count() <= 256);
|
|
|
|
SpinlockLocker locker(m_request_lock);
|
|
dbgln_if(PATA_DEBUG, "BMIDEChannel::ata_write_sectors ({} x {})", m_current_request->block_index(), m_current_request->block_count());
|
|
|
|
prdt().offset = m_dma_buffer_page->paddr().get();
|
|
prdt().size = m_current_request->buffer_size();
|
|
|
|
if (auto result = m_current_request->read_from_buffer(m_current_request->buffer(), m_dma_buffer_region->vaddr().as_ptr(), m_current_request->buffer_size()); result.is_error()) {
|
|
complete_current_request(AsyncDeviceRequest::MemoryFault);
|
|
return;
|
|
}
|
|
|
|
// Note: This is a fix for a quirk for an IDE controller on ICH7 machine.
|
|
// We need to select the drive and then we wait 10 microseconds... and it doesn't hurt anything
|
|
m_io_group.io_base().offset(ATA_REG_HDDEVSEL).out<u8>(0xA0 | ((slave_request ? 1 : 0) << 4));
|
|
IO::delay(10);
|
|
|
|
VERIFY(prdt().size <= PAGE_SIZE);
|
|
VERIFY(m_io_group.bus_master_base().has_value());
|
|
// Stop bus master
|
|
m_io_group.bus_master_base().value().out<u8>(0);
|
|
|
|
// Write the PRDT location
|
|
m_io_group.bus_master_base().value().offset(4).out<u32>(m_prdt_page->paddr().get());
|
|
|
|
// Turn on "Interrupt" and "Error" flag. The error flag should be cleared by hardware.
|
|
m_io_group.bus_master_base().value().offset(2).out<u8>(m_io_group.bus_master_base().value().offset(2).in<u8>() | 0x6);
|
|
|
|
ata_access(Direction::Write, slave_request, m_current_request->block_index(), m_current_request->block_count(), capabilities);
|
|
|
|
// Start bus master
|
|
m_io_group.bus_master_base().value().out<u8>(0x1);
|
|
}
|
|
|
|
void BMIDEChannel::send_ata_io_command(LBAMode lba_mode, Direction direction) const
|
|
{
|
|
if (lba_mode != LBAMode::FortyEightBit) {
|
|
m_io_group.io_base().offset(ATA_REG_COMMAND).out<u8>(direction == Direction::Read ? ATA_CMD_READ_DMA : ATA_CMD_WRITE_DMA);
|
|
} else {
|
|
m_io_group.io_base().offset(ATA_REG_COMMAND).out<u8>(direction == Direction::Read ? ATA_CMD_READ_DMA_EXT : ATA_CMD_WRITE_DMA_EXT);
|
|
}
|
|
}
|
|
|
|
void BMIDEChannel::ata_read_sectors(bool slave_request, u16 capabilities)
|
|
{
|
|
VERIFY(m_lock.is_locked());
|
|
VERIFY(!m_current_request.is_null());
|
|
VERIFY(m_current_request->block_count() <= 256);
|
|
|
|
SpinlockLocker locker(m_request_lock);
|
|
dbgln_if(PATA_DEBUG, "BMIDEChannel::ata_read_sectors ({} x {})", m_current_request->block_index(), m_current_request->block_count());
|
|
|
|
// Note: This is a fix for a quirk for an IDE controller on ICH7 machine.
|
|
// We need to select the drive and then we wait 10 microseconds... and it doesn't hurt anything
|
|
m_io_group.io_base().offset(ATA_REG_HDDEVSEL).out<u8>(0xA0 | ((slave_request ? 1 : 0) << 4));
|
|
IO::delay(10);
|
|
|
|
prdt().offset = m_dma_buffer_page->paddr().get();
|
|
prdt().size = m_current_request->buffer_size();
|
|
|
|
VERIFY(prdt().size <= PAGE_SIZE);
|
|
|
|
VERIFY(m_io_group.bus_master_base().has_value());
|
|
// Stop bus master
|
|
m_io_group.bus_master_base().value().out<u8>(0);
|
|
|
|
// Write the PRDT location
|
|
m_io_group.bus_master_base().value().offset(4).out<u32>(m_prdt_page->paddr().get());
|
|
|
|
// Set transfer direction
|
|
m_io_group.bus_master_base().value().out<u8>(0x8);
|
|
|
|
// Turn on "Interrupt" and "Error" flag. The error flag should be cleared by hardware.
|
|
m_io_group.bus_master_base().value().offset(2).out<u8>(m_io_group.bus_master_base().value().offset(2).in<u8>() | 0x6);
|
|
|
|
ata_access(Direction::Read, slave_request, m_current_request->block_index(), m_current_request->block_count(), capabilities);
|
|
|
|
// Start bus master
|
|
m_io_group.bus_master_base().value().out<u8>(0x9);
|
|
}
|
|
|
|
}
|