/* * Copyright (c) 2021, the SerenityOS developers. * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions are met: * * 1. Redistributions of source code must retain the above copyright notice, this * list of conditions and the following disclaimer. * * 2. Redistributions in binary form must reproduce the above copyright notice, * this list of conditions and the following disclaimer in the documentation * and/or other materials provided with the distribution. * * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE * DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER * CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ #include #include namespace Kernel { VirtIOQueue::VirtIOQueue(u16 queue_size, u16 notify_offset) : m_queue_size(queue_size) , m_notify_offset(notify_offset) , m_free_buffers(queue_size) { size_t size_of_descriptors = sizeof(VirtIOQueueDescriptor) * queue_size; size_t size_of_driver = sizeof(VirtIOQueueDriver) + queue_size * sizeof(u16); size_t size_of_device = sizeof(VirtIOQueueDevice) + queue_size * sizeof(VirtIOQueueDeviceItem); m_queue_region = MM.allocate_contiguous_kernel_region(page_round_up(size_of_descriptors + size_of_driver + size_of_device), "VirtIO Queue", Region::Access::Read | Region::Access::Write); VERIFY(m_queue_region); // TODO: ensure alignment!!! u8* ptr = m_queue_region->vaddr().as_ptr(); memset(ptr, 0, m_queue_region->size()); m_descriptors = reinterpret_cast(ptr); m_driver = reinterpret_cast(ptr + size_of_descriptors); m_device = reinterpret_cast(ptr + size_of_descriptors + size_of_driver); m_tokens.resize(queue_size); for (auto i = 0; i < queue_size; i++) { m_descriptors[i].next = (i + 1) % queue_size; // link all of the descriptors in a circle } enable_interrupts(); } VirtIOQueue::~VirtIOQueue() { } void VirtIOQueue::enable_interrupts() { m_driver->flags = 0; } void VirtIOQueue::disable_interrupts() { m_driver->flags = 1; } bool VirtIOQueue::supply_buffer(Badge, const ScatterGatherList& scatter_list, BufferType buffer_type, void* token) { VERIFY(scatter_list.length() && scatter_list.length() <= m_free_buffers); m_free_buffers -= scatter_list.length(); auto descriptor_index = m_free_head; auto last_index = descriptor_index; scatter_list.for_each_entry([&](auto paddr, auto size) { m_descriptors[descriptor_index].flags = static_cast(buffer_type) | VIRTQ_DESC_F_NEXT; m_descriptors[descriptor_index].address = static_cast(paddr); m_descriptors[descriptor_index].length = static_cast(size); last_index = descriptor_index; descriptor_index = m_descriptors[descriptor_index].next; // ensure we place the buffer in chain order }); m_descriptors[last_index].flags &= ~(VIRTQ_DESC_F_NEXT); // last descriptor in chain doesnt have a next descriptor m_driver->rings[m_driver_index_shadow % m_queue_size] = m_free_head; // m_driver_index_shadow is used to prevent accesses to index before the rings are updated m_tokens[m_free_head] = token; m_free_head = descriptor_index; full_memory_barrier(); m_driver_index_shadow++; m_driver->index++; full_memory_barrier(); auto device_flags = m_device->flags; return !(device_flags & 1); // if bit 1 is enabled the device disabled interrupts } bool VirtIOQueue::new_data_available() const { return m_device->index != m_used_tail; } void* VirtIOQueue::get_buffer(size_t* size) { if (!new_data_available()) { *size = 0; return nullptr; } full_memory_barrier(); auto descriptor_index = m_device->rings[m_used_tail % m_queue_size].index; *size = m_device->rings[m_used_tail % m_queue_size].length; m_used_tail++; auto token = m_tokens[descriptor_index]; pop_buffer(descriptor_index); return token; } void VirtIOQueue::discard_used_buffers() { size_t size; while (!get_buffer(&size)) { } } void VirtIOQueue::pop_buffer(u16 descriptor_index) { m_tokens[descriptor_index] = nullptr; auto i = descriptor_index; while (m_descriptors[i].flags & VIRTQ_DESC_F_NEXT) { m_free_buffers++; i = m_descriptors[i].next; } m_free_buffers++; // the last descriptor in the chain doesnt have the NEXT flag m_descriptors[i].next = m_free_head; // empend the popped descriptors to the free chain m_free_head = descriptor_index; } }