1
Fork 0
mirror of https://github.com/RGBCube/serenity synced 2025-07-08 22:37:35 +00:00

Kernel: Introduce shared locking mode

A Lock can now be held either in shared or exclusive mode. Multiple threads can
hold the same lock in shared mode at one time, but if any thread holds the lock
in exclusive mode, no other thread can hold it at the same time in either mode.
This commit is contained in:
Sergey Bugaev 2020-04-18 12:21:00 +03:00 committed by Andreas Kling
parent 2f15449531
commit 05ba4295e9
2 changed files with 62 additions and 20 deletions

View file

@ -30,8 +30,20 @@
namespace Kernel { namespace Kernel {
void Lock::lock() static bool modes_conflict(Lock::Mode mode1, Lock::Mode mode2)
{ {
if (mode1 == Lock::Mode::Unlocked || mode2 == Lock::Mode::Unlocked)
return false;
if (mode1 == Lock::Mode::Shared && mode2 == Lock::Mode::Shared)
return false;
return true;
}
void Lock::lock(Mode mode)
{
ASSERT(mode != Mode::Unlocked);
ASSERT(!Scheduler::is_active()); ASSERT(!Scheduler::is_active());
if (!are_interrupts_enabled()) { if (!are_interrupts_enabled()) {
klog() << "Interrupts disabled when trying to take Lock{" << m_name << "}"; klog() << "Interrupts disabled when trying to take Lock{" << m_name << "}";
@ -41,9 +53,15 @@ void Lock::lock()
for (;;) { for (;;) {
bool expected = false; bool expected = false;
if (m_lock.compare_exchange_strong(expected, true, AK::memory_order_acq_rel)) { if (m_lock.compare_exchange_strong(expected, true, AK::memory_order_acq_rel)) {
if (!m_holder || m_holder == Thread::current) { // FIXME: Do not add new readers if writers are queued.
bool modes_dont_conflict = !modes_conflict(m_mode, mode);
bool already_hold_exclusive_lock = m_mode == Mode::Exclusive && m_holder == Thread::current;
if (modes_dont_conflict || already_hold_exclusive_lock) {
// We got the lock!
if (!already_hold_exclusive_lock)
m_mode = mode;
m_holder = Thread::current; m_holder = Thread::current;
++m_level; m_times_locked++;
m_lock.store(false, AK::memory_order_release); m_lock.store(false, AK::memory_order_release);
return; return;
} }
@ -57,14 +75,20 @@ void Lock::unlock()
for (;;) { for (;;) {
bool expected = false; bool expected = false;
if (m_lock.compare_exchange_strong(expected, true, AK::memory_order_acq_rel)) { if (m_lock.compare_exchange_strong(expected, true, AK::memory_order_acq_rel)) {
ASSERT(m_holder == Thread::current); ASSERT(m_times_locked);
ASSERT(m_level); --m_times_locked;
--m_level;
if (m_level) { ASSERT(m_mode != Mode::Unlocked);
if (m_mode == Mode::Exclusive)
ASSERT(m_holder == Thread::current);
if (m_holder == Thread::current && (m_mode == Mode::Shared || m_times_locked == 0))
m_holder = nullptr;
if (m_times_locked > 0) {
m_lock.store(false, AK::memory_order_release); m_lock.store(false, AK::memory_order_release);
return; return;
} }
m_holder = nullptr; m_mode = Mode::Unlocked;
m_queue.wake_one(&m_lock); m_queue.wake_one(&m_lock);
return; return;
} }
@ -75,19 +99,21 @@ void Lock::unlock()
bool Lock::force_unlock_if_locked() bool Lock::force_unlock_if_locked()
{ {
ASSERT(m_mode != Mode::Shared);
InterruptDisabler disabler; InterruptDisabler disabler;
if (m_holder != Thread::current) if (m_holder != Thread::current)
return false; return false;
ASSERT(m_level == 1); ASSERT(m_times_locked == 1);
ASSERT(m_holder == Thread::current);
m_holder = nullptr; m_holder = nullptr;
--m_level; m_mode = Mode::Unlocked;
m_times_locked = 0;
m_queue.wake_one(); m_queue.wake_one();
return true; return true;
} }
void Lock::clear_waiters() void Lock::clear_waiters()
{ {
ASSERT(m_mode != Mode::Shared);
InterruptDisabler disabler; InterruptDisabler disabler;
m_queue.clear(); m_queue.clear();
} }

View file

@ -41,9 +41,15 @@ public:
: m_name(name) : m_name(name)
{ {
} }
~Lock() {} ~Lock() { }
void lock(); enum class Mode {
Unlocked,
Shared,
Exclusive
};
void lock(Mode = Mode::Exclusive);
void unlock(); void unlock();
bool force_unlock_if_locked(); bool force_unlock_if_locked();
bool is_locked() const { return m_holder; } bool is_locked() const { return m_holder; }
@ -53,33 +59,43 @@ public:
private: private:
Atomic<bool> m_lock { false }; Atomic<bool> m_lock { false };
u32 m_level { 0 };
Thread* m_holder { nullptr };
const char* m_name { nullptr }; const char* m_name { nullptr };
WaitQueue m_queue; WaitQueue m_queue;
Mode m_mode { Mode::Unlocked };
// When locked exclusively, only the thread already holding the lock can
// lock it again. When locked in shared mode, any thread can do that.
u32 m_times_locked { 0 };
// One of the threads that hold this lock, or nullptr. When locked in shared
// mode, this is stored on best effort basis: nullptr value does *not* mean
// the lock is unlocked, it just means we don't know which threads hold it.
// When locked exclusively, this is always the one thread that holds the
// lock.
Thread* m_holder { nullptr };
}; };
class Locker { class Locker {
public: public:
[[gnu::always_inline]] inline explicit Locker(Lock& l) [[gnu::always_inline]] inline explicit Locker(Lock& l, Lock::Mode mode = Lock::Mode::Exclusive)
: m_lock(l) : m_lock(l)
{ {
lock(); m_lock.lock(mode);
} }
[[gnu::always_inline]] inline ~Locker() { unlock(); } [[gnu::always_inline]] inline ~Locker() { unlock(); }
[[gnu::always_inline]] inline void unlock() { m_lock.unlock(); } [[gnu::always_inline]] inline void unlock() { m_lock.unlock(); }
[[gnu::always_inline]] inline void lock() { m_lock.lock(); } [[gnu::always_inline]] inline void lock(Lock::Mode mode = Lock::Mode::Exclusive) { m_lock.lock(mode); }
private: private:
Lock& m_lock; Lock& m_lock;
}; };
#define LOCKER(lock) Locker locker(lock) #define LOCKER(...) Locker locker(__VA_ARGS__)
template<typename T> template<typename T>
class Lockable { class Lockable {
public: public:
Lockable() {} Lockable() { }
Lockable(T&& resource) Lockable(T&& resource)
: m_resource(move(resource)) : m_resource(move(resource))
{ {