123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164 |
- /*
- * Copyright 2011 Tilera Corporation. All Rights Reserved.
- *
- * This program is free software; you can redistribute it and/or
- * modify it under the terms of the GNU General Public License
- * as published by the Free Software Foundation, version 2.
- *
- * This program is distributed in the hope that it will be useful, but
- * WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE, GOOD TITLE or
- * NON INFRINGEMENT. See the GNU General Public License for
- * more details.
- *
- * 64-bit SMP ticket spinlocks, allowing only a single CPU anywhere
- * (the type definitions are in asm/spinlock_types.h)
- */
- #ifndef _ASM_TILE_SPINLOCK_64_H
- #define _ASM_TILE_SPINLOCK_64_H
- #include <linux/compiler.h>
- /* Shifts and masks for the various fields in "lock". */
- #define __ARCH_SPIN_CURRENT_SHIFT 17
- #define __ARCH_SPIN_NEXT_MASK 0x7fff
- #define __ARCH_SPIN_NEXT_OVERFLOW 0x8000
- /*
- * Return the "current" portion of a ticket lock value,
- * i.e. the number that currently owns the lock.
- */
- static inline u32 arch_spin_current(u32 val)
- {
- return val >> __ARCH_SPIN_CURRENT_SHIFT;
- }
- /*
- * Return the "next" portion of a ticket lock value,
- * i.e. the number that the next task to try to acquire the lock will get.
- */
- static inline u32 arch_spin_next(u32 val)
- {
- return val & __ARCH_SPIN_NEXT_MASK;
- }
- /* The lock is locked if a task would have to wait to get it. */
- static inline int arch_spin_is_locked(arch_spinlock_t *lock)
- {
- /* Use READ_ONCE() to ensure that calling this in a loop is OK. */
- u32 val = READ_ONCE(lock->lock);
- return arch_spin_current(val) != arch_spin_next(val);
- }
- /* Bump the current ticket so the next task owns the lock. */
- static inline void arch_spin_unlock(arch_spinlock_t *lock)
- {
- wmb(); /* guarantee anything modified under the lock is visible */
- __insn_fetchadd4(&lock->lock, 1U << __ARCH_SPIN_CURRENT_SHIFT);
- }
- void arch_spin_unlock_wait(arch_spinlock_t *lock);
- void arch_spin_lock_slow(arch_spinlock_t *lock, u32 val);
- /* Grab the "next" ticket number and bump it atomically.
- * If the current ticket is not ours, go to the slow path.
- * We also take the slow path if the "next" value overflows.
- */
- static inline void arch_spin_lock(arch_spinlock_t *lock)
- {
- u32 val = __insn_fetchadd4(&lock->lock, 1);
- u32 ticket = val & (__ARCH_SPIN_NEXT_MASK | __ARCH_SPIN_NEXT_OVERFLOW);
- if (unlikely(arch_spin_current(val) != ticket))
- arch_spin_lock_slow(lock, ticket);
- }
- /* Try to get the lock, and return whether we succeeded. */
- int arch_spin_trylock(arch_spinlock_t *lock);
- /* We cannot take an interrupt after getting a ticket, so don't enable them. */
- #define arch_spin_lock_flags(lock, flags) arch_spin_lock(lock)
- /*
- * Read-write spinlocks, allowing multiple readers
- * but only one writer.
- *
- * We use fetchadd() for readers, and fetchor() with the sign bit
- * for writers.
- */
- #define __WRITE_LOCK_BIT (1 << 31)
- static inline int arch_write_val_locked(int val)
- {
- return val < 0; /* Optimize "val & __WRITE_LOCK_BIT". */
- }
- /**
- * read_can_lock - would read_trylock() succeed?
- * @lock: the rwlock in question.
- */
- static inline int arch_read_can_lock(arch_rwlock_t *rw)
- {
- return !arch_write_val_locked(rw->lock);
- }
- /**
- * write_can_lock - would write_trylock() succeed?
- * @lock: the rwlock in question.
- */
- static inline int arch_write_can_lock(arch_rwlock_t *rw)
- {
- return rw->lock == 0;
- }
- extern void __read_lock_failed(arch_rwlock_t *rw);
- static inline void arch_read_lock(arch_rwlock_t *rw)
- {
- u32 val = __insn_fetchaddgez4(&rw->lock, 1);
- if (unlikely(arch_write_val_locked(val)))
- __read_lock_failed(rw);
- }
- extern void __write_lock_failed(arch_rwlock_t *rw, u32 val);
- static inline void arch_write_lock(arch_rwlock_t *rw)
- {
- u32 val = __insn_fetchor4(&rw->lock, __WRITE_LOCK_BIT);
- if (unlikely(val != 0))
- __write_lock_failed(rw, val);
- }
- static inline void arch_read_unlock(arch_rwlock_t *rw)
- {
- __insn_mf();
- __insn_fetchadd4(&rw->lock, -1);
- }
- static inline void arch_write_unlock(arch_rwlock_t *rw)
- {
- __insn_mf();
- __insn_exch4(&rw->lock, 0); /* Avoid waiting in the write buffer. */
- }
- static inline int arch_read_trylock(arch_rwlock_t *rw)
- {
- return !arch_write_val_locked(__insn_fetchaddgez4(&rw->lock, 1));
- }
- static inline int arch_write_trylock(arch_rwlock_t *rw)
- {
- u32 val = __insn_fetchor4(&rw->lock, __WRITE_LOCK_BIT);
- if (likely(val == 0))
- return 1;
- if (!arch_write_val_locked(val))
- __insn_fetchand4(&rw->lock, ~__WRITE_LOCK_BIT);
- return 0;
- }
- #define arch_read_lock_flags(lock, flags) arch_read_lock(lock)
- #define arch_write_lock_flags(lock, flags) arch_write_lock(lock)
- #endif /* _ASM_TILE_SPINLOCK_64_H */
|