Re: [PATCH V8 10/10] csky: Add qspinlock support

From: Waiman Long
Date: Sun Jul 24 2022 - 22:09:04 EST


On 7/24/22 08:25, guoren@xxxxxxxxxx wrote:
From: Guo Ren <guoren@xxxxxxxxxxxxxxxxx>

Enable qspinlock by the requirements mentioned in a8ad07e5240c9
("asm-generic: qspinlock: Indicate the use of mixed-size atomics").

C-SKY only has "ldex/stex" for all atomic operations. So csky give a
strong forward guarantee for "ldex/stex." That means when ldex grabbed
the cache line into $L1, it would block other cores from snooping the
address with several cycles.

Signed-off-by: Guo Ren <guoren@xxxxxxxxxxxxxxxxx>
Signed-off-by: Guo Ren <guoren@xxxxxxxxxx>
---
arch/csky/Kconfig | 16 ++++++++++++++++
arch/csky/include/asm/Kbuild | 2 ++
arch/csky/include/asm/cmpxchg.h | 20 ++++++++++++++++++++
3 files changed, 38 insertions(+)

diff --git a/arch/csky/Kconfig b/arch/csky/Kconfig
index dfdb436b6078..09f7d1f06bca 100644
--- a/arch/csky/Kconfig
+++ b/arch/csky/Kconfig
@@ -354,6 +354,22 @@ config HAVE_EFFICIENT_UNALIGNED_STRING_OPS
Say Y here to enable EFFICIENT_UNALIGNED_STRING_OPS. Some CPU models could
deal with unaligned access by hardware.
+choice
+ prompt "C-SKY spinlock type"
+ default CSKY_TICKET_SPINLOCKS
+
+config CSKY_TICKET_SPINLOCKS
+ bool "Using ticket spinlock"
+
+config CSKY_QUEUED_SPINLOCKS
+ bool "Using queued spinlock"
+ depends on SMP
+ select ARCH_USE_QUEUED_SPINLOCKS
+ help
+ Make sure your micro arch LL/SC has a strong forward progress guarantee.
+ Otherwise, stay at ticket-lock/combo-lock.

"combo-lock"? It is a cut-and-paste error. Right?

Cheers,
Longman