Re: [PATCH v2 2/2] lockdep: Fix merging of hlocks with non-zero references

From: Peter Zijlstra
Date: Mon May 27 2019 - 12:09:53 EST


On Mon, May 27, 2019 at 05:14:38PM +0200, Peter Zijlstra wrote:
> On Fri, May 24, 2019 at 11:15:09PM +0300, Imre Deak wrote:
> > diff --git a/kernel/locking/lockdep.c b/kernel/locking/lockdep.c
> > index 967352d32af1..9e2a4ab6c731 100644
> > --- a/kernel/locking/lockdep.c
> > +++ b/kernel/locking/lockdep.c
> > @@ -3637,6 +3637,11 @@ print_lock_nested_lock_not_held(struct task_struct *curr,
> >
> > static int __lock_is_held(const struct lockdep_map *lock, int read);
> >
> > +static int hlock_reference(int reference)
> > +{
> > + return reference ? : 1;
> > +}
> > +
> > /*
> > * This gets called for every mutex_lock*()/spin_lock*() operation.
> > * We maintain the dependency maps and validate the locking attempt:
> > @@ -3702,17 +3707,15 @@ static int __lock_acquire(struct lockdep_map *lock, unsigned int subclass,
> > if (depth) {
> > hlock = curr->held_locks + depth - 1;
> > if (hlock->class_idx == class_idx && nest_lock) {
> > - if (hlock->references) {
> > - /*
> > - * Check: unsigned int references overflow.
> > - */
> > - if (DEBUG_LOCKS_WARN_ON(hlock->references == UINT_MAX))
>
> What tree is this against? Afaict this is still 12 bits ?!
>
> > - return 0;
> > + /*
> > + * Check: unsigned int references overflow.
> > + */
> > + if (DEBUG_LOCKS_WARN_ON(hlock_reference(hlock->references) >
> > + UINT_MAX - hlock_reference(references)))
>
> Idem. Also very weird overflow check..
>
> > + return 0;
> >
> > - hlock->references++;
> > - } else {
> > - hlock->references = 2;
> > - }
> > + hlock->references = hlock_reference(hlock->references) +
> > + hlock_reference(references);
> >
> > return 2;
> > }

I think you wanted something like this: ?

--- a/kernel/locking/lockdep.c
+++ b/kernel/locking/lockdep.c
@@ -3731,6 +3731,11 @@ print_lock_nested_lock_not_held(struct t

static int __lock_is_held(const struct lockdep_map *lock, int read);

+static inline int hlock_references(struct held_lock *hlock)
+{
+ return hlock->references ? : 1;
+}
+
/*
* This gets called for every mutex_lock*()/spin_lock*() operation.
* We maintain the dependency maps and validate the locking attempt:
@@ -3796,17 +3801,14 @@ static int __lock_acquire(struct lockdep
if (depth) {
hlock = curr->held_locks + depth - 1;
if (hlock->class_idx == class_idx && nest_lock) {
- if (hlock->references) {
- /*
- * Check: unsigned int references:12, overflow.
- */
- if (DEBUG_LOCKS_WARN_ON(hlock->references == (1 << 12)-1))
- return 0;
-
- hlock->references++;
- } else {
- hlock->references = 2;
- }
+ if (!references)
+ references++;
+
+ hlock->references = hlock_references(hlock) + references;
+
+ /* Overflow */
+ if (DEBUG_LOCKS_WARN_ON(hlock->references < references))
+ return 0;

return 2;
}