aboutsummaryrefslogtreecommitdiffstats
path: root/contrib/libs/clang18-rt/lib/scudo/standalone/mutex.h
blob: 843fb1ec021c22fa7967588c76fa5d0e75659c5f (plain) (blame)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
//===-- mutex.h -------------------------------------------------*- C++ -*-===//
//
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
// See https://llvm.org/LICENSE.txt for license information.
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
//
//===----------------------------------------------------------------------===//

#ifndef SCUDO_MUTEX_H_
#define SCUDO_MUTEX_H_

#include "atomic_helpers.h"
#include "common.h"
#include "thread_annotations.h"

#include <string.h>

#if SCUDO_FUCHSIA
#error #include <lib/sync/mutex.h> // for sync_mutex_t
#endif

namespace scudo {

class CAPABILITY("mutex") HybridMutex {
public:
  bool tryLock() TRY_ACQUIRE(true);
  NOINLINE void lock() ACQUIRE() {
    if (LIKELY(tryLock()))
      return;
      // The compiler may try to fully unroll the loop, ending up in a
      // NumberOfTries*NumberOfYields block of pauses mixed with tryLocks. This
      // is large, ugly and unneeded, a compact loop is better for our purpose
      // here. Use a pragma to tell the compiler not to unroll the loop.
#ifdef __clang__
#pragma nounroll
#endif
    for (u8 I = 0U; I < NumberOfTries; I++) {
      delayLoop();
      if (tryLock())
        return;
    }
    lockSlow();
  }
  void unlock() RELEASE();

  // TODO(chiahungduan): In general, we may want to assert the owner of lock as
  // well. Given the current uses of HybridMutex, it's acceptable without
  // asserting the owner. Re-evaluate this when we have certain scenarios which
  // requires a more fine-grained lock granularity.
  ALWAYS_INLINE void assertHeld() ASSERT_CAPABILITY(this) {
    if (SCUDO_DEBUG)
      assertHeldImpl();
  }

private:
  void delayLoop() {
    // The value comes from the average time spent in accessing caches (which
    // are the fastest operations) so that we are unlikely to wait too long for
    // fast operations.
    constexpr u32 SpinTimes = 16;
    volatile u32 V = 0;
    for (u32 I = 0; I < SpinTimes; ++I) {
      u32 Tmp = V + 1;
      V = Tmp;
    }
  }

  void assertHeldImpl();

  // TODO(chiahungduan): Adapt this value based on scenarios. E.g., primary and
  // secondary allocator have different allocation times.
  static constexpr u8 NumberOfTries = 32U;

#if SCUDO_LINUX
  atomic_u32 M = {};
#elif SCUDO_FUCHSIA
  sync_mutex_t M = {};
#endif

  void lockSlow() ACQUIRE();
};

class SCOPED_CAPABILITY ScopedLock {
public:
  explicit ScopedLock(HybridMutex &M) ACQUIRE(M) : Mutex(M) { Mutex.lock(); }
  ~ScopedLock() RELEASE() { Mutex.unlock(); }

private:
  HybridMutex &Mutex;

  ScopedLock(const ScopedLock &) = delete;
  void operator=(const ScopedLock &) = delete;
};

} // namespace scudo

#endif // SCUDO_MUTEX_H_