1 #define JEMALLOC_MUTEX_C_ 2 #include "jemalloc/internal/jemalloc_preamble.h" 3 #include "jemalloc/internal/jemalloc_internal_includes.h" 4 5 #include "jemalloc/internal/assert.h" 6 #include "jemalloc/internal/malloc_io.h" 7 #include "jemalloc/internal/spin.h" 8 9 #ifndef _CRT_SPINCOUNT 10 #define _CRT_SPINCOUNT 4000 11 #endif 12 13 /******************************************************************************/ 14 /* Data. */ 15 16 #ifdef JEMALLOC_LAZY_LOCK 17 bool isthreaded = false; 18 #endif 19 #ifdef JEMALLOC_MUTEX_INIT_CB 20 static bool postpone_init = true; 21 static malloc_mutex_t *postponed_mutexes = NULL; 22 #endif 23 24 /******************************************************************************/ 25 /* 26 * We intercept pthread_create() calls in order to toggle isthreaded if the 27 * process goes multi-threaded. 28 */ 29 30 #if defined(JEMALLOC_LAZY_LOCK) && !defined(_WIN32) 31 JEMALLOC_EXPORT int 32 pthread_create(pthread_t *__restrict thread, 33 const pthread_attr_t *__restrict attr, void *(*start_routine)(void *), 34 void *__restrict arg) { 35 return pthread_create_wrapper(thread, attr, start_routine, arg); 36 } 37 #endif 38 39 /******************************************************************************/ 40 41 #ifdef JEMALLOC_MUTEX_INIT_CB 42 JEMALLOC_EXPORT int _pthread_mutex_init_calloc_cb(pthread_mutex_t *mutex, 43 void *(calloc_cb)(size_t, size_t)); 44 #endif 45 46 void 47 malloc_mutex_lock_slow(malloc_mutex_t *mutex) { 48 mutex_prof_data_t *data = &mutex->prof_data; 49 UNUSED nstime_t before = NSTIME_ZERO_INITIALIZER; 50 51 if (ncpus == 1) { 52 goto label_spin_done; 53 } 54 55 int cnt = 0, max_cnt = MALLOC_MUTEX_MAX_SPIN; 56 do { 57 spin_cpu_spinwait(); 58 if (!malloc_mutex_trylock_final(mutex)) { 59 data->n_spin_acquired++; 60 return; 61 } 62 } while (cnt++ < max_cnt); 63 64 if (!config_stats) { 65 /* Only spin is useful when stats is off. */ 66 malloc_mutex_lock_final(mutex); 67 return; 68 } 69 label_spin_done: 70 nstime_update(&before); 71 /* Copy before to after to avoid clock skews. */ 72 nstime_t after; 73 nstime_copy(&after, &before); 74 uint32_t n_thds = atomic_fetch_add_u32(&data->n_waiting_thds, 1, 75 ATOMIC_RELAXED) + 1; 76 /* One last try as above two calls may take quite some cycles. */ 77 if (!malloc_mutex_trylock_final(mutex)) { 78 atomic_fetch_sub_u32(&data->n_waiting_thds, 1, ATOMIC_RELAXED); 79 data->n_spin_acquired++; 80 return; 81 } 82 83 /* True slow path. */ 84 malloc_mutex_lock_final(mutex); 85 /* Update more slow-path only counters. */ 86 atomic_fetch_sub_u32(&data->n_waiting_thds, 1, ATOMIC_RELAXED); 87 nstime_update(&after); 88 89 nstime_t delta; 90 nstime_copy(&delta, &after); 91 nstime_subtract(&delta, &before); 92 93 data->n_wait_times++; 94 nstime_add(&data->tot_wait_time, &delta); 95 if (nstime_compare(&data->max_wait_time, &delta) < 0) { 96 nstime_copy(&data->max_wait_time, &delta); 97 } 98 if (n_thds > data->max_n_thds) { 99 data->max_n_thds = n_thds; 100 } 101 } 102 103 static void 104 mutex_prof_data_init(mutex_prof_data_t *data) { 105 memset(data, 0, sizeof(mutex_prof_data_t)); 106 nstime_init(&data->max_wait_time, 0); 107 nstime_init(&data->tot_wait_time, 0); 108 data->prev_owner = NULL; 109 } 110 111 void 112 malloc_mutex_prof_data_reset(tsdn_t *tsdn, malloc_mutex_t *mutex) { 113 malloc_mutex_assert_owner(tsdn, mutex); 114 mutex_prof_data_init(&mutex->prof_data); 115 } 116 117 static int 118 mutex_addr_comp(const witness_t *witness1, void *mutex1, 119 const witness_t *witness2, void *mutex2) { 120 assert(mutex1 != NULL); 121 assert(mutex2 != NULL); 122 uintptr_t mu1int = (uintptr_t)mutex1; 123 uintptr_t mu2int = (uintptr_t)mutex2; 124 if (mu1int < mu2int) { 125 return -1; 126 } else if (mu1int == mu2int) { 127 return 0; 128 } else { 129 return 1; 130 } 131 } 132 133 bool 134 malloc_mutex_init(malloc_mutex_t *mutex, const char *name, 135 witness_rank_t rank, malloc_mutex_lock_order_t lock_order) { 136 mutex_prof_data_init(&mutex->prof_data); 137 #ifdef _WIN32 138 # if _WIN32_WINNT >= 0x0600 139 InitializeSRWLock(&mutex->lock); 140 # else 141 if (!InitializeCriticalSectionAndSpinCount(&mutex->lock, 142 _CRT_SPINCOUNT)) { 143 return true; 144 } 145 # endif 146 #elif (defined(JEMALLOC_OS_UNFAIR_LOCK)) 147 mutex->lock = OS_UNFAIR_LOCK_INIT; 148 #elif (defined(JEMALLOC_OSSPIN)) 149 mutex->lock = 0; 150 #elif (defined(JEMALLOC_MUTEX_INIT_CB)) 151 if (postpone_init) { 152 mutex->postponed_next = postponed_mutexes; 153 postponed_mutexes = mutex; 154 } else { 155 if (_pthread_mutex_init_calloc_cb(&mutex->lock, 156 bootstrap_calloc) != 0) { 157 return true; 158 } 159 } 160 #else 161 pthread_mutexattr_t attr; 162 163 if (pthread_mutexattr_init(&attr) != 0) { 164 return true; 165 } 166 pthread_mutexattr_settype(&attr, MALLOC_MUTEX_TYPE); 167 if (pthread_mutex_init(&mutex->lock, &attr) != 0) { 168 pthread_mutexattr_destroy(&attr); 169 return true; 170 } 171 pthread_mutexattr_destroy(&attr); 172 #endif 173 if (config_debug) { 174 mutex->lock_order = lock_order; 175 if (lock_order == malloc_mutex_address_ordered) { 176 witness_init(&mutex->witness, name, rank, 177 mutex_addr_comp, mutex); 178 } else { 179 witness_init(&mutex->witness, name, rank, NULL, NULL); 180 } 181 } 182 return false; 183 } 184 185 void 186 malloc_mutex_prefork(tsdn_t *tsdn, malloc_mutex_t *mutex) { 187 malloc_mutex_lock(tsdn, mutex); 188 } 189 190 void 191 malloc_mutex_postfork_parent(tsdn_t *tsdn, malloc_mutex_t *mutex) { 192 malloc_mutex_unlock(tsdn, mutex); 193 } 194 195 void 196 malloc_mutex_postfork_child(tsdn_t *tsdn, malloc_mutex_t *mutex) { 197 #ifdef JEMALLOC_MUTEX_INIT_CB 198 malloc_mutex_unlock(tsdn, mutex); 199 #else 200 if (malloc_mutex_init(mutex, mutex->witness.name, 201 mutex->witness.rank, mutex->lock_order)) { 202 malloc_printf("<jemalloc>: Error re-initializing mutex in " 203 "child\n"); 204 if (opt_abort) { 205 abort(); 206 } 207 } 208 #endif 209 } 210 211 bool 212 malloc_mutex_boot(void) { 213 #ifdef JEMALLOC_MUTEX_INIT_CB 214 postpone_init = false; 215 while (postponed_mutexes != NULL) { 216 if (_pthread_mutex_init_calloc_cb(&postponed_mutexes->lock, 217 bootstrap_calloc) != 0) { 218 return true; 219 } 220 postponed_mutexes = postponed_mutexes->postponed_next; 221 } 222 #endif 223 return false; 224 } 225