Mercurial
comparison third_party/libuv/src/unix/thread.c @ 160:948de3f54cea
[ThirdParty] Added libuv
| author | June Park <parkjune1995@gmail.com> |
|---|---|
| date | Wed, 14 Jan 2026 19:39:52 -0800 |
| parents | |
| children |
comparison
equal
deleted
inserted
replaced
| 159:05cf9467a1c3 | 160:948de3f54cea |
|---|---|
| 1 /* Copyright Joyent, Inc. and other Node contributors. All rights reserved. | |
| 2 * | |
| 3 * Permission is hereby granted, free of charge, to any person obtaining a copy | |
| 4 * of this software and associated documentation files (the "Software"), to | |
| 5 * deal in the Software without restriction, including without limitation the | |
| 6 * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or | |
| 7 * sell copies of the Software, and to permit persons to whom the Software is | |
| 8 * furnished to do so, subject to the following conditions: | |
| 9 * | |
| 10 * The above copyright notice and this permission notice shall be included in | |
| 11 * all copies or substantial portions of the Software. | |
| 12 * | |
| 13 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | |
| 14 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | |
| 15 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE | |
| 16 * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | |
| 17 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING | |
| 18 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS | |
| 19 * IN THE SOFTWARE. | |
| 20 */ | |
| 21 | |
| 22 #include "uv.h" | |
| 23 #include "internal.h" | |
| 24 | |
| 25 #include <pthread.h> | |
| 26 #ifdef __OpenBSD__ | |
| 27 #include <pthread_np.h> | |
| 28 #endif | |
| 29 #include <assert.h> | |
| 30 #include <errno.h> | |
| 31 | |
| 32 #include <sys/time.h> | |
| 33 #include <sys/resource.h> /* getrlimit() */ | |
| 34 #include <unistd.h> /* getpagesize() */ | |
| 35 | |
| 36 #include <limits.h> | |
| 37 | |
| 38 #ifdef __MVS__ | |
| 39 #include <sys/ipc.h> | |
| 40 #include <sys/sem.h> | |
| 41 #endif | |
| 42 | |
| 43 #if defined(__GLIBC__) && !defined(__UCLIBC__) | |
| 44 #include <gnu/libc-version.h> /* gnu_get_libc_version() */ | |
| 45 #endif | |
| 46 | |
| 47 #if defined(__linux__) | |
| 48 # include <sched.h> | |
| 49 # define uv__cpu_set_t cpu_set_t | |
| 50 #elif defined(__FreeBSD__) | |
| 51 # include <sys/param.h> | |
| 52 # include <sys/cpuset.h> | |
| 53 # include <pthread_np.h> | |
| 54 # define uv__cpu_set_t cpuset_t | |
| 55 #endif | |
| 56 | |
| 57 | |
| 58 #undef NANOSEC | |
| 59 #define NANOSEC ((uint64_t) 1e9) | |
| 60 | |
| 61 /* Musl's PTHREAD_STACK_MIN is 2 KB on all architectures, which is | |
| 62 * too small to safely receive signals on. | |
| 63 * | |
| 64 * Musl's PTHREAD_STACK_MIN + MINSIGSTKSZ == 8192 on arm64 (which has | |
| 65 * the largest MINSIGSTKSZ of the architectures that musl supports) so | |
| 66 * let's use that as a lower bound. | |
| 67 * | |
| 68 * We use a hardcoded value because PTHREAD_STACK_MIN + MINSIGSTKSZ | |
| 69 * is between 28 and 133 KB when compiling against glibc, depending | |
| 70 * on the architecture. | |
| 71 */ | |
| 72 static size_t uv__min_stack_size(void) { | |
| 73 static const size_t min = 8192; | |
| 74 | |
| 75 #ifdef PTHREAD_STACK_MIN /* Not defined on NetBSD. */ | |
| 76 if (min < (size_t) PTHREAD_STACK_MIN) | |
| 77 return PTHREAD_STACK_MIN; | |
| 78 #endif /* PTHREAD_STACK_MIN */ | |
| 79 | |
| 80 return min; | |
| 81 } | |
| 82 | |
| 83 | |
| 84 /* On Linux, threads created by musl have a much smaller stack than threads | |
| 85 * created by glibc (80 vs. 2048 or 4096 kB.) Follow glibc for consistency. | |
| 86 */ | |
| 87 static size_t uv__default_stack_size(void) { | |
| 88 #if !defined(__linux__) | |
| 89 return 0; | |
| 90 #elif defined(__PPC__) || defined(__ppc__) || defined(__powerpc__) | |
| 91 return 4 << 20; /* glibc default. */ | |
| 92 #else | |
| 93 return 2 << 20; /* glibc default. */ | |
| 94 #endif | |
| 95 } | |
| 96 | |
| 97 | |
| 98 /* On MacOS, threads other than the main thread are created with a reduced | |
| 99 * stack size by default. Adjust to RLIMIT_STACK aligned to the page size. | |
| 100 */ | |
| 101 size_t uv__thread_stack_size(void) { | |
| 102 #if defined(__APPLE__) || defined(__linux__) | |
| 103 struct rlimit lim; | |
| 104 | |
| 105 /* getrlimit() can fail on some aarch64 systems due to a glibc bug where | |
| 106 * the system call wrapper invokes the wrong system call. Don't treat | |
| 107 * that as fatal, just use the default stack size instead. | |
| 108 */ | |
| 109 if (getrlimit(RLIMIT_STACK, &lim)) | |
| 110 return uv__default_stack_size(); | |
| 111 | |
| 112 if (lim.rlim_cur == RLIM_INFINITY) | |
| 113 return uv__default_stack_size(); | |
| 114 | |
| 115 /* pthread_attr_setstacksize() expects page-aligned values. */ | |
| 116 lim.rlim_cur -= lim.rlim_cur % (rlim_t) getpagesize(); | |
| 117 | |
| 118 if (lim.rlim_cur >= (rlim_t) uv__min_stack_size()) | |
| 119 return lim.rlim_cur; | |
| 120 #endif | |
| 121 | |
| 122 return uv__default_stack_size(); | |
| 123 } | |
| 124 | |
| 125 | |
| 126 int uv_thread_create(uv_thread_t *tid, void (*entry)(void *arg), void *arg) { | |
| 127 uv_thread_options_t params; | |
| 128 params.flags = UV_THREAD_NO_FLAGS; | |
| 129 return uv_thread_create_ex(tid, ¶ms, entry, arg); | |
| 130 } | |
| 131 | |
| 132 | |
| 133 int uv_thread_detach(uv_thread_t *tid) { | |
| 134 return UV__ERR(pthread_detach(*tid)); | |
| 135 } | |
| 136 | |
| 137 | |
| 138 int uv_thread_create_ex(uv_thread_t* tid, | |
| 139 const uv_thread_options_t* params, | |
| 140 void (*entry)(void *arg), | |
| 141 void *arg) { | |
| 142 int err; | |
| 143 pthread_attr_t* attr; | |
| 144 pthread_attr_t attr_storage; | |
| 145 size_t pagesize; | |
| 146 size_t stack_size; | |
| 147 size_t min_stack_size; | |
| 148 | |
| 149 /* Used to squelch a -Wcast-function-type warning. */ | |
| 150 union { | |
| 151 void (*in)(void*); | |
| 152 void* (*out)(void*); | |
| 153 } f; | |
| 154 | |
| 155 stack_size = | |
| 156 params->flags & UV_THREAD_HAS_STACK_SIZE ? params->stack_size : 0; | |
| 157 | |
| 158 attr = NULL; | |
| 159 if (stack_size == 0) { | |
| 160 stack_size = uv__thread_stack_size(); | |
| 161 } else { | |
| 162 pagesize = (size_t)getpagesize(); | |
| 163 /* Round up to the nearest page boundary. */ | |
| 164 stack_size = (stack_size + pagesize - 1) &~ (pagesize - 1); | |
| 165 min_stack_size = uv__min_stack_size(); | |
| 166 if (stack_size < min_stack_size) | |
| 167 stack_size = min_stack_size; | |
| 168 } | |
| 169 | |
| 170 if (stack_size > 0) { | |
| 171 attr = &attr_storage; | |
| 172 | |
| 173 if (pthread_attr_init(attr)) | |
| 174 abort(); | |
| 175 | |
| 176 if (pthread_attr_setstacksize(attr, stack_size)) | |
| 177 abort(); | |
| 178 } | |
| 179 | |
| 180 f.in = entry; | |
| 181 err = pthread_create(tid, attr, f.out, arg); | |
| 182 | |
| 183 if (attr != NULL) | |
| 184 pthread_attr_destroy(attr); | |
| 185 | |
| 186 return UV__ERR(err); | |
| 187 } | |
| 188 | |
| 189 #if UV__CPU_AFFINITY_SUPPORTED | |
| 190 | |
| 191 int uv_thread_setaffinity(uv_thread_t* tid, | |
| 192 char* cpumask, | |
| 193 char* oldmask, | |
| 194 size_t mask_size) { | |
| 195 int i; | |
| 196 int r; | |
| 197 uv__cpu_set_t cpuset; | |
| 198 int cpumasksize; | |
| 199 | |
| 200 cpumasksize = uv_cpumask_size(); | |
| 201 if (cpumasksize < 0) | |
| 202 return cpumasksize; | |
| 203 if (mask_size < (size_t)cpumasksize) | |
| 204 return UV_EINVAL; | |
| 205 | |
| 206 if (oldmask != NULL) { | |
| 207 r = uv_thread_getaffinity(tid, oldmask, mask_size); | |
| 208 if (r < 0) | |
| 209 return r; | |
| 210 } | |
| 211 | |
| 212 CPU_ZERO(&cpuset); | |
| 213 for (i = 0; i < cpumasksize; i++) | |
| 214 if (cpumask[i]) | |
| 215 CPU_SET(i, &cpuset); | |
| 216 | |
| 217 #if defined(__ANDROID__) || defined(__OHOS__) | |
| 218 if (sched_setaffinity(pthread_gettid_np(*tid), sizeof(cpuset), &cpuset)) | |
| 219 r = errno; | |
| 220 else | |
| 221 r = 0; | |
| 222 #else | |
| 223 r = pthread_setaffinity_np(*tid, sizeof(cpuset), &cpuset); | |
| 224 #endif | |
| 225 | |
| 226 return UV__ERR(r); | |
| 227 } | |
| 228 | |
| 229 | |
| 230 int uv_thread_getaffinity(uv_thread_t* tid, | |
| 231 char* cpumask, | |
| 232 size_t mask_size) { | |
| 233 int r; | |
| 234 int i; | |
| 235 uv__cpu_set_t cpuset; | |
| 236 int cpumasksize; | |
| 237 | |
| 238 cpumasksize = uv_cpumask_size(); | |
| 239 if (cpumasksize < 0) | |
| 240 return cpumasksize; | |
| 241 if (mask_size < (size_t)cpumasksize) | |
| 242 return UV_EINVAL; | |
| 243 | |
| 244 CPU_ZERO(&cpuset); | |
| 245 #if defined(__ANDROID__) || defined(__OHOS__) | |
| 246 if (sched_getaffinity(pthread_gettid_np(*tid), sizeof(cpuset), &cpuset)) | |
| 247 r = errno; | |
| 248 else | |
| 249 r = 0; | |
| 250 #else | |
| 251 r = pthread_getaffinity_np(*tid, sizeof(cpuset), &cpuset); | |
| 252 #endif | |
| 253 if (r) | |
| 254 return UV__ERR(r); | |
| 255 for (i = 0; i < cpumasksize; i++) | |
| 256 cpumask[i] = !!CPU_ISSET(i, &cpuset); | |
| 257 | |
| 258 return 0; | |
| 259 } | |
| 260 #else | |
| 261 int uv_thread_setaffinity(uv_thread_t* tid, | |
| 262 char* cpumask, | |
| 263 char* oldmask, | |
| 264 size_t mask_size) { | |
| 265 return UV_ENOTSUP; | |
| 266 } | |
| 267 | |
| 268 | |
| 269 int uv_thread_getaffinity(uv_thread_t* tid, | |
| 270 char* cpumask, | |
| 271 size_t mask_size) { | |
| 272 return UV_ENOTSUP; | |
| 273 } | |
| 274 #endif /* defined(__linux__) || defined(UV_BSD_H) */ | |
| 275 | |
| 276 int uv_thread_getcpu(void) { | |
| 277 #if UV__CPU_AFFINITY_SUPPORTED | |
| 278 int cpu; | |
| 279 | |
| 280 cpu = sched_getcpu(); | |
| 281 if (cpu < 0) | |
| 282 return UV__ERR(errno); | |
| 283 | |
| 284 return cpu; | |
| 285 #else | |
| 286 return UV_ENOTSUP; | |
| 287 #endif | |
| 288 } | |
| 289 | |
| 290 uv_thread_t uv_thread_self(void) { | |
| 291 return pthread_self(); | |
| 292 } | |
| 293 | |
| 294 int uv_thread_join(uv_thread_t *tid) { | |
| 295 return UV__ERR(pthread_join(*tid, NULL)); | |
| 296 } | |
| 297 | |
| 298 | |
| 299 int uv_thread_equal(const uv_thread_t* t1, const uv_thread_t* t2) { | |
| 300 return pthread_equal(*t1, *t2); | |
| 301 } | |
| 302 | |
| 303 int uv_thread_setname(const char* name) { | |
| 304 if (name == NULL) | |
| 305 return UV_EINVAL; | |
| 306 return uv__thread_setname(name); | |
| 307 } | |
| 308 | |
| 309 int uv_thread_getname(uv_thread_t* tid, char* name, size_t size) { | |
| 310 if (name == NULL || size == 0) | |
| 311 return UV_EINVAL; | |
| 312 | |
| 313 return uv__thread_getname(tid, name, size); | |
| 314 } | |
| 315 | |
| 316 int uv_mutex_init(uv_mutex_t* mutex) { | |
| 317 #if defined(NDEBUG) || !defined(PTHREAD_MUTEX_ERRORCHECK) | |
| 318 return UV__ERR(pthread_mutex_init(mutex, NULL)); | |
| 319 #else | |
| 320 pthread_mutexattr_t attr; | |
| 321 int err; | |
| 322 | |
| 323 if (pthread_mutexattr_init(&attr)) | |
| 324 abort(); | |
| 325 | |
| 326 if (pthread_mutexattr_settype(&attr, PTHREAD_MUTEX_ERRORCHECK)) | |
| 327 abort(); | |
| 328 | |
| 329 err = pthread_mutex_init(mutex, &attr); | |
| 330 | |
| 331 if (pthread_mutexattr_destroy(&attr)) | |
| 332 abort(); | |
| 333 | |
| 334 return UV__ERR(err); | |
| 335 #endif | |
| 336 } | |
| 337 | |
| 338 | |
| 339 int uv_mutex_init_recursive(uv_mutex_t* mutex) { | |
| 340 pthread_mutexattr_t attr; | |
| 341 int err; | |
| 342 | |
| 343 if (pthread_mutexattr_init(&attr)) | |
| 344 abort(); | |
| 345 | |
| 346 if (pthread_mutexattr_settype(&attr, PTHREAD_MUTEX_RECURSIVE)) | |
| 347 abort(); | |
| 348 | |
| 349 err = pthread_mutex_init(mutex, &attr); | |
| 350 | |
| 351 if (pthread_mutexattr_destroy(&attr)) | |
| 352 abort(); | |
| 353 | |
| 354 return UV__ERR(err); | |
| 355 } | |
| 356 | |
| 357 | |
| 358 void uv_mutex_destroy(uv_mutex_t* mutex) { | |
| 359 if (pthread_mutex_destroy(mutex)) | |
| 360 abort(); | |
| 361 } | |
| 362 | |
| 363 | |
| 364 void uv_mutex_lock(uv_mutex_t* mutex) { | |
| 365 if (pthread_mutex_lock(mutex)) | |
| 366 abort(); | |
| 367 } | |
| 368 | |
| 369 | |
| 370 int uv_mutex_trylock(uv_mutex_t* mutex) { | |
| 371 int err; | |
| 372 | |
| 373 err = pthread_mutex_trylock(mutex); | |
| 374 if (err) { | |
| 375 if (err != EBUSY && err != EAGAIN) | |
| 376 abort(); | |
| 377 return UV_EBUSY; | |
| 378 } | |
| 379 | |
| 380 return 0; | |
| 381 } | |
| 382 | |
| 383 | |
| 384 void uv_mutex_unlock(uv_mutex_t* mutex) { | |
| 385 if (pthread_mutex_unlock(mutex)) | |
| 386 abort(); | |
| 387 } | |
| 388 | |
| 389 | |
| 390 int uv_rwlock_init(uv_rwlock_t* rwlock) { | |
| 391 return UV__ERR(pthread_rwlock_init(rwlock, NULL)); | |
| 392 } | |
| 393 | |
| 394 | |
| 395 void uv_rwlock_destroy(uv_rwlock_t* rwlock) { | |
| 396 if (pthread_rwlock_destroy(rwlock)) | |
| 397 abort(); | |
| 398 } | |
| 399 | |
| 400 | |
| 401 void uv_rwlock_rdlock(uv_rwlock_t* rwlock) { | |
| 402 if (pthread_rwlock_rdlock(rwlock)) | |
| 403 abort(); | |
| 404 } | |
| 405 | |
| 406 | |
| 407 int uv_rwlock_tryrdlock(uv_rwlock_t* rwlock) { | |
| 408 int err; | |
| 409 | |
| 410 err = pthread_rwlock_tryrdlock(rwlock); | |
| 411 if (err) { | |
| 412 if (err != EBUSY && err != EAGAIN) | |
| 413 abort(); | |
| 414 return UV_EBUSY; | |
| 415 } | |
| 416 | |
| 417 return 0; | |
| 418 } | |
| 419 | |
| 420 | |
| 421 void uv_rwlock_rdunlock(uv_rwlock_t* rwlock) { | |
| 422 if (pthread_rwlock_unlock(rwlock)) | |
| 423 abort(); | |
| 424 } | |
| 425 | |
| 426 | |
| 427 void uv_rwlock_wrlock(uv_rwlock_t* rwlock) { | |
| 428 if (pthread_rwlock_wrlock(rwlock)) | |
| 429 abort(); | |
| 430 } | |
| 431 | |
| 432 | |
| 433 int uv_rwlock_trywrlock(uv_rwlock_t* rwlock) { | |
| 434 int err; | |
| 435 | |
| 436 err = pthread_rwlock_trywrlock(rwlock); | |
| 437 if (err) { | |
| 438 if (err != EBUSY && err != EAGAIN) | |
| 439 abort(); | |
| 440 return UV_EBUSY; | |
| 441 } | |
| 442 | |
| 443 return 0; | |
| 444 } | |
| 445 | |
| 446 | |
| 447 void uv_rwlock_wrunlock(uv_rwlock_t* rwlock) { | |
| 448 if (pthread_rwlock_unlock(rwlock)) | |
| 449 abort(); | |
| 450 } | |
| 451 | |
| 452 | |
| 453 void uv_once(uv_once_t* guard, void (*callback)(void)) { | |
| 454 if (pthread_once(guard, callback)) | |
| 455 abort(); | |
| 456 } | |
| 457 | |
| 458 #if defined(__APPLE__) && defined(__MACH__) | |
| 459 | |
| 460 int uv_sem_init(uv_sem_t* sem, unsigned int value) { | |
| 461 kern_return_t err; | |
| 462 | |
| 463 err = semaphore_create(mach_task_self(), sem, SYNC_POLICY_FIFO, value); | |
| 464 if (err == KERN_SUCCESS) | |
| 465 return 0; | |
| 466 if (err == KERN_INVALID_ARGUMENT) | |
| 467 return UV_EINVAL; | |
| 468 if (err == KERN_RESOURCE_SHORTAGE) | |
| 469 return UV_ENOMEM; | |
| 470 | |
| 471 abort(); | |
| 472 return UV_EINVAL; /* Satisfy the compiler. */ | |
| 473 } | |
| 474 | |
| 475 | |
| 476 void uv_sem_destroy(uv_sem_t* sem) { | |
| 477 if (semaphore_destroy(mach_task_self(), *sem)) | |
| 478 abort(); | |
| 479 } | |
| 480 | |
| 481 | |
| 482 void uv_sem_post(uv_sem_t* sem) { | |
| 483 if (semaphore_signal(*sem)) | |
| 484 abort(); | |
| 485 } | |
| 486 | |
| 487 | |
| 488 void uv_sem_wait(uv_sem_t* sem) { | |
| 489 int r; | |
| 490 | |
| 491 do | |
| 492 r = semaphore_wait(*sem); | |
| 493 while (r == KERN_ABORTED); | |
| 494 | |
| 495 if (r != KERN_SUCCESS) | |
| 496 abort(); | |
| 497 } | |
| 498 | |
| 499 | |
| 500 int uv_sem_trywait(uv_sem_t* sem) { | |
| 501 mach_timespec_t interval; | |
| 502 kern_return_t err; | |
| 503 | |
| 504 interval.tv_sec = 0; | |
| 505 interval.tv_nsec = 0; | |
| 506 | |
| 507 err = semaphore_timedwait(*sem, interval); | |
| 508 if (err == KERN_SUCCESS) | |
| 509 return 0; | |
| 510 if (err == KERN_OPERATION_TIMED_OUT) | |
| 511 return UV_EAGAIN; | |
| 512 | |
| 513 abort(); | |
| 514 return UV_EINVAL; /* Satisfy the compiler. */ | |
| 515 } | |
| 516 | |
| 517 #else /* !(defined(__APPLE__) && defined(__MACH__)) */ | |
| 518 | |
| 519 #if defined(__GLIBC__) && !defined(__UCLIBC__) | |
| 520 | |
| 521 /* Hack around https://sourceware.org/bugzilla/show_bug.cgi?id=12674 | |
| 522 * by providing a custom implementation for glibc < 2.21 in terms of other | |
| 523 * concurrency primitives. | |
| 524 * Refs: https://github.com/nodejs/node/issues/19903 */ | |
| 525 | |
| 526 /* To preserve ABI compatibility, we treat the uv_sem_t as storage for | |
| 527 * a pointer to the actual struct we're using underneath. */ | |
| 528 | |
| 529 static uv_once_t glibc_version_check_once = UV_ONCE_INIT; | |
| 530 static int platform_needs_custom_semaphore = 0; | |
| 531 | |
| 532 static void glibc_version_check(void) { | |
| 533 const char* version = gnu_get_libc_version(); | |
| 534 platform_needs_custom_semaphore = | |
| 535 version[0] == '2' && version[1] == '.' && | |
| 536 atoi(version + 2) < 21; | |
| 537 } | |
| 538 | |
| 539 #elif defined(__MVS__) | |
| 540 | |
| 541 #define platform_needs_custom_semaphore 1 | |
| 542 | |
| 543 #else /* !defined(__GLIBC__) && !defined(__MVS__) */ | |
| 544 | |
| 545 #define platform_needs_custom_semaphore 0 | |
| 546 | |
| 547 #endif | |
| 548 | |
| 549 typedef struct uv_semaphore_s { | |
| 550 uv_mutex_t mutex; | |
| 551 uv_cond_t cond; | |
| 552 unsigned int value; | |
| 553 } uv_semaphore_t; | |
| 554 | |
| 555 #if (defined(__GLIBC__) && !defined(__UCLIBC__)) || \ | |
| 556 platform_needs_custom_semaphore | |
| 557 STATIC_ASSERT(sizeof(uv_sem_t) >= sizeof(uv_semaphore_t*)); | |
| 558 #endif | |
| 559 | |
| 560 static int uv__custom_sem_init(uv_sem_t* sem_, unsigned int value) { | |
| 561 int err; | |
| 562 uv_semaphore_t* sem; | |
| 563 | |
| 564 sem = uv__malloc(sizeof(*sem)); | |
| 565 if (sem == NULL) | |
| 566 return UV_ENOMEM; | |
| 567 | |
| 568 if ((err = uv_mutex_init(&sem->mutex)) != 0) { | |
| 569 uv__free(sem); | |
| 570 return err; | |
| 571 } | |
| 572 | |
| 573 if ((err = uv_cond_init(&sem->cond)) != 0) { | |
| 574 uv_mutex_destroy(&sem->mutex); | |
| 575 uv__free(sem); | |
| 576 return err; | |
| 577 } | |
| 578 | |
| 579 sem->value = value; | |
| 580 *(uv_semaphore_t**)sem_ = sem; | |
| 581 return 0; | |
| 582 } | |
| 583 | |
| 584 | |
| 585 static void uv__custom_sem_destroy(uv_sem_t* sem_) { | |
| 586 uv_semaphore_t* sem; | |
| 587 | |
| 588 sem = *(uv_semaphore_t**)sem_; | |
| 589 uv_cond_destroy(&sem->cond); | |
| 590 uv_mutex_destroy(&sem->mutex); | |
| 591 uv__free(sem); | |
| 592 } | |
| 593 | |
| 594 | |
| 595 static void uv__custom_sem_post(uv_sem_t* sem_) { | |
| 596 uv_semaphore_t* sem; | |
| 597 | |
| 598 sem = *(uv_semaphore_t**)sem_; | |
| 599 uv_mutex_lock(&sem->mutex); | |
| 600 sem->value++; | |
| 601 if (sem->value == 1) | |
| 602 uv_cond_signal(&sem->cond); /* Release one to replace us. */ | |
| 603 uv_mutex_unlock(&sem->mutex); | |
| 604 } | |
| 605 | |
| 606 | |
| 607 static void uv__custom_sem_wait(uv_sem_t* sem_) { | |
| 608 uv_semaphore_t* sem; | |
| 609 | |
| 610 sem = *(uv_semaphore_t**)sem_; | |
| 611 uv_mutex_lock(&sem->mutex); | |
| 612 while (sem->value == 0) | |
| 613 uv_cond_wait(&sem->cond, &sem->mutex); | |
| 614 sem->value--; | |
| 615 uv_mutex_unlock(&sem->mutex); | |
| 616 } | |
| 617 | |
| 618 | |
| 619 static int uv__custom_sem_trywait(uv_sem_t* sem_) { | |
| 620 uv_semaphore_t* sem; | |
| 621 | |
| 622 sem = *(uv_semaphore_t**)sem_; | |
| 623 if (uv_mutex_trylock(&sem->mutex) != 0) | |
| 624 return UV_EAGAIN; | |
| 625 | |
| 626 if (sem->value == 0) { | |
| 627 uv_mutex_unlock(&sem->mutex); | |
| 628 return UV_EAGAIN; | |
| 629 } | |
| 630 | |
| 631 sem->value--; | |
| 632 uv_mutex_unlock(&sem->mutex); | |
| 633 | |
| 634 return 0; | |
| 635 } | |
| 636 | |
| 637 static int uv__sem_init(uv_sem_t* sem, unsigned int value) { | |
| 638 if (sem_init(sem, 0, value)) | |
| 639 return UV__ERR(errno); | |
| 640 return 0; | |
| 641 } | |
| 642 | |
| 643 | |
| 644 static void uv__sem_destroy(uv_sem_t* sem) { | |
| 645 if (sem_destroy(sem)) | |
| 646 abort(); | |
| 647 } | |
| 648 | |
| 649 | |
| 650 static void uv__sem_post(uv_sem_t* sem) { | |
| 651 if (sem_post(sem)) | |
| 652 abort(); | |
| 653 } | |
| 654 | |
| 655 | |
| 656 static void uv__sem_wait(uv_sem_t* sem) { | |
| 657 int r; | |
| 658 | |
| 659 do | |
| 660 r = sem_wait(sem); | |
| 661 while (r == -1 && errno == EINTR); | |
| 662 | |
| 663 if (r) | |
| 664 abort(); | |
| 665 } | |
| 666 | |
| 667 | |
| 668 static int uv__sem_trywait(uv_sem_t* sem) { | |
| 669 int r; | |
| 670 | |
| 671 do | |
| 672 r = sem_trywait(sem); | |
| 673 while (r == -1 && errno == EINTR); | |
| 674 | |
| 675 if (r) { | |
| 676 if (errno == EAGAIN) | |
| 677 return UV_EAGAIN; | |
| 678 abort(); | |
| 679 } | |
| 680 | |
| 681 return 0; | |
| 682 } | |
| 683 | |
| 684 int uv_sem_init(uv_sem_t* sem, unsigned int value) { | |
| 685 #if defined(__GLIBC__) && !defined(__UCLIBC__) | |
| 686 uv_once(&glibc_version_check_once, glibc_version_check); | |
| 687 #endif | |
| 688 | |
| 689 if (platform_needs_custom_semaphore) | |
| 690 return uv__custom_sem_init(sem, value); | |
| 691 else | |
| 692 return uv__sem_init(sem, value); | |
| 693 } | |
| 694 | |
| 695 | |
| 696 void uv_sem_destroy(uv_sem_t* sem) { | |
| 697 if (platform_needs_custom_semaphore) | |
| 698 uv__custom_sem_destroy(sem); | |
| 699 else | |
| 700 uv__sem_destroy(sem); | |
| 701 } | |
| 702 | |
| 703 | |
| 704 void uv_sem_post(uv_sem_t* sem) { | |
| 705 if (platform_needs_custom_semaphore) | |
| 706 uv__custom_sem_post(sem); | |
| 707 else | |
| 708 uv__sem_post(sem); | |
| 709 } | |
| 710 | |
| 711 | |
| 712 void uv_sem_wait(uv_sem_t* sem) { | |
| 713 if (platform_needs_custom_semaphore) | |
| 714 uv__custom_sem_wait(sem); | |
| 715 else | |
| 716 uv__sem_wait(sem); | |
| 717 } | |
| 718 | |
| 719 | |
| 720 int uv_sem_trywait(uv_sem_t* sem) { | |
| 721 if (platform_needs_custom_semaphore) | |
| 722 return uv__custom_sem_trywait(sem); | |
| 723 else | |
| 724 return uv__sem_trywait(sem); | |
| 725 } | |
| 726 | |
| 727 #endif /* defined(__APPLE__) && defined(__MACH__) */ | |
| 728 | |
| 729 | |
| 730 #if defined(__APPLE__) && defined(__MACH__) || defined(__MVS__) | |
| 731 | |
| 732 int uv_cond_init(uv_cond_t* cond) { | |
| 733 return UV__ERR(pthread_cond_init(cond, NULL)); | |
| 734 } | |
| 735 | |
| 736 #else /* !(defined(__APPLE__) && defined(__MACH__)) */ | |
| 737 | |
| 738 int uv_cond_init(uv_cond_t* cond) { | |
| 739 pthread_condattr_t attr; | |
| 740 int err; | |
| 741 | |
| 742 err = pthread_condattr_init(&attr); | |
| 743 if (err) | |
| 744 return UV__ERR(err); | |
| 745 | |
| 746 err = pthread_condattr_setclock(&attr, CLOCK_MONOTONIC); | |
| 747 if (err) | |
| 748 goto error2; | |
| 749 | |
| 750 err = pthread_cond_init(cond, &attr); | |
| 751 if (err) | |
| 752 goto error2; | |
| 753 | |
| 754 err = pthread_condattr_destroy(&attr); | |
| 755 if (err) | |
| 756 goto error; | |
| 757 | |
| 758 return 0; | |
| 759 | |
| 760 error: | |
| 761 pthread_cond_destroy(cond); | |
| 762 error2: | |
| 763 pthread_condattr_destroy(&attr); | |
| 764 return UV__ERR(err); | |
| 765 } | |
| 766 | |
| 767 #endif /* defined(__APPLE__) && defined(__MACH__) */ | |
| 768 | |
| 769 void uv_cond_destroy(uv_cond_t* cond) { | |
| 770 #if defined(__APPLE__) && defined(__MACH__) | |
| 771 /* It has been reported that destroying condition variables that have been | |
| 772 * signalled but not waited on can sometimes result in application crashes. | |
| 773 * See https://codereview.chromium.org/1323293005. | |
| 774 */ | |
| 775 pthread_mutex_t mutex; | |
| 776 struct timespec ts; | |
| 777 int err; | |
| 778 | |
| 779 if (pthread_mutex_init(&mutex, NULL)) | |
| 780 abort(); | |
| 781 | |
| 782 if (pthread_mutex_lock(&mutex)) | |
| 783 abort(); | |
| 784 | |
| 785 ts.tv_sec = 0; | |
| 786 ts.tv_nsec = 1; | |
| 787 | |
| 788 err = pthread_cond_timedwait_relative_np(cond, &mutex, &ts); | |
| 789 if (err != 0 && err != ETIMEDOUT) | |
| 790 abort(); | |
| 791 | |
| 792 if (pthread_mutex_unlock(&mutex)) | |
| 793 abort(); | |
| 794 | |
| 795 if (pthread_mutex_destroy(&mutex)) | |
| 796 abort(); | |
| 797 #endif /* defined(__APPLE__) && defined(__MACH__) */ | |
| 798 | |
| 799 if (pthread_cond_destroy(cond)) | |
| 800 abort(); | |
| 801 } | |
| 802 | |
| 803 void uv_cond_signal(uv_cond_t* cond) { | |
| 804 if (pthread_cond_signal(cond)) | |
| 805 abort(); | |
| 806 } | |
| 807 | |
| 808 void uv_cond_broadcast(uv_cond_t* cond) { | |
| 809 if (pthread_cond_broadcast(cond)) | |
| 810 abort(); | |
| 811 } | |
| 812 | |
| 813 #if defined(__APPLE__) && defined(__MACH__) | |
| 814 | |
| 815 void uv_cond_wait(uv_cond_t* cond, uv_mutex_t* mutex) { | |
| 816 int r; | |
| 817 | |
| 818 errno = 0; | |
| 819 r = pthread_cond_wait(cond, mutex); | |
| 820 | |
| 821 /* Workaround for a bug in OS X at least up to 13.6 | |
| 822 * See https://github.com/libuv/libuv/issues/4165 | |
| 823 */ | |
| 824 if (r == EINVAL) | |
| 825 if (errno == EBUSY) | |
| 826 return; | |
| 827 | |
| 828 if (r) | |
| 829 abort(); | |
| 830 } | |
| 831 | |
| 832 #else /* !(defined(__APPLE__) && defined(__MACH__)) */ | |
| 833 | |
| 834 void uv_cond_wait(uv_cond_t* cond, uv_mutex_t* mutex) { | |
| 835 if (pthread_cond_wait(cond, mutex)) | |
| 836 abort(); | |
| 837 } | |
| 838 | |
| 839 #endif | |
| 840 | |
| 841 int uv_cond_timedwait(uv_cond_t* cond, uv_mutex_t* mutex, uint64_t timeout) { | |
| 842 int r; | |
| 843 struct timespec ts; | |
| 844 #if defined(__MVS__) | |
| 845 struct timeval tv; | |
| 846 #endif | |
| 847 | |
| 848 #if defined(__APPLE__) && defined(__MACH__) | |
| 849 ts.tv_sec = timeout / NANOSEC; | |
| 850 ts.tv_nsec = timeout % NANOSEC; | |
| 851 r = pthread_cond_timedwait_relative_np(cond, mutex, &ts); | |
| 852 #else | |
| 853 #if defined(__MVS__) | |
| 854 if (gettimeofday(&tv, NULL)) | |
| 855 abort(); | |
| 856 timeout += tv.tv_sec * NANOSEC + tv.tv_usec * 1e3; | |
| 857 #else | |
| 858 timeout += uv__hrtime(UV_CLOCK_PRECISE); | |
| 859 #endif | |
| 860 ts.tv_sec = timeout / NANOSEC; | |
| 861 ts.tv_nsec = timeout % NANOSEC; | |
| 862 r = pthread_cond_timedwait(cond, mutex, &ts); | |
| 863 #endif | |
| 864 | |
| 865 | |
| 866 if (r == 0) | |
| 867 return 0; | |
| 868 | |
| 869 if (r == ETIMEDOUT) | |
| 870 return UV_ETIMEDOUT; | |
| 871 | |
| 872 abort(); | |
| 873 #ifndef __SUNPRO_C | |
| 874 return UV_EINVAL; /* Satisfy the compiler. */ | |
| 875 #endif | |
| 876 } | |
| 877 | |
| 878 | |
| 879 int uv_key_create(uv_key_t* key) { | |
| 880 return UV__ERR(pthread_key_create(key, NULL)); | |
| 881 } | |
| 882 | |
| 883 | |
| 884 void uv_key_delete(uv_key_t* key) { | |
| 885 if (pthread_key_delete(*key)) | |
| 886 abort(); | |
| 887 } | |
| 888 | |
| 889 | |
| 890 void* uv_key_get(uv_key_t* key) { | |
| 891 return pthread_getspecific(*key); | |
| 892 } | |
| 893 | |
| 894 | |
| 895 void uv_key_set(uv_key_t* key, void* value) { | |
| 896 if (pthread_setspecific(*key, value)) | |
| 897 abort(); | |
| 898 } | |
| 899 | |
| 900 #if defined(_AIX) || defined(__MVS__) || defined(__PASE__) | |
| 901 int uv__thread_setname(const char* name) { | |
| 902 return UV_ENOSYS; | |
| 903 } | |
| 904 #elif defined(__APPLE__) | |
| 905 int uv__thread_setname(const char* name) { | |
| 906 char namebuf[UV_PTHREAD_MAX_NAMELEN_NP]; | |
| 907 strncpy(namebuf, name, sizeof(namebuf) - 1); | |
| 908 namebuf[sizeof(namebuf) - 1] = '\0'; | |
| 909 int err = pthread_setname_np(namebuf); | |
| 910 if (err) | |
| 911 return UV__ERR(errno); | |
| 912 return 0; | |
| 913 } | |
| 914 #elif defined(__NetBSD__) | |
| 915 int uv__thread_setname(const char* name) { | |
| 916 char namebuf[UV_PTHREAD_MAX_NAMELEN_NP]; | |
| 917 strncpy(namebuf, name, sizeof(namebuf) - 1); | |
| 918 namebuf[sizeof(namebuf) - 1] = '\0'; | |
| 919 return UV__ERR(pthread_setname_np(pthread_self(), "%s", namebuf)); | |
| 920 } | |
| 921 #elif defined(__OpenBSD__) | |
| 922 int uv__thread_setname(const char* name) { | |
| 923 char namebuf[UV_PTHREAD_MAX_NAMELEN_NP]; | |
| 924 strncpy(namebuf, name, sizeof(namebuf) - 1); | |
| 925 namebuf[sizeof(namebuf) - 1] = '\0'; | |
| 926 pthread_set_name_np(pthread_self(), namebuf); | |
| 927 return 0; | |
| 928 } | |
| 929 #else | |
| 930 int uv__thread_setname(const char* name) { | |
| 931 char namebuf[UV_PTHREAD_MAX_NAMELEN_NP]; | |
| 932 strncpy(namebuf, name, sizeof(namebuf) - 1); | |
| 933 namebuf[sizeof(namebuf) - 1] = '\0'; | |
| 934 return UV__ERR(pthread_setname_np(pthread_self(), namebuf)); | |
| 935 } | |
| 936 #endif | |
| 937 | |
| 938 #if (defined(__ANDROID_API__) && __ANDROID_API__ < 26) || \ | |
| 939 defined(_AIX) || \ | |
| 940 defined(__MVS__) || \ | |
| 941 defined(__PASE__) | |
| 942 int uv__thread_getname(uv_thread_t* tid, char* name, size_t size) { | |
| 943 return UV_ENOSYS; | |
| 944 } | |
| 945 #elif defined(__OpenBSD__) | |
| 946 int uv__thread_getname(uv_thread_t* tid, char* name, size_t size) { | |
| 947 char thread_name[UV_PTHREAD_MAX_NAMELEN_NP]; | |
| 948 pthread_get_name_np(*tid, thread_name, sizeof(thread_name)); | |
| 949 strncpy(name, thread_name, size - 1); | |
| 950 name[size - 1] = '\0'; | |
| 951 return 0; | |
| 952 } | |
| 953 #elif defined(__APPLE__) | |
| 954 int uv__thread_getname(uv_thread_t* tid, char* name, size_t size) { | |
| 955 char thread_name[UV_PTHREAD_MAX_NAMELEN_NP]; | |
| 956 if (pthread_getname_np(*tid, thread_name, sizeof(thread_name)) != 0) | |
| 957 return UV__ERR(errno); | |
| 958 | |
| 959 strncpy(name, thread_name, size - 1); | |
| 960 name[size - 1] = '\0'; | |
| 961 return 0; | |
| 962 } | |
| 963 #else | |
| 964 int uv__thread_getname(uv_thread_t* tid, char* name, size_t size) { | |
| 965 int r; | |
| 966 char thread_name[UV_PTHREAD_MAX_NAMELEN_NP]; | |
| 967 r = pthread_getname_np(*tid, thread_name, sizeof(thread_name)); | |
| 968 if (r != 0) | |
| 969 return UV__ERR(r); | |
| 970 | |
| 971 strncpy(name, thread_name, size - 1); | |
| 972 name[size - 1] = '\0'; | |
| 973 return 0; | |
| 974 } | |
| 975 #endif |