]> git.proxmox.com Git - mirror_qemu.git/blob - util/qemu-thread-posix.c
Merge remote-tracking branch 'remotes/xtensa/tags/20191023-xtensa' into staging
[mirror_qemu.git] / util / qemu-thread-posix.c
1 /*
2 * Wrappers around mutex/cond/thread functions
3 *
4 * Copyright Red Hat, Inc. 2009
5 *
6 * Author:
7 * Marcelo Tosatti <mtosatti@redhat.com>
8 *
9 * This work is licensed under the terms of the GNU GPL, version 2 or later.
10 * See the COPYING file in the top-level directory.
11 *
12 */
13 #include "qemu/osdep.h"
14 #include "qemu/thread.h"
15 #include "qemu/atomic.h"
16 #include "qemu/notify.h"
17 #include "qemu-thread-common.h"
18
19 static bool name_threads;
20
21 void qemu_thread_naming(bool enable)
22 {
23 name_threads = enable;
24
25 #ifndef CONFIG_THREAD_SETNAME_BYTHREAD
26 /* This is a debugging option, not fatal */
27 if (enable) {
28 fprintf(stderr, "qemu: thread naming not supported on this host\n");
29 }
30 #endif
31 }
32
33 static void error_exit(int err, const char *msg)
34 {
35 fprintf(stderr, "qemu: %s: %s\n", msg, strerror(err));
36 abort();
37 }
38
39 static void compute_abs_deadline(struct timespec *ts, int ms)
40 {
41 struct timeval tv;
42 gettimeofday(&tv, NULL);
43 ts->tv_nsec = tv.tv_usec * 1000 + (ms % 1000) * 1000000;
44 ts->tv_sec = tv.tv_sec + ms / 1000;
45 if (ts->tv_nsec >= 1000000000) {
46 ts->tv_sec++;
47 ts->tv_nsec -= 1000000000;
48 }
49 }
50
51 void qemu_mutex_init(QemuMutex *mutex)
52 {
53 int err;
54
55 err = pthread_mutex_init(&mutex->lock, NULL);
56 if (err)
57 error_exit(err, __func__);
58 qemu_mutex_post_init(mutex);
59 }
60
61 void qemu_mutex_destroy(QemuMutex *mutex)
62 {
63 int err;
64
65 assert(mutex->initialized);
66 mutex->initialized = false;
67 err = pthread_mutex_destroy(&mutex->lock);
68 if (err)
69 error_exit(err, __func__);
70 }
71
72 void qemu_mutex_lock_impl(QemuMutex *mutex, const char *file, const int line)
73 {
74 int err;
75
76 assert(mutex->initialized);
77 qemu_mutex_pre_lock(mutex, file, line);
78 err = pthread_mutex_lock(&mutex->lock);
79 if (err)
80 error_exit(err, __func__);
81 qemu_mutex_post_lock(mutex, file, line);
82 }
83
84 int qemu_mutex_trylock_impl(QemuMutex *mutex, const char *file, const int line)
85 {
86 int err;
87
88 assert(mutex->initialized);
89 err = pthread_mutex_trylock(&mutex->lock);
90 if (err == 0) {
91 qemu_mutex_post_lock(mutex, file, line);
92 return 0;
93 }
94 if (err != EBUSY) {
95 error_exit(err, __func__);
96 }
97 return -EBUSY;
98 }
99
100 void qemu_mutex_unlock_impl(QemuMutex *mutex, const char *file, const int line)
101 {
102 int err;
103
104 assert(mutex->initialized);
105 qemu_mutex_pre_unlock(mutex, file, line);
106 err = pthread_mutex_unlock(&mutex->lock);
107 if (err)
108 error_exit(err, __func__);
109 }
110
111 void qemu_rec_mutex_init(QemuRecMutex *mutex)
112 {
113 int err;
114 pthread_mutexattr_t attr;
115
116 pthread_mutexattr_init(&attr);
117 pthread_mutexattr_settype(&attr, PTHREAD_MUTEX_RECURSIVE);
118 err = pthread_mutex_init(&mutex->lock, &attr);
119 pthread_mutexattr_destroy(&attr);
120 if (err) {
121 error_exit(err, __func__);
122 }
123 mutex->initialized = true;
124 }
125
126 void qemu_cond_init(QemuCond *cond)
127 {
128 int err;
129
130 err = pthread_cond_init(&cond->cond, NULL);
131 if (err)
132 error_exit(err, __func__);
133 cond->initialized = true;
134 }
135
136 void qemu_cond_destroy(QemuCond *cond)
137 {
138 int err;
139
140 assert(cond->initialized);
141 cond->initialized = false;
142 err = pthread_cond_destroy(&cond->cond);
143 if (err)
144 error_exit(err, __func__);
145 }
146
147 void qemu_cond_signal(QemuCond *cond)
148 {
149 int err;
150
151 assert(cond->initialized);
152 err = pthread_cond_signal(&cond->cond);
153 if (err)
154 error_exit(err, __func__);
155 }
156
157 void qemu_cond_broadcast(QemuCond *cond)
158 {
159 int err;
160
161 assert(cond->initialized);
162 err = pthread_cond_broadcast(&cond->cond);
163 if (err)
164 error_exit(err, __func__);
165 }
166
167 void qemu_cond_wait_impl(QemuCond *cond, QemuMutex *mutex, const char *file, const int line)
168 {
169 int err;
170
171 assert(cond->initialized);
172 qemu_mutex_pre_unlock(mutex, file, line);
173 err = pthread_cond_wait(&cond->cond, &mutex->lock);
174 qemu_mutex_post_lock(mutex, file, line);
175 if (err)
176 error_exit(err, __func__);
177 }
178
179 bool qemu_cond_timedwait_impl(QemuCond *cond, QemuMutex *mutex, int ms,
180 const char *file, const int line)
181 {
182 int err;
183 struct timespec ts;
184
185 assert(cond->initialized);
186 trace_qemu_mutex_unlock(mutex, file, line);
187 compute_abs_deadline(&ts, ms);
188 err = pthread_cond_timedwait(&cond->cond, &mutex->lock, &ts);
189 trace_qemu_mutex_locked(mutex, file, line);
190 if (err && err != ETIMEDOUT) {
191 error_exit(err, __func__);
192 }
193 return err != ETIMEDOUT;
194 }
195
196 void qemu_sem_init(QemuSemaphore *sem, int init)
197 {
198 int rc;
199
200 #ifndef CONFIG_SEM_TIMEDWAIT
201 rc = pthread_mutex_init(&sem->lock, NULL);
202 if (rc != 0) {
203 error_exit(rc, __func__);
204 }
205 rc = pthread_cond_init(&sem->cond, NULL);
206 if (rc != 0) {
207 error_exit(rc, __func__);
208 }
209 if (init < 0) {
210 error_exit(EINVAL, __func__);
211 }
212 sem->count = init;
213 #else
214 rc = sem_init(&sem->sem, 0, init);
215 if (rc < 0) {
216 error_exit(errno, __func__);
217 }
218 #endif
219 sem->initialized = true;
220 }
221
222 void qemu_sem_destroy(QemuSemaphore *sem)
223 {
224 int rc;
225
226 assert(sem->initialized);
227 sem->initialized = false;
228 #ifndef CONFIG_SEM_TIMEDWAIT
229 rc = pthread_cond_destroy(&sem->cond);
230 if (rc < 0) {
231 error_exit(rc, __func__);
232 }
233 rc = pthread_mutex_destroy(&sem->lock);
234 if (rc < 0) {
235 error_exit(rc, __func__);
236 }
237 #else
238 rc = sem_destroy(&sem->sem);
239 if (rc < 0) {
240 error_exit(errno, __func__);
241 }
242 #endif
243 }
244
245 void qemu_sem_post(QemuSemaphore *sem)
246 {
247 int rc;
248
249 assert(sem->initialized);
250 #ifndef CONFIG_SEM_TIMEDWAIT
251 pthread_mutex_lock(&sem->lock);
252 if (sem->count == UINT_MAX) {
253 rc = EINVAL;
254 } else {
255 sem->count++;
256 rc = pthread_cond_signal(&sem->cond);
257 }
258 pthread_mutex_unlock(&sem->lock);
259 if (rc != 0) {
260 error_exit(rc, __func__);
261 }
262 #else
263 rc = sem_post(&sem->sem);
264 if (rc < 0) {
265 error_exit(errno, __func__);
266 }
267 #endif
268 }
269
270 int qemu_sem_timedwait(QemuSemaphore *sem, int ms)
271 {
272 int rc;
273 struct timespec ts;
274
275 assert(sem->initialized);
276 #ifndef CONFIG_SEM_TIMEDWAIT
277 rc = 0;
278 compute_abs_deadline(&ts, ms);
279 pthread_mutex_lock(&sem->lock);
280 while (sem->count == 0) {
281 rc = pthread_cond_timedwait(&sem->cond, &sem->lock, &ts);
282 if (rc == ETIMEDOUT) {
283 break;
284 }
285 if (rc != 0) {
286 error_exit(rc, __func__);
287 }
288 }
289 if (rc != ETIMEDOUT) {
290 --sem->count;
291 }
292 pthread_mutex_unlock(&sem->lock);
293 return (rc == ETIMEDOUT ? -1 : 0);
294 #else
295 if (ms <= 0) {
296 /* This is cheaper than sem_timedwait. */
297 do {
298 rc = sem_trywait(&sem->sem);
299 } while (rc == -1 && errno == EINTR);
300 if (rc == -1 && errno == EAGAIN) {
301 return -1;
302 }
303 } else {
304 compute_abs_deadline(&ts, ms);
305 do {
306 rc = sem_timedwait(&sem->sem, &ts);
307 } while (rc == -1 && errno == EINTR);
308 if (rc == -1 && errno == ETIMEDOUT) {
309 return -1;
310 }
311 }
312 if (rc < 0) {
313 error_exit(errno, __func__);
314 }
315 return 0;
316 #endif
317 }
318
319 void qemu_sem_wait(QemuSemaphore *sem)
320 {
321 int rc;
322
323 assert(sem->initialized);
324 #ifndef CONFIG_SEM_TIMEDWAIT
325 pthread_mutex_lock(&sem->lock);
326 while (sem->count == 0) {
327 rc = pthread_cond_wait(&sem->cond, &sem->lock);
328 if (rc != 0) {
329 error_exit(rc, __func__);
330 }
331 }
332 --sem->count;
333 pthread_mutex_unlock(&sem->lock);
334 #else
335 do {
336 rc = sem_wait(&sem->sem);
337 } while (rc == -1 && errno == EINTR);
338 if (rc < 0) {
339 error_exit(errno, __func__);
340 }
341 #endif
342 }
343
344 #ifdef __linux__
345 #include "qemu/futex.h"
346 #else
347 static inline void qemu_futex_wake(QemuEvent *ev, int n)
348 {
349 assert(ev->initialized);
350 pthread_mutex_lock(&ev->lock);
351 if (n == 1) {
352 pthread_cond_signal(&ev->cond);
353 } else {
354 pthread_cond_broadcast(&ev->cond);
355 }
356 pthread_mutex_unlock(&ev->lock);
357 }
358
359 static inline void qemu_futex_wait(QemuEvent *ev, unsigned val)
360 {
361 assert(ev->initialized);
362 pthread_mutex_lock(&ev->lock);
363 if (ev->value == val) {
364 pthread_cond_wait(&ev->cond, &ev->lock);
365 }
366 pthread_mutex_unlock(&ev->lock);
367 }
368 #endif
369
370 /* Valid transitions:
371 * - free->set, when setting the event
372 * - busy->set, when setting the event, followed by qemu_futex_wake
373 * - set->free, when resetting the event
374 * - free->busy, when waiting
375 *
376 * set->busy does not happen (it can be observed from the outside but
377 * it really is set->free->busy).
378 *
379 * busy->free provably cannot happen; to enforce it, the set->free transition
380 * is done with an OR, which becomes a no-op if the event has concurrently
381 * transitioned to free or busy.
382 */
383
384 #define EV_SET 0
385 #define EV_FREE 1
386 #define EV_BUSY -1
387
388 void qemu_event_init(QemuEvent *ev, bool init)
389 {
390 #ifndef __linux__
391 pthread_mutex_init(&ev->lock, NULL);
392 pthread_cond_init(&ev->cond, NULL);
393 #endif
394
395 ev->value = (init ? EV_SET : EV_FREE);
396 ev->initialized = true;
397 }
398
399 void qemu_event_destroy(QemuEvent *ev)
400 {
401 assert(ev->initialized);
402 ev->initialized = false;
403 #ifndef __linux__
404 pthread_mutex_destroy(&ev->lock);
405 pthread_cond_destroy(&ev->cond);
406 #endif
407 }
408
409 void qemu_event_set(QemuEvent *ev)
410 {
411 /* qemu_event_set has release semantics, but because it *loads*
412 * ev->value we need a full memory barrier here.
413 */
414 assert(ev->initialized);
415 smp_mb();
416 if (atomic_read(&ev->value) != EV_SET) {
417 if (atomic_xchg(&ev->value, EV_SET) == EV_BUSY) {
418 /* There were waiters, wake them up. */
419 qemu_futex_wake(ev, INT_MAX);
420 }
421 }
422 }
423
424 void qemu_event_reset(QemuEvent *ev)
425 {
426 unsigned value;
427
428 assert(ev->initialized);
429 value = atomic_read(&ev->value);
430 smp_mb_acquire();
431 if (value == EV_SET) {
432 /*
433 * If there was a concurrent reset (or even reset+wait),
434 * do nothing. Otherwise change EV_SET->EV_FREE.
435 */
436 atomic_or(&ev->value, EV_FREE);
437 }
438 }
439
440 void qemu_event_wait(QemuEvent *ev)
441 {
442 unsigned value;
443
444 assert(ev->initialized);
445 value = atomic_read(&ev->value);
446 smp_mb_acquire();
447 if (value != EV_SET) {
448 if (value == EV_FREE) {
449 /*
450 * Leave the event reset and tell qemu_event_set that there
451 * are waiters. No need to retry, because there cannot be
452 * a concurrent busy->free transition. After the CAS, the
453 * event will be either set or busy.
454 */
455 if (atomic_cmpxchg(&ev->value, EV_FREE, EV_BUSY) == EV_SET) {
456 return;
457 }
458 }
459 qemu_futex_wait(ev, EV_BUSY);
460 }
461 }
462
463 static __thread NotifierList thread_exit;
464
465 /*
466 * Note that in this implementation you can register a thread-exit
467 * notifier for the main thread, but it will never be called.
468 * This is OK because main thread exit can only happen when the
469 * entire process is exiting, and the API allows notifiers to not
470 * be called on process exit.
471 */
472 void qemu_thread_atexit_add(Notifier *notifier)
473 {
474 notifier_list_add(&thread_exit, notifier);
475 }
476
477 void qemu_thread_atexit_remove(Notifier *notifier)
478 {
479 notifier_remove(notifier);
480 }
481
482 static void qemu_thread_atexit_notify(void *arg)
483 {
484 /*
485 * Called when non-main thread exits (via qemu_thread_exit()
486 * or by returning from its start routine.)
487 */
488 notifier_list_notify(&thread_exit, NULL);
489 }
490
491 typedef struct {
492 void *(*start_routine)(void *);
493 void *arg;
494 char *name;
495 } QemuThreadArgs;
496
497 static void *qemu_thread_start(void *args)
498 {
499 QemuThreadArgs *qemu_thread_args = args;
500 void *(*start_routine)(void *) = qemu_thread_args->start_routine;
501 void *arg = qemu_thread_args->arg;
502 void *r;
503
504 #ifdef CONFIG_THREAD_SETNAME_BYTHREAD
505 /* Attempt to set the threads name; note that this is for debug, so
506 * we're not going to fail if we can't set it.
507 */
508 if (name_threads && qemu_thread_args->name) {
509 # if defined(CONFIG_PTHREAD_SETNAME_NP_W_TID)
510 pthread_setname_np(pthread_self(), qemu_thread_args->name);
511 # elif defined(CONFIG_PTHREAD_SETNAME_NP_WO_TID)
512 pthread_setname_np(qemu_thread_args->name);
513 # endif
514 }
515 #endif
516 g_free(qemu_thread_args->name);
517 g_free(qemu_thread_args);
518 pthread_cleanup_push(qemu_thread_atexit_notify, NULL);
519 r = start_routine(arg);
520 pthread_cleanup_pop(1);
521 return r;
522 }
523
524 void qemu_thread_create(QemuThread *thread, const char *name,
525 void *(*start_routine)(void*),
526 void *arg, int mode)
527 {
528 sigset_t set, oldset;
529 int err;
530 pthread_attr_t attr;
531 QemuThreadArgs *qemu_thread_args;
532
533 err = pthread_attr_init(&attr);
534 if (err) {
535 error_exit(err, __func__);
536 }
537
538 if (mode == QEMU_THREAD_DETACHED) {
539 pthread_attr_setdetachstate(&attr, PTHREAD_CREATE_DETACHED);
540 }
541
542 /* Leave signal handling to the iothread. */
543 sigfillset(&set);
544 /* Blocking the signals can result in undefined behaviour. */
545 sigdelset(&set, SIGSEGV);
546 sigdelset(&set, SIGFPE);
547 sigdelset(&set, SIGILL);
548 /* TODO avoid SIGBUS loss on macOS */
549 pthread_sigmask(SIG_SETMASK, &set, &oldset);
550
551 qemu_thread_args = g_new0(QemuThreadArgs, 1);
552 qemu_thread_args->name = g_strdup(name);
553 qemu_thread_args->start_routine = start_routine;
554 qemu_thread_args->arg = arg;
555
556 err = pthread_create(&thread->thread, &attr,
557 qemu_thread_start, qemu_thread_args);
558
559 if (err)
560 error_exit(err, __func__);
561
562 pthread_sigmask(SIG_SETMASK, &oldset, NULL);
563
564 pthread_attr_destroy(&attr);
565 }
566
567 void qemu_thread_get_self(QemuThread *thread)
568 {
569 thread->thread = pthread_self();
570 }
571
572 bool qemu_thread_is_self(QemuThread *thread)
573 {
574 return pthread_equal(pthread_self(), thread->thread);
575 }
576
577 void qemu_thread_exit(void *retval)
578 {
579 pthread_exit(retval);
580 }
581
582 void *qemu_thread_join(QemuThread *thread)
583 {
584 int err;
585 void *ret;
586
587 err = pthread_join(thread->thread, &ret);
588 if (err) {
589 error_exit(err, __func__);
590 }
591 return ret;
592 }