]> git.proxmox.com Git - mirror_qemu.git/blob - util/qemu-thread-posix.c
target-i386: fix "info lapic" segfault on isapc
[mirror_qemu.git] / util / qemu-thread-posix.c
1 /*
2 * Wrappers around mutex/cond/thread functions
3 *
4 * Copyright Red Hat, Inc. 2009
5 *
6 * Author:
7 * Marcelo Tosatti <mtosatti@redhat.com>
8 *
9 * This work is licensed under the terms of the GNU GPL, version 2 or later.
10 * See the COPYING file in the top-level directory.
11 *
12 */
13 #include "qemu/osdep.h"
14 #include "qemu/thread.h"
15 #include "qemu/atomic.h"
16 #include "qemu/notify.h"
17
18 static bool name_threads;
19
20 void qemu_thread_naming(bool enable)
21 {
22 name_threads = enable;
23
24 #ifndef CONFIG_THREAD_SETNAME_BYTHREAD
25 /* This is a debugging option, not fatal */
26 if (enable) {
27 fprintf(stderr, "qemu: thread naming not supported on this host\n");
28 }
29 #endif
30 }
31
32 static void error_exit(int err, const char *msg)
33 {
34 fprintf(stderr, "qemu: %s: %s\n", msg, strerror(err));
35 abort();
36 }
37
38 void qemu_mutex_init(QemuMutex *mutex)
39 {
40 int err;
41
42 err = pthread_mutex_init(&mutex->lock, NULL);
43 if (err)
44 error_exit(err, __func__);
45 }
46
47 void qemu_mutex_destroy(QemuMutex *mutex)
48 {
49 int err;
50
51 err = pthread_mutex_destroy(&mutex->lock);
52 if (err)
53 error_exit(err, __func__);
54 }
55
56 void qemu_mutex_lock(QemuMutex *mutex)
57 {
58 int err;
59
60 err = pthread_mutex_lock(&mutex->lock);
61 if (err)
62 error_exit(err, __func__);
63 }
64
65 int qemu_mutex_trylock(QemuMutex *mutex)
66 {
67 return pthread_mutex_trylock(&mutex->lock);
68 }
69
70 void qemu_mutex_unlock(QemuMutex *mutex)
71 {
72 int err;
73
74 err = pthread_mutex_unlock(&mutex->lock);
75 if (err)
76 error_exit(err, __func__);
77 }
78
79 void qemu_rec_mutex_init(QemuRecMutex *mutex)
80 {
81 int err;
82 pthread_mutexattr_t attr;
83
84 pthread_mutexattr_init(&attr);
85 pthread_mutexattr_settype(&attr, PTHREAD_MUTEX_RECURSIVE);
86 err = pthread_mutex_init(&mutex->lock, &attr);
87 pthread_mutexattr_destroy(&attr);
88 if (err) {
89 error_exit(err, __func__);
90 }
91 }
92
93 void qemu_cond_init(QemuCond *cond)
94 {
95 int err;
96
97 err = pthread_cond_init(&cond->cond, NULL);
98 if (err)
99 error_exit(err, __func__);
100 }
101
102 void qemu_cond_destroy(QemuCond *cond)
103 {
104 int err;
105
106 err = pthread_cond_destroy(&cond->cond);
107 if (err)
108 error_exit(err, __func__);
109 }
110
111 void qemu_cond_signal(QemuCond *cond)
112 {
113 int err;
114
115 err = pthread_cond_signal(&cond->cond);
116 if (err)
117 error_exit(err, __func__);
118 }
119
120 void qemu_cond_broadcast(QemuCond *cond)
121 {
122 int err;
123
124 err = pthread_cond_broadcast(&cond->cond);
125 if (err)
126 error_exit(err, __func__);
127 }
128
129 void qemu_cond_wait(QemuCond *cond, QemuMutex *mutex)
130 {
131 int err;
132
133 err = pthread_cond_wait(&cond->cond, &mutex->lock);
134 if (err)
135 error_exit(err, __func__);
136 }
137
138 void qemu_sem_init(QemuSemaphore *sem, int init)
139 {
140 int rc;
141
142 #if defined(__APPLE__) || defined(__NetBSD__)
143 rc = pthread_mutex_init(&sem->lock, NULL);
144 if (rc != 0) {
145 error_exit(rc, __func__);
146 }
147 rc = pthread_cond_init(&sem->cond, NULL);
148 if (rc != 0) {
149 error_exit(rc, __func__);
150 }
151 if (init < 0) {
152 error_exit(EINVAL, __func__);
153 }
154 sem->count = init;
155 #else
156 rc = sem_init(&sem->sem, 0, init);
157 if (rc < 0) {
158 error_exit(errno, __func__);
159 }
160 #endif
161 }
162
163 void qemu_sem_destroy(QemuSemaphore *sem)
164 {
165 int rc;
166
167 #if defined(__APPLE__) || defined(__NetBSD__)
168 rc = pthread_cond_destroy(&sem->cond);
169 if (rc < 0) {
170 error_exit(rc, __func__);
171 }
172 rc = pthread_mutex_destroy(&sem->lock);
173 if (rc < 0) {
174 error_exit(rc, __func__);
175 }
176 #else
177 rc = sem_destroy(&sem->sem);
178 if (rc < 0) {
179 error_exit(errno, __func__);
180 }
181 #endif
182 }
183
184 void qemu_sem_post(QemuSemaphore *sem)
185 {
186 int rc;
187
188 #if defined(__APPLE__) || defined(__NetBSD__)
189 pthread_mutex_lock(&sem->lock);
190 if (sem->count == UINT_MAX) {
191 rc = EINVAL;
192 } else {
193 sem->count++;
194 rc = pthread_cond_signal(&sem->cond);
195 }
196 pthread_mutex_unlock(&sem->lock);
197 if (rc != 0) {
198 error_exit(rc, __func__);
199 }
200 #else
201 rc = sem_post(&sem->sem);
202 if (rc < 0) {
203 error_exit(errno, __func__);
204 }
205 #endif
206 }
207
208 static void compute_abs_deadline(struct timespec *ts, int ms)
209 {
210 struct timeval tv;
211 gettimeofday(&tv, NULL);
212 ts->tv_nsec = tv.tv_usec * 1000 + (ms % 1000) * 1000000;
213 ts->tv_sec = tv.tv_sec + ms / 1000;
214 if (ts->tv_nsec >= 1000000000) {
215 ts->tv_sec++;
216 ts->tv_nsec -= 1000000000;
217 }
218 }
219
220 int qemu_sem_timedwait(QemuSemaphore *sem, int ms)
221 {
222 int rc;
223 struct timespec ts;
224
225 #if defined(__APPLE__) || defined(__NetBSD__)
226 rc = 0;
227 compute_abs_deadline(&ts, ms);
228 pthread_mutex_lock(&sem->lock);
229 while (sem->count == 0) {
230 rc = pthread_cond_timedwait(&sem->cond, &sem->lock, &ts);
231 if (rc == ETIMEDOUT) {
232 break;
233 }
234 if (rc != 0) {
235 error_exit(rc, __func__);
236 }
237 }
238 if (rc != ETIMEDOUT) {
239 --sem->count;
240 }
241 pthread_mutex_unlock(&sem->lock);
242 return (rc == ETIMEDOUT ? -1 : 0);
243 #else
244 if (ms <= 0) {
245 /* This is cheaper than sem_timedwait. */
246 do {
247 rc = sem_trywait(&sem->sem);
248 } while (rc == -1 && errno == EINTR);
249 if (rc == -1 && errno == EAGAIN) {
250 return -1;
251 }
252 } else {
253 compute_abs_deadline(&ts, ms);
254 do {
255 rc = sem_timedwait(&sem->sem, &ts);
256 } while (rc == -1 && errno == EINTR);
257 if (rc == -1 && errno == ETIMEDOUT) {
258 return -1;
259 }
260 }
261 if (rc < 0) {
262 error_exit(errno, __func__);
263 }
264 return 0;
265 #endif
266 }
267
268 void qemu_sem_wait(QemuSemaphore *sem)
269 {
270 int rc;
271
272 #if defined(__APPLE__) || defined(__NetBSD__)
273 pthread_mutex_lock(&sem->lock);
274 while (sem->count == 0) {
275 rc = pthread_cond_wait(&sem->cond, &sem->lock);
276 if (rc != 0) {
277 error_exit(rc, __func__);
278 }
279 }
280 --sem->count;
281 pthread_mutex_unlock(&sem->lock);
282 #else
283 do {
284 rc = sem_wait(&sem->sem);
285 } while (rc == -1 && errno == EINTR);
286 if (rc < 0) {
287 error_exit(errno, __func__);
288 }
289 #endif
290 }
291
292 #ifdef __linux__
293 #include "qemu/futex.h"
294 #else
295 static inline void qemu_futex_wake(QemuEvent *ev, int n)
296 {
297 pthread_mutex_lock(&ev->lock);
298 if (n == 1) {
299 pthread_cond_signal(&ev->cond);
300 } else {
301 pthread_cond_broadcast(&ev->cond);
302 }
303 pthread_mutex_unlock(&ev->lock);
304 }
305
306 static inline void qemu_futex_wait(QemuEvent *ev, unsigned val)
307 {
308 pthread_mutex_lock(&ev->lock);
309 if (ev->value == val) {
310 pthread_cond_wait(&ev->cond, &ev->lock);
311 }
312 pthread_mutex_unlock(&ev->lock);
313 }
314 #endif
315
316 /* Valid transitions:
317 * - free->set, when setting the event
318 * - busy->set, when setting the event, followed by qemu_futex_wake
319 * - set->free, when resetting the event
320 * - free->busy, when waiting
321 *
322 * set->busy does not happen (it can be observed from the outside but
323 * it really is set->free->busy).
324 *
325 * busy->free provably cannot happen; to enforce it, the set->free transition
326 * is done with an OR, which becomes a no-op if the event has concurrently
327 * transitioned to free or busy.
328 */
329
330 #define EV_SET 0
331 #define EV_FREE 1
332 #define EV_BUSY -1
333
334 void qemu_event_init(QemuEvent *ev, bool init)
335 {
336 #ifndef __linux__
337 pthread_mutex_init(&ev->lock, NULL);
338 pthread_cond_init(&ev->cond, NULL);
339 #endif
340
341 ev->value = (init ? EV_SET : EV_FREE);
342 }
343
344 void qemu_event_destroy(QemuEvent *ev)
345 {
346 #ifndef __linux__
347 pthread_mutex_destroy(&ev->lock);
348 pthread_cond_destroy(&ev->cond);
349 #endif
350 }
351
352 void qemu_event_set(QemuEvent *ev)
353 {
354 /* qemu_event_set has release semantics, but because it *loads*
355 * ev->value we need a full memory barrier here.
356 */
357 smp_mb();
358 if (atomic_read(&ev->value) != EV_SET) {
359 if (atomic_xchg(&ev->value, EV_SET) == EV_BUSY) {
360 /* There were waiters, wake them up. */
361 qemu_futex_wake(ev, INT_MAX);
362 }
363 }
364 }
365
366 void qemu_event_reset(QemuEvent *ev)
367 {
368 unsigned value;
369
370 value = atomic_read(&ev->value);
371 smp_mb_acquire();
372 if (value == EV_SET) {
373 /*
374 * If there was a concurrent reset (or even reset+wait),
375 * do nothing. Otherwise change EV_SET->EV_FREE.
376 */
377 atomic_or(&ev->value, EV_FREE);
378 }
379 }
380
381 void qemu_event_wait(QemuEvent *ev)
382 {
383 unsigned value;
384
385 value = atomic_read(&ev->value);
386 smp_mb_acquire();
387 if (value != EV_SET) {
388 if (value == EV_FREE) {
389 /*
390 * Leave the event reset and tell qemu_event_set that there
391 * are waiters. No need to retry, because there cannot be
392 * a concurrent busy->free transition. After the CAS, the
393 * event will be either set or busy.
394 */
395 if (atomic_cmpxchg(&ev->value, EV_FREE, EV_BUSY) == EV_SET) {
396 return;
397 }
398 }
399 qemu_futex_wait(ev, EV_BUSY);
400 }
401 }
402
403 static pthread_key_t exit_key;
404
405 union NotifierThreadData {
406 void *ptr;
407 NotifierList list;
408 };
409 QEMU_BUILD_BUG_ON(sizeof(union NotifierThreadData) != sizeof(void *));
410
411 void qemu_thread_atexit_add(Notifier *notifier)
412 {
413 union NotifierThreadData ntd;
414 ntd.ptr = pthread_getspecific(exit_key);
415 notifier_list_add(&ntd.list, notifier);
416 pthread_setspecific(exit_key, ntd.ptr);
417 }
418
419 void qemu_thread_atexit_remove(Notifier *notifier)
420 {
421 union NotifierThreadData ntd;
422 ntd.ptr = pthread_getspecific(exit_key);
423 notifier_remove(notifier);
424 pthread_setspecific(exit_key, ntd.ptr);
425 }
426
427 static void qemu_thread_atexit_run(void *arg)
428 {
429 union NotifierThreadData ntd = { .ptr = arg };
430 notifier_list_notify(&ntd.list, NULL);
431 }
432
433 static void __attribute__((constructor)) qemu_thread_atexit_init(void)
434 {
435 pthread_key_create(&exit_key, qemu_thread_atexit_run);
436 }
437
438
439 /* Attempt to set the threads name; note that this is for debug, so
440 * we're not going to fail if we can't set it.
441 */
442 static void qemu_thread_set_name(QemuThread *thread, const char *name)
443 {
444 #ifdef CONFIG_PTHREAD_SETNAME_NP
445 pthread_setname_np(thread->thread, name);
446 #endif
447 }
448
449 void qemu_thread_create(QemuThread *thread, const char *name,
450 void *(*start_routine)(void*),
451 void *arg, int mode)
452 {
453 sigset_t set, oldset;
454 int err;
455 pthread_attr_t attr;
456
457 err = pthread_attr_init(&attr);
458 if (err) {
459 error_exit(err, __func__);
460 }
461
462 /* Leave signal handling to the iothread. */
463 sigfillset(&set);
464 pthread_sigmask(SIG_SETMASK, &set, &oldset);
465 err = pthread_create(&thread->thread, &attr, start_routine, arg);
466 if (err)
467 error_exit(err, __func__);
468
469 if (name_threads) {
470 qemu_thread_set_name(thread, name);
471 }
472
473 if (mode == QEMU_THREAD_DETACHED) {
474 err = pthread_detach(thread->thread);
475 if (err) {
476 error_exit(err, __func__);
477 }
478 }
479 pthread_sigmask(SIG_SETMASK, &oldset, NULL);
480
481 pthread_attr_destroy(&attr);
482 }
483
484 void qemu_thread_get_self(QemuThread *thread)
485 {
486 thread->thread = pthread_self();
487 }
488
489 bool qemu_thread_is_self(QemuThread *thread)
490 {
491 return pthread_equal(pthread_self(), thread->thread);
492 }
493
494 void qemu_thread_exit(void *retval)
495 {
496 pthread_exit(retval);
497 }
498
499 void *qemu_thread_join(QemuThread *thread)
500 {
501 int err;
502 void *ret;
503
504 err = pthread_join(thread->thread, &ret);
505 if (err) {
506 error_exit(err, __func__);
507 }
508 return ret;
509 }