1 //===-- tsan_interceptors.cc ----------------------------------------------===//
3 // The LLVM Compiler Infrastructure
5 // This file is distributed under the University of Illinois Open Source
6 // License. See LICENSE.TXT for details.
8 //===----------------------------------------------------------------------===//
10 // This file is a part of ThreadSanitizer (TSan), a race detector.
12 // FIXME: move as many interceptors as possible into
13 // sanitizer_common/sanitizer_common_interceptors.inc
14 //===----------------------------------------------------------------------===//
16 #include "sanitizer_common/sanitizer_atomic.h"
17 #include "sanitizer_common/sanitizer_libc.h"
18 #include "sanitizer_common/sanitizer_linux.h"
19 #include "sanitizer_common/sanitizer_platform_limits_posix.h"
20 #include "sanitizer_common/sanitizer_placement_new.h"
21 #include "sanitizer_common/sanitizer_stacktrace.h"
22 #include "sanitizer_common/sanitizer_tls_get_addr.h"
23 #include "interception/interception.h"
24 #include "tsan_interceptors.h"
25 #include "tsan_interface.h"
26 #include "tsan_platform.h"
27 #include "tsan_suppressions.h"
29 #include "tsan_mman.h"
33 #include "sanitizer_common/sanitizer_posix.h"
36 using namespace __tsan
; // NOLINT
38 #if SANITIZER_FREEBSD || SANITIZER_MAC
39 #define __errno_location __error
40 #define stdout __stdoutp
41 #define stderr __stderrp
45 #define __errno_location __errno
49 #if SANITIZER_LINUX || SANITIZER_FREEBSD
50 #define PTHREAD_CREATE_DETACHED 1
52 #define PTHREAD_CREATE_DETACHED 2
57 const int kSigCount
= 129;
59 const int kSigCount
= 65;
63 // The size is determined by looking at sizeof of real siginfo_t on linux.
64 u64 opaque
[128 / sizeof(u64
)];
69 u64 opaque
[768 / sizeof(u64
) + 1];
73 // The size is determined by looking at sizeof of real ucontext_t on linux.
74 u64 opaque
[936 / sizeof(u64
) + 1];
78 #if defined(__x86_64__) || defined(__mips__) || SANITIZER_PPC64V1
79 #define PTHREAD_ABI_BASE "GLIBC_2.3.2"
80 #elif defined(__aarch64__) || SANITIZER_PPC64V2
81 #define PTHREAD_ABI_BASE "GLIBC_2.17"
84 extern "C" int pthread_attr_init(void *attr
);
85 extern "C" int pthread_attr_destroy(void *attr
);
86 DECLARE_REAL(int, pthread_attr_getdetachstate
, void *, void *)
87 extern "C" int pthread_attr_setstacksize(void *attr
, uptr stacksize
);
88 extern "C" int pthread_key_create(unsigned *key
, void (*destructor
)(void* v
));
89 extern "C" int pthread_setspecific(unsigned key
, const void *v
);
90 DECLARE_REAL(int, pthread_mutexattr_gettype
, void *, void *)
91 DECLARE_REAL(int, fflush
, __sanitizer_FILE
*fp
)
92 DECLARE_REAL_AND_INTERCEPTOR(void *, malloc
, uptr size
)
93 DECLARE_REAL_AND_INTERCEPTOR(void, free
, void *ptr
)
94 extern "C" void *pthread_self();
95 extern "C" void _exit(int status
);
96 extern "C" int *__errno_location();
97 extern "C" int fileno_unlocked(void *stream
);
98 extern "C" int dirfd(void *dirp
);
99 #if !SANITIZER_FREEBSD && !SANITIZER_ANDROID
100 extern "C" int mallopt(int param
, int value
);
102 extern __sanitizer_FILE
*stdout
, *stderr
;
103 #if !SANITIZER_FREEBSD && !SANITIZER_MAC
104 const int PTHREAD_MUTEX_RECURSIVE
= 1;
105 const int PTHREAD_MUTEX_RECURSIVE_NP
= 1;
107 const int PTHREAD_MUTEX_RECURSIVE
= 2;
108 const int PTHREAD_MUTEX_RECURSIVE_NP
= 2;
110 const int EINVAL
= 22;
111 const int EBUSY
= 16;
112 const int EOWNERDEAD
= 130;
113 #if !SANITIZER_FREEBSD && !SANITIZER_MAC
114 const int EPOLL_CTL_ADD
= 1;
116 const int SIGILL
= 4;
117 const int SIGABRT
= 6;
118 const int SIGFPE
= 8;
119 const int SIGSEGV
= 11;
120 const int SIGPIPE
= 13;
121 const int SIGTERM
= 15;
122 #if defined(__mips__) || SANITIZER_FREEBSD || SANITIZER_MAC
123 const int SIGBUS
= 10;
124 const int SIGSYS
= 12;
126 const int SIGBUS
= 7;
127 const int SIGSYS
= 31;
129 void *const MAP_FAILED
= (void*)-1;
131 const int PTHREAD_BARRIER_SERIAL_THREAD
= -1;
133 const int MAP_FIXED
= 0x10;
134 typedef long long_t
; // NOLINT
136 // From /usr/include/unistd.h
137 # define F_ULOCK 0 /* Unlock a previously locked region. */
138 # define F_LOCK 1 /* Lock a region for exclusive use. */
139 # define F_TLOCK 2 /* Test and lock a region for exclusive use. */
140 # define F_TEST 3 /* Test a region for other processes locks. */
142 #define errno (*__errno_location())
144 typedef void (*sighandler_t
)(int sig
);
145 typedef void (*sigactionhandler_t
)(int sig
, my_siginfo_t
*siginfo
, void *uctx
);
147 #if SANITIZER_ANDROID
151 sighandler_t sa_handler
;
152 sigactionhandler_t sa_sigaction
;
154 __sanitizer_sigset_t sa_mask
;
155 void (*sa_restorer
)();
163 sighandler_t sa_handler
;
164 sigactionhandler_t sa_sigaction
;
166 #if SANITIZER_FREEBSD
168 __sanitizer_sigset_t sa_mask
;
170 __sanitizer_sigset_t sa_mask
;
173 __sanitizer_sigset_t sa_mask
;
177 void (*sa_restorer
)();
182 const sighandler_t SIG_DFL
= (sighandler_t
)0;
183 const sighandler_t SIG_IGN
= (sighandler_t
)1;
184 const sighandler_t SIG_ERR
= (sighandler_t
)-1;
185 #if SANITIZER_FREEBSD || SANITIZER_MAC
186 const int SA_SIGINFO
= 0x40;
187 const int SIG_SETMASK
= 3;
188 #elif defined(__mips__)
189 const int SA_SIGINFO
= 8;
190 const int SIG_SETMASK
= 3;
192 const int SA_SIGINFO
= 4;
193 const int SIG_SETMASK
= 2;
196 #define COMMON_INTERCEPTOR_NOTHING_IS_INITIALIZED \
197 (!cur_thread()->is_inited)
199 static sigaction_t sigactions
[kSigCount
];
205 my_siginfo_t siginfo
;
209 struct ThreadSignalContext
{
211 atomic_uintptr_t in_blocking_func
;
212 atomic_uintptr_t have_pending_signals
;
213 SignalDesc pending_signals
[kSigCount
];
214 // emptyset and oldset are too big for stack.
215 __sanitizer_sigset_t emptyset
;
216 __sanitizer_sigset_t oldset
;
219 // The object is 64-byte aligned, because we want hot data to be located in
220 // a single cache line if possible (it's accessed in every interceptor).
221 static ALIGNED(64) char libignore_placeholder
[sizeof(LibIgnore
)];
222 static LibIgnore
*libignore() {
223 return reinterpret_cast<LibIgnore
*>(&libignore_placeholder
[0]);
226 void InitializeLibIgnore() {
227 const SuppressionContext
&supp
= *Suppressions();
228 const uptr n
= supp
.SuppressionCount();
229 for (uptr i
= 0; i
< n
; i
++) {
230 const Suppression
*s
= supp
.SuppressionAt(i
);
231 if (0 == internal_strcmp(s
->type
, kSuppressionLib
))
232 libignore()->AddIgnoredLibrary(s
->templ
);
234 if (flags()->ignore_noninstrumented_modules
)
235 libignore()->IgnoreNoninstrumentedModules(true);
236 libignore()->OnLibraryLoaded(0);
239 } // namespace __tsan
241 static ThreadSignalContext
*SigCtx(ThreadState
*thr
) {
242 ThreadSignalContext
*ctx
= (ThreadSignalContext
*)thr
->signal_ctx
;
243 if (ctx
== 0 && !thr
->is_dead
) {
244 ctx
= (ThreadSignalContext
*)MmapOrDie(sizeof(*ctx
), "ThreadSignalContext");
245 MemoryResetRange(thr
, (uptr
)&SigCtx
, (uptr
)ctx
, sizeof(*ctx
));
246 thr
->signal_ctx
= ctx
;
252 static unsigned g_thread_finalize_key
;
255 ScopedInterceptor::ScopedInterceptor(ThreadState
*thr
, const char *fname
,
257 : thr_(thr
), pc_(pc
), in_ignored_lib_(false), ignoring_(false) {
259 if (!thr_
->is_inited
) return;
260 if (!thr_
->ignore_interceptors
) FuncEntry(thr
, pc
);
261 DPrintf("#%d: intercept %s()\n", thr_
->tid
, fname
);
263 !thr_
->in_ignored_lib
&& (flags()->ignore_interceptors_accesses
||
264 libignore()->IsIgnored(pc
, &in_ignored_lib_
));
268 ScopedInterceptor::~ScopedInterceptor() {
269 if (!thr_
->is_inited
) return;
271 if (!thr_
->ignore_interceptors
) {
272 ProcessPendingSignals(thr_
);
278 void ScopedInterceptor::EnableIgnores() {
280 ThreadIgnoreBegin(thr_
, pc_
);
281 if (in_ignored_lib_
) {
282 DCHECK(!thr_
->in_ignored_lib
);
283 thr_
->in_ignored_lib
= true;
288 void ScopedInterceptor::DisableIgnores() {
290 ThreadIgnoreEnd(thr_
, pc_
);
291 if (in_ignored_lib_
) {
292 DCHECK(thr_
->in_ignored_lib
);
293 thr_
->in_ignored_lib
= false;
298 #define TSAN_INTERCEPT(func) INTERCEPT_FUNCTION(func)
299 #if SANITIZER_FREEBSD
300 # define TSAN_INTERCEPT_VER(func, ver) INTERCEPT_FUNCTION(func)
302 # define TSAN_INTERCEPT_VER(func, ver) INTERCEPT_FUNCTION_VER(func, ver)
305 #define READ_STRING_OF_LEN(thr, pc, s, len, n) \
306 MemoryAccessRange((thr), (pc), (uptr)(s), \
307 common_flags()->strict_string_checks ? (len) + 1 : (n), false)
309 #define READ_STRING(thr, pc, s, n) \
310 READ_STRING_OF_LEN((thr), (pc), (s), internal_strlen(s), (n))
312 #define BLOCK_REAL(name) (BlockingCall(thr), REAL(name))
314 struct BlockingCall
{
315 explicit BlockingCall(ThreadState
*thr
)
319 atomic_store(&ctx
->in_blocking_func
, 1, memory_order_relaxed
);
320 if (atomic_load(&ctx
->have_pending_signals
, memory_order_relaxed
) == 0)
322 atomic_store(&ctx
->in_blocking_func
, 0, memory_order_relaxed
);
323 ProcessPendingSignals(thr
);
325 // When we are in a "blocking call", we process signals asynchronously
326 // (right when they arrive). In this context we do not expect to be
327 // executing any user/runtime code. The known interceptor sequence when
328 // this is not true is: pthread_join -> munmap(stack). It's fine
329 // to ignore munmap in this case -- we handle stack shadow separately.
330 thr
->ignore_interceptors
++;
334 thr
->ignore_interceptors
--;
335 atomic_store(&ctx
->in_blocking_func
, 0, memory_order_relaxed
);
339 ThreadSignalContext
*ctx
;
342 TSAN_INTERCEPTOR(unsigned, sleep
, unsigned sec
) {
343 SCOPED_TSAN_INTERCEPTOR(sleep
, sec
);
344 unsigned res
= BLOCK_REAL(sleep
)(sec
);
349 TSAN_INTERCEPTOR(int, usleep
, long_t usec
) {
350 SCOPED_TSAN_INTERCEPTOR(usleep
, usec
);
351 int res
= BLOCK_REAL(usleep
)(usec
);
356 TSAN_INTERCEPTOR(int, nanosleep
, void *req
, void *rem
) {
357 SCOPED_TSAN_INTERCEPTOR(nanosleep
, req
, rem
);
358 int res
= BLOCK_REAL(nanosleep
)(req
, rem
);
363 // The sole reason tsan wraps atexit callbacks is to establish synchronization
364 // between callback setup and callback execution.
370 static void at_exit_wrapper(void *arg
) {
371 ThreadState
*thr
= cur_thread();
373 Acquire(thr
, pc
, (uptr
)arg
);
374 AtExitCtx
*ctx
= (AtExitCtx
*)arg
;
375 ((void(*)(void *arg
))ctx
->f
)(ctx
->arg
);
379 static int setup_at_exit_wrapper(ThreadState
*thr
, uptr pc
, void(*f
)(),
380 void *arg
, void *dso
);
382 #if !SANITIZER_ANDROID
383 TSAN_INTERCEPTOR(int, atexit
, void (*f
)()) {
384 if (cur_thread()->in_symbolizer
)
386 // We want to setup the atexit callback even if we are in ignored lib
388 SCOPED_INTERCEPTOR_RAW(atexit
, f
);
389 return setup_at_exit_wrapper(thr
, pc
, (void(*)())f
, 0, 0);
393 TSAN_INTERCEPTOR(int, __cxa_atexit
, void (*f
)(void *a
), void *arg
, void *dso
) {
394 if (cur_thread()->in_symbolizer
)
396 SCOPED_TSAN_INTERCEPTOR(__cxa_atexit
, f
, arg
, dso
);
397 return setup_at_exit_wrapper(thr
, pc
, (void(*)())f
, arg
, dso
);
400 static int setup_at_exit_wrapper(ThreadState
*thr
, uptr pc
, void(*f
)(),
401 void *arg
, void *dso
) {
402 AtExitCtx
*ctx
= (AtExitCtx
*)InternalAlloc(sizeof(AtExitCtx
));
405 Release(thr
, pc
, (uptr
)ctx
);
406 // Memory allocation in __cxa_atexit will race with free during exit,
407 // because we do not see synchronization around atexit callback list.
408 ThreadIgnoreBegin(thr
, pc
);
409 int res
= REAL(__cxa_atexit
)(at_exit_wrapper
, ctx
, dso
);
410 ThreadIgnoreEnd(thr
, pc
);
415 static void on_exit_wrapper(int status
, void *arg
) {
416 ThreadState
*thr
= cur_thread();
418 Acquire(thr
, pc
, (uptr
)arg
);
419 AtExitCtx
*ctx
= (AtExitCtx
*)arg
;
420 ((void(*)(int status
, void *arg
))ctx
->f
)(status
, ctx
->arg
);
424 TSAN_INTERCEPTOR(int, on_exit
, void(*f
)(int, void*), void *arg
) {
425 if (cur_thread()->in_symbolizer
)
427 SCOPED_TSAN_INTERCEPTOR(on_exit
, f
, arg
);
428 AtExitCtx
*ctx
= (AtExitCtx
*)InternalAlloc(sizeof(AtExitCtx
));
429 ctx
->f
= (void(*)())f
;
431 Release(thr
, pc
, (uptr
)ctx
);
432 // Memory allocation in __cxa_atexit will race with free during exit,
433 // because we do not see synchronization around atexit callback list.
434 ThreadIgnoreBegin(thr
, pc
);
435 int res
= REAL(on_exit
)(on_exit_wrapper
, ctx
);
436 ThreadIgnoreEnd(thr
, pc
);
442 static void JmpBufGarbageCollect(ThreadState
*thr
, uptr sp
) {
443 for (uptr i
= 0; i
< thr
->jmp_bufs
.Size(); i
++) {
444 JmpBuf
*buf
= &thr
->jmp_bufs
[i
];
446 uptr sz
= thr
->jmp_bufs
.Size();
447 internal_memcpy(buf
, &thr
->jmp_bufs
[sz
- 1], sizeof(*buf
));
448 thr
->jmp_bufs
.PopBack();
454 static void SetJmp(ThreadState
*thr
, uptr sp
, uptr mangled_sp
) {
455 if (!thr
->is_inited
) // called from libc guts during bootstrap
458 JmpBufGarbageCollect(thr
, sp
);
460 JmpBuf
*buf
= thr
->jmp_bufs
.PushBack();
462 buf
->mangled_sp
= mangled_sp
;
463 buf
->shadow_stack_pos
= thr
->shadow_stack_pos
;
464 ThreadSignalContext
*sctx
= SigCtx(thr
);
465 buf
->int_signal_send
= sctx
? sctx
->int_signal_send
: 0;
466 buf
->in_blocking_func
= sctx
?
467 atomic_load(&sctx
->in_blocking_func
, memory_order_relaxed
) :
469 buf
->in_signal_handler
= atomic_load(&thr
->in_signal_handler
,
470 memory_order_relaxed
);
473 static void LongJmp(ThreadState
*thr
, uptr
*env
) {
475 uptr mangled_sp
= env
[0];
476 #elif SANITIZER_FREEBSD || SANITIZER_MAC
477 uptr mangled_sp
= env
[2];
478 #elif defined(SANITIZER_LINUX)
480 uptr mangled_sp
= env
[13];
481 # elif defined(__mips64)
482 uptr mangled_sp
= env
[1];
484 uptr mangled_sp
= env
[6];
487 // Find the saved buf by mangled_sp.
488 for (uptr i
= 0; i
< thr
->jmp_bufs
.Size(); i
++) {
489 JmpBuf
*buf
= &thr
->jmp_bufs
[i
];
490 if (buf
->mangled_sp
== mangled_sp
) {
491 CHECK_GE(thr
->shadow_stack_pos
, buf
->shadow_stack_pos
);
493 while (thr
->shadow_stack_pos
> buf
->shadow_stack_pos
)
495 ThreadSignalContext
*sctx
= SigCtx(thr
);
497 sctx
->int_signal_send
= buf
->int_signal_send
;
498 atomic_store(&sctx
->in_blocking_func
, buf
->in_blocking_func
,
499 memory_order_relaxed
);
501 atomic_store(&thr
->in_signal_handler
, buf
->in_signal_handler
,
502 memory_order_relaxed
);
503 JmpBufGarbageCollect(thr
, buf
->sp
- 1); // do not collect buf->sp
507 Printf("ThreadSanitizer: can't find longjmp buf\n");
511 // FIXME: put everything below into a common extern "C" block?
512 extern "C" void __tsan_setjmp(uptr sp
, uptr mangled_sp
) {
513 SetJmp(cur_thread(), sp
, mangled_sp
);
517 TSAN_INTERCEPTOR(int, setjmp
, void *env
);
518 TSAN_INTERCEPTOR(int, _setjmp
, void *env
);
519 TSAN_INTERCEPTOR(int, sigsetjmp
, void *env
);
520 #else // SANITIZER_MAC
521 // Not called. Merely to satisfy TSAN_INTERCEPT().
522 extern "C" SANITIZER_INTERFACE_ATTRIBUTE
523 int __interceptor_setjmp(void *env
);
524 extern "C" int __interceptor_setjmp(void *env
) {
529 // FIXME: any reason to have a separate declaration?
530 extern "C" SANITIZER_INTERFACE_ATTRIBUTE
531 int __interceptor__setjmp(void *env
);
532 extern "C" int __interceptor__setjmp(void *env
) {
537 extern "C" SANITIZER_INTERFACE_ATTRIBUTE
538 int __interceptor_sigsetjmp(void *env
);
539 extern "C" int __interceptor_sigsetjmp(void *env
) {
544 extern "C" SANITIZER_INTERFACE_ATTRIBUTE
545 int __interceptor___sigsetjmp(void *env
);
546 extern "C" int __interceptor___sigsetjmp(void *env
) {
551 extern "C" int setjmp(void *env
);
552 extern "C" int _setjmp(void *env
);
553 extern "C" int sigsetjmp(void *env
);
554 extern "C" int __sigsetjmp(void *env
);
555 DEFINE_REAL(int, setjmp
, void *env
)
556 DEFINE_REAL(int, _setjmp
, void *env
)
557 DEFINE_REAL(int, sigsetjmp
, void *env
)
558 DEFINE_REAL(int, __sigsetjmp
, void *env
)
559 #endif // SANITIZER_MAC
561 TSAN_INTERCEPTOR(void, longjmp
, uptr
*env
, int val
) {
562 // Note: if we call REAL(longjmp) in the context of ScopedInterceptor,
563 // bad things will happen. We will jump over ScopedInterceptor dtor and can
564 // leave thr->in_ignored_lib set.
566 SCOPED_INTERCEPTOR_RAW(longjmp
, env
, val
);
568 LongJmp(cur_thread(), env
);
569 REAL(longjmp
)(env
, val
);
572 TSAN_INTERCEPTOR(void, siglongjmp
, uptr
*env
, int val
) {
574 SCOPED_INTERCEPTOR_RAW(siglongjmp
, env
, val
);
576 LongJmp(cur_thread(), env
);
577 REAL(siglongjmp
)(env
, val
);
581 TSAN_INTERCEPTOR(void*, malloc
, uptr size
) {
582 if (cur_thread()->in_symbolizer
)
583 return InternalAlloc(size
);
586 SCOPED_INTERCEPTOR_RAW(malloc
, size
);
587 p
= user_alloc(thr
, pc
, size
);
589 invoke_malloc_hook(p
, size
);
593 TSAN_INTERCEPTOR(void*, __libc_memalign
, uptr align
, uptr sz
) {
594 SCOPED_TSAN_INTERCEPTOR(__libc_memalign
, align
, sz
);
595 return user_alloc(thr
, pc
, sz
, align
);
598 TSAN_INTERCEPTOR(void*, calloc
, uptr size
, uptr n
) {
599 if (cur_thread()->in_symbolizer
)
600 return InternalCalloc(size
, n
);
603 SCOPED_INTERCEPTOR_RAW(calloc
, size
, n
);
604 p
= user_calloc(thr
, pc
, size
, n
);
606 invoke_malloc_hook(p
, n
* size
);
610 TSAN_INTERCEPTOR(void*, realloc
, void *p
, uptr size
) {
611 if (cur_thread()->in_symbolizer
)
612 return InternalRealloc(p
, size
);
616 SCOPED_INTERCEPTOR_RAW(realloc
, p
, size
);
617 p
= user_realloc(thr
, pc
, p
, size
);
619 invoke_malloc_hook(p
, size
);
623 TSAN_INTERCEPTOR(void, free
, void *p
) {
626 if (cur_thread()->in_symbolizer
)
627 return InternalFree(p
);
629 SCOPED_INTERCEPTOR_RAW(free
, p
);
630 user_free(thr
, pc
, p
);
633 TSAN_INTERCEPTOR(void, cfree
, void *p
) {
636 if (cur_thread()->in_symbolizer
)
637 return InternalFree(p
);
639 SCOPED_INTERCEPTOR_RAW(cfree
, p
);
640 user_free(thr
, pc
, p
);
643 TSAN_INTERCEPTOR(uptr
, malloc_usable_size
, void *p
) {
644 SCOPED_INTERCEPTOR_RAW(malloc_usable_size
, p
);
645 return user_alloc_usable_size(p
);
649 TSAN_INTERCEPTOR(char*, strcpy
, char *dst
, const char *src
) { // NOLINT
650 SCOPED_TSAN_INTERCEPTOR(strcpy
, dst
, src
); // NOLINT
651 uptr srclen
= internal_strlen(src
);
652 MemoryAccessRange(thr
, pc
, (uptr
)dst
, srclen
+ 1, true);
653 MemoryAccessRange(thr
, pc
, (uptr
)src
, srclen
+ 1, false);
654 return REAL(strcpy
)(dst
, src
); // NOLINT
657 TSAN_INTERCEPTOR(char*, strncpy
, char *dst
, char *src
, uptr n
) {
658 SCOPED_TSAN_INTERCEPTOR(strncpy
, dst
, src
, n
);
659 uptr srclen
= internal_strnlen(src
, n
);
660 MemoryAccessRange(thr
, pc
, (uptr
)dst
, n
, true);
661 MemoryAccessRange(thr
, pc
, (uptr
)src
, min(srclen
+ 1, n
), false);
662 return REAL(strncpy
)(dst
, src
, n
);
665 TSAN_INTERCEPTOR(char*, strdup
, const char *str
) {
666 SCOPED_TSAN_INTERCEPTOR(strdup
, str
);
667 // strdup will call malloc, so no instrumentation is required here.
668 return REAL(strdup
)(str
);
671 static bool fix_mmap_addr(void **addr
, long_t sz
, int flags
) {
673 if (!IsAppMem((uptr
)*addr
) || !IsAppMem((uptr
)*addr
+ sz
- 1)) {
674 if (flags
& MAP_FIXED
) {
685 TSAN_INTERCEPTOR(void *, mmap
, void *addr
, SIZE_T sz
, int prot
, int flags
,
687 SCOPED_TSAN_INTERCEPTOR(mmap
, addr
, sz
, prot
, flags
, fd
, off
);
688 if (!fix_mmap_addr(&addr
, sz
, flags
))
690 void *res
= REAL(mmap
)(addr
, sz
, prot
, flags
, fd
, off
);
691 if (res
!= MAP_FAILED
) {
693 FdAccess(thr
, pc
, fd
);
695 if (thr
->ignore_reads_and_writes
== 0)
696 MemoryRangeImitateWrite(thr
, pc
, (uptr
)res
, sz
);
698 MemoryResetRange(thr
, pc
, (uptr
)res
, sz
);
704 TSAN_INTERCEPTOR(void *, mmap64
, void *addr
, SIZE_T sz
, int prot
, int flags
,
705 int fd
, OFF64_T off
) {
706 SCOPED_TSAN_INTERCEPTOR(mmap64
, addr
, sz
, prot
, flags
, fd
, off
);
707 if (!fix_mmap_addr(&addr
, sz
, flags
))
709 void *res
= REAL(mmap64
)(addr
, sz
, prot
, flags
, fd
, off
);
710 if (res
!= MAP_FAILED
) {
712 FdAccess(thr
, pc
, fd
);
714 if (thr
->ignore_reads_and_writes
== 0)
715 MemoryRangeImitateWrite(thr
, pc
, (uptr
)res
, sz
);
717 MemoryResetRange(thr
, pc
, (uptr
)res
, sz
);
721 #define TSAN_MAYBE_INTERCEPT_MMAP64 TSAN_INTERCEPT(mmap64)
723 #define TSAN_MAYBE_INTERCEPT_MMAP64
726 TSAN_INTERCEPTOR(int, munmap
, void *addr
, long_t sz
) {
727 SCOPED_TSAN_INTERCEPTOR(munmap
, addr
, sz
);
729 // If sz == 0, munmap will return EINVAL and don't unmap any memory.
730 DontNeedShadowFor((uptr
)addr
, sz
);
731 ScopedGlobalProcessor sgp
;
732 ctx
->metamap
.ResetRange(thr
->proc(), (uptr
)addr
, (uptr
)sz
);
734 int res
= REAL(munmap
)(addr
, sz
);
739 TSAN_INTERCEPTOR(void*, memalign
, uptr align
, uptr sz
) {
740 SCOPED_INTERCEPTOR_RAW(memalign
, align
, sz
);
741 return user_alloc(thr
, pc
, sz
, align
);
743 #define TSAN_MAYBE_INTERCEPT_MEMALIGN TSAN_INTERCEPT(memalign)
745 #define TSAN_MAYBE_INTERCEPT_MEMALIGN
749 TSAN_INTERCEPTOR(void*, aligned_alloc
, uptr align
, uptr sz
) {
750 SCOPED_INTERCEPTOR_RAW(memalign
, align
, sz
);
751 return user_alloc(thr
, pc
, sz
, align
);
754 TSAN_INTERCEPTOR(void*, valloc
, uptr sz
) {
755 SCOPED_INTERCEPTOR_RAW(valloc
, sz
);
756 return user_alloc(thr
, pc
, sz
, GetPageSizeCached());
761 TSAN_INTERCEPTOR(void*, pvalloc
, uptr sz
) {
762 SCOPED_INTERCEPTOR_RAW(pvalloc
, sz
);
763 sz
= RoundUp(sz
, GetPageSizeCached());
764 return user_alloc(thr
, pc
, sz
, GetPageSizeCached());
766 #define TSAN_MAYBE_INTERCEPT_PVALLOC TSAN_INTERCEPT(pvalloc)
768 #define TSAN_MAYBE_INTERCEPT_PVALLOC
772 TSAN_INTERCEPTOR(int, posix_memalign
, void **memptr
, uptr align
, uptr sz
) {
773 SCOPED_INTERCEPTOR_RAW(posix_memalign
, memptr
, align
, sz
);
774 *memptr
= user_alloc(thr
, pc
, sz
, align
);
779 // __cxa_guard_acquire and friends need to be intercepted in a special way -
780 // regular interceptors will break statically-linked libstdc++. Linux
781 // interceptors are especially defined as weak functions (so that they don't
782 // cause link errors when user defines them as well). So they silently
783 // auto-disable themselves when such symbol is already present in the binary. If
784 // we link libstdc++ statically, it will bring own __cxa_guard_acquire which
785 // will silently replace our interceptor. That's why on Linux we simply export
786 // these interceptors with INTERFACE_ATTRIBUTE.
787 // On OS X, we don't support statically linking, so we just use a regular
790 #define STDCXX_INTERCEPTOR TSAN_INTERCEPTOR
792 #define STDCXX_INTERCEPTOR(rettype, name, ...) \
793 extern "C" rettype INTERFACE_ATTRIBUTE name(__VA_ARGS__)
796 // Used in thread-safe function static initialization.
797 STDCXX_INTERCEPTOR(int, __cxa_guard_acquire
, atomic_uint32_t
*g
) {
798 SCOPED_INTERCEPTOR_RAW(__cxa_guard_acquire
, g
);
800 u32 cmp
= atomic_load(g
, memory_order_acquire
);
802 if (atomic_compare_exchange_strong(g
, &cmp
, 1<<16, memory_order_relaxed
))
804 } else if (cmp
== 1) {
805 Acquire(thr
, pc
, (uptr
)g
);
808 internal_sched_yield();
813 STDCXX_INTERCEPTOR(void, __cxa_guard_release
, atomic_uint32_t
*g
) {
814 SCOPED_INTERCEPTOR_RAW(__cxa_guard_release
, g
);
815 Release(thr
, pc
, (uptr
)g
);
816 atomic_store(g
, 1, memory_order_release
);
819 STDCXX_INTERCEPTOR(void, __cxa_guard_abort
, atomic_uint32_t
*g
) {
820 SCOPED_INTERCEPTOR_RAW(__cxa_guard_abort
, g
);
821 atomic_store(g
, 0, memory_order_relaxed
);
825 void DestroyThreadState() {
826 ThreadState
*thr
= cur_thread();
827 Processor
*proc
= thr
->proc();
829 ProcUnwire(proc
, thr
);
831 ThreadSignalContext
*sctx
= thr
->signal_ctx
;
834 UnmapOrDie(sctx
, sizeof(*sctx
));
837 cur_thread_finalize();
839 } // namespace __tsan
842 static void thread_finalize(void *v
) {
845 if (pthread_setspecific(g_thread_finalize_key
, (void*)(iter
- 1))) {
846 Printf("ThreadSanitizer: failed to set thread key\n");
851 DestroyThreadState();
857 void* (*callback
)(void *arg
);
859 atomic_uintptr_t tid
;
862 extern "C" void *__tsan_thread_start_func(void *arg
) {
863 ThreadParam
*p
= (ThreadParam
*)arg
;
864 void* (*callback
)(void *arg
) = p
->callback
;
865 void *param
= p
->param
;
868 ThreadState
*thr
= cur_thread();
869 // Thread-local state is not initialized yet.
870 ScopedIgnoreInterceptors ignore
;
872 ThreadIgnoreBegin(thr
, 0);
873 if (pthread_setspecific(g_thread_finalize_key
,
874 (void *)GetPthreadDestructorIterations())) {
875 Printf("ThreadSanitizer: failed to set thread key\n");
878 ThreadIgnoreEnd(thr
, 0);
880 while ((tid
= atomic_load(&p
->tid
, memory_order_acquire
)) == 0)
881 internal_sched_yield();
882 Processor
*proc
= ProcCreate();
884 ThreadStart(thr
, tid
, GetTid());
885 atomic_store(&p
->tid
, 0, memory_order_release
);
887 void *res
= callback(param
);
888 // Prevent the callback from being tail called,
889 // it mixes up stack traces.
890 volatile int foo
= 42;
895 TSAN_INTERCEPTOR(int, pthread_create
,
896 void *th
, void *attr
, void *(*callback
)(void*), void * param
) {
897 SCOPED_INTERCEPTOR_RAW(pthread_create
, th
, attr
, callback
, param
);
898 if (ctx
->after_multithreaded_fork
) {
899 if (flags()->die_after_fork
) {
900 Report("ThreadSanitizer: starting new threads after multi-threaded "
901 "fork is not supported. Dying (set die_after_fork=0 to override)\n");
904 VPrintf(1, "ThreadSanitizer: starting new threads after multi-threaded "
905 "fork is not supported (pid %d). Continuing because of "
906 "die_after_fork=0, but you are on your own\n", internal_getpid());
909 __sanitizer_pthread_attr_t myattr
;
911 pthread_attr_init(&myattr
);
915 REAL(pthread_attr_getdetachstate
)(attr
, &detached
);
916 AdjustStackSize(attr
);
919 p
.callback
= callback
;
921 atomic_store(&p
.tid
, 0, memory_order_relaxed
);
924 // Otherwise we see false positives in pthread stack manipulation.
925 ScopedIgnoreInterceptors ignore
;
926 ThreadIgnoreBegin(thr
, pc
);
927 res
= REAL(pthread_create
)(th
, attr
, __tsan_thread_start_func
, &p
);
928 ThreadIgnoreEnd(thr
, pc
);
931 int tid
= ThreadCreate(thr
, pc
, *(uptr
*)th
,
932 detached
== PTHREAD_CREATE_DETACHED
);
934 // Synchronization on p.tid serves two purposes:
935 // 1. ThreadCreate must finish before the new thread starts.
936 // Otherwise the new thread can call pthread_detach, but the pthread_t
937 // identifier is not yet registered in ThreadRegistry by ThreadCreate.
938 // 2. ThreadStart must finish before this thread continues.
939 // Otherwise, this thread can call pthread_detach and reset thr->sync
940 // before the new thread got a chance to acquire from it in ThreadStart.
941 atomic_store(&p
.tid
, tid
, memory_order_release
);
942 while (atomic_load(&p
.tid
, memory_order_acquire
) != 0)
943 internal_sched_yield();
946 pthread_attr_destroy(&myattr
);
950 TSAN_INTERCEPTOR(int, pthread_join
, void *th
, void **ret
) {
951 SCOPED_INTERCEPTOR_RAW(pthread_join
, th
, ret
);
952 int tid
= ThreadTid(thr
, pc
, (uptr
)th
);
953 ThreadIgnoreBegin(thr
, pc
);
954 int res
= BLOCK_REAL(pthread_join
)(th
, ret
);
955 ThreadIgnoreEnd(thr
, pc
);
957 ThreadJoin(thr
, pc
, tid
);
962 DEFINE_REAL_PTHREAD_FUNCTIONS
964 TSAN_INTERCEPTOR(int, pthread_detach
, void *th
) {
965 SCOPED_TSAN_INTERCEPTOR(pthread_detach
, th
);
966 int tid
= ThreadTid(thr
, pc
, (uptr
)th
);
967 int res
= REAL(pthread_detach
)(th
);
969 ThreadDetach(thr
, pc
, tid
);
975 // NPTL implementation of pthread_cond has 2 versions (2.2.5 and 2.3.2).
976 // pthread_cond_t has different size in the different versions.
977 // If call new REAL functions for old pthread_cond_t, they will corrupt memory
978 // after pthread_cond_t (old cond is smaller).
979 // If we call old REAL functions for new pthread_cond_t, we will lose some
980 // functionality (e.g. old functions do not support waiting against
982 // Proper handling would require to have 2 versions of interceptors as well.
983 // But this is messy, in particular requires linker scripts when sanitizer
984 // runtime is linked into a shared library.
985 // Instead we assume we don't have dynamic libraries built against old
986 // pthread (2.2.5 is dated by 2002). And provide legacy_pthread_cond flag
987 // that allows to work with old libraries (but this mode does not support
988 // some features, e.g. pthread_condattr_getpshared).
989 static void *init_cond(void *c
, bool force
= false) {
990 // sizeof(pthread_cond_t) >= sizeof(uptr) in both versions.
991 // So we allocate additional memory on the side large enough to hold
992 // any pthread_cond_t object. Always call new REAL functions, but pass
993 // the aux object to them.
994 // Note: the code assumes that PTHREAD_COND_INITIALIZER initializes
995 // first word of pthread_cond_t to zero.
996 // It's all relevant only for linux.
997 if (!common_flags()->legacy_pthread_cond
)
999 atomic_uintptr_t
*p
= (atomic_uintptr_t
*)c
;
1000 uptr cond
= atomic_load(p
, memory_order_acquire
);
1001 if (!force
&& cond
!= 0)
1003 void *newcond
= WRAP(malloc
)(pthread_cond_t_sz
);
1004 internal_memset(newcond
, 0, pthread_cond_t_sz
);
1005 if (atomic_compare_exchange_strong(p
, &cond
, (uptr
)newcond
,
1006 memory_order_acq_rel
))
1008 WRAP(free
)(newcond
);
1012 struct CondMutexUnlockCtx
{
1013 ScopedInterceptor
*si
;
1019 static void cond_mutex_unlock(CondMutexUnlockCtx
*arg
) {
1020 // pthread_cond_wait interceptor has enabled async signal delivery
1021 // (see BlockingCall below). Disable async signals since we are running
1022 // tsan code. Also ScopedInterceptor and BlockingCall destructors won't run
1023 // since the thread is cancelled, so we have to manually execute them
1024 // (the thread still can run some user code due to pthread_cleanup_push).
1025 ThreadSignalContext
*ctx
= SigCtx(arg
->thr
);
1026 CHECK_EQ(atomic_load(&ctx
->in_blocking_func
, memory_order_relaxed
), 1);
1027 atomic_store(&ctx
->in_blocking_func
, 0, memory_order_relaxed
);
1028 MutexLock(arg
->thr
, arg
->pc
, (uptr
)arg
->m
);
1029 // Undo BlockingCall ctor effects.
1030 arg
->thr
->ignore_interceptors
--;
1031 arg
->si
->~ScopedInterceptor();
1034 INTERCEPTOR(int, pthread_cond_init
, void *c
, void *a
) {
1035 void *cond
= init_cond(c
, true);
1036 SCOPED_TSAN_INTERCEPTOR(pthread_cond_init
, cond
, a
);
1037 MemoryAccessRange(thr
, pc
, (uptr
)c
, sizeof(uptr
), true);
1038 return REAL(pthread_cond_init
)(cond
, a
);
1041 static int cond_wait(ThreadState
*thr
, uptr pc
, ScopedInterceptor
*si
,
1042 int (*fn
)(void *c
, void *m
, void *abstime
), void *c
,
1044 MemoryAccessRange(thr
, pc
, (uptr
)c
, sizeof(uptr
), false);
1045 MutexUnlock(thr
, pc
, (uptr
)m
);
1046 CondMutexUnlockCtx arg
= {si
, thr
, pc
, m
};
1048 // This ensures that we handle mutex lock even in case of pthread_cancel.
1049 // See test/tsan/cond_cancel.cc.
1051 // Enable signal delivery while the thread is blocked.
1052 BlockingCall
bc(thr
);
1053 res
= call_pthread_cancel_with_cleanup(
1054 fn
, c
, m
, t
, (void (*)(void *arg
))cond_mutex_unlock
, &arg
);
1056 if (res
== errno_EOWNERDEAD
) MutexRepair(thr
, pc
, (uptr
)m
);
1057 MutexLock(thr
, pc
, (uptr
)m
);
1061 INTERCEPTOR(int, pthread_cond_wait
, void *c
, void *m
) {
1062 void *cond
= init_cond(c
);
1063 SCOPED_TSAN_INTERCEPTOR(pthread_cond_wait
, cond
, m
);
1064 return cond_wait(thr
, pc
, &si
, (int (*)(void *c
, void *m
, void *abstime
))REAL(
1069 INTERCEPTOR(int, pthread_cond_timedwait
, void *c
, void *m
, void *abstime
) {
1070 void *cond
= init_cond(c
);
1071 SCOPED_TSAN_INTERCEPTOR(pthread_cond_timedwait
, cond
, m
, abstime
);
1072 return cond_wait(thr
, pc
, &si
, REAL(pthread_cond_timedwait
), cond
, m
,
1077 INTERCEPTOR(int, pthread_cond_timedwait_relative_np
, void *c
, void *m
,
1079 void *cond
= init_cond(c
);
1080 SCOPED_TSAN_INTERCEPTOR(pthread_cond_timedwait_relative_np
, cond
, m
, reltime
);
1081 return cond_wait(thr
, pc
, &si
, REAL(pthread_cond_timedwait_relative_np
), cond
,
1086 INTERCEPTOR(int, pthread_cond_signal
, void *c
) {
1087 void *cond
= init_cond(c
);
1088 SCOPED_TSAN_INTERCEPTOR(pthread_cond_signal
, cond
);
1089 MemoryAccessRange(thr
, pc
, (uptr
)c
, sizeof(uptr
), false);
1090 return REAL(pthread_cond_signal
)(cond
);
1093 INTERCEPTOR(int, pthread_cond_broadcast
, void *c
) {
1094 void *cond
= init_cond(c
);
1095 SCOPED_TSAN_INTERCEPTOR(pthread_cond_broadcast
, cond
);
1096 MemoryAccessRange(thr
, pc
, (uptr
)c
, sizeof(uptr
), false);
1097 return REAL(pthread_cond_broadcast
)(cond
);
1100 INTERCEPTOR(int, pthread_cond_destroy
, void *c
) {
1101 void *cond
= init_cond(c
);
1102 SCOPED_TSAN_INTERCEPTOR(pthread_cond_destroy
, cond
);
1103 MemoryAccessRange(thr
, pc
, (uptr
)c
, sizeof(uptr
), true);
1104 int res
= REAL(pthread_cond_destroy
)(cond
);
1105 if (common_flags()->legacy_pthread_cond
) {
1106 // Free our aux cond and zero the pointer to not leave dangling pointers.
1108 atomic_store((atomic_uintptr_t
*)c
, 0, memory_order_relaxed
);
1113 TSAN_INTERCEPTOR(int, pthread_mutex_init
, void *m
, void *a
) {
1114 SCOPED_TSAN_INTERCEPTOR(pthread_mutex_init
, m
, a
);
1115 int res
= REAL(pthread_mutex_init
)(m
, a
);
1117 bool recursive
= false;
1120 if (REAL(pthread_mutexattr_gettype
)(a
, &type
) == 0)
1121 recursive
= (type
== PTHREAD_MUTEX_RECURSIVE
1122 || type
== PTHREAD_MUTEX_RECURSIVE_NP
);
1124 MutexCreate(thr
, pc
, (uptr
)m
, false, recursive
, false);
1129 TSAN_INTERCEPTOR(int, pthread_mutex_destroy
, void *m
) {
1130 SCOPED_TSAN_INTERCEPTOR(pthread_mutex_destroy
, m
);
1131 int res
= REAL(pthread_mutex_destroy
)(m
);
1132 if (res
== 0 || res
== EBUSY
) {
1133 MutexDestroy(thr
, pc
, (uptr
)m
);
1138 TSAN_INTERCEPTOR(int, pthread_mutex_trylock
, void *m
) {
1139 SCOPED_TSAN_INTERCEPTOR(pthread_mutex_trylock
, m
);
1140 int res
= REAL(pthread_mutex_trylock
)(m
);
1141 if (res
== EOWNERDEAD
)
1142 MutexRepair(thr
, pc
, (uptr
)m
);
1143 if (res
== 0 || res
== EOWNERDEAD
)
1144 MutexLock(thr
, pc
, (uptr
)m
, /*rec=*/1, /*try_lock=*/true);
1149 TSAN_INTERCEPTOR(int, pthread_mutex_timedlock
, void *m
, void *abstime
) {
1150 SCOPED_TSAN_INTERCEPTOR(pthread_mutex_timedlock
, m
, abstime
);
1151 int res
= REAL(pthread_mutex_timedlock
)(m
, abstime
);
1153 MutexLock(thr
, pc
, (uptr
)m
);
1160 TSAN_INTERCEPTOR(int, pthread_spin_init
, void *m
, int pshared
) {
1161 SCOPED_TSAN_INTERCEPTOR(pthread_spin_init
, m
, pshared
);
1162 int res
= REAL(pthread_spin_init
)(m
, pshared
);
1164 MutexCreate(thr
, pc
, (uptr
)m
, false, false, false);
1169 TSAN_INTERCEPTOR(int, pthread_spin_destroy
, void *m
) {
1170 SCOPED_TSAN_INTERCEPTOR(pthread_spin_destroy
, m
);
1171 int res
= REAL(pthread_spin_destroy
)(m
);
1173 MutexDestroy(thr
, pc
, (uptr
)m
);
1178 TSAN_INTERCEPTOR(int, pthread_spin_lock
, void *m
) {
1179 SCOPED_TSAN_INTERCEPTOR(pthread_spin_lock
, m
);
1180 int res
= REAL(pthread_spin_lock
)(m
);
1182 MutexLock(thr
, pc
, (uptr
)m
);
1187 TSAN_INTERCEPTOR(int, pthread_spin_trylock
, void *m
) {
1188 SCOPED_TSAN_INTERCEPTOR(pthread_spin_trylock
, m
);
1189 int res
= REAL(pthread_spin_trylock
)(m
);
1191 MutexLock(thr
, pc
, (uptr
)m
, /*rec=*/1, /*try_lock=*/true);
1196 TSAN_INTERCEPTOR(int, pthread_spin_unlock
, void *m
) {
1197 SCOPED_TSAN_INTERCEPTOR(pthread_spin_unlock
, m
);
1198 MutexUnlock(thr
, pc
, (uptr
)m
);
1199 int res
= REAL(pthread_spin_unlock
)(m
);
1204 TSAN_INTERCEPTOR(int, pthread_rwlock_init
, void *m
, void *a
) {
1205 SCOPED_TSAN_INTERCEPTOR(pthread_rwlock_init
, m
, a
);
1206 int res
= REAL(pthread_rwlock_init
)(m
, a
);
1208 MutexCreate(thr
, pc
, (uptr
)m
, true, false, false);
1213 TSAN_INTERCEPTOR(int, pthread_rwlock_destroy
, void *m
) {
1214 SCOPED_TSAN_INTERCEPTOR(pthread_rwlock_destroy
, m
);
1215 int res
= REAL(pthread_rwlock_destroy
)(m
);
1217 MutexDestroy(thr
, pc
, (uptr
)m
);
1222 TSAN_INTERCEPTOR(int, pthread_rwlock_rdlock
, void *m
) {
1223 SCOPED_TSAN_INTERCEPTOR(pthread_rwlock_rdlock
, m
);
1224 int res
= REAL(pthread_rwlock_rdlock
)(m
);
1226 MutexReadLock(thr
, pc
, (uptr
)m
);
1231 TSAN_INTERCEPTOR(int, pthread_rwlock_tryrdlock
, void *m
) {
1232 SCOPED_TSAN_INTERCEPTOR(pthread_rwlock_tryrdlock
, m
);
1233 int res
= REAL(pthread_rwlock_tryrdlock
)(m
);
1235 MutexReadLock(thr
, pc
, (uptr
)m
, /*try_lock=*/true);
1241 TSAN_INTERCEPTOR(int, pthread_rwlock_timedrdlock
, void *m
, void *abstime
) {
1242 SCOPED_TSAN_INTERCEPTOR(pthread_rwlock_timedrdlock
, m
, abstime
);
1243 int res
= REAL(pthread_rwlock_timedrdlock
)(m
, abstime
);
1245 MutexReadLock(thr
, pc
, (uptr
)m
);
1251 TSAN_INTERCEPTOR(int, pthread_rwlock_wrlock
, void *m
) {
1252 SCOPED_TSAN_INTERCEPTOR(pthread_rwlock_wrlock
, m
);
1253 int res
= REAL(pthread_rwlock_wrlock
)(m
);
1255 MutexLock(thr
, pc
, (uptr
)m
);
1260 TSAN_INTERCEPTOR(int, pthread_rwlock_trywrlock
, void *m
) {
1261 SCOPED_TSAN_INTERCEPTOR(pthread_rwlock_trywrlock
, m
);
1262 int res
= REAL(pthread_rwlock_trywrlock
)(m
);
1264 MutexLock(thr
, pc
, (uptr
)m
, /*rec=*/1, /*try_lock=*/true);
1270 TSAN_INTERCEPTOR(int, pthread_rwlock_timedwrlock
, void *m
, void *abstime
) {
1271 SCOPED_TSAN_INTERCEPTOR(pthread_rwlock_timedwrlock
, m
, abstime
);
1272 int res
= REAL(pthread_rwlock_timedwrlock
)(m
, abstime
);
1274 MutexLock(thr
, pc
, (uptr
)m
);
1280 TSAN_INTERCEPTOR(int, pthread_rwlock_unlock
, void *m
) {
1281 SCOPED_TSAN_INTERCEPTOR(pthread_rwlock_unlock
, m
);
1282 MutexReadOrWriteUnlock(thr
, pc
, (uptr
)m
);
1283 int res
= REAL(pthread_rwlock_unlock
)(m
);
1288 TSAN_INTERCEPTOR(int, pthread_barrier_init
, void *b
, void *a
, unsigned count
) {
1289 SCOPED_TSAN_INTERCEPTOR(pthread_barrier_init
, b
, a
, count
);
1290 MemoryWrite(thr
, pc
, (uptr
)b
, kSizeLog1
);
1291 int res
= REAL(pthread_barrier_init
)(b
, a
, count
);
1295 TSAN_INTERCEPTOR(int, pthread_barrier_destroy
, void *b
) {
1296 SCOPED_TSAN_INTERCEPTOR(pthread_barrier_destroy
, b
);
1297 MemoryWrite(thr
, pc
, (uptr
)b
, kSizeLog1
);
1298 int res
= REAL(pthread_barrier_destroy
)(b
);
1302 TSAN_INTERCEPTOR(int, pthread_barrier_wait
, void *b
) {
1303 SCOPED_TSAN_INTERCEPTOR(pthread_barrier_wait
, b
);
1304 Release(thr
, pc
, (uptr
)b
);
1305 MemoryRead(thr
, pc
, (uptr
)b
, kSizeLog1
);
1306 int res
= REAL(pthread_barrier_wait
)(b
);
1307 MemoryRead(thr
, pc
, (uptr
)b
, kSizeLog1
);
1308 if (res
== 0 || res
== PTHREAD_BARRIER_SERIAL_THREAD
) {
1309 Acquire(thr
, pc
, (uptr
)b
);
1315 TSAN_INTERCEPTOR(int, pthread_once
, void *o
, void (*f
)()) {
1316 SCOPED_INTERCEPTOR_RAW(pthread_once
, o
, f
);
1317 if (o
== 0 || f
== 0)
1321 a
= static_cast<atomic_uint32_t
*>(o
);
1322 else // On OS X, pthread_once_t has a header with a long-sized signature.
1323 a
= static_cast<atomic_uint32_t
*>((void *)((char *)o
+ sizeof(long_t
)));
1324 u32 v
= atomic_load(a
, memory_order_acquire
);
1325 if (v
== 0 && atomic_compare_exchange_strong(a
, &v
, 1,
1326 memory_order_relaxed
)) {
1328 if (!thr
->in_ignored_lib
)
1329 Release(thr
, pc
, (uptr
)o
);
1330 atomic_store(a
, 2, memory_order_release
);
1333 internal_sched_yield();
1334 v
= atomic_load(a
, memory_order_acquire
);
1336 if (!thr
->in_ignored_lib
)
1337 Acquire(thr
, pc
, (uptr
)o
);
1342 #if SANITIZER_LINUX && !SANITIZER_ANDROID
1343 TSAN_INTERCEPTOR(int, __fxstat
, int version
, int fd
, void *buf
) {
1344 SCOPED_TSAN_INTERCEPTOR(__fxstat
, version
, fd
, buf
);
1346 FdAccess(thr
, pc
, fd
);
1347 return REAL(__fxstat
)(version
, fd
, buf
);
1349 #define TSAN_MAYBE_INTERCEPT___FXSTAT TSAN_INTERCEPT(__fxstat)
1351 #define TSAN_MAYBE_INTERCEPT___FXSTAT
1354 TSAN_INTERCEPTOR(int, fstat
, int fd
, void *buf
) {
1355 #if SANITIZER_FREEBSD || SANITIZER_MAC || SANITIZER_ANDROID
1356 SCOPED_TSAN_INTERCEPTOR(fstat
, fd
, buf
);
1358 FdAccess(thr
, pc
, fd
);
1359 return REAL(fstat
)(fd
, buf
);
1361 SCOPED_TSAN_INTERCEPTOR(__fxstat
, 0, fd
, buf
);
1363 FdAccess(thr
, pc
, fd
);
1364 return REAL(__fxstat
)(0, fd
, buf
);
1368 #if SANITIZER_LINUX && !SANITIZER_ANDROID
1369 TSAN_INTERCEPTOR(int, __fxstat64
, int version
, int fd
, void *buf
) {
1370 SCOPED_TSAN_INTERCEPTOR(__fxstat64
, version
, fd
, buf
);
1372 FdAccess(thr
, pc
, fd
);
1373 return REAL(__fxstat64
)(version
, fd
, buf
);
1375 #define TSAN_MAYBE_INTERCEPT___FXSTAT64 TSAN_INTERCEPT(__fxstat64)
1377 #define TSAN_MAYBE_INTERCEPT___FXSTAT64
1380 #if SANITIZER_LINUX && !SANITIZER_ANDROID
1381 TSAN_INTERCEPTOR(int, fstat64
, int fd
, void *buf
) {
1382 SCOPED_TSAN_INTERCEPTOR(__fxstat64
, 0, fd
, buf
);
1384 FdAccess(thr
, pc
, fd
);
1385 return REAL(__fxstat64
)(0, fd
, buf
);
1387 #define TSAN_MAYBE_INTERCEPT_FSTAT64 TSAN_INTERCEPT(fstat64)
1389 #define TSAN_MAYBE_INTERCEPT_FSTAT64
1392 TSAN_INTERCEPTOR(int, open
, const char *name
, int flags
, int mode
) {
1393 SCOPED_TSAN_INTERCEPTOR(open
, name
, flags
, mode
);
1394 READ_STRING(thr
, pc
, name
, 0);
1395 int fd
= REAL(open
)(name
, flags
, mode
);
1397 FdFileCreate(thr
, pc
, fd
);
1402 TSAN_INTERCEPTOR(int, open64
, const char *name
, int flags
, int mode
) {
1403 SCOPED_TSAN_INTERCEPTOR(open64
, name
, flags
, mode
);
1404 READ_STRING(thr
, pc
, name
, 0);
1405 int fd
= REAL(open64
)(name
, flags
, mode
);
1407 FdFileCreate(thr
, pc
, fd
);
1410 #define TSAN_MAYBE_INTERCEPT_OPEN64 TSAN_INTERCEPT(open64)
1412 #define TSAN_MAYBE_INTERCEPT_OPEN64
1415 TSAN_INTERCEPTOR(int, creat
, const char *name
, int mode
) {
1416 SCOPED_TSAN_INTERCEPTOR(creat
, name
, mode
);
1417 READ_STRING(thr
, pc
, name
, 0);
1418 int fd
= REAL(creat
)(name
, mode
);
1420 FdFileCreate(thr
, pc
, fd
);
1425 TSAN_INTERCEPTOR(int, creat64
, const char *name
, int mode
) {
1426 SCOPED_TSAN_INTERCEPTOR(creat64
, name
, mode
);
1427 READ_STRING(thr
, pc
, name
, 0);
1428 int fd
= REAL(creat64
)(name
, mode
);
1430 FdFileCreate(thr
, pc
, fd
);
1433 #define TSAN_MAYBE_INTERCEPT_CREAT64 TSAN_INTERCEPT(creat64)
1435 #define TSAN_MAYBE_INTERCEPT_CREAT64
1438 TSAN_INTERCEPTOR(int, dup
, int oldfd
) {
1439 SCOPED_TSAN_INTERCEPTOR(dup
, oldfd
);
1440 int newfd
= REAL(dup
)(oldfd
);
1441 if (oldfd
>= 0 && newfd
>= 0 && newfd
!= oldfd
)
1442 FdDup(thr
, pc
, oldfd
, newfd
, true);
1446 TSAN_INTERCEPTOR(int, dup2
, int oldfd
, int newfd
) {
1447 SCOPED_TSAN_INTERCEPTOR(dup2
, oldfd
, newfd
);
1448 int newfd2
= REAL(dup2
)(oldfd
, newfd
);
1449 if (oldfd
>= 0 && newfd2
>= 0 && newfd2
!= oldfd
)
1450 FdDup(thr
, pc
, oldfd
, newfd2
, false);
1455 TSAN_INTERCEPTOR(int, dup3
, int oldfd
, int newfd
, int flags
) {
1456 SCOPED_TSAN_INTERCEPTOR(dup3
, oldfd
, newfd
, flags
);
1457 int newfd2
= REAL(dup3
)(oldfd
, newfd
, flags
);
1458 if (oldfd
>= 0 && newfd2
>= 0 && newfd2
!= oldfd
)
1459 FdDup(thr
, pc
, oldfd
, newfd2
, false);
1465 TSAN_INTERCEPTOR(int, eventfd
, unsigned initval
, int flags
) {
1466 SCOPED_TSAN_INTERCEPTOR(eventfd
, initval
, flags
);
1467 int fd
= REAL(eventfd
)(initval
, flags
);
1469 FdEventCreate(thr
, pc
, fd
);
1472 #define TSAN_MAYBE_INTERCEPT_EVENTFD TSAN_INTERCEPT(eventfd)
1474 #define TSAN_MAYBE_INTERCEPT_EVENTFD
1478 TSAN_INTERCEPTOR(int, signalfd
, int fd
, void *mask
, int flags
) {
1479 SCOPED_TSAN_INTERCEPTOR(signalfd
, fd
, mask
, flags
);
1481 FdClose(thr
, pc
, fd
);
1482 fd
= REAL(signalfd
)(fd
, mask
, flags
);
1484 FdSignalCreate(thr
, pc
, fd
);
1487 #define TSAN_MAYBE_INTERCEPT_SIGNALFD TSAN_INTERCEPT(signalfd)
1489 #define TSAN_MAYBE_INTERCEPT_SIGNALFD
1493 TSAN_INTERCEPTOR(int, inotify_init
, int fake
) {
1494 SCOPED_TSAN_INTERCEPTOR(inotify_init
, fake
);
1495 int fd
= REAL(inotify_init
)(fake
);
1497 FdInotifyCreate(thr
, pc
, fd
);
1500 #define TSAN_MAYBE_INTERCEPT_INOTIFY_INIT TSAN_INTERCEPT(inotify_init)
1502 #define TSAN_MAYBE_INTERCEPT_INOTIFY_INIT
1506 TSAN_INTERCEPTOR(int, inotify_init1
, int flags
) {
1507 SCOPED_TSAN_INTERCEPTOR(inotify_init1
, flags
);
1508 int fd
= REAL(inotify_init1
)(flags
);
1510 FdInotifyCreate(thr
, pc
, fd
);
1513 #define TSAN_MAYBE_INTERCEPT_INOTIFY_INIT1 TSAN_INTERCEPT(inotify_init1)
1515 #define TSAN_MAYBE_INTERCEPT_INOTIFY_INIT1
1518 TSAN_INTERCEPTOR(int, socket
, int domain
, int type
, int protocol
) {
1519 SCOPED_TSAN_INTERCEPTOR(socket
, domain
, type
, protocol
);
1520 int fd
= REAL(socket
)(domain
, type
, protocol
);
1522 FdSocketCreate(thr
, pc
, fd
);
1526 TSAN_INTERCEPTOR(int, socketpair
, int domain
, int type
, int protocol
, int *fd
) {
1527 SCOPED_TSAN_INTERCEPTOR(socketpair
, domain
, type
, protocol
, fd
);
1528 int res
= REAL(socketpair
)(domain
, type
, protocol
, fd
);
1529 if (res
== 0 && fd
[0] >= 0 && fd
[1] >= 0)
1530 FdPipeCreate(thr
, pc
, fd
[0], fd
[1]);
1534 TSAN_INTERCEPTOR(int, connect
, int fd
, void *addr
, unsigned addrlen
) {
1535 SCOPED_TSAN_INTERCEPTOR(connect
, fd
, addr
, addrlen
);
1536 FdSocketConnecting(thr
, pc
, fd
);
1537 int res
= REAL(connect
)(fd
, addr
, addrlen
);
1538 if (res
== 0 && fd
>= 0)
1539 FdSocketConnect(thr
, pc
, fd
);
1543 TSAN_INTERCEPTOR(int, bind
, int fd
, void *addr
, unsigned addrlen
) {
1544 SCOPED_TSAN_INTERCEPTOR(bind
, fd
, addr
, addrlen
);
1545 int res
= REAL(bind
)(fd
, addr
, addrlen
);
1546 if (fd
> 0 && res
== 0)
1547 FdAccess(thr
, pc
, fd
);
1551 TSAN_INTERCEPTOR(int, listen
, int fd
, int backlog
) {
1552 SCOPED_TSAN_INTERCEPTOR(listen
, fd
, backlog
);
1553 int res
= REAL(listen
)(fd
, backlog
);
1554 if (fd
> 0 && res
== 0)
1555 FdAccess(thr
, pc
, fd
);
1559 TSAN_INTERCEPTOR(int, close
, int fd
) {
1560 SCOPED_TSAN_INTERCEPTOR(close
, fd
);
1562 FdClose(thr
, pc
, fd
);
1563 return REAL(close
)(fd
);
1567 TSAN_INTERCEPTOR(int, __close
, int fd
) {
1568 SCOPED_TSAN_INTERCEPTOR(__close
, fd
);
1570 FdClose(thr
, pc
, fd
);
1571 return REAL(__close
)(fd
);
1573 #define TSAN_MAYBE_INTERCEPT___CLOSE TSAN_INTERCEPT(__close)
1575 #define TSAN_MAYBE_INTERCEPT___CLOSE
1579 #if SANITIZER_LINUX && !SANITIZER_ANDROID
1580 TSAN_INTERCEPTOR(void, __res_iclose
, void *state
, bool free_addr
) {
1581 SCOPED_TSAN_INTERCEPTOR(__res_iclose
, state
, free_addr
);
1583 int cnt
= ExtractResolvFDs(state
, fds
, ARRAY_SIZE(fds
));
1584 for (int i
= 0; i
< cnt
; i
++) {
1586 FdClose(thr
, pc
, fds
[i
]);
1588 REAL(__res_iclose
)(state
, free_addr
);
1590 #define TSAN_MAYBE_INTERCEPT___RES_ICLOSE TSAN_INTERCEPT(__res_iclose)
1592 #define TSAN_MAYBE_INTERCEPT___RES_ICLOSE
1595 TSAN_INTERCEPTOR(int, pipe
, int *pipefd
) {
1596 SCOPED_TSAN_INTERCEPTOR(pipe
, pipefd
);
1597 int res
= REAL(pipe
)(pipefd
);
1598 if (res
== 0 && pipefd
[0] >= 0 && pipefd
[1] >= 0)
1599 FdPipeCreate(thr
, pc
, pipefd
[0], pipefd
[1]);
1604 TSAN_INTERCEPTOR(int, pipe2
, int *pipefd
, int flags
) {
1605 SCOPED_TSAN_INTERCEPTOR(pipe2
, pipefd
, flags
);
1606 int res
= REAL(pipe2
)(pipefd
, flags
);
1607 if (res
== 0 && pipefd
[0] >= 0 && pipefd
[1] >= 0)
1608 FdPipeCreate(thr
, pc
, pipefd
[0], pipefd
[1]);
1613 TSAN_INTERCEPTOR(int, unlink
, char *path
) {
1614 SCOPED_TSAN_INTERCEPTOR(unlink
, path
);
1615 Release(thr
, pc
, File2addr(path
));
1616 int res
= REAL(unlink
)(path
);
1620 TSAN_INTERCEPTOR(void*, tmpfile
, int fake
) {
1621 SCOPED_TSAN_INTERCEPTOR(tmpfile
, fake
);
1622 void *res
= REAL(tmpfile
)(fake
);
1624 int fd
= fileno_unlocked(res
);
1626 FdFileCreate(thr
, pc
, fd
);
1632 TSAN_INTERCEPTOR(void*, tmpfile64
, int fake
) {
1633 SCOPED_TSAN_INTERCEPTOR(tmpfile64
, fake
);
1634 void *res
= REAL(tmpfile64
)(fake
);
1636 int fd
= fileno_unlocked(res
);
1638 FdFileCreate(thr
, pc
, fd
);
1642 #define TSAN_MAYBE_INTERCEPT_TMPFILE64 TSAN_INTERCEPT(tmpfile64)
1644 #define TSAN_MAYBE_INTERCEPT_TMPFILE64
1647 TSAN_INTERCEPTOR(uptr
, fread
, void *ptr
, uptr size
, uptr nmemb
, void *f
) {
1648 // libc file streams can call user-supplied functions, see fopencookie.
1650 SCOPED_TSAN_INTERCEPTOR(fread
, ptr
, size
, nmemb
, f
);
1651 MemoryAccessRange(thr
, pc
, (uptr
)ptr
, size
* nmemb
, true);
1653 return REAL(fread
)(ptr
, size
, nmemb
, f
);
1656 TSAN_INTERCEPTOR(uptr
, fwrite
, const void *p
, uptr size
, uptr nmemb
, void *f
) {
1657 // libc file streams can call user-supplied functions, see fopencookie.
1659 SCOPED_TSAN_INTERCEPTOR(fwrite
, p
, size
, nmemb
, f
);
1660 MemoryAccessRange(thr
, pc
, (uptr
)p
, size
* nmemb
, false);
1662 return REAL(fwrite
)(p
, size
, nmemb
, f
);
1665 static void FlushStreams() {
1666 // Flushing all the streams here may freeze the process if a child thread is
1667 // performing file stream operations at the same time.
1668 REAL(fflush
)(stdout
);
1669 REAL(fflush
)(stderr
);
1672 TSAN_INTERCEPTOR(void, abort
, int fake
) {
1673 SCOPED_TSAN_INTERCEPTOR(abort
, fake
);
1678 TSAN_INTERCEPTOR(int, puts
, const char *s
) {
1679 SCOPED_TSAN_INTERCEPTOR(puts
, s
);
1680 MemoryAccessRange(thr
, pc
, (uptr
)s
, internal_strlen(s
), false);
1681 return REAL(puts
)(s
);
1684 TSAN_INTERCEPTOR(int, rmdir
, char *path
) {
1685 SCOPED_TSAN_INTERCEPTOR(rmdir
, path
);
1686 Release(thr
, pc
, Dir2addr(path
));
1687 int res
= REAL(rmdir
)(path
);
1691 TSAN_INTERCEPTOR(int, closedir
, void *dirp
) {
1692 SCOPED_TSAN_INTERCEPTOR(closedir
, dirp
);
1694 int fd
= dirfd(dirp
);
1695 FdClose(thr
, pc
, fd
);
1697 return REAL(closedir
)(dirp
);
1701 TSAN_INTERCEPTOR(int, epoll_create
, int size
) {
1702 SCOPED_TSAN_INTERCEPTOR(epoll_create
, size
);
1703 int fd
= REAL(epoll_create
)(size
);
1705 FdPollCreate(thr
, pc
, fd
);
1709 TSAN_INTERCEPTOR(int, epoll_create1
, int flags
) {
1710 SCOPED_TSAN_INTERCEPTOR(epoll_create1
, flags
);
1711 int fd
= REAL(epoll_create1
)(flags
);
1713 FdPollCreate(thr
, pc
, fd
);
1717 TSAN_INTERCEPTOR(int, epoll_ctl
, int epfd
, int op
, int fd
, void *ev
) {
1718 SCOPED_TSAN_INTERCEPTOR(epoll_ctl
, epfd
, op
, fd
, ev
);
1720 FdAccess(thr
, pc
, epfd
);
1721 if (epfd
>= 0 && fd
>= 0)
1722 FdAccess(thr
, pc
, fd
);
1723 if (op
== EPOLL_CTL_ADD
&& epfd
>= 0)
1724 FdRelease(thr
, pc
, epfd
);
1725 int res
= REAL(epoll_ctl
)(epfd
, op
, fd
, ev
);
1729 TSAN_INTERCEPTOR(int, epoll_wait
, int epfd
, void *ev
, int cnt
, int timeout
) {
1730 SCOPED_TSAN_INTERCEPTOR(epoll_wait
, epfd
, ev
, cnt
, timeout
);
1732 FdAccess(thr
, pc
, epfd
);
1733 int res
= BLOCK_REAL(epoll_wait
)(epfd
, ev
, cnt
, timeout
);
1734 if (res
> 0 && epfd
>= 0)
1735 FdAcquire(thr
, pc
, epfd
);
1739 TSAN_INTERCEPTOR(int, epoll_pwait
, int epfd
, void *ev
, int cnt
, int timeout
,
1741 SCOPED_TSAN_INTERCEPTOR(epoll_pwait
, epfd
, ev
, cnt
, timeout
, sigmask
);
1743 FdAccess(thr
, pc
, epfd
);
1744 int res
= BLOCK_REAL(epoll_pwait
)(epfd
, ev
, cnt
, timeout
, sigmask
);
1745 if (res
> 0 && epfd
>= 0)
1746 FdAcquire(thr
, pc
, epfd
);
1750 #define TSAN_MAYBE_INTERCEPT_EPOLL \
1751 TSAN_INTERCEPT(epoll_create); \
1752 TSAN_INTERCEPT(epoll_create1); \
1753 TSAN_INTERCEPT(epoll_ctl); \
1754 TSAN_INTERCEPT(epoll_wait); \
1755 TSAN_INTERCEPT(epoll_pwait)
1757 #define TSAN_MAYBE_INTERCEPT_EPOLL
1760 // The following functions are intercepted merely to process pending signals.
1761 // If program blocks signal X, we must deliver the signal before the function
1762 // returns. Similarly, if program unblocks a signal (or returns from sigsuspend)
1763 // it's better to deliver the signal straight away.
1764 TSAN_INTERCEPTOR(int, sigsuspend
, const __sanitizer_sigset_t
*mask
) {
1765 SCOPED_TSAN_INTERCEPTOR(sigsuspend
, mask
);
1766 return REAL(sigsuspend
)(mask
);
1769 TSAN_INTERCEPTOR(int, sigblock
, int mask
) {
1770 SCOPED_TSAN_INTERCEPTOR(sigblock
, mask
);
1771 return REAL(sigblock
)(mask
);
1774 TSAN_INTERCEPTOR(int, sigsetmask
, int mask
) {
1775 SCOPED_TSAN_INTERCEPTOR(sigsetmask
, mask
);
1776 return REAL(sigsetmask
)(mask
);
1779 TSAN_INTERCEPTOR(int, pthread_sigmask
, int how
, const __sanitizer_sigset_t
*set
,
1780 __sanitizer_sigset_t
*oldset
) {
1781 SCOPED_TSAN_INTERCEPTOR(pthread_sigmask
, how
, set
, oldset
);
1782 return REAL(pthread_sigmask
)(how
, set
, oldset
);
1787 static void CallUserSignalHandler(ThreadState
*thr
, bool sync
, bool acquire
,
1788 bool sigact
, int sig
, my_siginfo_t
*info
, void *uctx
) {
1790 Acquire(thr
, 0, (uptr
)&sigactions
[sig
]);
1791 // Signals are generally asynchronous, so if we receive a signals when
1792 // ignores are enabled we should disable ignores. This is critical for sync
1793 // and interceptors, because otherwise we can miss syncronization and report
1795 int ignore_reads_and_writes
= thr
->ignore_reads_and_writes
;
1796 int ignore_interceptors
= thr
->ignore_interceptors
;
1797 int ignore_sync
= thr
->ignore_sync
;
1798 if (!ctx
->after_multithreaded_fork
) {
1799 thr
->ignore_reads_and_writes
= 0;
1800 thr
->fast_state
.ClearIgnoreBit();
1801 thr
->ignore_interceptors
= 0;
1802 thr
->ignore_sync
= 0;
1804 // Ensure that the handler does not spoil errno.
1805 const int saved_errno
= errno
;
1807 // This code races with sigaction. Be careful to not read sa_sigaction twice.
1808 // Also need to remember pc for reporting before the call,
1809 // because the handler can reset it.
1810 volatile uptr pc
= sigact
?
1811 (uptr
)sigactions
[sig
].sa_sigaction
:
1812 (uptr
)sigactions
[sig
].sa_handler
;
1813 if (pc
!= (uptr
)SIG_DFL
&& pc
!= (uptr
)SIG_IGN
) {
1815 ((sigactionhandler_t
)pc
)(sig
, info
, uctx
);
1817 ((sighandler_t
)pc
)(sig
);
1819 if (!ctx
->after_multithreaded_fork
) {
1820 thr
->ignore_reads_and_writes
= ignore_reads_and_writes
;
1821 if (ignore_reads_and_writes
)
1822 thr
->fast_state
.SetIgnoreBit();
1823 thr
->ignore_interceptors
= ignore_interceptors
;
1824 thr
->ignore_sync
= ignore_sync
;
1826 // We do not detect errno spoiling for SIGTERM,
1827 // because some SIGTERM handlers do spoil errno but reraise SIGTERM,
1828 // tsan reports false positive in such case.
1829 // It's difficult to properly detect this situation (reraise),
1830 // because in async signal processing case (when handler is called directly
1831 // from rtl_generic_sighandler) we have not yet received the reraised
1832 // signal; and it looks too fragile to intercept all ways to reraise a signal.
1833 if (flags()->report_bugs
&& !sync
&& sig
!= SIGTERM
&& errno
!= 99) {
1834 VarSizeStackTrace stack
;
1835 // StackTrace::GetNestInstructionPc(pc) is used because return address is
1836 // expected, OutputReport() will undo this.
1837 ObtainCurrentStack(thr
, StackTrace::GetNextInstructionPc(pc
), &stack
);
1838 ThreadRegistryLock
l(ctx
->thread_registry
);
1839 ScopedReport
rep(ReportTypeErrnoInSignal
);
1840 if (!IsFiredSuppression(ctx
, ReportTypeErrnoInSignal
, stack
)) {
1841 rep
.AddStack(stack
, true);
1842 OutputReport(thr
, rep
);
1845 errno
= saved_errno
;
1848 void ProcessPendingSignals(ThreadState
*thr
) {
1849 ThreadSignalContext
*sctx
= SigCtx(thr
);
1851 atomic_load(&sctx
->have_pending_signals
, memory_order_relaxed
) == 0)
1853 atomic_store(&sctx
->have_pending_signals
, 0, memory_order_relaxed
);
1854 atomic_fetch_add(&thr
->in_signal_handler
, 1, memory_order_relaxed
);
1855 internal_sigfillset(&sctx
->emptyset
);
1856 int res
= REAL(pthread_sigmask
)(SIG_SETMASK
, &sctx
->emptyset
, &sctx
->oldset
);
1858 for (int sig
= 0; sig
< kSigCount
; sig
++) {
1859 SignalDesc
*signal
= &sctx
->pending_signals
[sig
];
1860 if (signal
->armed
) {
1861 signal
->armed
= false;
1862 CallUserSignalHandler(thr
, false, true, signal
->sigaction
, sig
,
1863 &signal
->siginfo
, &signal
->ctx
);
1866 res
= REAL(pthread_sigmask
)(SIG_SETMASK
, &sctx
->oldset
, 0);
1868 atomic_fetch_add(&thr
->in_signal_handler
, -1, memory_order_relaxed
);
1871 } // namespace __tsan
1873 static bool is_sync_signal(ThreadSignalContext
*sctx
, int sig
) {
1874 return sig
== SIGSEGV
|| sig
== SIGBUS
|| sig
== SIGILL
||
1875 sig
== SIGABRT
|| sig
== SIGFPE
|| sig
== SIGPIPE
|| sig
== SIGSYS
||
1876 // If we are sending signal to ourselves, we must process it now.
1877 (sctx
&& sig
== sctx
->int_signal_send
);
1880 void ALWAYS_INLINE
rtl_generic_sighandler(bool sigact
, int sig
,
1881 my_siginfo_t
*info
, void *ctx
) {
1882 ThreadState
*thr
= cur_thread();
1883 ThreadSignalContext
*sctx
= SigCtx(thr
);
1884 if (sig
< 0 || sig
>= kSigCount
) {
1885 VPrintf(1, "ThreadSanitizer: ignoring signal %d\n", sig
);
1888 // Don't mess with synchronous signals.
1889 const bool sync
= is_sync_signal(sctx
, sig
);
1891 // If we are in blocking function, we can safely process it now
1892 // (but check if we are in a recursive interceptor,
1893 // i.e. pthread_join()->munmap()).
1894 (sctx
&& atomic_load(&sctx
->in_blocking_func
, memory_order_relaxed
))) {
1895 atomic_fetch_add(&thr
->in_signal_handler
, 1, memory_order_relaxed
);
1896 if (sctx
&& atomic_load(&sctx
->in_blocking_func
, memory_order_relaxed
)) {
1897 atomic_store(&sctx
->in_blocking_func
, 0, memory_order_relaxed
);
1898 CallUserSignalHandler(thr
, sync
, true, sigact
, sig
, info
, ctx
);
1899 atomic_store(&sctx
->in_blocking_func
, 1, memory_order_relaxed
);
1901 // Be very conservative with when we do acquire in this case.
1902 // It's unsafe to do acquire in async handlers, because ThreadState
1903 // can be in inconsistent state.
1904 // SIGSYS looks relatively safe -- it's synchronous and can actually
1905 // need some global state.
1906 bool acq
= (sig
== SIGSYS
);
1907 CallUserSignalHandler(thr
, sync
, acq
, sigact
, sig
, info
, ctx
);
1909 atomic_fetch_add(&thr
->in_signal_handler
, -1, memory_order_relaxed
);
1915 SignalDesc
*signal
= &sctx
->pending_signals
[sig
];
1916 if (signal
->armed
== false) {
1917 signal
->armed
= true;
1918 signal
->sigaction
= sigact
;
1920 internal_memcpy(&signal
->siginfo
, info
, sizeof(*info
));
1922 internal_memcpy(&signal
->ctx
, ctx
, sizeof(signal
->ctx
));
1923 atomic_store(&sctx
->have_pending_signals
, 1, memory_order_relaxed
);
1927 static void rtl_sighandler(int sig
) {
1928 rtl_generic_sighandler(false, sig
, 0, 0);
1931 static void rtl_sigaction(int sig
, my_siginfo_t
*info
, void *ctx
) {
1932 rtl_generic_sighandler(true, sig
, info
, ctx
);
1935 TSAN_INTERCEPTOR(int, sigaction
, int sig
, sigaction_t
*act
, sigaction_t
*old
) {
1936 // Note: if we call REAL(sigaction) directly for any reason without proxying
1937 // the signal handler through rtl_sigaction, very bad things will happen.
1938 // The handler will run synchronously and corrupt tsan per-thread state.
1939 SCOPED_INTERCEPTOR_RAW(sigaction
, sig
, act
, old
);
1941 internal_memcpy(old
, &sigactions
[sig
], sizeof(*old
));
1944 // Copy act into sigactions[sig].
1945 // Can't use struct copy, because compiler can emit call to memcpy.
1946 // Can't use internal_memcpy, because it copies byte-by-byte,
1947 // and signal handler reads the sa_handler concurrently. It it can read
1948 // some bytes from old value and some bytes from new value.
1949 // Use volatile to prevent insertion of memcpy.
1950 sigactions
[sig
].sa_handler
= *(volatile sighandler_t
*)&act
->sa_handler
;
1951 sigactions
[sig
].sa_flags
= *(volatile int*)&act
->sa_flags
;
1952 internal_memcpy(&sigactions
[sig
].sa_mask
, &act
->sa_mask
,
1953 sizeof(sigactions
[sig
].sa_mask
));
1954 #if !SANITIZER_FREEBSD && !SANITIZER_MAC
1955 sigactions
[sig
].sa_restorer
= act
->sa_restorer
;
1958 internal_memcpy(&newact
, act
, sizeof(newact
));
1959 internal_sigfillset(&newact
.sa_mask
);
1960 if (act
->sa_handler
!= SIG_IGN
&& act
->sa_handler
!= SIG_DFL
) {
1961 if (newact
.sa_flags
& SA_SIGINFO
)
1962 newact
.sa_sigaction
= rtl_sigaction
;
1964 newact
.sa_handler
= rtl_sighandler
;
1966 ReleaseStore(thr
, pc
, (uptr
)&sigactions
[sig
]);
1967 int res
= REAL(sigaction
)(sig
, &newact
, 0);
1971 TSAN_INTERCEPTOR(sighandler_t
, signal
, int sig
, sighandler_t h
) {
1974 internal_memset(&act
.sa_mask
, -1, sizeof(act
.sa_mask
));
1977 int res
= sigaction(sig
, &act
, &old
);
1980 return old
.sa_handler
;
1983 TSAN_INTERCEPTOR(int, raise
, int sig
) {
1984 SCOPED_TSAN_INTERCEPTOR(raise
, sig
);
1985 ThreadSignalContext
*sctx
= SigCtx(thr
);
1987 int prev
= sctx
->int_signal_send
;
1988 sctx
->int_signal_send
= sig
;
1989 int res
= REAL(raise
)(sig
);
1990 CHECK_EQ(sctx
->int_signal_send
, sig
);
1991 sctx
->int_signal_send
= prev
;
1995 TSAN_INTERCEPTOR(int, kill
, int pid
, int sig
) {
1996 SCOPED_TSAN_INTERCEPTOR(kill
, pid
, sig
);
1997 ThreadSignalContext
*sctx
= SigCtx(thr
);
1999 int prev
= sctx
->int_signal_send
;
2000 if (pid
== (int)internal_getpid()) {
2001 sctx
->int_signal_send
= sig
;
2003 int res
= REAL(kill
)(pid
, sig
);
2004 if (pid
== (int)internal_getpid()) {
2005 CHECK_EQ(sctx
->int_signal_send
, sig
);
2006 sctx
->int_signal_send
= prev
;
2011 TSAN_INTERCEPTOR(int, pthread_kill
, void *tid
, int sig
) {
2012 SCOPED_TSAN_INTERCEPTOR(pthread_kill
, tid
, sig
);
2013 ThreadSignalContext
*sctx
= SigCtx(thr
);
2015 int prev
= sctx
->int_signal_send
;
2016 if (tid
== pthread_self()) {
2017 sctx
->int_signal_send
= sig
;
2019 int res
= REAL(pthread_kill
)(tid
, sig
);
2020 if (tid
== pthread_self()) {
2021 CHECK_EQ(sctx
->int_signal_send
, sig
);
2022 sctx
->int_signal_send
= prev
;
2027 TSAN_INTERCEPTOR(int, gettimeofday
, void *tv
, void *tz
) {
2028 SCOPED_TSAN_INTERCEPTOR(gettimeofday
, tv
, tz
);
2029 // It's intercepted merely to process pending signals.
2030 return REAL(gettimeofday
)(tv
, tz
);
2033 TSAN_INTERCEPTOR(int, getaddrinfo
, void *node
, void *service
,
2034 void *hints
, void *rv
) {
2035 SCOPED_TSAN_INTERCEPTOR(getaddrinfo
, node
, service
, hints
, rv
);
2036 // We miss atomic synchronization in getaddrinfo,
2037 // and can report false race between malloc and free
2038 // inside of getaddrinfo. So ignore memory accesses.
2039 ThreadIgnoreBegin(thr
, pc
);
2040 int res
= REAL(getaddrinfo
)(node
, service
, hints
, rv
);
2041 ThreadIgnoreEnd(thr
, pc
);
2045 TSAN_INTERCEPTOR(int, fork
, int fake
) {
2046 if (cur_thread()->in_symbolizer
)
2047 return REAL(fork
)(fake
);
2048 SCOPED_INTERCEPTOR_RAW(fork
, fake
);
2049 ForkBefore(thr
, pc
);
2052 // On OS X, REAL(fork) can call intercepted functions (OSSpinLockLock), and
2053 // we'll assert in CheckNoLocks() unless we ignore interceptors.
2054 ScopedIgnoreInterceptors ignore
;
2055 pid
= REAL(fork
)(fake
);
2059 ForkChildAfter(thr
, pc
);
2061 } else if (pid
> 0) {
2063 ForkParentAfter(thr
, pc
);
2066 ForkParentAfter(thr
, pc
);
2071 TSAN_INTERCEPTOR(int, vfork
, int fake
) {
2072 // Some programs (e.g. openjdk) call close for all file descriptors
2073 // in the child process. Under tsan it leads to false positives, because
2074 // address space is shared, so the parent process also thinks that
2075 // the descriptors are closed (while they are actually not).
2076 // This leads to false positives due to missed synchronization.
2077 // Strictly saying this is undefined behavior, because vfork child is not
2078 // allowed to call any functions other than exec/exit. But this is what
2079 // openjdk does, so we want to handle it.
2080 // We could disable interceptors in the child process. But it's not possible
2081 // to simply intercept and wrap vfork, because vfork child is not allowed
2082 // to return from the function that calls vfork, and that's exactly what
2083 // we would do. So this would require some assembly trickery as well.
2084 // Instead we simply turn vfork into fork.
2085 return WRAP(fork
)(fake
);
2088 #if !SANITIZER_MAC && !SANITIZER_ANDROID
2089 typedef int (*dl_iterate_phdr_cb_t
)(__sanitizer_dl_phdr_info
*info
, SIZE_T size
,
2091 struct dl_iterate_phdr_data
{
2094 dl_iterate_phdr_cb_t cb
;
2098 static bool IsAppNotRodata(uptr addr
) {
2099 return IsAppMem(addr
) && *(u64
*)MemToShadow(addr
) != kShadowRodata
;
2102 static int dl_iterate_phdr_cb(__sanitizer_dl_phdr_info
*info
, SIZE_T size
,
2104 dl_iterate_phdr_data
*cbdata
= (dl_iterate_phdr_data
*)data
;
2105 // dlopen/dlclose allocate/free dynamic-linker-internal memory, which is later
2106 // accessible in dl_iterate_phdr callback. But we don't see synchronization
2107 // inside of dynamic linker, so we "unpoison" it here in order to not
2108 // produce false reports. Ignoring malloc/free in dlopen/dlclose is not enough
2109 // because some libc functions call __libc_dlopen.
2110 if (info
&& IsAppNotRodata((uptr
)info
->dlpi_name
))
2111 MemoryResetRange(cbdata
->thr
, cbdata
->pc
, (uptr
)info
->dlpi_name
,
2112 internal_strlen(info
->dlpi_name
));
2113 int res
= cbdata
->cb(info
, size
, cbdata
->data
);
2114 // Perform the check one more time in case info->dlpi_name was overwritten
2115 // by user callback.
2116 if (info
&& IsAppNotRodata((uptr
)info
->dlpi_name
))
2117 MemoryResetRange(cbdata
->thr
, cbdata
->pc
, (uptr
)info
->dlpi_name
,
2118 internal_strlen(info
->dlpi_name
));
2122 TSAN_INTERCEPTOR(int, dl_iterate_phdr
, dl_iterate_phdr_cb_t cb
, void *data
) {
2123 SCOPED_TSAN_INTERCEPTOR(dl_iterate_phdr
, cb
, data
);
2124 dl_iterate_phdr_data cbdata
;
2129 int res
= REAL(dl_iterate_phdr
)(dl_iterate_phdr_cb
, &cbdata
);
2134 static int OnExit(ThreadState
*thr
) {
2135 int status
= Finalize(thr
);
2140 struct TsanInterceptorContext
{
2142 const uptr caller_pc
;
2147 static void HandleRecvmsg(ThreadState
*thr
, uptr pc
,
2148 __sanitizer_msghdr
*msg
) {
2150 int cnt
= ExtractRecvmsgFDs(msg
, fds
, ARRAY_SIZE(fds
));
2151 for (int i
= 0; i
< cnt
; i
++)
2152 FdEventCreate(thr
, pc
, fds
[i
]);
2156 #include "sanitizer_common/sanitizer_platform_interceptors.h"
2157 // Causes interceptor recursion (getaddrinfo() and fopen())
2158 #undef SANITIZER_INTERCEPT_GETADDRINFO
2159 // There interceptors do not seem to be strictly necessary for tsan.
2160 // But we see cases where the interceptors consume 70% of execution time.
2161 // Memory blocks passed to fgetgrent_r are "written to" by tsan several times.
2162 // First, there is some recursion (getgrnam_r calls fgetgrent_r), and each
2163 // function "writes to" the buffer. Then, the same memory is "written to"
2164 // twice, first as buf and then as pwbufp (both of them refer to the same
2166 #undef SANITIZER_INTERCEPT_GETPWENT
2167 #undef SANITIZER_INTERCEPT_GETPWENT_R
2168 #undef SANITIZER_INTERCEPT_FGETPWENT
2169 #undef SANITIZER_INTERCEPT_GETPWNAM_AND_FRIENDS
2170 #undef SANITIZER_INTERCEPT_GETPWNAM_R_AND_FRIENDS
2171 // We define our own.
2172 #if SANITIZER_INTERCEPT_TLS_GET_ADDR
2173 #define NEED_TLS_GET_ADDR
2175 #undef SANITIZER_INTERCEPT_TLS_GET_ADDR
2177 #define COMMON_INTERCEPT_FUNCTION(name) INTERCEPT_FUNCTION(name)
2178 #define COMMON_INTERCEPT_FUNCTION_VER(name, ver) \
2179 INTERCEPT_FUNCTION_VER(name, ver)
2181 #define COMMON_INTERCEPTOR_WRITE_RANGE(ctx, ptr, size) \
2182 MemoryAccessRange(((TsanInterceptorContext *)ctx)->thr, \
2183 ((TsanInterceptorContext *)ctx)->pc, (uptr)ptr, size, \
2186 #define COMMON_INTERCEPTOR_READ_RANGE(ctx, ptr, size) \
2187 MemoryAccessRange(((TsanInterceptorContext *) ctx)->thr, \
2188 ((TsanInterceptorContext *) ctx)->pc, (uptr) ptr, size, \
2191 #define COMMON_INTERCEPTOR_ENTER(ctx, func, ...) \
2192 SCOPED_TSAN_INTERCEPTOR(func, __VA_ARGS__); \
2193 TsanInterceptorContext _ctx = {thr, caller_pc, pc}; \
2194 ctx = (void *)&_ctx; \
2197 #define COMMON_INTERCEPTOR_ENTER_NOIGNORE(ctx, func, ...) \
2198 SCOPED_INTERCEPTOR_RAW(func, __VA_ARGS__); \
2199 TsanInterceptorContext _ctx = {thr, caller_pc, pc}; \
2200 ctx = (void *)&_ctx; \
2203 #define COMMON_INTERCEPTOR_FILE_OPEN(ctx, file, path) \
2204 Acquire(thr, pc, File2addr(path)); \
2206 int fd = fileno_unlocked(file); \
2207 if (fd >= 0) FdFileCreate(thr, pc, fd); \
2210 #define COMMON_INTERCEPTOR_FILE_CLOSE(ctx, file) \
2212 int fd = fileno_unlocked(file); \
2213 if (fd >= 0) FdClose(thr, pc, fd); \
2216 #define COMMON_INTERCEPTOR_LIBRARY_LOADED(filename, handle) \
2217 libignore()->OnLibraryLoaded(filename)
2219 #define COMMON_INTERCEPTOR_LIBRARY_UNLOADED() \
2220 libignore()->OnLibraryUnloaded()
2222 #define COMMON_INTERCEPTOR_ACQUIRE(ctx, u) \
2223 Acquire(((TsanInterceptorContext *) ctx)->thr, pc, u)
2225 #define COMMON_INTERCEPTOR_RELEASE(ctx, u) \
2226 Release(((TsanInterceptorContext *) ctx)->thr, pc, u)
2228 #define COMMON_INTERCEPTOR_DIR_ACQUIRE(ctx, path) \
2229 Acquire(((TsanInterceptorContext *) ctx)->thr, pc, Dir2addr(path))
2231 #define COMMON_INTERCEPTOR_FD_ACQUIRE(ctx, fd) \
2232 FdAcquire(((TsanInterceptorContext *) ctx)->thr, pc, fd)
2234 #define COMMON_INTERCEPTOR_FD_RELEASE(ctx, fd) \
2235 FdRelease(((TsanInterceptorContext *) ctx)->thr, pc, fd)
2237 #define COMMON_INTERCEPTOR_FD_ACCESS(ctx, fd) \
2238 FdAccess(((TsanInterceptorContext *) ctx)->thr, pc, fd)
2240 #define COMMON_INTERCEPTOR_FD_SOCKET_ACCEPT(ctx, fd, newfd) \
2241 FdSocketAccept(((TsanInterceptorContext *) ctx)->thr, pc, fd, newfd)
2243 #define COMMON_INTERCEPTOR_SET_THREAD_NAME(ctx, name) \
2244 ThreadSetName(((TsanInterceptorContext *) ctx)->thr, name)
2246 #define COMMON_INTERCEPTOR_SET_PTHREAD_NAME(ctx, thread, name) \
2247 __tsan::ctx->thread_registry->SetThreadNameByUserId(thread, name)
2249 #define COMMON_INTERCEPTOR_BLOCK_REAL(name) BLOCK_REAL(name)
2251 #define COMMON_INTERCEPTOR_ON_EXIT(ctx) \
2252 OnExit(((TsanInterceptorContext *) ctx)->thr)
2254 #define COMMON_INTERCEPTOR_MUTEX_LOCK(ctx, m) \
2255 MutexLock(((TsanInterceptorContext *)ctx)->thr, \
2256 ((TsanInterceptorContext *)ctx)->pc, (uptr)m)
2258 #define COMMON_INTERCEPTOR_MUTEX_UNLOCK(ctx, m) \
2259 MutexUnlock(((TsanInterceptorContext *)ctx)->thr, \
2260 ((TsanInterceptorContext *)ctx)->pc, (uptr)m)
2262 #define COMMON_INTERCEPTOR_MUTEX_REPAIR(ctx, m) \
2263 MutexRepair(((TsanInterceptorContext *)ctx)->thr, \
2264 ((TsanInterceptorContext *)ctx)->pc, (uptr)m)
2266 #define COMMON_INTERCEPTOR_MUTEX_INVALID(ctx, m) \
2267 MutexInvalidAccess(((TsanInterceptorContext *)ctx)->thr, \
2268 ((TsanInterceptorContext *)ctx)->pc, (uptr)m)
2271 #define COMMON_INTERCEPTOR_HANDLE_RECVMSG(ctx, msg) \
2272 HandleRecvmsg(((TsanInterceptorContext *)ctx)->thr, \
2273 ((TsanInterceptorContext *)ctx)->pc, msg)
2276 #define COMMON_INTERCEPTOR_GET_TLS_RANGE(begin, end) \
2277 if (TsanThread *t = GetCurrentThread()) { \
2278 *begin = t->tls_begin(); \
2279 *end = t->tls_end(); \
2281 *begin = *end = 0; \
2284 #define COMMON_INTERCEPTOR_USER_CALLBACK_START() \
2285 SCOPED_TSAN_INTERCEPTOR_USER_CALLBACK_START()
2287 #define COMMON_INTERCEPTOR_USER_CALLBACK_END() \
2288 SCOPED_TSAN_INTERCEPTOR_USER_CALLBACK_END()
2290 #include "sanitizer_common/sanitizer_common_interceptors.inc"
2292 #define TSAN_SYSCALL() \
2293 ThreadState *thr = cur_thread(); \
2294 if (thr->ignore_interceptors) \
2296 ScopedSyscall scoped_syscall(thr) \
2299 struct ScopedSyscall
{
2302 explicit ScopedSyscall(ThreadState
*thr
)
2308 ProcessPendingSignals(thr
);
2312 #if !SANITIZER_FREEBSD && !SANITIZER_MAC
2313 static void syscall_access_range(uptr pc
, uptr p
, uptr s
, bool write
) {
2315 MemoryAccessRange(thr
, pc
, p
, s
, write
);
2318 static void syscall_acquire(uptr pc
, uptr addr
) {
2320 Acquire(thr
, pc
, addr
);
2321 DPrintf("syscall_acquire(%p)\n", addr
);
2324 static void syscall_release(uptr pc
, uptr addr
) {
2326 DPrintf("syscall_release(%p)\n", addr
);
2327 Release(thr
, pc
, addr
);
2330 static void syscall_fd_close(uptr pc
, int fd
) {
2332 FdClose(thr
, pc
, fd
);
2335 static USED
void syscall_fd_acquire(uptr pc
, int fd
) {
2337 FdAcquire(thr
, pc
, fd
);
2338 DPrintf("syscall_fd_acquire(%p)\n", fd
);
2341 static USED
void syscall_fd_release(uptr pc
, int fd
) {
2343 DPrintf("syscall_fd_release(%p)\n", fd
);
2344 FdRelease(thr
, pc
, fd
);
2347 static void syscall_pre_fork(uptr pc
) {
2349 ForkBefore(thr
, pc
);
2352 static void syscall_post_fork(uptr pc
, int pid
) {
2356 ForkChildAfter(thr
, pc
);
2358 } else if (pid
> 0) {
2360 ForkParentAfter(thr
, pc
);
2363 ForkParentAfter(thr
, pc
);
2368 #define COMMON_SYSCALL_PRE_READ_RANGE(p, s) \
2369 syscall_access_range(GET_CALLER_PC(), (uptr)(p), (uptr)(s), false)
2371 #define COMMON_SYSCALL_PRE_WRITE_RANGE(p, s) \
2372 syscall_access_range(GET_CALLER_PC(), (uptr)(p), (uptr)(s), true)
2374 #define COMMON_SYSCALL_POST_READ_RANGE(p, s) \
2380 #define COMMON_SYSCALL_POST_WRITE_RANGE(p, s) \
2386 #define COMMON_SYSCALL_ACQUIRE(addr) \
2387 syscall_acquire(GET_CALLER_PC(), (uptr)(addr))
2389 #define COMMON_SYSCALL_RELEASE(addr) \
2390 syscall_release(GET_CALLER_PC(), (uptr)(addr))
2392 #define COMMON_SYSCALL_FD_CLOSE(fd) syscall_fd_close(GET_CALLER_PC(), fd)
2394 #define COMMON_SYSCALL_FD_ACQUIRE(fd) syscall_fd_acquire(GET_CALLER_PC(), fd)
2396 #define COMMON_SYSCALL_FD_RELEASE(fd) syscall_fd_release(GET_CALLER_PC(), fd)
2398 #define COMMON_SYSCALL_PRE_FORK() \
2399 syscall_pre_fork(GET_CALLER_PC())
2401 #define COMMON_SYSCALL_POST_FORK(res) \
2402 syscall_post_fork(GET_CALLER_PC(), res)
2404 #include "sanitizer_common/sanitizer_common_syscalls.inc"
2406 #ifdef NEED_TLS_GET_ADDR
2407 // Define own interceptor instead of sanitizer_common's for three reasons:
2408 // 1. It must not process pending signals.
2409 // Signal handlers may contain MOVDQA instruction (see below).
2410 // 2. It must be as simple as possible to not contain MOVDQA.
2411 // 3. Sanitizer_common version uses COMMON_INTERCEPTOR_INITIALIZE_RANGE which
2412 // is empty for tsan (meant only for msan).
2413 // Note: __tls_get_addr can be called with mis-aligned stack due to:
2414 // https://gcc.gnu.org/bugzilla/show_bug.cgi?id=58066
2415 // So the interceptor must work with mis-aligned stack, in particular, does not
2416 // execute MOVDQA with stack addresses.
2417 TSAN_INTERCEPTOR(void *, __tls_get_addr
, void *arg
) {
2418 void *res
= REAL(__tls_get_addr
)(arg
);
2419 ThreadState
*thr
= cur_thread();
2422 DTLS::DTV
*dtv
= DTLS_on_tls_get_addr(arg
, res
, thr
->tls_addr
, thr
->tls_size
);
2425 // New DTLS block has been allocated.
2426 MemoryResetRange(thr
, 0, dtv
->beg
, dtv
->size
);
2433 static void finalize(void *arg
) {
2434 ThreadState
*thr
= cur_thread();
2435 int status
= Finalize(thr
);
2436 // Make sure the output is not lost.
2442 #if !SANITIZER_MAC && !SANITIZER_ANDROID
2443 static void unreachable() {
2444 Report("FATAL: ThreadSanitizer: unreachable called\n");
2449 void InitializeInterceptors() {
2451 // We need to setup it early, because functions like dlsym() can call it.
2452 REAL(memset
) = internal_memset
;
2453 REAL(memcpy
) = internal_memcpy
;
2456 // Instruct libc malloc to consume less memory.
2458 mallopt(1, 0); // M_MXFAST
2459 mallopt(-3, 32*1024); // M_MMAP_THRESHOLD
2462 InitializeCommonInterceptors();
2465 // We can not use TSAN_INTERCEPT to get setjmp addr,
2466 // because it does &setjmp and setjmp is not present in some versions of libc.
2467 using __interception::GetRealFunctionAddress
;
2468 GetRealFunctionAddress("setjmp", (uptr
*)&REAL(setjmp
), 0, 0);
2469 GetRealFunctionAddress("_setjmp", (uptr
*)&REAL(_setjmp
), 0, 0);
2470 GetRealFunctionAddress("sigsetjmp", (uptr
*)&REAL(sigsetjmp
), 0, 0);
2471 GetRealFunctionAddress("__sigsetjmp", (uptr
*)&REAL(__sigsetjmp
), 0, 0);
2474 TSAN_INTERCEPT(longjmp
);
2475 TSAN_INTERCEPT(siglongjmp
);
2477 TSAN_INTERCEPT(malloc
);
2478 TSAN_INTERCEPT(__libc_memalign
);
2479 TSAN_INTERCEPT(calloc
);
2480 TSAN_INTERCEPT(realloc
);
2481 TSAN_INTERCEPT(free
);
2482 TSAN_INTERCEPT(cfree
);
2483 TSAN_INTERCEPT(mmap
);
2484 TSAN_MAYBE_INTERCEPT_MMAP64
;
2485 TSAN_INTERCEPT(munmap
);
2486 TSAN_MAYBE_INTERCEPT_MEMALIGN
;
2487 TSAN_INTERCEPT(valloc
);
2488 TSAN_MAYBE_INTERCEPT_PVALLOC
;
2489 TSAN_INTERCEPT(posix_memalign
);
2491 TSAN_INTERCEPT(strcpy
); // NOLINT
2492 TSAN_INTERCEPT(strncpy
);
2493 TSAN_INTERCEPT(strdup
);
2495 TSAN_INTERCEPT(pthread_create
);
2496 TSAN_INTERCEPT(pthread_join
);
2497 TSAN_INTERCEPT(pthread_detach
);
2499 TSAN_INTERCEPT_VER(pthread_cond_init
, PTHREAD_ABI_BASE
);
2500 TSAN_INTERCEPT_VER(pthread_cond_signal
, PTHREAD_ABI_BASE
);
2501 TSAN_INTERCEPT_VER(pthread_cond_broadcast
, PTHREAD_ABI_BASE
);
2502 TSAN_INTERCEPT_VER(pthread_cond_wait
, PTHREAD_ABI_BASE
);
2503 TSAN_INTERCEPT_VER(pthread_cond_timedwait
, PTHREAD_ABI_BASE
);
2504 TSAN_INTERCEPT_VER(pthread_cond_destroy
, PTHREAD_ABI_BASE
);
2506 TSAN_INTERCEPT(pthread_mutex_init
);
2507 TSAN_INTERCEPT(pthread_mutex_destroy
);
2508 TSAN_INTERCEPT(pthread_mutex_trylock
);
2509 TSAN_INTERCEPT(pthread_mutex_timedlock
);
2511 TSAN_INTERCEPT(pthread_spin_init
);
2512 TSAN_INTERCEPT(pthread_spin_destroy
);
2513 TSAN_INTERCEPT(pthread_spin_lock
);
2514 TSAN_INTERCEPT(pthread_spin_trylock
);
2515 TSAN_INTERCEPT(pthread_spin_unlock
);
2517 TSAN_INTERCEPT(pthread_rwlock_init
);
2518 TSAN_INTERCEPT(pthread_rwlock_destroy
);
2519 TSAN_INTERCEPT(pthread_rwlock_rdlock
);
2520 TSAN_INTERCEPT(pthread_rwlock_tryrdlock
);
2521 TSAN_INTERCEPT(pthread_rwlock_timedrdlock
);
2522 TSAN_INTERCEPT(pthread_rwlock_wrlock
);
2523 TSAN_INTERCEPT(pthread_rwlock_trywrlock
);
2524 TSAN_INTERCEPT(pthread_rwlock_timedwrlock
);
2525 TSAN_INTERCEPT(pthread_rwlock_unlock
);
2527 TSAN_INTERCEPT(pthread_barrier_init
);
2528 TSAN_INTERCEPT(pthread_barrier_destroy
);
2529 TSAN_INTERCEPT(pthread_barrier_wait
);
2531 TSAN_INTERCEPT(pthread_once
);
2533 TSAN_INTERCEPT(fstat
);
2534 TSAN_MAYBE_INTERCEPT___FXSTAT
;
2535 TSAN_MAYBE_INTERCEPT_FSTAT64
;
2536 TSAN_MAYBE_INTERCEPT___FXSTAT64
;
2537 TSAN_INTERCEPT(open
);
2538 TSAN_MAYBE_INTERCEPT_OPEN64
;
2539 TSAN_INTERCEPT(creat
);
2540 TSAN_MAYBE_INTERCEPT_CREAT64
;
2541 TSAN_INTERCEPT(dup
);
2542 TSAN_INTERCEPT(dup2
);
2543 TSAN_INTERCEPT(dup3
);
2544 TSAN_MAYBE_INTERCEPT_EVENTFD
;
2545 TSAN_MAYBE_INTERCEPT_SIGNALFD
;
2546 TSAN_MAYBE_INTERCEPT_INOTIFY_INIT
;
2547 TSAN_MAYBE_INTERCEPT_INOTIFY_INIT1
;
2548 TSAN_INTERCEPT(socket
);
2549 TSAN_INTERCEPT(socketpair
);
2550 TSAN_INTERCEPT(connect
);
2551 TSAN_INTERCEPT(bind
);
2552 TSAN_INTERCEPT(listen
);
2553 TSAN_MAYBE_INTERCEPT_EPOLL
;
2554 TSAN_INTERCEPT(close
);
2555 TSAN_MAYBE_INTERCEPT___CLOSE
;
2556 TSAN_MAYBE_INTERCEPT___RES_ICLOSE
;
2557 TSAN_INTERCEPT(pipe
);
2558 TSAN_INTERCEPT(pipe2
);
2560 TSAN_INTERCEPT(unlink
);
2561 TSAN_INTERCEPT(tmpfile
);
2562 TSAN_MAYBE_INTERCEPT_TMPFILE64
;
2563 TSAN_INTERCEPT(fread
);
2564 TSAN_INTERCEPT(fwrite
);
2565 TSAN_INTERCEPT(abort
);
2566 TSAN_INTERCEPT(puts
);
2567 TSAN_INTERCEPT(rmdir
);
2568 TSAN_INTERCEPT(closedir
);
2570 TSAN_INTERCEPT(sigaction
);
2571 TSAN_INTERCEPT(signal
);
2572 TSAN_INTERCEPT(sigsuspend
);
2573 TSAN_INTERCEPT(sigblock
);
2574 TSAN_INTERCEPT(sigsetmask
);
2575 TSAN_INTERCEPT(pthread_sigmask
);
2576 TSAN_INTERCEPT(raise
);
2577 TSAN_INTERCEPT(kill
);
2578 TSAN_INTERCEPT(pthread_kill
);
2579 TSAN_INTERCEPT(sleep
);
2580 TSAN_INTERCEPT(usleep
);
2581 TSAN_INTERCEPT(nanosleep
);
2582 TSAN_INTERCEPT(gettimeofday
);
2583 TSAN_INTERCEPT(getaddrinfo
);
2585 TSAN_INTERCEPT(fork
);
2586 TSAN_INTERCEPT(vfork
);
2587 #if !SANITIZER_ANDROID
2588 TSAN_INTERCEPT(dl_iterate_phdr
);
2590 TSAN_INTERCEPT(on_exit
);
2591 TSAN_INTERCEPT(__cxa_atexit
);
2592 TSAN_INTERCEPT(_exit
);
2594 #ifdef NEED_TLS_GET_ADDR
2595 TSAN_INTERCEPT(__tls_get_addr
);
2598 #if !SANITIZER_MAC && !SANITIZER_ANDROID
2599 // Need to setup it, because interceptors check that the function is resolved.
2600 // But atexit is emitted directly into the module, so can't be resolved.
2601 REAL(atexit
) = (int(*)(void(*)()))unreachable
;
2604 if (REAL(__cxa_atexit
)(&finalize
, 0, 0)) {
2605 Printf("ThreadSanitizer: failed to setup atexit callback\n");
2610 if (pthread_key_create(&g_thread_finalize_key
, &thread_finalize
)) {
2611 Printf("ThreadSanitizer: failed to create thread key\n");
2619 } // namespace __tsan
2621 // Invisible barrier for tests.
2622 // There were several unsuccessful iterations for this functionality:
2623 // 1. Initially it was implemented in user code using
2624 // REAL(pthread_barrier_wait). But pthread_barrier_wait is not supported on
2625 // MacOS. Futexes are linux-specific for this matter.
2626 // 2. Then we switched to atomics+usleep(10). But usleep produced parasitic
2627 // "as-if synchronized via sleep" messages in reports which failed some
2629 // 3. Then we switched to atomics+sched_yield. But this produced tons of tsan-
2630 // visible events, which lead to "failed to restore stack trace" failures.
2631 // Note that no_sanitize_thread attribute does not turn off atomic interception
2632 // so attaching it to the function defined in user code does not help.
2633 // That's why we now have what we have.
2634 extern "C" SANITIZER_INTERFACE_ATTRIBUTE
2635 void __tsan_testonly_barrier_init(u64
*barrier
, u32 count
) {
2636 if (count
>= (1 << 8)) {
2637 Printf("barrier_init: count is too large (%d)\n", count
);
2640 // 8 lsb is thread count, the remaining are count of entered threads.
2644 extern "C" SANITIZER_INTERFACE_ATTRIBUTE
2645 void __tsan_testonly_barrier_wait(u64
*barrier
) {
2646 unsigned old
= __atomic_fetch_add(barrier
, 1 << 8, __ATOMIC_RELAXED
);
2647 unsigned old_epoch
= (old
>> 8) / (old
& 0xff);
2649 unsigned cur
= __atomic_load_n(barrier
, __ATOMIC_RELAXED
);
2650 unsigned cur_epoch
= (cur
>> 8) / (cur
& 0xff);
2651 if (cur_epoch
!= old_epoch
)
2653 internal_sched_yield();