From 74483c5955a632af5d9a4783cc2b541764450551 Mon Sep 17 00:00:00 2001 From: Rich Felker Date: Thu, 15 Oct 2015 12:04:48 -0400 Subject: [PATCH] mark arm thread-pointer-loading inline asm as volatile this builds on commits a603a75a72bb469c6be4963ed1b55fabe675fe15 and 0ba35d69c0e77b225ec640d2bd112ff6d9d3b2af to ensure that a compiler cannot conclude that it's valid to reorder the asm to a point before the thread pointer is set up, or to treat the inline function as if it were declared with attribute((const)). other archs already use volatile asm for thread pointer loading. --- arch/arm/pthread_arch.h | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/arch/arm/pthread_arch.h b/arch/arm/pthread_arch.h index 5cbb209e..4a4dd09e 100644 --- a/arch/arm/pthread_arch.h +++ b/arch/arm/pthread_arch.h @@ -4,7 +4,7 @@ static inline pthread_t __pthread_self() { char *p; - __asm__( "mrc p15,0,%0,c13,c0,3" : "=r"(p) ); + __asm__ __volatile__ ( "mrc p15,0,%0,c13,c0,3" : "=r"(p) ); return (void *)(p+8-sizeof(struct pthread)); } @@ -14,10 +14,10 @@ static inline pthread_t __pthread_self() { #ifdef __clang__ char *p; - __asm__( "bl __a_gettp\n\tmov %0,r0" : "=r"(p) : : "cc", "r0", "lr" ); + __asm__ __volatile__ ( "bl __a_gettp\n\tmov %0,r0" : "=r"(p) : : "cc", "r0", "lr" ); #else register char *p __asm__("r0"); - __asm__( "bl __a_gettp" : "=r"(p) : : "cc", "lr" ); + __asm__ __volatile__ ( "bl __a_gettp" : "=r"(p) : : "cc", "lr" ); #endif return (void *)(p+8-sizeof(struct pthread)); } -- 2.25.1