/* $OpenBSD: atomic.h,v 1.13 2014/09/11 20:32:16 kettenis Exp $ */ /* $NetBSD: atomic.h,v 1.1.2.2 2000/02/21 18:54:07 sommerfeld Exp $ */ /*- * Copyright (c) 2000 The NetBSD Foundation, Inc. * All rights reserved. * * This code is derived from software contributed to The NetBSD Foundation * by RedBack Networks Inc. * * Author: Bill Sommerfeld * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE * POSSIBILITY OF SUCH DAMAGE. */ #ifndef _MACHINE_ATOMIC_H_ #define _MACHINE_ATOMIC_H_ /* * Perform atomic operations on memory. Should be atomic with respect * to interrupts and multiple processors. * * void atomic_setbits_int(volatile u_int *a, u_int mask) { *a |= mask; } * void atomic_clearbits_int(volatile u_int *a, u_int mas) { *a &= ~mask; } */ #if defined(_KERNEL) && !defined(_LOCORE) #ifdef MULTIPROCESSOR #define LOCK "lock" #else #define LOCK #endif static __inline u_int64_t i386_atomic_testset_uq(volatile u_int64_t *ptr, u_int64_t val) { __asm__ volatile ("\n1:\t" LOCK " cmpxchg8b (%1); jnz 1b" : "+A" (val) : "r" (ptr), "b" ((u_int32_t)val), "c" ((u_int32_t)(val >> 32))); return val; } static __inline u_int32_t i386_atomic_testset_ul(volatile u_int32_t *ptr, unsigned long val) { __asm__ volatile ("xchgl %0,(%2)" :"=r" (val):"0" (val),"r" (ptr)); return val; } static __inline int i386_atomic_testset_i(volatile int *ptr, unsigned long val) { __asm__ volatile ("xchgl %0,(%2)" :"=r" (val):"0" (val),"r" (ptr)); return val; } static __inline void i386_atomic_setbits_l(volatile u_int32_t *ptr, unsigned long bits) { __asm volatile(LOCK " orl %1,%0" : "=m" (*ptr) : "ir" (bits)); } static __inline void i386_atomic_clearbits_l(volatile u_int32_t *ptr, unsigned long bits) { bits = ~bits; __asm volatile(LOCK " andl %1,%0" : "=m" (*ptr) : "ir" (bits)); } /* * cas = compare and set */ static __inline int i486_atomic_cas_int(volatile u_int *ptr, u_int expect, u_int set) { int res; __asm volatile(LOCK " cmpxchgl %2, %1" : "=a" (res), "=m" (*ptr) : "r" (set), "a" (expect), "m" (*ptr) : "memory"); return (res); } static __inline int i386_atomic_cas_int32(volatile int32_t *ptr, int32_t expect, int32_t set) { int res; __asm volatile(LOCK " cmpxchgl %2, %1" : "=a" (res), "=m" (*ptr) : "r" (set), "a" (expect), "m" (*ptr) : "memory"); return (res); } /* * The IA-32 architecture is rather strongly ordered. When accessing * normal write-back cachable memory, only reads may be reordered with * older writes to different locations. There are a few instructions * (clfush, non-temporal move instructions) that obey weaker ordering * rules, but those instructions will only be used in (inline) * assembly code where we can add the necessary fence instructions * ourselves. */ #define __membar(_f) do { __asm __volatile(_f ::: "memory"); } while (0) #ifdef MULTIPROCESSOR #define membar_enter() __membar("lock; addl $0,0(%%esp)") #define membar_exit() __membar("") #define membar_producer() __membar("") #define membar_consumer() __membar("") #define membar_sync() __membar("lock; addl $0,0(%%esp)") #else #define membar_enter() __membar("") #define membar_exit() __membar("") #define membar_producer() __membar("") #define membar_consumer() __membar("") #define membar_sync() __membar("") #endif int ucas_32(volatile int32_t *, int32_t, int32_t); #define futex_atomic_ucas_int32 ucas_32 #define atomic_setbits_int i386_atomic_setbits_l #define atomic_clearbits_int i386_atomic_clearbits_l #undef LOCK #endif /* defined(_KERNEL) && !defined(_LOCORE) */ #endif /* _MACHINE_ATOMIC_H_ */