| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522 |
- /** BEGIN COPYRIGHT BLOCK
- * This Program is free software; you can redistribute it and/or modify it under
- * the terms of the GNU General Public License as published by the Free Software
- * Foundation; version 2 of the License.
- *
- * This Program is distributed in the hope that it will be useful, but WITHOUT
- * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
- * FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details.
- *
- * You should have received a copy of the GNU General Public License along with
- * this Program; if not, write to the Free Software Foundation, Inc., 59 Temple
- * Place, Suite 330, Boston, MA 02111-1307 USA.
- *
- * In addition, as a special exception, Red Hat, Inc. gives You the additional
- * right to link the code of this Program with code not covered under the GNU
- * General Public License ("Non-GPL Code") and to distribute linked combinations
- * including the two, subject to the limitations in this paragraph. Non-GPL Code
- * permitted under this exception must only link to the code of this Program
- * through those well defined interfaces identified in the file named EXCEPTION
- * found in the source code files (the "Approved Interfaces"). The files of
- * Non-GPL Code may instantiate templates or use macros or inline functions from
- * the Approved Interfaces without causing the resulting work to be covered by
- * the GNU General Public License. Only Red Hat, Inc. may make changes or
- * additions to the list of Approved Interfaces. You must obey the GNU General
- * Public License in all respects for all of the Program code and other code used
- * in conjunction with the Program except the Non-GPL Code covered by this
- * exception. If you modify this file, you may extend this exception to your
- * version of the file, but you are not obligated to do so. If you do not wish to
- * provide this exception without modification, you must delete this exception
- * statement from your version and license this file solely under the GPL without
- * exception.
- *
- *
- * Copyright (C) 2008 Red Hat, Inc.
- * All rights reserved.
- * END COPYRIGHT BLOCK **/
- #ifdef HAVE_CONFIG_H
- # include <config.h>
- #endif
- #include "slap.h"
- #ifdef SOLARIS
- PRUint64 _sparcv9_AtomicSet_il(PRUint64 *address, PRUint64 newval);
- PRUint64 _sparcv9_AtomicAdd_il(PRUint64 *address, PRUint64 val);
- PRUint64 _sparcv9_AtomicSub_il(PRUint64 *address, PRUint64 val);
- #endif
- #ifdef HPUX
- #ifdef ATOMIC_64BIT_OPERATIONS
- #include <machine/sys/inline.h>
- #endif
- #endif
- #if defined LINUX && (defined CPU_x86 || !HAVE_DECL___SYNC_ADD_AND_FETCH)
- /* On systems that don't have the 64-bit GCC atomic builtins, we need to
- * implement our own atomic functions using inline assembly code. */
- PRUint64 __sync_add_and_fetch_8(PRUint64 *ptr, PRUint64 addval);
- PRUint64 __sync_sub_and_fetch_8(PRUint64 *ptr, PRUint64 subval);
- #endif
- #if defined LINUX && !HAVE_DECL___SYNC_ADD_AND_FETCH
- /* Systems that have the atomic builtins defined, but don't have
- * implementations for 64-bit values will automatically try to
- * call the __sync_*_8 versions we provide. If the atomic builtins
- * are not defined at all, we define them here to use our local
- * functions. */
- #define __sync_add_and_fetch __sync_add_and_fetch_8
- #define __sync_sub_and_fetch __sync_sub_and_fetch_8
- #endif
- /*
- * Counter Structure
- */
- typedef struct slapi_counter {
- PRUint64 value;
- #ifndef ATOMIC_64BIT_OPERATIONS
- Slapi_Mutex *lock;
- #endif
- } slapi_counter;
- /*
- * slapi_counter_new()
- *
- * Allocates and initializes a new Slapi_Counter.
- */
- Slapi_Counter *slapi_counter_new()
- {
- Slapi_Counter *counter = NULL;
- counter = (Slapi_Counter *)slapi_ch_calloc(1, sizeof(Slapi_Counter));
- if (counter != NULL) {
- slapi_counter_init(counter);
- }
- return counter;
- }
- /*
- * slapi_counter_init()
- *
- * Initializes a Slapi_Counter.
- */
- void slapi_counter_init(Slapi_Counter *counter)
- {
- if (counter != NULL) {
- #ifndef ATOMIC_64BIT_OPERATIONS
- /* Create the lock if necessary. */
- if (counter->lock == NULL) {
- counter->lock = slapi_new_mutex();
- }
- #endif
-
- /* Set the value to 0. */
- slapi_counter_set_value(counter, 0);
- }
- }
- /*
- * slapi_counter_destroy()
- *
- * Destroy's a Slapi_Counter and sets the
- * pointer to NULL to prevent reuse.
- */
- void slapi_counter_destroy(Slapi_Counter **counter)
- {
- if ((counter != NULL) && (*counter != NULL)) {
- #ifndef ATOMIC_64BIT_OPERATIONS
- slapi_destroy_mutex((*counter)->lock);
- #endif
- slapi_ch_free((void **)counter);
- }
- }
- /*
- * slapi_counter_increment()
- *
- * Atomically increments a Slapi_Counter.
- */
- PRUint64 slapi_counter_increment(Slapi_Counter *counter)
- {
- return slapi_counter_add(counter, 1);
- }
- /*
- * slapi_counter_decrement()
- *
- * Atomically decrements a Slapi_Counter. Note
- * that this will not prevent you from wrapping
- * around 0.
- */
- PRUint64 slapi_counter_decrement(Slapi_Counter *counter)
- {
- return slapi_counter_subtract(counter, 1);
- }
- /*
- * slapi_counter_add()
- *
- * Atomically add a value to a Slapi_Counter.
- */
- PRUint64 slapi_counter_add(Slapi_Counter *counter, PRUint64 addvalue)
- {
- PRUint64 newvalue = 0;
- #ifdef HPUX
- PRUint64 prev = 0;
- #endif
- if (counter == NULL) {
- return newvalue;
- }
- #ifndef ATOMIC_64BIT_OPERATIONS
- slapi_lock_mutex(counter->lock);
- counter->value += addvalue;
- newvalue = counter->value;
- slapi_unlock_mutex(counter->lock);
- #else
- #ifdef LINUX
- newvalue = __sync_add_and_fetch(&(counter->value), addvalue);
- #elif defined(SOLARIS)
- newvalue = _sparcv9_AtomicAdd(&(counter->value), addvalue);
- #elif defined(HPUX)
- /* fetchadd only works with values of 1, 4, 8, and 16. In addition, it requires
- * it's argument to be an integer constant. */
- if (addvalue == 1) {
- newvalue = _Asm_fetchadd(_FASZ_D, _SEM_ACQ, &(counter->value), 1, _LDHINT_NONE);
- newvalue += 1;
- } else if (addvalue == 4) {
- newvalue = _Asm_fetchadd(_FASZ_D, _SEM_ACQ, &(counter->value), 4, _LDHINT_NONE);
- newvalue += 4;
- } else if (addvalue == 8) {
- newvalue = _Asm_fetchadd(_FASZ_D, _SEM_ACQ, &(counter->value), 8, _LDHINT_NONE);
- newvalue += 8;
- } else if (addvalue == 16) {
- newvalue = _Asm_fetchadd(_FASZ_D, _SEM_ACQ, &(counter->value), 16, _LDHINT_NONE);
- newvalue += 16;
- } else {
- /* For other values, we have to use cmpxchg. */
- do {
- prev = slapi_counter_get_value(counter);
- newvalue = prev + addvalue;
- /* Put prev in a register for cmpxchg to compare against */
- _Asm_mov_to_ar(_AREG_CCV, prev);
- } while (prev != _Asm_cmpxchg(_FASZ_D, _SEM_ACQ, &(counter->value), newvalue, _LDHINT_NONE));
- }
- #endif
- #endif /* ATOMIC_64BIT_OPERATIONS */
- return newvalue;
- }
- /*
- * slapi_counter_subtract()
- *
- * Atomically subtract a value from a Slapi_Counter. Note
- * that this will not prevent you from wrapping around 0.
- */
- PRUint64 slapi_counter_subtract(Slapi_Counter *counter, PRUint64 subvalue)
- {
- PRUint64 newvalue = 0;
- #ifdef HPUX
- PRUint64 prev = 0;
- #endif
- if (counter == NULL) {
- return newvalue;
- }
- #ifndef ATOMIC_64BIT_OPERATIONS
- slapi_lock_mutex(counter->lock);
- counter->value -= subvalue;
- newvalue = counter->value;
- slapi_unlock_mutex(counter->lock);
- #else
- #ifdef LINUX
- newvalue = __sync_sub_and_fetch(&(counter->value), subvalue);
- #elif defined(SOLARIS)
- newvalue = _sparcv9_AtomicSub(&(counter->value), subvalue);
- #elif defined(HPUX)
- /* fetchadd only works with values of -1, -4, -8, and -16. In addition, it requires
- * it's argument to be an integer constant. */
- if (subvalue == 1) {
- newvalue = _Asm_fetchadd(_FASZ_D, _SEM_ACQ, &(counter->value), -1, _LDHINT_NONE);
- newvalue -= 1;
- } else if (subvalue == 4) {
- newvalue = _Asm_fetchadd(_FASZ_D, _SEM_ACQ, &(counter->value), -4, _LDHINT_NONE);
- newvalue -= 4;
- } else if (subvalue == 8) {
- newvalue = _Asm_fetchadd(_FASZ_D, _SEM_ACQ, &(counter->value), -8, _LDHINT_NONE);
- newvalue -= 8;
- } else if (subvalue == 16) {
- newvalue = _Asm_fetchadd(_FASZ_D, _SEM_ACQ, &(counter->value), -16, _LDHINT_NONE);
- newvalue -= 16;
- } else {
- /* For other values, we have to use cmpxchg. */
- do {
- prev = slapi_counter_get_value(counter);
- newvalue = prev - subvalue;
- /* Put prev in a register for cmpxchg to compare against */
- _Asm_mov_to_ar(_AREG_CCV, prev);
- } while (prev != _Asm_cmpxchg(_FASZ_D, _SEM_ACQ, &(counter->value), newvalue, _LDHINT_NONE));
- }
- #endif
- #endif /* ATOMIC_64BIT_OPERATIONS */
- return newvalue;
- }
- /*
- * slapi_counter_set_value()
- *
- * Atomically sets the value of a Slapi_Counter.
- */
- PRUint64 slapi_counter_set_value(Slapi_Counter *counter, PRUint64 newvalue)
- {
- PRUint64 value = 0;
- if (counter == NULL) {
- return value;
- }
- #ifndef ATOMIC_64BIT_OPERATIONS
- slapi_lock_mutex(counter->lock);
- counter->value = newvalue;
- slapi_unlock_mutex(counter->lock);
- return newvalue;
- #else
- #ifdef LINUX
- /* Use our own inline assembly for an atomic set if
- * the builtins aren't available. */
- #if defined CPU_x86 || !HAVE_DECL___SYNC_ADD_AND_FETCH
- /*
- * %0 = counter->value
- * %1 = newvalue
- */
- __asm__ __volatile__(
- #ifdef CPU_x86
- /* Save the PIC register */
- " pushl %%ebx;"
- #endif /* CPU_x86 */
- /* Put value of counter->value in EDX:EAX */
- "retryset: movl %0, %%eax;"
- " movl 4%0, %%edx;"
- /* Put newval in ECX:EBX */
- " movl %1, %%ebx;"
- " movl 4+%1, %%ecx;"
- /* If EDX:EAX and counter-> are the same,
- * replace *ptr with ECX:EBX */
- " lock; cmpxchg8b %0;"
- " jnz retryset;"
- #ifdef CPU_x86
- /* Restore the PIC register */
- " popl %%ebx"
- #endif /* CPU_x86 */
- : "+o" (counter->value)
- : "m" (newvalue)
- #ifdef CPU_x86
- : "memory", "eax", "ecx", "edx", "cc");
- #else
- : "memory", "eax", "ebx", "ecx", "edx", "cc");
- #endif
- return newvalue;
- #else
- while (1) {
- value = counter->value;
- if (__sync_bool_compare_and_swap(&(counter->value), value, newvalue)) {
- return newvalue;
- }
- }
- #endif /* CPU_x86 || !HAVE_DECL___SYNC_ADD_AND_FETCH */
- #elif defined(SOLARIS)
- _sparcv9_AtomicSet(&(counter->value), newvalue);
- return newvalue;
- #elif defined(HPUX)
- do {
- value = counter->value;
- /* Put value in a register for cmpxchg to compare against */
- _Asm_mov_to_ar(_AREG_CCV, value);
- } while (value != _Asm_cmpxchg(_FASZ_D, _SEM_ACQ, &(counter->value), newvalue, _LDHINT_NONE));
- return newvalue;
- #endif
- #endif /* ATOMIC_64BIT_OPERATIONS */
- }
- /*
- * slapi_counter_get_value()
- *
- * Returns the value of a Slapi_Counter.
- */
- PRUint64 slapi_counter_get_value(Slapi_Counter *counter)
- {
- PRUint64 value = 0;
- if (counter == NULL) {
- return value;
- }
- #ifndef ATOMIC_64BIT_OPERATIONS
- slapi_lock_mutex(counter->lock);
- value = counter->value;
- slapi_unlock_mutex(counter->lock);
- #else
- #ifdef LINUX
- /* Use our own inline assembly for an atomic get if
- * the builtins aren't available. */
- #if defined CPU_x86 || !HAVE_DECL___SYNC_ADD_AND_FETCH
- /*
- * %0 = counter->value
- * %1 = value
- */
- __asm__ __volatile__(
- #ifdef CPU_x86
- /* Save the PIC register */
- " pushl %%ebx;"
- #endif /* CPU_x86 */
- /* Put value of counter->value in EDX:EAX */
- "retryget: movl %0, %%eax;"
- " movl 4%0, %%edx;"
- /* Copy EDX:EAX to ECX:EBX */
- " movl %%eax, %%ebx;"
- " movl %%edx, %%ecx;"
- /* If EDX:EAX and counter->value are the same,
- * replace *ptr with ECX:EBX */
- " lock; cmpxchg8b %0;"
- " jnz retryget;"
- /* Put retreived value into value */
- " movl %%ebx, %1;"
- " movl %%ecx, 4%1;"
- #ifdef CPU_x86
- /* Restore the PIC register */
- " popl %%ebx"
- #endif /* CPU_x86 */
- : "+o" (counter->value), "=m" (value)
- :
- #ifdef CPU_x86
- : "memory", "eax", "ecx", "edx", "cc");
- #else
- : "memory", "eax", "ebx", "ecx", "edx", "cc");
- #endif
- #else
- while (1) {
- value = counter->value;
- if (__sync_bool_compare_and_swap(&(counter->value), value, value)) {
- break;
- }
- }
- #endif /* CPU_x86 || !HAVE_DECL___SYNC_ADD_AND_FETCH */
- #elif defined(SOLARIS)
- while (1) {
- value = counter->value;
- if (value == _sparcv9_AtomicSet(&(counter->value), value)) {
- break;
- }
- }
- #elif defined(HPUX)
- do {
- value = counter->value;
- /* Put value in a register for cmpxchg to compare against */
- _Asm_mov_to_ar(_AREG_CCV, value);
- } while (value != _Asm_cmpxchg(_FASZ_D, _SEM_ACQ, &(counter->value), value, _LDHINT_NONE));
- #endif
- #endif /* ATOMIC_64BIT_OPERATIONS */
- return value;
- }
- #if defined LINUX && (defined CPU_x86 || !HAVE_DECL___SYNC_ADD_AND_FETCH)
- /* On systems that don't have the 64-bit GCC atomic builtins, we need to
- * implement our own atomic add and subtract functions using inline
- * assembly code. */
- PRUint64 __sync_add_and_fetch_8(PRUint64 *ptr, PRUint64 addval)
- {
- PRUint64 retval = 0;
- /*
- * %0 = *ptr
- * %1 = retval
- * %2 = addval
- */
- __asm__ __volatile__(
- #ifdef CPU_x86
- /* Save the PIC register */
- " pushl %%ebx;"
- #endif /* CPU_x86 */
- /* Put value of *ptr in EDX:EAX */
- "retryadd: movl %0, %%eax;"
- " movl 4%0, %%edx;"
- /* Put addval in ECX:EBX */
- " movl %2, %%ebx;"
- " movl 4+%2, %%ecx;"
- /* Add value from EDX:EAX to value in ECX:EBX */
- " addl %%eax, %%ebx;"
- " adcl %%edx, %%ecx;"
- /* If EDX:EAX and *ptr are the same, replace ptr with ECX:EBX */
- " lock; cmpxchg8b %0;"
- " jnz retryadd;"
- /* Put new value into retval */
- " movl %%ebx, %1;"
- " movl %%ecx, 4%1;"
- #ifdef CPU_x86
- /* Restore the PIC register */
- " popl %%ebx"
- #endif /* CPU_x86 */
- : "+o" (*ptr), "=m" (retval)
- : "m" (addval)
- #ifdef CPU_x86
- : "memory", "eax", "ecx", "edx", "cc");
- #else
- : "memory", "eax", "ebx", "ecx", "edx", "cc");
- #endif
- return retval;
- }
- PRUint64 __sync_sub_and_fetch_8(PRUint64 *ptr, PRUint64 subval)
- {
- PRUint64 retval = 0;
- /*
- * %0 = *ptr
- * %1 = retval
- * %2 = subval
- */
- __asm__ __volatile__(
- #ifdef CPU_x86
- /* Save the PIC register */
- " pushl %%ebx;"
- #endif /* CPU_x86 */
- /* Put value of *ptr in EDX:EAX */
- "retrysub: movl %0, %%eax;"
- " movl 4%0, %%edx;"
- /* Copy EDX:EAX to ECX:EBX */
- " movl %%eax, %%ebx;"
- " movl %%edx, %%ecx;"
- /* Subtract subval from value in ECX:EBX */
- " subl %2, %%ebx;"
- " sbbl 4+%2, %%ecx;"
- /* If EDX:EAX and ptr are the same, replace *ptr with ECX:EBX */
- " lock; cmpxchg8b %0;"
- " jnz retrysub;"
- /* Put new value into retval */
- " movl %%ebx, %1;"
- " movl %%ecx, 4%1;"
- #ifdef CPU_x86
- /* Restore the PIC register */
- " popl %%ebx"
- #endif /* CPU_x86 */
- : "+o" (*ptr), "=m" (retval)
- : "m" (subval)
- #ifdef CPU_x86
- : "memory", "eax", "ecx", "edx", "cc");
- #else
- : "memory", "eax", "ebx", "ecx", "edx", "cc");
- #endif
- return retval;
- }
- #endif /* LINUX && (defined CPU_x86 || !HAVE_DECL___SYNC_ADD_AND_FETCH) */
|