[16] | 1 | /*************************************************************************** |
---|
| 2 | * |
---|
| 3 | * Copyright (C) 2001 International Business Machines |
---|
| 4 | * All rights reserved. |
---|
| 5 | * |
---|
| 6 | * This file is part of the GPFS mmfslinux kernel module. |
---|
| 7 | * |
---|
| 8 | * Redistribution and use in source and binary forms, with or without |
---|
| 9 | * modification, are permitted provided that the following conditions |
---|
| 10 | * are met: |
---|
| 11 | * |
---|
| 12 | * 1. Redistributions of source code must retain the above copyright notice, |
---|
| 13 | * this list of conditions and the following disclaimer. |
---|
| 14 | * 2. Redistributions in binary form must reproduce the above copyright |
---|
| 15 | * notice, this list of conditions and the following disclaimer in the |
---|
| 16 | * documentation and/or other materials provided with the distribution. |
---|
| 17 | * 3. The name of the author may not be used to endorse or promote products |
---|
| 18 | * derived from this software without specific prior written |
---|
| 19 | * permission. |
---|
| 20 | * |
---|
| 21 | * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR |
---|
| 22 | * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES |
---|
| 23 | * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. |
---|
| 24 | * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, |
---|
| 25 | * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, |
---|
| 26 | * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; |
---|
| 27 | * OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, |
---|
| 28 | * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR |
---|
| 29 | * OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF |
---|
| 30 | * ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
---|
| 31 | * |
---|
| 32 | *************************************************************************** */ |
---|
| 33 | /* @(#)98 1.20 src/avs/fs/mmfs/ts/kernext/ibm-linux/cxiAtomic-plat.h, mmfs, avs_rgpfs24, rgpfs240610b 6/24/05 15:57:45 */ |
---|
| 34 | /* |
---|
| 35 | * Platform specific synchronization/atomic operations for Linux |
---|
| 36 | * |
---|
| 37 | * Note that these should not be directly invoked; instead use the |
---|
| 38 | * ATOMIC_XXX and ATOMIC_XXXLP macros from <cxiAtomic.h>. |
---|
| 39 | * |
---|
| 40 | * Definitions for baseline atomic operations (long/pointer variants) |
---|
| 41 | * comp_and_swap(lp) |
---|
| 42 | * |
---|
| 43 | * Definitions for atomic operations |
---|
| 44 | * fetch_and_add(lp) |
---|
| 45 | * fetch_and_and(lp) |
---|
| 46 | * fetch_and_or(lp) |
---|
| 47 | * compare_and_swap(lp) |
---|
| 48 | * _check_lock |
---|
| 49 | * |
---|
| 50 | */ |
---|
| 51 | |
---|
| 52 | #ifndef _h_cxiAtomic_plat |
---|
| 53 | #define _h_cxiAtomic_plat |
---|
| 54 | |
---|
| 55 | #ifndef _h_cxiAtomic |
---|
| 56 | #error Platform header (XXX-plat.h) should not be included directly |
---|
| 57 | #endif |
---|
| 58 | |
---|
| 59 | /* NOTE: need to further split this file into architecture specific headers. */ |
---|
| 60 | |
---|
| 61 | #include <cxiTypes.h> |
---|
| 62 | |
---|
| 63 | |
---|
| 64 | /* Memory fencing operations for various architectures */ |
---|
| 65 | |
---|
| 66 | #if defined(GPFS_ARCH_POWER) || defined(GPFS_ARCH_PPC64) |
---|
| 67 | #ifndef CONFIG_UP |
---|
| 68 | #define IMPORT_FENCE __asm__ __volatile__ ("isync" : : ) |
---|
| 69 | #define EXPORT_FENCE __asm__ __volatile__ ("sync" : : ) |
---|
| 70 | #else |
---|
| 71 | #define IMPORT_FENCE ((void)0) |
---|
| 72 | #define EXPORT_FENCE ((void)0) |
---|
| 73 | #endif |
---|
| 74 | /* A complete fence is defined as insuring that the most recently preceding |
---|
| 75 | store is visible to all processors before any subsequent access completes |
---|
| 76 | in storage. For PowerPC MP, the implementations of COMPLETE_FENCE and |
---|
| 77 | EXPORT_FENCE are the same. */ |
---|
| 78 | #define COMPLETE_FENCE EXPORT_FENCE |
---|
| 79 | #endif /* GPFS_ARCH_POWER */ |
---|
| 80 | |
---|
| 81 | #ifdef GPFS_ARCH_I386 |
---|
| 82 | /* Memory in the I386 architecture is always consistent from all processors, |
---|
| 83 | so explicit fence instructions are not needed. */ |
---|
| 84 | #define IMPORT_FENCE ((void)0) |
---|
| 85 | #define EXPORT_FENCE ((void)0) |
---|
| 86 | #define COMPLETE_FENCE ((void)0) |
---|
| 87 | #endif /* GPFS_ARCH_I386 */ |
---|
| 88 | |
---|
| 89 | #ifdef GPFS_ARCH_IA64 |
---|
| 90 | /* Only full/complete memory fence available */ |
---|
| 91 | #define IMPORT_FENCE __asm__ __volatile__ ("mf" : : ) |
---|
| 92 | #define EXPORT_FENCE __asm__ __volatile__ ("mf" : : ) |
---|
| 93 | #define COMPLETE_FENCE __asm__ __volatile__ ("mf" : : ) |
---|
| 94 | #endif |
---|
| 95 | |
---|
| 96 | #ifdef GPFS_ARCH_X86_64 |
---|
| 97 | #define IMPORT_FENCE __asm__ __volatile__ ("sfence":::"memory") |
---|
| 98 | #define EXPORT_FENCE __asm__ __volatile__ ("mfence":::"memory") |
---|
| 99 | #define COMPLETE_FENCE EXPORT_FENCE |
---|
| 100 | #endif |
---|
| 101 | |
---|
| 102 | |
---|
| 103 | /* Baseline atomic operation for i386: comp_and_swap */ |
---|
| 104 | |
---|
| 105 | #if defined(GPFS_ARCH_I386) |
---|
| 106 | |
---|
| 107 | /* Compare the contents of word_addr with the contents of old_val_addr. |
---|
| 108 | If the values are equal, store new_val in word_addr and return 1. |
---|
| 109 | Otherwise, set old_val_addr to the current value of word_addr and |
---|
| 110 | return 0. See ppc64 comp_and_swaplp for details on exception table |
---|
| 111 | code . */ |
---|
| 112 | static inline int |
---|
| 113 | comp_and_swap(volatile int *word_addr, int *old_val_addr, int new_val) |
---|
| 114 | { |
---|
| 115 | unsigned char result; |
---|
| 116 | |
---|
| 117 | __asm__ __volatile__( |
---|
| 118 | "1: lock; cmpxchg %3,%0 \n\ |
---|
| 119 | 2: setz %2 \n\ |
---|
| 120 | .section .fixup, \"ax\" \n\ |
---|
| 121 | 3: jmp 2b \n\ |
---|
| 122 | .previous \n\ |
---|
| 123 | .section __ex_table, \"a\" \n\ |
---|
| 124 | .align 4 \n\ |
---|
| 125 | .long 1b, 3b \n\ |
---|
| 126 | .previous" |
---|
| 127 | |
---|
| 128 | :"=m" (*word_addr), |
---|
| 129 | "=a" (*old_val_addr), |
---|
| 130 | "=&b" (result) |
---|
| 131 | |
---|
| 132 | :"r" (new_val), |
---|
| 133 | "a" (*old_val_addr) |
---|
| 134 | |
---|
| 135 | :"cc", |
---|
| 136 | "memory"); |
---|
| 137 | return result; |
---|
| 138 | } |
---|
| 139 | |
---|
| 140 | #endif /* GPFS_ARCH_I386 */ |
---|
| 141 | |
---|
| 142 | /* Baseline atomic operations for x86_64: comp_and_swap and comp_and_swaplp . |
---|
| 143 | See ppc64 comp_and_swaplp for details on exception table code. */ |
---|
| 144 | |
---|
| 145 | #ifdef GPFS_ARCH_X86_64 |
---|
| 146 | |
---|
| 147 | /* Compare the contents of word_addr with the contents of old_val_addr. |
---|
| 148 | If the values are equal, store new_val in word_addr and return 1. |
---|
| 149 | Otherwise, set old_val_addr to the current value of word_addr and |
---|
| 150 | return 0. See ppc64 comp_and_swaplp for details on exception table |
---|
| 151 | code . */ |
---|
| 152 | static inline int |
---|
| 153 | comp_and_swap(volatile int *word_addr, int *old_val_addr, int new_val) |
---|
| 154 | { |
---|
| 155 | unsigned char result; |
---|
| 156 | |
---|
| 157 | __asm__ __volatile__( |
---|
| 158 | "lock; cmpxchg %3,%0 \n\ |
---|
| 159 | setz %2" |
---|
| 160 | |
---|
| 161 | :"=m" (*word_addr), |
---|
| 162 | "=a" (*old_val_addr), |
---|
| 163 | "=&b" (result) |
---|
| 164 | |
---|
| 165 | :"r" (new_val), |
---|
| 166 | "a" (*old_val_addr) |
---|
| 167 | |
---|
| 168 | :"cc", |
---|
| 169 | "memory"); |
---|
| 170 | return result; |
---|
| 171 | } |
---|
| 172 | |
---|
| 173 | static inline int |
---|
| 174 | comp_and_swaplp(volatile long *word_addr, long *old_val_addr, long new_val) |
---|
| 175 | { |
---|
| 176 | char result; |
---|
| 177 | |
---|
| 178 | __asm__ __volatile__( |
---|
| 179 | "1: lock; cmpxchgq %3,%0 \n\ |
---|
| 180 | 2: setz %2 \n\ |
---|
| 181 | .section .fixup, \"ax\" \n\ |
---|
| 182 | 3: jmp 2b \n\ |
---|
| 183 | .previous \n\ |
---|
| 184 | .section __ex_table, \"a\" \n\ |
---|
| 185 | .align 8 \n\ |
---|
| 186 | .quad 1b, 3b \n\ |
---|
| 187 | .previous" |
---|
| 188 | |
---|
| 189 | :"=m" (*word_addr), |
---|
| 190 | "=a" (*old_val_addr), |
---|
| 191 | "=q" (result) |
---|
| 192 | |
---|
| 193 | :"r" (new_val), |
---|
| 194 | "a" (*old_val_addr) |
---|
| 195 | |
---|
| 196 | :"cc", |
---|
| 197 | "memory"); |
---|
| 198 | return result; |
---|
| 199 | } |
---|
| 200 | #endif /* GPFS_ARCH_X86_64 */ |
---|
| 201 | |
---|
| 202 | |
---|
| 203 | /* Baseline atomic operation for power: comp_and_swap */ |
---|
| 204 | |
---|
| 205 | #if defined(GPFS_ARCH_POWER) || defined(GPFS_ARCH_PPC64) |
---|
| 206 | |
---|
| 207 | /* Compare the contents of word_addr with the contents of old_val_addr. |
---|
| 208 | If the values are equal, store new_val in word_addr and return 1. |
---|
| 209 | Otherwise, set old_val_addr to the current value of word_addr and |
---|
| 210 | return 0. */ |
---|
| 211 | static inline int |
---|
| 212 | comp_and_swap(volatile int *word_addr, int *old_val_addr, int new_val) |
---|
| 213 | { |
---|
| 214 | int result; |
---|
| 215 | |
---|
| 216 | __asm__ __volatile__( |
---|
| 217 | "1: lwarx %0,0,%4 # result = *word_addr \n\ |
---|
| 218 | cmplw cr0,%0,%3 # compare result to *old_val_addr \n\ |
---|
| 219 | bne- 2f # skip to 2: if mismatch \n\ |
---|
| 220 | stwcx. %2,0,%4 # *word_addr = new_val \n\ |
---|
| 221 | bne- 1b # repeat if reservation lost \n\ |
---|
| 222 | li %0,1 # result = 1 \n\ |
---|
| 223 | b 3f \n\ |
---|
| 224 | 2: stw %0,%1 # *old_val_addr = result \n\ |
---|
| 225 | li %0,0 # result = 0 \n\ |
---|
| 226 | 3:" |
---|
| 227 | // output values |
---|
| 228 | : "=&r" (result), // %0 return value + temp variable |
---|
| 229 | "=m" (*old_val_addr) // %1 changed if mismatch |
---|
| 230 | |
---|
| 231 | // input values |
---|
| 232 | : "r" (new_val), // %2 |
---|
| 233 | "r" (*old_val_addr), // %3 |
---|
| 234 | "r" (word_addr) // %4 |
---|
| 235 | : "cr0", "memory"); // "memory" because we modify *word_addr |
---|
| 236 | |
---|
| 237 | return result; |
---|
| 238 | } |
---|
| 239 | #endif /* GPFS_ARCH_POWER */ |
---|
| 240 | |
---|
| 241 | #ifdef GPFS_ARCH_PPC64 |
---|
| 242 | /* This is a regular comp_and_swap function, but with an added twist. |
---|
| 243 | * In SLES9 SP1 ppc64, a patch has been added that modifies the page |
---|
| 244 | * fault handler to search the exceptions table _before_ an actual |
---|
| 245 | * exception happens, in the course of handling a minor page fault |
---|
| 246 | * triggered by a store to a userspace address. If the offending |
---|
| 247 | * instruction is not found in the module exception table, the page |
---|
| 248 | * fault will result in an Oops, even though the dereferenced address |
---|
| 249 | * is actually OK, and would have resulted in a successful store had |
---|
| 250 | * it been given a chance to proceed. This problems occurs in |
---|
| 251 | * internalAcquire, where we have to do some atomic store operations on |
---|
| 252 | * the lockWord that may be located in the userspace. To work around |
---|
| 253 | * this check, we add exception handling code to all ppc64 atomic ops. |
---|
| 254 | * This exception code does absolutely nothing (it transfers control |
---|
| 255 | * back to the instruction following the one that triggered the fault), |
---|
| 256 | * but that doesn't really matter, as we do not expect the exception |
---|
| 257 | * handling code to ever be invoked, we only want search_exception_tables() |
---|
| 258 | * not to return false. If a bad address is passed to internalAcquire, |
---|
| 259 | * we'll get an Oops or assert before getting a chance to run any atomic |
---|
| 260 | * ops. See LTC bugzilla 14533 for more details. */ |
---|
| 261 | static inline int |
---|
| 262 | comp_and_swaplp(volatile long *word_addr, long *old_val_addr, long new_val) |
---|
| 263 | { |
---|
| 264 | long result; |
---|
| 265 | |
---|
| 266 | __asm__ __volatile__( |
---|
| 267 | "1: ldarx %0,0,%4 # result = *word_addr \n\ |
---|
| 268 | 8: cmpld cr0,%0,%3 # compare result to *old_val_addr \n\ |
---|
| 269 | bne- 2f # skip to 2: if mismatch \n\ |
---|
| 270 | 4: stdcx. %2,0,%4 # *word_addr = new_val \n\ |
---|
| 271 | .section .fixup, \"ax\" \n\ |
---|
| 272 | 5: b 6f \n\ |
---|
| 273 | 7: b 8b \n\ |
---|
| 274 | .previous \n\ |
---|
| 275 | .section __ex_table, \"a\" \n\ |
---|
| 276 | .align 3 \n\ |
---|
| 277 | .llong 4b, 5b \n\ |
---|
| 278 | .llong 1b, 7b \n\ |
---|
| 279 | .previous \n\ |
---|
| 280 | 6: bne- 1b # repeat if reservation lost \n\ |
---|
| 281 | li %0,1 # result = 1 \n\ |
---|
| 282 | b 3f \n\ |
---|
| 283 | 2: std %0,%1 # *old_val_addr = result \n\ |
---|
| 284 | li %0,0 # result = 0 \n\ |
---|
| 285 | 3:" |
---|
| 286 | // output values |
---|
| 287 | : "=&r" (result), // %0 return value + temp variable |
---|
| 288 | "=m" (*old_val_addr) // %1 changed if mismatch |
---|
| 289 | |
---|
| 290 | // input values |
---|
| 291 | : "r" (new_val), // %2 |
---|
| 292 | "r" (*old_val_addr), // %3 |
---|
| 293 | "r" (word_addr) // %4 |
---|
| 294 | : "cr0", "memory"); // "memory" because we modify *word_addr |
---|
| 295 | |
---|
| 296 | return (int)result; |
---|
| 297 | } |
---|
| 298 | #endif |
---|
| 299 | |
---|
| 300 | |
---|
| 301 | /* Baseline atomic operations for ia64: comp_and_swap and comp_and_swaplp */ |
---|
| 302 | |
---|
| 303 | /* Found the HP IA64 ISA guide very useful here: |
---|
| 304 | http://devresource.hp.com/devresource/Docs/Refs/IA64ISA/ */ |
---|
| 305 | |
---|
| 306 | #ifdef GPFS_ARCH_IA64 |
---|
| 307 | |
---|
| 308 | #define MASK_LOWER32 0x00000000FFFFFFFFULL |
---|
| 309 | |
---|
| 310 | /* Compare the contents of word_addr with the contents of old_val_addr. |
---|
| 311 | If the values are equal, store new_val in word_addr and return 1. |
---|
| 312 | Otherwise, set old_val_addr to the current value of word_addr and |
---|
| 313 | return 0. */ |
---|
| 314 | |
---|
| 315 | /* compare and swap 4-byte halfword */ |
---|
| 316 | static inline int |
---|
| 317 | comp_and_swap(volatile int *word_addr, int *old_val_addr, int new_val) |
---|
| 318 | { |
---|
| 319 | UInt64 old_val = ((UInt64)*old_val_addr) & MASK_LOWER32; |
---|
| 320 | UInt64 ret_val; |
---|
| 321 | |
---|
| 322 | /* Ensure mov-to-AR[CCV] in separate instruction group/bundle from cmpxchg |
---|
| 323 | to handle RAW dependency */ |
---|
| 324 | __asm__ __volatile__ ("mov ar.ccv=%0\n\ |
---|
| 325 | ;;" |
---|
| 326 | : |
---|
| 327 | : "rO"(old_val)); |
---|
| 328 | /* Use acquire consistancy sem with cmpxchg |
---|
| 329 | (memory write visible to all subsequent data memory accesses) */ |
---|
| 330 | __asm__ __volatile__ ("cmpxchg4.acq %0=[%1],%2,ar.ccv" |
---|
| 331 | |
---|
| 332 | : "=r"(ret_val) |
---|
| 333 | |
---|
| 334 | : "r"(word_addr), |
---|
| 335 | "r"(new_val) |
---|
| 336 | |
---|
| 337 | : "memory"); |
---|
| 338 | |
---|
| 339 | if (ret_val == old_val) |
---|
| 340 | return 1; |
---|
| 341 | else |
---|
| 342 | { |
---|
| 343 | *old_val_addr = (int)ret_val; |
---|
| 344 | return 0; |
---|
| 345 | } |
---|
| 346 | } |
---|
| 347 | |
---|
| 348 | /* compare and swap natural 8-byte word */ |
---|
| 349 | static inline int |
---|
| 350 | comp_and_swaplp(volatile long *word_addr, long *old_val_addr, long new_val) |
---|
| 351 | { |
---|
| 352 | long ret; |
---|
| 353 | |
---|
| 354 | /* Ensure mov-to-AR[CCV] in separate instruction group/bundle from cmpxchg |
---|
| 355 | to handle RAW dependency */ |
---|
| 356 | __asm__ __volatile__ ("mov ar.ccv=%0\n\ |
---|
| 357 | ;;" |
---|
| 358 | : |
---|
| 359 | : "rO"(*old_val_addr)); |
---|
| 360 | |
---|
| 361 | /* Use acquire consistancy sem with cmpxchg |
---|
| 362 | (memory write visible to all subsequent data memory accesses) */ |
---|
| 363 | __asm__ __volatile__ ("cmpxchg8.acq %0=[%1],%2,ar.ccv" |
---|
| 364 | |
---|
| 365 | : "=r"(ret) |
---|
| 366 | |
---|
| 367 | : "r"(word_addr), |
---|
| 368 | "r"(new_val) |
---|
| 369 | |
---|
| 370 | : "memory"); |
---|
| 371 | |
---|
| 372 | if (ret == *old_val_addr) |
---|
| 373 | return 1; |
---|
| 374 | else |
---|
| 375 | { |
---|
| 376 | *old_val_addr = ret; |
---|
| 377 | return 0; |
---|
| 378 | } |
---|
| 379 | } |
---|
| 380 | |
---|
| 381 | #endif /* GPFS_ARCH_IA64 */ |
---|
| 382 | |
---|
| 383 | |
---|
| 384 | /* fetch_and_XXX and fetch_and_XXXlp operations */ |
---|
| 385 | |
---|
| 386 | /* With inlined functions we cannot use the standard trace statements, so |
---|
| 387 | for the atomic operations the USE_LOCK_TRACE must be toggled on to |
---|
| 388 | debug these operations (which fortunately shouldn't happen often). */ |
---|
| 389 | #undef USE_LOCK_TRACE |
---|
| 390 | |
---|
| 391 | #ifdef USE_LOCK_TRACE |
---|
| 392 | #ifdef _KERNEL |
---|
| 393 | #define LOCK_TRACE printk |
---|
| 394 | #else |
---|
| 395 | #define LOCK_TRACE printf |
---|
| 396 | #endif /* _KERNEL */ |
---|
| 397 | #else |
---|
| 398 | #define LOCK_TRACE(X1,X2,X3,X4,X5,X6) |
---|
| 399 | #endif /* USE_LOCK_TRACE */ |
---|
| 400 | |
---|
| 401 | static inline int |
---|
| 402 | fetch_and_add(atomic_p wd, int i) |
---|
| 403 | { |
---|
| 404 | int ret, oldVal, newVal; |
---|
| 405 | oldVal = cxiSafeGetInt(wd); |
---|
| 406 | |
---|
| 407 | do |
---|
| 408 | { |
---|
| 409 | newVal = oldVal + i; |
---|
| 410 | ret = comp_and_swap((volatile int *)wd, &oldVal, newVal); |
---|
| 411 | |
---|
| 412 | LOCK_TRACE( |
---|
| 413 | "fetch_and_add: wd 0x%lX *wd 0x%lX old 0x%lX new 0x%lX ret %d\n", |
---|
| 414 | wd, *wd, oldVal, newVal, ret); |
---|
| 415 | |
---|
| 416 | } while (ret == 0); |
---|
| 417 | |
---|
| 418 | return oldVal; |
---|
| 419 | } |
---|
| 420 | |
---|
| 421 | #ifdef __64BIT__ |
---|
| 422 | static inline long |
---|
| 423 | fetch_and_addlp(atomic_l wd, long i) |
---|
| 424 | { |
---|
| 425 | long oldVal, newVal; |
---|
| 426 | int ret; |
---|
| 427 | |
---|
| 428 | oldVal = cxiSafeGetLong(wd); |
---|
| 429 | |
---|
| 430 | do |
---|
| 431 | { |
---|
| 432 | newVal = oldVal + i; |
---|
| 433 | ret = comp_and_swaplp((volatile long *)wd, &oldVal, newVal); |
---|
| 434 | |
---|
| 435 | LOCK_TRACE( |
---|
| 436 | "fetch_and_addlp: wd 0x%lX *wd 0x%lX old 0x%lX new 0x%lX ret %d\n", |
---|
| 437 | wd, *wd, oldVal, newVal, ret); |
---|
| 438 | |
---|
| 439 | } while (ret == 0); |
---|
| 440 | |
---|
| 441 | return oldVal; |
---|
| 442 | } |
---|
| 443 | #endif /* __64BIT__ */ |
---|
| 444 | |
---|
| 445 | static inline int |
---|
| 446 | fetch_and_and(atomic_p wd, uint mask) |
---|
| 447 | { |
---|
| 448 | int ret, oldVal,newVal; |
---|
| 449 | oldVal = cxiSafeGetInt(wd); |
---|
| 450 | |
---|
| 451 | do |
---|
| 452 | { |
---|
| 453 | newVal = oldVal & mask; |
---|
| 454 | ret = comp_and_swap((volatile int *)wd, &oldVal, newVal); |
---|
| 455 | |
---|
| 456 | LOCK_TRACE( |
---|
| 457 | "fetch_and_and: wd 0x%lX *wd 0x%lX old 0x%lX new 0x%lX ret %d\n", |
---|
| 458 | wd, *wd, oldVal, newVal, ret); |
---|
| 459 | |
---|
| 460 | } while (ret == 0); |
---|
| 461 | |
---|
| 462 | return oldVal; |
---|
| 463 | } |
---|
| 464 | |
---|
| 465 | #ifdef __64BIT__ |
---|
| 466 | static inline long |
---|
| 467 | fetch_and_andlp(atomic_l wd, ulong mask) |
---|
| 468 | { |
---|
| 469 | long oldVal,newVal; |
---|
| 470 | int ret; |
---|
| 471 | oldVal = cxiSafeGetLong(wd); |
---|
| 472 | |
---|
| 473 | do |
---|
| 474 | { |
---|
| 475 | newVal = oldVal & mask; |
---|
| 476 | ret = comp_and_swaplp((volatile long *)wd, &oldVal, newVal); |
---|
| 477 | |
---|
| 478 | LOCK_TRACE( |
---|
| 479 | "fetch_and_andlp: wd 0x%lX *wd 0x%lX old 0x%lX new 0x%lX ret %d\n", |
---|
| 480 | wd, *wd, oldVal, newVal, ret); |
---|
| 481 | |
---|
| 482 | } while (ret == 0); |
---|
| 483 | |
---|
| 484 | return oldVal; |
---|
| 485 | } |
---|
| 486 | #endif /* __64BIT__ */ |
---|
| 487 | |
---|
| 488 | static inline int |
---|
| 489 | fetch_and_or(atomic_p wd, uint mask) |
---|
| 490 | { |
---|
| 491 | int ret, oldVal,newVal; |
---|
| 492 | oldVal = cxiSafeGetInt(wd); |
---|
| 493 | |
---|
| 494 | do |
---|
| 495 | { |
---|
| 496 | newVal = oldVal | mask; |
---|
| 497 | ret = comp_and_swap((volatile int *)wd, &oldVal, newVal); |
---|
| 498 | |
---|
| 499 | LOCK_TRACE( |
---|
| 500 | "fetch_and_or: wd 0x%lX *wd 0x%lX old 0x%lX new 0x%lX ret %d\n", |
---|
| 501 | wd, *wd, oldVal, newVal, ret); |
---|
| 502 | |
---|
| 503 | } while (ret == 0); |
---|
| 504 | |
---|
| 505 | return oldVal; |
---|
| 506 | } |
---|
| 507 | |
---|
| 508 | #ifdef __64BIT__ |
---|
| 509 | static inline long |
---|
| 510 | fetch_and_orlp(atomic_l wd, ulong mask) |
---|
| 511 | { |
---|
| 512 | long oldVal,newVal; |
---|
| 513 | int ret; |
---|
| 514 | oldVal = cxiSafeGetLong(wd); |
---|
| 515 | |
---|
| 516 | do |
---|
| 517 | { |
---|
| 518 | newVal = oldVal | mask; |
---|
| 519 | ret = comp_and_swaplp((volatile long *)wd, &oldVal, newVal); |
---|
| 520 | |
---|
| 521 | LOCK_TRACE( |
---|
| 522 | "fetch_and_orlp: wd 0x%lX *wd 0x%lX old 0x%lX new 0x%lX ret %d\n", |
---|
| 523 | wd, *wd, oldVal, newVal, ret); |
---|
| 524 | |
---|
| 525 | } while (ret == 0); |
---|
| 526 | |
---|
| 527 | return oldVal; |
---|
| 528 | } |
---|
| 529 | #endif /* __64BIT__ */ |
---|
| 530 | |
---|
| 531 | static inline Boolean |
---|
| 532 | compare_and_swap(atomic_p wd, int *oldVal, int newVal) |
---|
| 533 | { |
---|
| 534 | Boolean ret; |
---|
| 535 | |
---|
| 536 | ret = comp_and_swap((volatile int *)wd, oldVal, newVal); |
---|
| 537 | |
---|
| 538 | LOCK_TRACE( |
---|
| 539 | "compare_and_swap out: wd 0x%lX *wd 0x%lX old 0x%lX " |
---|
| 540 | "new 0x%lX ret %d\n", wd, *wd, *oldVal, newVal, ret); |
---|
| 541 | return ret; |
---|
| 542 | } |
---|
| 543 | |
---|
| 544 | #ifdef __64BIT__ |
---|
| 545 | static inline Boolean |
---|
| 546 | compare_and_swaplp(atomic_l wd, long *oldVal, long newVal) |
---|
| 547 | { |
---|
| 548 | Boolean ret; |
---|
| 549 | |
---|
| 550 | ret = comp_and_swaplp((volatile long *)wd, oldVal, newVal); |
---|
| 551 | |
---|
| 552 | LOCK_TRACE( |
---|
| 553 | "compare_and_swaplp out: wd 0x%lX *wd 0x%lX old 0x%lX " |
---|
| 554 | "new 0x%lX ret %d\n", wd, *wd, *oldVal, newVal, ret); |
---|
| 555 | return ret; |
---|
| 556 | } |
---|
| 557 | #endif /* __64BIT__ */ |
---|
| 558 | |
---|
| 559 | static inline Boolean |
---|
| 560 | _check_lock(atomic_p wd, int oldVal, int newVal) |
---|
| 561 | { |
---|
| 562 | int old_val_addr = oldVal; |
---|
| 563 | Boolean ret; |
---|
| 564 | |
---|
| 565 | ret = comp_and_swap((volatile int *) wd, &old_val_addr, newVal); |
---|
| 566 | |
---|
| 567 | LOCK_TRACE( |
---|
| 568 | "_check_lock: wd 0x%X *wd 0x%X old 0x%X new 0x%X ret %d\n", |
---|
| 569 | wd, *wd, old_val_addr, newVal, ret); |
---|
| 570 | |
---|
| 571 | if (ret) |
---|
| 572 | { |
---|
| 573 | IMPORT_FENCE; |
---|
| 574 | return 0; |
---|
| 575 | } |
---|
| 576 | else |
---|
| 577 | return 1; |
---|
| 578 | } |
---|
| 579 | |
---|
| 580 | #ifdef __64BIT__ |
---|
| 581 | static inline Boolean |
---|
| 582 | _check_locklp(atomic_l wd, long oldVal, long newVal) |
---|
| 583 | { |
---|
| 584 | long old_val_addr = oldVal; |
---|
| 585 | Boolean ret; |
---|
| 586 | |
---|
| 587 | ret = comp_and_swaplp((volatile long *) wd, &old_val_addr, newVal); |
---|
| 588 | |
---|
| 589 | LOCK_TRACE( |
---|
| 590 | "_check_locklp: wd 0x%lX *wd 0x%lX old 0x%lX new 0x%lX ret %d\n", |
---|
| 591 | wd, *wd, old_val_addr, newVal, ret); |
---|
| 592 | |
---|
| 593 | if (ret) |
---|
| 594 | { |
---|
| 595 | IMPORT_FENCE; |
---|
| 596 | return 0; |
---|
| 597 | } |
---|
| 598 | else |
---|
| 599 | return 1; |
---|
| 600 | } |
---|
| 601 | #endif /* __64BIT__ */ |
---|
| 602 | |
---|
| 603 | #endif /* _h_cxiAtomic_plat */ |
---|