4 /* ECC atomic, DMA, SMP and interrupt safe scrub function */
6 static inline void atomic_scrub(void *va
, u32 size
)
8 unsigned long *virt_addr
= va
;
12 for (i
= 0; i
< size
/ sizeof(unsigned long); i
++, virt_addr
++) {
15 * Very carefully read and write to memory atomically
16 * so we are interrupt, DMA and SMP safe.
18 * Intel: asm("lock; addl $0, %0"::"m"(*virt_addr));
21 __asm__
__volatile__ (
23 "1: ll %0, %1 # atomic_add \n"
24 " ll %0, %1 # atomic_add \n"
29 : "=&r" (temp
), "=m" (*virt_addr
)
This page took 0.034025 seconds and 6 git commands to generate.