| #ifndef __ASM_SH_ATOMIC_GRB_H |
| #define __ASM_SH_ATOMIC_GRB_H |
| |
| #define ATOMIC_OP(op) \ |
| static inline void atomic_##op(int i, atomic_t *v) \ |
| { \ |
| int tmp; \ |
| \ |
| __asm__ __volatile__ ( \ |
| " .align 2 \n\t" \ |
| " mova 1f, r0 \n\t" /* r0 = end point */ \ |
| " mov r15, r1 \n\t" /* r1 = saved sp */ \ |
| " mov #-6, r15 \n\t" /* LOGIN: r15 = size */ \ |
| " mov.l @%1, %0 \n\t" /* load old value */ \ |
| " " #op " %2, %0 \n\t" /* $op */ \ |
| " mov.l %0, @%1 \n\t" /* store new value */ \ |
| "1: mov r1, r15 \n\t" /* LOGOUT */ \ |
| : "=&r" (tmp), \ |
| "+r" (v) \ |
| : "r" (i) \ |
| : "memory" , "r0", "r1"); \ |
| } \ |
| |
| #define ATOMIC_OP_RETURN(op) \ |
| static inline int atomic_##op##_return(int i, atomic_t *v) \ |
| { \ |
| int tmp; \ |
| \ |
| __asm__ __volatile__ ( \ |
| " .align 2 \n\t" \ |
| " mova 1f, r0 \n\t" /* r0 = end point */ \ |
| " mov r15, r1 \n\t" /* r1 = saved sp */ \ |
| " mov #-6, r15 \n\t" /* LOGIN: r15 = size */ \ |
| " mov.l @%1, %0 \n\t" /* load old value */ \ |
| " " #op " %2, %0 \n\t" /* $op */ \ |
| " mov.l %0, @%1 \n\t" /* store new value */ \ |
| "1: mov r1, r15 \n\t" /* LOGOUT */ \ |
| : "=&r" (tmp), \ |
| "+r" (v) \ |
| : "r" (i) \ |
| : "memory" , "r0", "r1"); \ |
| \ |
| return tmp; \ |
| } |
| |
| #define ATOMIC_OPS(op) ATOMIC_OP(op) ATOMIC_OP_RETURN(op) |
| |
| ATOMIC_OPS(add) |
| ATOMIC_OPS(sub) |
| |
| ATOMIC_OP(and) |
| ATOMIC_OP(or) |
| ATOMIC_OP(xor) |
| |
| #undef ATOMIC_OPS |
| #undef ATOMIC_OP_RETURN |
| #undef ATOMIC_OP |
| |
| #endif /* __ASM_SH_ATOMIC_GRB_H */ |