123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273 |
- /*
- * Copyright (C) 1994 - 1999 by Ralf Baechle
- * Copyright (C) 1996 by Paul M. Antoine
- * Copyright (C) 1994 - 1999 by Ralf Baechle
- *
- * Changed set_except_vector declaration to allow return of previous
- * vector address value - necessary for "borrowing" vectors.
- *
- * Kevin D. Kissell, kevink@mips.org and Carsten Langgaard, carstenl@mips.com
- * Copyright (C) 2000 MIPS Technologies, Inc.
- *
- * SPDX-License-Identifier: GPL-2.0
- */
- #ifndef _ASM_SYSTEM_H
- #define _ASM_SYSTEM_H
- #include <asm/sgidefs.h>
- #include <asm/ptrace.h>
- #if 0
- #include <linux/kernel.h>
- #endif
- static __inline__ void
- __sti(void)
- {
- __asm__ __volatile__(
- ".set\tpush\n\t"
- ".set\treorder\n\t"
- ".set\tnoat\n\t"
- "mfc0\t$1,$12\n\t"
- "ori\t$1,0x1f\n\t"
- "xori\t$1,0x1e\n\t"
- "mtc0\t$1,$12\n\t"
- ".set\tpop\n\t"
- : /* no outputs */
- : /* no inputs */
- : "$1", "memory");
- }
- /*
- * For cli() we have to insert nops to make shure that the new value
- * has actually arrived in the status register before the end of this
- * macro.
- * R4000/R4400 need three nops, the R4600 two nops and the R10000 needs
- * no nops at all.
- */
- static __inline__ void
- __cli(void)
- {
- __asm__ __volatile__(
- ".set\tpush\n\t"
- ".set\treorder\n\t"
- ".set\tnoat\n\t"
- "mfc0\t$1,$12\n\t"
- "ori\t$1,1\n\t"
- "xori\t$1,1\n\t"
- ".set\tnoreorder\n\t"
- "mtc0\t$1,$12\n\t"
- "nop\n\t"
- "nop\n\t"
- "nop\n\t"
- ".set\tpop\n\t"
- : /* no outputs */
- : /* no inputs */
- : "$1", "memory");
- }
- #define __save_flags(x) \
- __asm__ __volatile__( \
- ".set\tpush\n\t" \
- ".set\treorder\n\t" \
- "mfc0\t%0,$12\n\t" \
- ".set\tpop\n\t" \
- : "=r" (x))
- #define __save_and_cli(x) \
- __asm__ __volatile__( \
- ".set\tpush\n\t" \
- ".set\treorder\n\t" \
- ".set\tnoat\n\t" \
- "mfc0\t%0,$12\n\t" \
- "ori\t$1,%0,1\n\t" \
- "xori\t$1,1\n\t" \
- ".set\tnoreorder\n\t" \
- "mtc0\t$1,$12\n\t" \
- "nop\n\t" \
- "nop\n\t" \
- "nop\n\t" \
- ".set\tpop\n\t" \
- : "=r" (x) \
- : /* no inputs */ \
- : "$1", "memory")
- #define __restore_flags(flags) \
- do { \
- unsigned long __tmp1; \
- \
- __asm__ __volatile__( \
- ".set\tnoreorder\t\t\t# __restore_flags\n\t" \
- ".set\tnoat\n\t" \
- "mfc0\t$1, $12\n\t" \
- "andi\t%0, 1\n\t" \
- "ori\t$1, 1\n\t" \
- "xori\t$1, 1\n\t" \
- "or\t%0, $1\n\t" \
- "mtc0\t%0, $12\n\t" \
- "nop\n\t" \
- "nop\n\t" \
- "nop\n\t" \
- ".set\tat\n\t" \
- ".set\treorder" \
- : "=r" (__tmp1) \
- : "0" (flags) \
- : "$1", "memory"); \
- } while(0)
- #ifdef CONFIG_SMP
- extern void __global_sti(void);
- extern void __global_cli(void);
- extern unsigned long __global_save_flags(void);
- extern void __global_restore_flags(unsigned long);
- # define sti() __global_sti()
- # define cli() __global_cli()
- # define save_flags(x) do { x = __global_save_flags(); } while (0)
- # define restore_flags(x) __global_restore_flags(x)
- # define save_and_cli(x) do { save_flags(x); cli(); } while(0)
- #else /* Single processor */
- # define sti() __sti()
- # define cli() __cli()
- # define save_flags(x) __save_flags(x)
- # define save_and_cli(x) __save_and_cli(x)
- # define restore_flags(x) __restore_flags(x)
- #endif /* SMP */
- /* For spinlocks etc */
- #define local_irq_save(x) __save_and_cli(x);
- #define local_irq_restore(x) __restore_flags(x);
- #define local_irq_disable() __cli();
- #define local_irq_enable() __sti();
- /*
- * These are probably defined overly paranoid ...
- */
- #ifdef CONFIG_CPU_HAS_WB
- #include <asm/wbflush.h>
- #define rmb() do { } while(0)
- #define wmb() wbflush()
- #define mb() wbflush()
- #else /* CONFIG_CPU_HAS_WB */
- #define mb() \
- __asm__ __volatile__( \
- "# prevent instructions being moved around\n\t" \
- ".set\tnoreorder\n\t" \
- "# 8 nops to fool the R4400 pipeline\n\t" \
- "nop;nop;nop;nop;nop;nop;nop;nop\n\t" \
- ".set\treorder" \
- : /* no output */ \
- : /* no input */ \
- : "memory")
- #define rmb() mb()
- #define wmb() mb()
- #endif /* CONFIG_CPU_HAS_WB */
- #ifdef CONFIG_SMP
- #define smp_mb() mb()
- #define smp_rmb() rmb()
- #define smp_wmb() wmb()
- #else
- #define smp_mb() barrier()
- #define smp_rmb() barrier()
- #define smp_wmb() barrier()
- #endif
- #define set_mb(var, value) \
- do { var = value; mb(); } while (0)
- #define set_wmb(var, value) \
- do { var = value; wmb(); } while (0)
- #if !defined (_LANGUAGE_ASSEMBLY)
- /*
- * switch_to(n) should switch tasks to task nr n, first
- * checking that n isn't the current task, in which case it does nothing.
- */
- #if 0
- extern asmlinkage void *resume(void *last, void *next);
- #endif
- #endif /* !defined (_LANGUAGE_ASSEMBLY) */
- #define prepare_to_switch() do { } while(0)
- #define switch_to(prev,next,last) \
- do { \
- (last) = resume(prev, next); \
- } while(0)
- /*
- * For 32 and 64 bit operands we can take advantage of ll and sc.
- * FIXME: This doesn't work for R3000 machines.
- */
- static __inline__ unsigned long xchg_u32(volatile int * m, unsigned long val)
- {
- #ifdef CONFIG_CPU_HAS_LLSC
- unsigned long dummy;
- __asm__ __volatile__(
- ".set\tnoreorder\t\t\t# xchg_u32\n\t"
- ".set\tnoat\n\t"
- "ll\t%0, %3\n"
- "1:\tmove\t$1, %2\n\t"
- "sc\t$1, %1\n\t"
- "beqzl\t$1, 1b\n\t"
- " ll\t%0, %3\n\t"
- ".set\tat\n\t"
- ".set\treorder"
- : "=r" (val), "=o" (*m), "=r" (dummy)
- : "o" (*m), "2" (val)
- : "memory");
- return val;
- #else
- unsigned long flags, retval;
- save_flags(flags);
- cli();
- retval = *m;
- *m = val;
- restore_flags(flags);
- return retval;
- #endif /* Processor-dependent optimization */
- }
- #define xchg(ptr,x) ((__typeof__(*(ptr)))__xchg((unsigned long)(x),(ptr),sizeof(*(ptr))))
- #define tas(ptr) (xchg((ptr),1))
- static __inline__ unsigned long
- __xchg(unsigned long x, volatile void * ptr, int size)
- {
- switch (size) {
- case 4:
- return xchg_u32(ptr, x);
- }
- return x;
- }
- extern void *set_except_vector(int n, void *addr);
- extern void __die(const char *, struct pt_regs *, const char *where,
- unsigned long line) __attribute__((noreturn));
- extern void __die_if_kernel(const char *, struct pt_regs *, const char *where,
- unsigned long line);
- #define die(msg, regs) \
- __die(msg, regs, __FILE__ ":"__FUNCTION__, __LINE__)
- #define die_if_kernel(msg, regs) \
- __die_if_kernel(msg, regs, __FILE__ ":"__FUNCTION__, __LINE__)
- static inline void execution_hazard_barrier(void)
- {
- __asm__ __volatile__(
- ".set noreorder\n"
- "ehb\n"
- ".set reorder");
- }
- #endif /* _ASM_SYSTEM_H */
|