patch-2.4.20 linux-2.4.20/include/asm-x86_64/softirq.h

Next file: linux-2.4.20/include/asm-x86_64/spinlock.h
Previous file: linux-2.4.20/include/asm-x86_64/sockios.h
Back to the patch index
Back to the overall index

diff -urN linux-2.4.19/include/asm-x86_64/softirq.h linux-2.4.20/include/asm-x86_64/softirq.h
@@ -0,0 +1,51 @@
+#ifndef __ASM_SOFTIRQ_H
+#define __ASM_SOFTIRQ_H
+
+#include <asm/atomic.h>
+#include <asm/hardirq.h>
+
+#define __cpu_bh_enable(cpu) \
+		do { barrier(); local_bh_count(cpu)--; } while (0)
+#define cpu_bh_disable(cpu) \
+		do { local_bh_count(cpu)++; barrier(); } while (0)
+
+#define local_bh_disable()	cpu_bh_disable(smp_processor_id())
+#define __local_bh_enable()	__cpu_bh_enable(smp_processor_id())
+
+#define in_softirq() (local_bh_count(smp_processor_id()) != 0)
+
+/*
+ * NOTE: this assembly code assumes:
+ *
+ *    (char *)&local_bh_count - 8 == (char *)&softirq_pending
+ *
+ * If you change the offsets in irq_stat then you have to
+ * update this code as well.
+ */
+#define local_bh_enable()						\
+do {									\
+	unsigned int *ptr = &local_bh_count(smp_processor_id());	\
+									\
+	barrier();							\
+	if (!--*ptr)							\
+		__asm__ __volatile__ (					\
+			"cmpl $0, -8(%0);"				\
+			"jnz 2f;"					\
+			"1:;"						\
+									\
+                   ".subsection 1\n" \
+                   ".ifndef _text_lock_" __stringify(KBUILD_BASENAME) "\n" \
+                   "_text_lock_" __stringify(KBUILD_BASENAME) ":\n" \
+                   ".endif\n" \
+			"2:"	\
+			"call do_softirq_thunk;"		\
+			""		\
+			"jmp 1b;"					\
+			".subsection 0;"				\
+									\
+		: /* no output */					\
+		: "r" (ptr)				\
+		/* no registers clobbered */ );				\
+} while (0)
+
+#endif	/* __ASM_SOFTIRQ_H */

FUNET's LINUX-ADM group, linux-adm@nic.funet.fi
TCL-scripts by Sam Shen (who was at: slshen@lbl.gov)