Index: arch/ia32/include/atomic.h
===================================================================
--- arch/ia32/include/atomic.h	(revision 53f98219259f3ae305745185632cbe185fb7c871)
+++ arch/ia32/include/atomic.h	(revision d71007efd9bf1adac91d39ec37befb57a9123574)
@@ -33,16 +33,5 @@
 #include <arch/barrier.h>
 #include <preemption.h>
-
-typedef struct { volatile __u32 count; } atomic_t;
-
-static inline void atomic_set(atomic_t *val, __u32 i)
-{
-	val->count = i;
-}
-
-static inline __u32 atomic_get(atomic_t *val)
-{
-	return val->count;
-}
+#include <typedefs.h>
 
 static inline void atomic_inc(atomic_t *val) {
@@ -62,7 +51,7 @@
 }
 
-static inline count_t atomic_postinc(atomic_t *val) 
+static inline long atomic_postinc(atomic_t *val) 
 {
-	count_t r;
+	long r;
 
 	__asm__ volatile (
@@ -75,7 +64,7 @@
 }
 
-static inline count_t atomic_postdec(atomic_t *val) 
+static inline long atomic_postdec(atomic_t *val) 
 {
-	count_t r;
+	long r;
 	
 	__asm__ volatile (
@@ -103,5 +92,5 @@
 }
 
-/** Ia32 specific fast spinlock */
+/** ia32 specific fast spinlock */
 static inline void atomic_lock_arch(atomic_t *val)
 {
@@ -116,5 +105,5 @@
 		"mov %0, %1;"
 		"testl %1, %1;"
-		"jnz 0b;"       /* Leightweight looping on locked spinlock */
+		"jnz 0b;"       /* Lightweight looping on locked spinlock */
 		
 		"incl %1;"      /* now use the atomic operation */
