Index: generic/src/synch/spinlock.c
===================================================================
--- generic/src/synch/spinlock.c	(revision 5c9a08b46f360925ecb95360eb2b35b88a5f8087)
+++ generic/src/synch/spinlock.c	(revision 741ade3f36e60a1a38af7db1c310804979268e06)
@@ -52,5 +52,4 @@
 }
 
-#ifdef CONFIG_DEBUG_SPINLOCK
 /** Lock spinlock
  *
@@ -61,5 +60,6 @@
  * @param sl Pointer to spinlock_t structure.
  */
-void spinlock_lock(spinlock_t *sl)
+#ifdef CONFIG_DEBUG_SPINLOCK
+void spinlock_lock_debug(spinlock_t *sl)
 {
 	count_t i = 0;
@@ -83,29 +83,4 @@
 	if (deadlock_reported)
 		printf("cpu%d: not deadlocked\n", CPU->id);
-
-	/*
-	 * Prevent critical section code from bleeding out this way up.
-	 */
-	CS_ENTER_BARRIER();
-
-}
-
-#else
-
-/** Lock spinlock
- *
- * Lock spinlock.
- *
- * @param sl Pointer to spinlock_t structure.
- */
-void spinlock_lock(spinlock_t *sl)
-{
-	preemption_disable();
-
-	/*
-	 * Each architecture has its own efficient/recommended
-	 * implementation of spinlock.
-	 */
-	spinlock_arch(&sl->val);
 
 	/*
@@ -144,22 +119,3 @@
 }
 
-/** Unlock spinlock
- *
- * Unlock spinlock.
- *
- * @param sl Pointer to spinlock_t structure.
- */
-void spinlock_unlock(spinlock_t *sl)
-{
-	ASSERT(atomic_get(&sl->val) != 0);
-
-	/*
-	 * Prevent critical section code from bleeding out this way down.
-	 */
-	CS_LEAVE_BARRIER();
-	
-	atomic_set(&sl->val,0);
-	preemption_enable();
-}
-
 #endif
