Index: uspace/lib/c/arch/amd64/include/libarch/atomic.h
===================================================================
--- uspace/lib/c/arch/amd64/include/libarch/atomic.h	(revision e0a4686ebb68dd7d42dd083c5ea6c37d6ced027d)
+++ uspace/lib/c/arch/amd64/include/libarch/atomic.h	(revision 57d44dd9c17ddb49818d70775e58b45ccd3511fd)
@@ -46,11 +46,11 @@
 #ifdef __PCC__
 	asm volatile (
-		"lock incq %0\n"
-		: "+m" (val->count)
+	    "lock incq %0\n"
+	    : "+m" (val->count)
 	);
 #else
 	asm volatile (
-		"lock incq %[count]\n"
-		: [count] "+m" (val->count)
+	    "lock incq %[count]\n"
+	    : [count] "+m" (val->count)
 	);
 #endif
@@ -61,11 +61,11 @@
 #ifdef __PCC__
 	asm volatile (
-		"lock decq %0\n"
-		: "+m" (val->count)
+	    "lock decq %0\n"
+	    : "+m" (val->count)
 	);
 #else
 	asm volatile (
-		"lock decq %[count]\n"
-		: [count] "+m" (val->count)
+	    "lock decq %[count]\n"
+	    : [count] "+m" (val->count)
 	);
 #endif
@@ -78,13 +78,13 @@
 #ifdef __PCC__
 	asm volatile (
-		"lock xaddq %1, %0\n"
-		: "+m" (val->count),
-		  "+r" (r)
+	    "lock xaddq %1, %0\n"
+	    : "+m" (val->count),
+	      "+r" (r)
 	);
 #else
 	asm volatile (
-		"lock xaddq %[r], %[count]\n"
-		: [count] "+m" (val->count),
-		  [r] "+r" (r)
+	    "lock xaddq %[r], %[count]\n"
+	    : [count] "+m" (val->count),
+	      [r] "+r" (r)
 	);
 #endif
@@ -99,13 +99,13 @@
 #ifdef __PCC__
 	asm volatile (
-		"lock xaddq %1, %0\n"
-		: "+m" (val->count),
-		  "+r" (r)
+	    "lock xaddq %1, %0\n"
+	    : "+m" (val->count),
+	      "+r" (r)
 	);
 #else
 	asm volatile (
-		"lock xaddq %[r], %[count]\n"
-		: [count] "+m" (val->count),
-		  [r] "+r" (r)
+	    "lock xaddq %[r], %[count]\n"
+	    : [count] "+m" (val->count),
+	      [r] "+r" (r)
 	);
 #endif
