Index: kernel/arch/amd64/include/arch/atomic.h
===================================================================
--- kernel/arch/amd64/include/arch/atomic.h	(revision 013e5d32c9af4ea5eff1d29e41946a398895b296)
+++ kernel/arch/amd64/include/arch/atomic.h	(revision 105fcf0714aa35d68140847af774ff8916200a43)
@@ -45,11 +45,11 @@
 #ifdef CONFIG_SMP
 	asm volatile (
-		"lock incq %[count]\n"
-		: [count] "+m" (val->count)
-	);
-#else
-	asm volatile (
-		"incq %[count]\n"
-		: [count] "+m" (val->count)
+	    "lock incq %[count]\n"
+	    : [count] "+m" (val->count)
+	);
+#else
+	asm volatile (
+	    "incq %[count]\n"
+	    : [count] "+m" (val->count)
 	);
 #endif /* CONFIG_SMP */
@@ -60,11 +60,11 @@
 #ifdef CONFIG_SMP
 	asm volatile (
-		"lock decq %[count]\n"
-		: [count] "+m" (val->count)
-	);
-#else
-	asm volatile (
-		"decq %[count]\n"
-		: [count] "+m" (val->count)
+	    "lock decq %[count]\n"
+	    : [count] "+m" (val->count)
+	);
+#else
+	asm volatile (
+	    "decq %[count]\n"
+	    : [count] "+m" (val->count)
 	);
 #endif /* CONFIG_SMP */
@@ -76,7 +76,7 @@
 
 	asm volatile (
-		"lock xaddq %[r], %[count]\n"
-		: [count] "+m" (val->count),
-		  [r] "+r" (r)
+	    "lock xaddq %[r], %[count]\n"
+	    : [count] "+m" (val->count),
+	      [r] "+r" (r)
 	);
 
@@ -89,7 +89,7 @@
 
 	asm volatile (
-		"lock xaddq %[r], %[count]\n"
-		: [count] "+m" (val->count),
-		  [r] "+r" (r)
+	    "lock xaddq %[r], %[count]\n"
+	    : [count] "+m" (val->count),
+	      [r] "+r" (r)
 	);
 
@@ -105,7 +105,7 @@
 
 	asm volatile (
-		"xchgq %[v], %[count]\n"
-		: [v] "+r" (v),
-		  [count] "+m" (val->count)
+	    "xchgq %[v], %[count]\n"
+	    : [v] "+r" (v),
+	      [count] "+m" (val->count)
 	);
 
@@ -120,16 +120,16 @@
 	preemption_disable();
 	asm volatile (
-		"0:\n"
-		"	pause\n"
-		"	mov %[count], %[tmp]\n"
-		"	testq %[tmp], %[tmp]\n"
-		"	jnz 0b\n"       /* lightweight looping on locked spinlock */
-
-		"	incq %[tmp]\n"  /* now use the atomic operation */
-		"	xchgq %[count], %[tmp]\n"
-		"	testq %[tmp], %[tmp]\n"
-		"	jnz 0b\n"
-		: [count] "+m" (val->count),
-		  [tmp] "=&r" (tmp)
+	    "0:\n"
+	    "	pause\n"
+	    "	mov %[count], %[tmp]\n"
+	    "	testq %[tmp], %[tmp]\n"
+	    "	jnz 0b\n"       /* lightweight looping on locked spinlock */
+
+	    "	incq %[tmp]\n"  /* now use the atomic operation */
+	    "	xchgq %[count], %[tmp]\n"
+	    "	testq %[tmp], %[tmp]\n"
+	    "	jnz 0b\n"
+	    : [count] "+m" (val->count),
+	      [tmp] "=&r" (tmp)
 	);
 
Index: kernel/arch/amd64/include/arch/cpuid.h
===================================================================
--- kernel/arch/amd64/include/arch/cpuid.h	(revision 013e5d32c9af4ea5eff1d29e41946a398895b296)
+++ kernel/arch/amd64/include/arch/cpuid.h	(revision 105fcf0714aa35d68140847af774ff8916200a43)
@@ -55,5 +55,5 @@
 	uint32_t cpuid_ecx;
 	uint32_t cpuid_edx;
-} __attribute__ ((packed)) cpu_info_t;
+} __attribute__((packed)) cpu_info_t;
 
 extern int has_cpuid(void);
Index: kernel/arch/amd64/include/arch/cycle.h
===================================================================
--- kernel/arch/amd64/include/arch/cycle.h	(revision 013e5d32c9af4ea5eff1d29e41946a398895b296)
+++ kernel/arch/amd64/include/arch/cycle.h	(revision 105fcf0714aa35d68140847af774ff8916200a43)
@@ -44,7 +44,7 @@
 
 	asm volatile (
-		"rdtsc\n"
-		: "=a" (lower),
-		  "=d" (upper)
+	    "rdtsc\n"
+	    : "=a" (lower),
+	      "=d" (upper)
 	);
 
Index: kernel/arch/amd64/include/arch/mm/page.h
===================================================================
--- kernel/arch/amd64/include/arch/mm/page.h	(revision 013e5d32c9af4ea5eff1d29e41946a398895b296)
+++ kernel/arch/amd64/include/arch/mm/page.h	(revision 105fcf0714aa35d68140847af774ff8916200a43)
@@ -198,5 +198,5 @@
 	unsigned int accessed : 1;
 	unsigned int dirty : 1;
-	unsigned int unused: 1;
+	unsigned int unused : 1;
 	unsigned int global : 1;
 	unsigned int soft_valid : 1;  /**< Valid content even if present bit is cleared. */
@@ -205,5 +205,5 @@
 	unsigned int addr_32_51 : 21;
 	unsigned int no_execute : 1;
-} __attribute__ ((packed)) pte_t;
+} __attribute__((packed)) pte_t;
 
 NO_TRACE static inline unsigned int get_pt_flags(pte_t *pt, size_t i)
Index: kernel/arch/amd64/include/arch/pm.h
===================================================================
--- kernel/arch/amd64/include/arch/pm.h	(revision 013e5d32c9af4ea5eff1d29e41946a398895b296)
+++ kernel/arch/amd64/include/arch/pm.h	(revision 105fcf0714aa35d68140847af774ff8916200a43)
@@ -37,6 +37,6 @@
 
 #ifndef __ASSEMBLER__
-	#include <typedefs.h>
-	#include <arch/context.h>
+#include <typedefs.h>
+#include <arch/context.h>
 #endif
 
@@ -91,55 +91,55 @@
 
 typedef struct {
-	unsigned limit_0_15: 16;
-	unsigned base_0_15: 16;
-	unsigned base_16_23: 8;
-	unsigned access: 8;
-	unsigned limit_16_19: 4;
-	unsigned available: 1;
-	unsigned longmode: 1;
-	unsigned special: 1;
+	unsigned limit_0_15 : 16;
+	unsigned base_0_15 : 16;
+	unsigned base_16_23 : 8;
+	unsigned access : 8;
+	unsigned limit_16_19 : 4;
+	unsigned available : 1;
+	unsigned longmode : 1;
+	unsigned special : 1;
 	unsigned granularity : 1;
-	unsigned base_24_31: 8;
-} __attribute__ ((packed)) descriptor_t;
+	unsigned base_24_31 : 8;
+} __attribute__((packed)) descriptor_t;
 
 typedef struct {
-	unsigned limit_0_15: 16;
-	unsigned base_0_15: 16;
-	unsigned base_16_23: 8;
-	unsigned type: 4;
+	unsigned limit_0_15 : 16;
+	unsigned base_0_15 : 16;
+	unsigned base_16_23 : 8;
+	unsigned type : 4;
 	unsigned : 1;
 	unsigned dpl : 2;
 	unsigned present : 1;
-	unsigned limit_16_19: 4;
-	unsigned available: 1;
+	unsigned limit_16_19 : 4;
+	unsigned available : 1;
 	unsigned : 2;
 	unsigned granularity : 1;
-	unsigned base_24_31: 8;
+	unsigned base_24_31 : 8;
 	unsigned base_32_63 : 32;
 	unsigned  : 32;
-} __attribute__ ((packed)) tss_descriptor_t;
+} __attribute__((packed)) tss_descriptor_t;
 
 typedef struct {
-	unsigned offset_0_15: 16;
-	unsigned selector: 16;
-	unsigned ist:3;
-	unsigned unused: 5;
-	unsigned type: 5;
-	unsigned dpl: 2;
+	unsigned offset_0_15 : 16;
+	unsigned selector : 16;
+	unsigned ist : 3;
+	unsigned unused : 5;
+	unsigned type : 5;
+	unsigned dpl : 2;
 	unsigned present : 1;
-	unsigned offset_16_31: 16;
-	unsigned offset_32_63: 32;
+	unsigned offset_16_31 : 16;
+	unsigned offset_32_63 : 32;
 	unsigned  : 32;
-} __attribute__ ((packed)) idescriptor_t;
+} __attribute__((packed)) idescriptor_t;
 
 typedef struct {
 	uint16_t limit;
 	uint64_t base;
-} __attribute__ ((packed)) ptr_16_64_t;
+} __attribute__((packed)) ptr_16_64_t;
 
 typedef struct {
 	uint16_t limit;
 	uint32_t base;
-} __attribute__ ((packed)) ptr_16_32_t;
+} __attribute__((packed)) ptr_16_32_t;
 
 typedef struct {
@@ -160,5 +160,5 @@
 	uint16_t iomap_base;
 	uint8_t iomap[TSS_IOMAP_SIZE];
-} __attribute__ ((packed)) tss_t;
+} __attribute__((packed)) tss_t;
 
 extern tss_t *tss_p;
