Index: kernel/generic/src/synch/rcu.c
===================================================================
--- kernel/generic/src/synch/rcu.c	(revision fbe17545297ef393a9495bfdf765d286198df065)
+++ kernel/generic/src/synch/rcu.c	(revision 2708f6aa00d00b5d4b4fa38d184dbe4b24ff32f1)
@@ -237,5 +237,5 @@
 #endif
 	
-	rcu.detector_thr = 0;
+	rcu.detector_thr = NULL;
 	
 	rcu.stat_expedited_cnt = 0;
@@ -260,9 +260,9 @@
 #endif
 	
-	CPU->rcu.cur_cbs = 0;
+	CPU->rcu.cur_cbs = NULL;
 	CPU->rcu.cur_cbs_cnt = 0;
-	CPU->rcu.next_cbs = 0;
+	CPU->rcu.next_cbs = NULL;
 	CPU->rcu.next_cbs_cnt = 0;
-	CPU->rcu.arriving_cbs = 0;
+	CPU->rcu.arriving_cbs = NULL;
 	CPU->rcu.parriving_cbs_tail = &CPU->rcu.arriving_cbs;
 	CPU->rcu.arriving_cbs_cnt = 0;
@@ -275,5 +275,5 @@
 	/* BSP creates reclaimer threads before AP's rcu_cpu_init() runs. */
 	if (config.cpu_active == 1)
-		CPU->rcu.reclaimer_thr = 0;
+		CPU->rcu.reclaimer_thr = NULL;
 	
 	CPU->rcu.stat_max_cbs = 0;
@@ -317,5 +317,5 @@
 	/* Stop and wait for reclaimers. */
 	for (unsigned int cpu_id = 0; cpu_id < config.cpu_active; ++cpu_id) {
-		ASSERT(cpus[cpu_id].rcu.reclaimer_thr != 0);
+		ASSERT(cpus[cpu_id].rcu.reclaimer_thr != NULL);
 	
 		if (cpus[cpu_id].rcu.reclaimer_thr) {
@@ -323,5 +323,5 @@
 			thread_join(cpus[cpu_id].rcu.reclaimer_thr);
 			thread_detach(cpus[cpu_id].rcu.reclaimer_thr);
-			cpus[cpu_id].rcu.reclaimer_thr = 0;
+			cpus[cpu_id].rcu.reclaimer_thr = NULL;
 		}
 	}
@@ -333,5 +333,5 @@
 		thread_join(rcu.detector_thr);
 		thread_detach(rcu.detector_thr);
-		rcu.detector_thr = 0;
+		rcu.detector_thr = NULL;
 	}
 #endif
@@ -357,5 +357,5 @@
 		
 		cpus[cpu_id].rcu.reclaimer_thr = 
-			thread_create(reclaimer, 0, TASK, THREAD_FLAG_NONE, name);
+			thread_create(reclaimer, NULL, TASK, THREAD_FLAG_NONE, name);
 
 		if (!cpus[cpu_id].rcu.reclaimer_thr) 
@@ -373,5 +373,5 @@
 {
 	rcu.detector_thr = 
-		thread_create(detector, 0, TASK, THREAD_FLAG_NONE, "rcu-det");
+		thread_create(detector, NULL, TASK, THREAD_FLAG_NONE, "rcu-det");
 	
 	if (!rcu.detector_thr) 
@@ -410,5 +410,5 @@
 		 * to finish. 
 		 * 
-		 * Note that THREAD may be 0 in scheduler() and not just during boot.
+		 * Note that THREAD may be NULL in scheduler() and not just during boot.
 		 */
 		if ((THREAD && THREAD->rcu.was_preempted) || CPU->rcu.is_delaying_gp) {
@@ -522,5 +522,5 @@
 	
 	cpu_mask_for_each(*cpu_mask, cpu_id) {
-		smp_call(cpu_id, add_barrier_cb, 0);
+		smp_call(cpu_id, add_barrier_cb, NULL);
 	}
 	
@@ -583,5 +583,5 @@
 	
 	rcu_item->func = func;
-	rcu_item->next = 0;
+	rcu_item->next = NULL;
 	
 	preemption_disable();
@@ -611,5 +611,5 @@
 {
 	ASSERT(THREAD && THREAD->wired);
-	return 0 == CPU->rcu.cur_cbs;
+	return NULL == CPU->rcu.cur_cbs;
 }
 
@@ -617,5 +617,5 @@
 {
 	ASSERT(THREAD && THREAD->wired);
-	return 0 == CPU->rcu.next_cbs;
+	return NULL == CPU->rcu.next_cbs;
 }
 
@@ -628,5 +628,5 @@
 	 * a false negative if we race with a local interrupt handler.
 	 */
-	return 0 == CPU->rcu.arriving_cbs;
+	return NULL == CPU->rcu.arriving_cbs;
 }
 
@@ -741,5 +741,5 @@
 	}
 	
-	*phead = 0;
+	*phead = NULL;
 }
 
@@ -779,5 +779,5 @@
 	CPU->rcu.next_cbs_cnt = CPU->rcu.arriving_cbs_cnt;
 	
-	CPU->rcu.arriving_cbs = 0;
+	CPU->rcu.arriving_cbs = NULL;
 	CPU->rcu.parriving_cbs_tail = &CPU->rcu.arriving_cbs;
 	CPU->rcu.arriving_cbs_cnt = 0;
@@ -1289,5 +1289,5 @@
 	atomic_set(&rcu.delaying_cpu_cnt, 0);
 	
-	sample_cpus(cpu_mask, 0);
+	sample_cpus(cpu_mask, NULL);
 }
 
@@ -1452,5 +1452,5 @@
 void rcu_thread_exiting(void)
 {
-	ASSERT(THREAD != 0);
+	ASSERT(THREAD != NULL);
 	ASSERT(THREAD->state == Exiting);
 	ASSERT(PREEMPTION_DISABLED || interrupts_disabled());
