sgi-xp: replace AMO_t typedef by struct amo

Replace the AMO_t typedef by a direct reference to 'struct amo'.

Signed-off-by: Dean Nelson <dcn@sgi.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
diff --git a/drivers/misc/sgi-xp/xpc_sn2.c b/drivers/misc/sgi-xp/xpc_sn2.c
index 0fef7d8..01dd40e 100644
--- a/drivers/misc/sgi-xp/xpc_sn2.c
+++ b/drivers/misc/sgi-xp/xpc_sn2.c
@@ -111,13 +111,14 @@
  */
 
 static u64
-xpc_receive_IRQ_amo_sn2(AMO_t *amo)
+xpc_receive_IRQ_amo_sn2(struct amo *amo)
 {
 	return FETCHOP_LOAD_OP(TO_AMO((u64)&amo->variable), FETCHOP_CLEAR);
 }
 
 static enum xp_retval
-xpc_send_IRQ_sn2(AMO_t *amo, u64 flag, int nasid, int phys_cpuid, int vector)
+xpc_send_IRQ_sn2(struct amo *amo, u64 flag, int nasid, int phys_cpuid,
+		 int vector)
 {
 	int ret = 0;
 	unsigned long irq_flags;
@@ -131,7 +132,7 @@
 	 * We must always use the nofault function regardless of whether we
 	 * are on a Shub 1.1 system or a Shub 1.2 slice 0xc processor. If we
 	 * didn't, we'd never know that the other partition is down and would
-	 * keep sending IRQs and AMOs to it until the heartbeat times out.
+	 * keep sending IRQs and amos to it until the heartbeat times out.
 	 */
 	ret = xp_nofault_PIOR((u64 *)GLOBAL_MMR_ADDR(NASID_GET(&amo->variable),
 						     xp_nofault_PIOR_target));
@@ -141,12 +142,12 @@
 	return ((ret == 0) ? xpSuccess : xpPioReadError);
 }
 
-static AMO_t *
+static struct amo *
 xpc_init_IRQ_amo_sn2(int index)
 {
-	AMO_t *amo = xpc_vars->amos_page + index;
+	struct amo *amo = xpc_vars->amos_page + index;
 
-	(void)xpc_receive_IRQ_amo_sn2(amo);	/* clear AMO variable */
+	(void)xpc_receive_IRQ_amo_sn2(amo);	/* clear amo variable */
 	return amo;
 }
 
@@ -166,7 +167,7 @@
 }
 
 /*
- * Flag the appropriate AMO variable and send an IRQ to the specified node.
+ * Flag the appropriate amo variable and send an IRQ to the specified node.
  */
 static void
 xpc_send_activate_IRQ_sn2(u64 amos_page_pa, int from_nasid, int to_nasid,
@@ -174,8 +175,9 @@
 {
 	int w_index = XPC_NASID_W_INDEX(from_nasid);
 	int b_index = XPC_NASID_B_INDEX(from_nasid);
-	AMO_t *amos = (AMO_t *)__va(amos_page_pa +
-				    (XPC_ACTIVATE_IRQ_AMOS * sizeof(AMO_t)));
+	struct amo *amos = (struct amo *)__va(amos_page_pa +
+					      (XPC_ACTIVATE_IRQ_AMOS *
+					      sizeof(struct amo)));
 
 	(void)xpc_send_IRQ_sn2(&amos[w_index], (1UL << b_index), to_nasid,
 			       to_phys_cpuid, SGI_XPC_ACTIVATE);
@@ -186,8 +188,9 @@
 {
 	int w_index = XPC_NASID_W_INDEX(from_nasid);
 	int b_index = XPC_NASID_B_INDEX(from_nasid);
-	AMO_t *amos = (AMO_t *)__va(xpc_vars->amos_page_pa +
-				    (XPC_ACTIVATE_IRQ_AMOS * sizeof(AMO_t)));
+	struct amo *amos = (struct amo *)__va(xpc_vars->amos_page_pa +
+					      (XPC_ACTIVATE_IRQ_AMOS *
+					      sizeof(struct amo)));
 
 	/* fake the sending and receipt of an activate IRQ from remote nasid */
 	FETCHOP_STORE_OP(TO_AMO((u64)&amos[w_index].variable), FETCHOP_OR,
@@ -227,7 +230,7 @@
 /*
  * Handle the receipt of a SGI_XPC_NOTIFY IRQ by seeing whether the specified
  * partition actually sent it. Since SGI_XPC_NOTIFY IRQs may be shared by more
- * than one partition, we use an AMO_t structure per partition to indicate
+ * than one partition, we use an amo structure per partition to indicate
  * whether a partition has sent an IRQ or not.  If it has, then wake up the
  * associated kthread to handle it.
  *
@@ -391,20 +394,20 @@
 xpc_indicate_partition_engaged_sn2(struct xpc_partition *part)
 {
 	unsigned long irq_flags;
-	AMO_t *amo = (AMO_t *)__va(part->sn.sn2.remote_amos_page_pa +
-				   (XPC_ENGAGED_PARTITIONS_AMO *
-				    sizeof(AMO_t)));
+	struct amo *amo = (struct amo *)__va(part->sn.sn2.remote_amos_page_pa +
+					     (XPC_ENGAGED_PARTITIONS_AMO *
+					     sizeof(struct amo)));
 
 	local_irq_save(irq_flags);
 
-	/* set bit corresponding to our partid in remote partition's AMO */
+	/* set bit corresponding to our partid in remote partition's amo */
 	FETCHOP_STORE_OP(TO_AMO((u64)&amo->variable), FETCHOP_OR,
 			 (1UL << sn_partition_id));
 	/*
 	 * We must always use the nofault function regardless of whether we
 	 * are on a Shub 1.1 system or a Shub 1.2 slice 0xc processor. If we
 	 * didn't, we'd never know that the other partition is down and would
-	 * keep sending IRQs and AMOs to it until the heartbeat times out.
+	 * keep sending IRQs and amos to it until the heartbeat times out.
 	 */
 	(void)xp_nofault_PIOR((u64 *)GLOBAL_MMR_ADDR(NASID_GET(&amo->
 							       variable),
@@ -418,20 +421,20 @@
 {
 	struct xpc_partition_sn2 *part_sn2 = &part->sn.sn2;
 	unsigned long irq_flags;
-	AMO_t *amo = (AMO_t *)__va(part_sn2->remote_amos_page_pa +
-				   (XPC_ENGAGED_PARTITIONS_AMO *
-				    sizeof(AMO_t)));
+	struct amo *amo = (struct amo *)__va(part_sn2->remote_amos_page_pa +
+					     (XPC_ENGAGED_PARTITIONS_AMO *
+					     sizeof(struct amo)));
 
 	local_irq_save(irq_flags);
 
-	/* clear bit corresponding to our partid in remote partition's AMO */
+	/* clear bit corresponding to our partid in remote partition's amo */
 	FETCHOP_STORE_OP(TO_AMO((u64)&amo->variable), FETCHOP_AND,
 			 ~(1UL << sn_partition_id));
 	/*
 	 * We must always use the nofault function regardless of whether we
 	 * are on a Shub 1.1 system or a Shub 1.2 slice 0xc processor. If we
 	 * didn't, we'd never know that the other partition is down and would
-	 * keep sending IRQs and AMOs to it until the heartbeat times out.
+	 * keep sending IRQs and amos to it until the heartbeat times out.
 	 */
 	(void)xp_nofault_PIOR((u64 *)GLOBAL_MMR_ADDR(NASID_GET(&amo->
 							       variable),
@@ -441,7 +444,7 @@
 
 	/*
 	 * Send activate IRQ to get other side to see that we've cleared our
-	 * bit in their engaged partitions AMO.
+	 * bit in their engaged partitions amo.
 	 */
 	xpc_send_activate_IRQ_sn2(part_sn2->remote_amos_page_pa,
 				  cnodeid_to_nasid(0),
@@ -452,9 +455,9 @@
 static int
 xpc_partition_engaged_sn2(short partid)
 {
-	AMO_t *amo = xpc_vars->amos_page + XPC_ENGAGED_PARTITIONS_AMO;
+	struct amo *amo = xpc_vars->amos_page + XPC_ENGAGED_PARTITIONS_AMO;
 
-	/* our partition's AMO variable ANDed with partid mask */
+	/* our partition's amo variable ANDed with partid mask */
 	return (FETCHOP_LOAD_OP(TO_AMO((u64)&amo->variable), FETCHOP_LOAD) &
 		(1UL << partid)) != 0;
 }
@@ -462,18 +465,18 @@
 static int
 xpc_any_partition_engaged_sn2(void)
 {
-	AMO_t *amo = xpc_vars->amos_page + XPC_ENGAGED_PARTITIONS_AMO;
+	struct amo *amo = xpc_vars->amos_page + XPC_ENGAGED_PARTITIONS_AMO;
 
-	/* our partition's AMO variable */
+	/* our partition's amo variable */
 	return FETCHOP_LOAD_OP(TO_AMO((u64)&amo->variable), FETCHOP_LOAD) != 0;
 }
 
 static void
 xpc_assume_partition_disengaged_sn2(short partid)
 {
-	AMO_t *amo = xpc_vars->amos_page + XPC_ENGAGED_PARTITIONS_AMO;
+	struct amo *amo = xpc_vars->amos_page + XPC_ENGAGED_PARTITIONS_AMO;
 
-	/* clear bit(s) based on partid mask in our partition's AMO */
+	/* clear bit(s) based on partid mask in our partition's amo */
 	FETCHOP_STORE_OP(TO_AMO((u64)&amo->variable), FETCHOP_AND,
 			 ~(1UL << partid));
 }
@@ -482,10 +485,10 @@
 static u64 xpc_prot_vec_sn2[MAX_NUMNODES];
 
 /*
- * Change protections to allow AMO operations on non-Shub 1.1 systems.
+ * Change protections to allow amo operations on non-Shub 1.1 systems.
  */
 static enum xp_retval
-xpc_allow_AMO_ops_sn2(AMO_t *amos_page)
+xpc_allow_amo_ops_sn2(struct amo *amos_page)
 {
 	u64 nasid_array = 0;
 	int ret;
@@ -493,7 +496,7 @@
 	/*
 	 * On SHUB 1.1, we cannot call sn_change_memprotect() since the BIST
 	 * collides with memory operations. On those systems we call
-	 * xpc_allow_AMO_ops_shub_wars_1_1_sn2() instead.
+	 * xpc_allow_amo_ops_shub_wars_1_1_sn2() instead.
 	 */
 	if (!enable_shub_wars_1_1()) {
 		ret = sn_change_memprotect(ia64_tpa((u64)amos_page), PAGE_SIZE,
@@ -506,10 +509,10 @@
 }
 
 /*
- * Change protections to allow AMO operations on Shub 1.1 systems.
+ * Change protections to allow amo operations on Shub 1.1 systems.
  */
 static void
-xpc_allow_AMO_ops_shub_wars_1_1_sn2(void)
+xpc_allow_amo_ops_shub_wars_1_1_sn2(void)
 {
 	int node;
 	int nasid;
@@ -536,7 +539,7 @@
 static enum xp_retval
 xpc_rsvd_page_init_sn2(struct xpc_rsvd_page *rp)
 {
-	AMO_t *amos_page;
+	struct amo *amos_page;
 	int i;
 	int ret;
 
@@ -549,32 +552,32 @@
 						     XPC_RP_VARS_SIZE);
 
 	/*
-	 * Before clearing xpc_vars, see if a page of AMOs had been previously
+	 * Before clearing xpc_vars, see if a page of amos had been previously
 	 * allocated. If not we'll need to allocate one and set permissions
-	 * so that cross-partition AMOs are allowed.
+	 * so that cross-partition amos are allowed.
 	 *
-	 * The allocated AMO page needs MCA reporting to remain disabled after
+	 * The allocated amo page needs MCA reporting to remain disabled after
 	 * XPC has unloaded.  To make this work, we keep a copy of the pointer
 	 * to this page (i.e., amos_page) in the struct xpc_vars structure,
 	 * which is pointed to by the reserved page, and re-use that saved copy
-	 * on subsequent loads of XPC. This AMO page is never freed, and its
+	 * on subsequent loads of XPC. This amo page is never freed, and its
 	 * memory protections are never restricted.
 	 */
 	amos_page = xpc_vars->amos_page;
 	if (amos_page == NULL) {
-		amos_page = (AMO_t *)TO_AMO(uncached_alloc_page(0, 1));
+		amos_page = (struct amo *)TO_AMO(uncached_alloc_page(0, 1));
 		if (amos_page == NULL) {
-			dev_err(xpc_part, "can't allocate page of AMOs\n");
+			dev_err(xpc_part, "can't allocate page of amos\n");
 			return xpNoMemory;
 		}
 
 		/*
-		 * Open up AMO-R/W to cpu.  This is done on Shub 1.1 systems
-		 * when xpc_allow_AMO_ops_shub_wars_1_1_sn2() is called.
+		 * Open up amo-R/W to cpu.  This is done on Shub 1.1 systems
+		 * when xpc_allow_amo_ops_shub_wars_1_1_sn2() is called.
 		 */
-		ret = xpc_allow_AMO_ops_sn2(amos_page);
+		ret = xpc_allow_amo_ops_sn2(amos_page);
 		if (ret != xpSuccess) {
-			dev_err(xpc_part, "can't allow AMO operations\n");
+			dev_err(xpc_part, "can't allow amo operations\n");
 			uncached_free_page(__IA64_UNCACHED_OFFSET |
 					   TO_PHYS((u64)amos_page), 1);
 			return ret;
@@ -595,11 +598,11 @@
 	memset((u64 *)xpc_vars_part, 0, sizeof(struct xpc_vars_part_sn2) *
 	       xp_max_npartitions);
 
-	/* initialize the activate IRQ related AMO variables */
+	/* initialize the activate IRQ related amo variables */
 	for (i = 0; i < xp_nasid_mask_words; i++)
 		(void)xpc_init_IRQ_amo_sn2(XPC_ACTIVATE_IRQ_AMOS + i);
 
-	/* initialize the engaged remote partitions related AMO variables */
+	/* initialize the engaged remote partitions related amo variables */
 	(void)xpc_init_IRQ_amo_sn2(XPC_ENGAGED_PARTITIONS_AMO);
 	(void)xpc_init_IRQ_amo_sn2(XPC_DEACTIVATE_REQUEST_AMO);
 
@@ -745,19 +748,20 @@
 {
 	struct xpc_partition_sn2 *part_sn2 = &part->sn.sn2;
 	unsigned long irq_flags;
-	AMO_t *amo = (AMO_t *)__va(part_sn2->remote_amos_page_pa +
-				  (XPC_DEACTIVATE_REQUEST_AMO * sizeof(AMO_t)));
+	struct amo *amo = (struct amo *)__va(part_sn2->remote_amos_page_pa +
+					     (XPC_DEACTIVATE_REQUEST_AMO *
+					     sizeof(struct amo)));
 
 	local_irq_save(irq_flags);
 
-	/* set bit corresponding to our partid in remote partition's AMO */
+	/* set bit corresponding to our partid in remote partition's amo */
 	FETCHOP_STORE_OP(TO_AMO((u64)&amo->variable), FETCHOP_OR,
 			 (1UL << sn_partition_id));
 	/*
 	 * We must always use the nofault function regardless of whether we
 	 * are on a Shub 1.1 system or a Shub 1.2 slice 0xc processor. If we
 	 * didn't, we'd never know that the other partition is down and would
-	 * keep sending IRQs and AMOs to it until the heartbeat times out.
+	 * keep sending IRQs and amos to it until the heartbeat times out.
 	 */
 	(void)xp_nofault_PIOR((u64 *)GLOBAL_MMR_ADDR(NASID_GET(&amo->
 							       variable),
@@ -767,7 +771,7 @@
 
 	/*
 	 * Send activate IRQ to get other side to see that we've set our
-	 * bit in their deactivate request AMO.
+	 * bit in their deactivate request amo.
 	 */
 	xpc_send_activate_IRQ_sn2(part_sn2->remote_amos_page_pa,
 				  cnodeid_to_nasid(0),
@@ -779,19 +783,20 @@
 xpc_cancel_partition_deactivation_request_sn2(struct xpc_partition *part)
 {
 	unsigned long irq_flags;
-	AMO_t *amo = (AMO_t *)__va(part->sn.sn2.remote_amos_page_pa +
-				  (XPC_DEACTIVATE_REQUEST_AMO * sizeof(AMO_t)));
+	struct amo *amo = (struct amo *)__va(part->sn.sn2.remote_amos_page_pa +
+					     (XPC_DEACTIVATE_REQUEST_AMO *
+					     sizeof(struct amo)));
 
 	local_irq_save(irq_flags);
 
-	/* clear bit corresponding to our partid in remote partition's AMO */
+	/* clear bit corresponding to our partid in remote partition's amo */
 	FETCHOP_STORE_OP(TO_AMO((u64)&amo->variable), FETCHOP_AND,
 			 ~(1UL << sn_partition_id));
 	/*
 	 * We must always use the nofault function regardless of whether we
 	 * are on a Shub 1.1 system or a Shub 1.2 slice 0xc processor. If we
 	 * didn't, we'd never know that the other partition is down and would
-	 * keep sending IRQs and AMOs to it until the heartbeat times out.
+	 * keep sending IRQs and amos to it until the heartbeat times out.
 	 */
 	(void)xp_nofault_PIOR((u64 *)GLOBAL_MMR_ADDR(NASID_GET(&amo->
 							       variable),
@@ -803,9 +808,9 @@
 static int
 xpc_partition_deactivation_requested_sn2(short partid)
 {
-	AMO_t *amo = xpc_vars->amos_page + XPC_DEACTIVATE_REQUEST_AMO;
+	struct amo *amo = xpc_vars->amos_page + XPC_DEACTIVATE_REQUEST_AMO;
 
-	/* our partition's AMO variable ANDed with partid mask */
+	/* our partition's amo variable ANDed with partid mask */
 	return (FETCHOP_LOAD_OP(TO_AMO((u64)&amo->variable), FETCHOP_LOAD) &
 		(1UL << partid)) != 0;
 }
@@ -976,7 +981,7 @@
 }
 
 /*
- * Loop through the activation AMO variables and process any bits
+ * Loop through the activation amo variables and process any bits
  * which are set.  Each bit indicates a nasid sending a partition
  * activation or deactivation request.
  *
@@ -989,11 +994,11 @@
 	u64 nasid_mask;
 	u64 nasid;		/* remote nasid */
 	int n_IRQs_detected = 0;
-	AMO_t *act_amos;
+	struct amo *act_amos;
 
 	act_amos = xpc_vars->amos_page + XPC_ACTIVATE_IRQ_AMOS;
 
-	/* scan through act AMO variable looking for non-zero entries */
+	/* scan through act amo variable looking for non-zero entries */
 	for (word = 0; word < xp_nasid_mask_words; word++) {
 
 		if (xpc_exiting)
@@ -1005,7 +1010,7 @@
 			continue;
 		}
 
-		dev_dbg(xpc_part, "AMO[%d] gave back 0x%lx\n", word,
+		dev_dbg(xpc_part, "amo[%d] gave back 0x%lx\n", word,
 			nasid_mask);
 
 		/*
@@ -1038,7 +1043,7 @@
 
 	n_IRQs_detected = xpc_identify_activate_IRQ_sender_sn2();
 	if (n_IRQs_detected < n_IRQs_expected) {
-		/* retry once to help avoid missing AMO */
+		/* retry once to help avoid missing amo */
 		(void)xpc_identify_activate_IRQ_sender_sn2();
 	}
 }
@@ -1386,7 +1391,7 @@
 		part_sn2->remote_openclose_args_pa =
 		    pulled_entry->openclose_args_pa;
 		part_sn2->remote_chctl_amo_va =
-		    (AMO_t *)__va(pulled_entry->chctl_amo_pa);
+		    (struct amo *)__va(pulled_entry->chctl_amo_pa);
 		part_sn2->notify_IRQ_nasid = pulled_entry->notify_IRQ_nasid;
 		part_sn2->notify_IRQ_phys_cpuid =
 		    pulled_entry->notify_IRQ_phys_cpuid;
@@ -1417,7 +1422,7 @@
 	enum xp_retval ret;
 
 	/*
-	 * Register the remote partition's AMOs with SAL so it can handle
+	 * Register the remote partition's amos with SAL so it can handle
 	 * and cleanup errors within that address range should the remote
 	 * partition go down. We don't unregister this range because it is
 	 * difficult to tell when outstanding writes to the remote partition
@@ -2192,9 +2197,9 @@
 	xpc_send_msg = xpc_send_msg_sn2;
 	xpc_received_msg = xpc_received_msg_sn2;
 
-	/* open up protections for IPI and [potentially] AMO operations */
+	/* open up protections for IPI and [potentially] amo operations */
 	xpc_allow_IPI_ops_sn2();
-	xpc_allow_AMO_ops_shub_wars_1_1_sn2();
+	xpc_allow_amo_ops_shub_wars_1_1_sn2();
 
 	/*
 	 * This is safe to do before the xpc_hb_checker thread has started