[rtems commit] aarch64: whitespace fixes in start.S

Gedare Bloom gedare at rtems.org
Thu Jun 24 18:58:08 UTC 2021


Module:    rtems
Branch:    master
Commit:    17a9103c53ab1363d8e0e245a0cd3d77b828b177
Changeset: http://git.rtems.org/rtems/commit/?id=17a9103c53ab1363d8e0e245a0cd3d77b828b177

Author:    Gedare Bloom <gedare at rtems.org>
Date:      Thu Jun 24 10:02:53 2021 -0600

aarch64: whitespace fixes in start.S

---

 bsps/aarch64/shared/start/start.S | 332 +++++++++++++++++++-------------------
 1 file changed, 166 insertions(+), 166 deletions(-)

diff --git a/bsps/aarch64/shared/start/start.S b/bsps/aarch64/shared/start/start.S
index bc1fb00..f4c39da 100644
--- a/bsps/aarch64/shared/start/start.S
+++ b/bsps/aarch64/shared/start/start.S
@@ -39,64 +39,64 @@
 
 #include <bspopts.h>
 
-	/* Global symbols */
-	.globl	_start
-	.section	".bsp_start_text", "ax"
+  /* Global symbols */
+  .globl _start
+  .section ".bsp_start_text", "ax"
 
 /* Start entry */
 
 _start:
 
-	/*
-	 * We do not save the context since we do not return to the boot
-	 * loader but preserve x1 and x2 to allow access to bootloader parameters
-	 */
+  /*
+   * We do not save the context since we do not return to the boot
+   * loader but preserve x1 and x2 to allow access to bootloader parameters
+   */
 #ifndef BSP_START_NEEDS_REGISTER_INITIALIZATION
-	mov	x5, x1		/* machine type number or ~0 for DT boot */
-	mov	x6, x2		/* physical address of ATAGs or DTB */
+  mov x5, x1    /* machine type number or ~0 for DT boot */
+  mov x6, x2    /* physical address of ATAGs or DTB */
 #else /* BSP_START_NEEDS_REGISTER_INITIALIZATION */
-	mov	x0, XZR
-	mov	x1, XZR
-	mov	x2, XZR
-	mov	x3, XZR
-	mov	x4, XZR
-	mov	x5, XZR
-	mov	x6, XZR
-	mov	x7, XZR
-	mov	x8, XZR
-	mov	x9, XZR
-	mov	x10, XZR
-	mov	x11, XZR
-	mov	x12, XZR
-	mov	x13, XZR
-	mov	x14, XZR
-	mov	x15, XZR
-	mov	x16, XZR
-	mov	x17, XZR
-	mov	x18, XZR
-	mov	x19, XZR
-	mov	x20, XZR
-	mov	x21, XZR
-	mov	x22, XZR
-	mov	x23, XZR
-	mov	x24, XZR
-	mov	x25, XZR
-	mov	x26, XZR
-	mov	x27, XZR
-	mov	x28, XZR
-	mov	x29, XZR
-	mov	x30, XZR
+  mov x0, XZR
+  mov x1, XZR
+  mov x2, XZR
+  mov x3, XZR
+  mov x4, XZR
+  mov x5, XZR
+  mov x6, XZR
+  mov x7, XZR
+  mov x8, XZR
+  mov x9, XZR
+  mov x10, XZR
+  mov x11, XZR
+  mov x12, XZR
+  mov x13, XZR
+  mov x14, XZR
+  mov x15, XZR
+  mov x16, XZR
+  mov x17, XZR
+  mov x18, XZR
+  mov x19, XZR
+  mov x20, XZR
+  mov x21, XZR
+  mov x22, XZR
+  mov x23, XZR
+  mov x24, XZR
+  mov x25, XZR
+  mov x26, XZR
+  mov x27, XZR
+  mov x28, XZR
+  mov x29, XZR
+  mov x30, XZR
 #ifdef AARCH64_MULTILIB_VFP
 #endif
 #endif
 
-	/* Initialize SCTLR_EL1 */
-	mov x0, XZR
+  /* Initialize SCTLR_EL1 */
+  mov x0, XZR
 #if defined(RTEMS_DEBUG)
-	/* Enable Stack alignment checking */
-	orr x0, x0, #(1<<3)
+  /* Enable Stack alignment checking */
+  orr x0, x0, #(1<<3)
 #endif
-	msr SCTLR_EL1, x0
+  msr SCTLR_EL1, x0
 
   mrs x0, CurrentEL
   cmp x0, #(1<<2)
@@ -105,7 +105,7 @@ _start:
   b.eq _el2_start
 
 _el3_start:
-	/* Drop from EL3 to EL2 */
+  /* Drop from EL3 to EL2 */
 
   /* Initialize HCR_EL2 and SCTLR_EL2 */
   msr HCR_EL2, XZR
@@ -139,170 +139,170 @@ _el3_start:
   eret
 
 _el2_start:
-	/* Drop from EL2 to EL1 */
-
-	/* Configure HCR_EL2 */
-	mrs x0, HCR_EL2
-	/* Set EL1 Execution state to AArch64 */
-	orr x0, x0, #(1<<31)
-	/* Disable ID traps */
-	bic x0, x0, #(1<<15)
-	bic x0, x0, #(1<<16)
-	bic x0, x0, #(1<<17)
-	bic x0, x0, #(1<<18)
-	msr HCR_EL2, x0
-
-	/* Set to EL1h mode for eret */
-	mov x0, #0b00101
-	msr SPSR_EL2, x0
-
-	/* Set EL1 entry point */
-	adr x0, _el1_start
-	msr ELR_EL2, x0
-	eret
+  /* Drop from EL2 to EL1 */
+
+  /* Configure HCR_EL2 */
+  mrs x0, HCR_EL2
+  /* Set EL1 Execution state to AArch64 */
+  orr x0, x0, #(1<<31)
+  /* Disable ID traps */
+  bic x0, x0, #(1<<15)
+  bic x0, x0, #(1<<16)
+  bic x0, x0, #(1<<17)
+  bic x0, x0, #(1<<18)
+  msr HCR_EL2, x0
+
+  /* Set to EL1h mode for eret */
+  mov x0, #0b00101
+  msr SPSR_EL2, x0
+
+  /* Set EL1 entry point */
+  adr x0, _el1_start
+  msr ELR_EL2, x0
+  eret
 
 _el1_start:
 
 #ifdef RTEMS_SMP
-	/* Read MPIDR and get current processor index */
-        mrs	x7, mpidr_el1
-	and	x7, #0xff
+  /* Read MPIDR and get current processor index */
+  mrs x7, mpidr_el1
+  and x7, #0xff
 #endif
 
 #ifdef RTEMS_SMP
-        /*
-         * Get current per-CPU control and store it in PL1 only Thread ID
-         * Register (TPIDRPRW).
-         */
+  /*
+   * Get current per-CPU control and store it in PL1 only Thread ID
+   * Register (TPIDRPRW).
+   */
 #ifdef AARCH64_MULTILIB_ARCH_V8_ILP32
-	ldr	w1, =_Per_CPU_Information
+  ldr w1, =_Per_CPU_Information
 #else
-	ldr	x1, =_Per_CPU_Information
+  ldr x1, =_Per_CPU_Information
 #endif
-	add	x1, x1, x7, asl #PER_CPU_CONTROL_SIZE_LOG2
-	mcr	p15, 0, x1, c13, c0, 4
+  add x1, x1, x7, asl #PER_CPU_CONTROL_SIZE_LOG2
+  mcr p15, 0, x1, c13, c0, 4
 
 #endif
 
-	/* Calculate interrupt stack area end for current processor */
+  /* Calculate interrupt stack area end for current processor */
 #ifdef AARCH64_MULTILIB_ARCH_V8_ILP32
-	ldr	w1, =_ISR_Stack_size
+  ldr w1, =_ISR_Stack_size
 #else
-	ldr	x1, =_ISR_Stack_size
+  ldr x1, =_ISR_Stack_size
 #endif
 #ifdef RTEMS_SMP
-	add	x3, x7, #1
-	mul	x1, x1, x3
+  add x3, x7, #1
+  mul x1, x1, x3
 #endif
 #ifdef AARCH64_MULTILIB_ARCH_V8_ILP32
-	ldr	w2, =_ISR_Stack_area_begin
+  ldr w2, =_ISR_Stack_area_begin
 #else
-	ldr	x2, =_ISR_Stack_area_begin
+  ldr x2, =_ISR_Stack_area_begin
 #endif
-	add	x3, x1, x2
+  add x3, x1, x2
 
-	/* Save original DAIF value */
-	mrs	x4, DAIF
+  /* Save original DAIF value */
+  mrs  x4, DAIF
 
 #ifdef BSP_START_NEEDS_REGISTER_INITIALIZATION
-	mov	x8, XZR
-	mov	x9, XZR
-	mov	x10, XZR
-	mov	x11, XZR
-	mov	x12, XZR
-	mov	x13, XZR
-	mov	x14, XZR
-	mov	x15, XZR
+  mov x8, XZR
+  mov x9, XZR
+  mov x10, XZR
+  mov x11, XZR
+  mov x12, XZR
+  mov x13, XZR
+  mov x14, XZR
+  mov x15, XZR
 #endif
 
-	/*
-	 * SPx: the stack pointer corresponding to the current exception level
-	 * Normal operation for RTEMS on AArch64 uses SPx and runs on EL1
-	 * Exception operation (synchronous errors, IRQ, FIQ, System Errors) uses SP0
-	*/
+  /*
+   * SPx: the stack pointer corresponding to the current exception level
+   * Normal operation for RTEMS on AArch64 uses SPx and runs on EL1
+   * Exception operation (synchronous errors, IRQ, FIQ, System Errors) uses SP0
+  */
 #ifdef AARCH64_MULTILIB_ARCH_V8_ILP32
-	ldr	w1, =bsp_stack_exception_size
+  ldr w1, =bsp_stack_exception_size
 #else
-	ldr	x1, =bsp_stack_exception_size
+  ldr x1, =bsp_stack_exception_size
 #endif
-	/* Switch to SP0 and set exception stack */
-	msr	spsel, #0
-	mov	sp, x3
-	/* Switch back to SPx for normal operation */
-	msr	spsel, #1
-	sub	x3, x3, x1
+  /* Switch to SP0 and set exception stack */
+  msr spsel, #0
+  mov sp, x3
+  /* Switch back to SPx for normal operation */
+  msr spsel, #1
+  sub x3, x3, x1
 
-	/* Set SP1 stack used for normal operation */
-	mov	sp, x3
+  /* Set SP1 stack used for normal operation */
+  mov sp, x3
 
-	/* Stay in EL1 mode */
+  /* Stay in EL1 mode */
 
 #ifdef AARCH64_MULTILIB_VFP
 #ifdef AARCH64_MULTILIB_HAS_CPACR
-	/* Read CPACR */
-	mrs x0, CPACR_EL1
+  /* Read CPACR */
+  mrs x0, CPACR_EL1
 
-	/* Enable EL1 access permissions for CP10 */
-	orr x0, x0, #(1 << 20)
+  /* Enable EL1 access permissions for CP10 */
+  orr x0, x0, #(1 << 20)
 
-	/* Write CPACR */
-	msr CPACR_EL1, x0
-	isb
+  /* Write CPACR */
+  msr CPACR_EL1, x0
+  isb
 #endif
 
-	/* FPU does not need to be enabled on AArch64 */
+  /* FPU does not need to be enabled on AArch64 */
 
 #ifdef BSP_START_NEEDS_REGISTER_INITIALIZATION
-	mov	x0, #0
-	mov	CPTR_EL3, XZR
-	mov	CPTR_EL2, XZR
-	mov	d0, XZR
-	mov	d1, XZR
-	mov	d2, XZR
-	mov	d3, XZR
-	mov	d4, XZR
-	mov	d5, XZR
-	mov	d6, XZR
-	mov	d7, XZR
-	mov	d8, XZR
-	mov	d9, XZR
-	mov	d10, XZR
-	mov	d11, XZR
-	mov	d12, XZR
-	mov	d13, XZR
-	mov	d14, XZR
-	mov	d15, XZR
-	mov	d16, XZR
-	mov	d17, XZR
-	mov	d18, XZR
-	mov	d19, XZR
-	mov	d20, XZR
-	mov	d21, XZR
-	mov	d22, XZR
-	mov	d23, XZR
-	mov	d24, XZR
-	mov	d25, XZR
-	mov	d26, XZR
-	mov	d27, XZR
-	mov	d28, XZR
-	mov	d29, XZR
-	mov	d30, XZR
-	mov	d31, XZR
+  mov x0, #0
+  mov CPTR_EL3, XZR
+  mov CPTR_EL2, XZR
+  mov d0, XZR
+  mov d1, XZR
+  mov d2, XZR
+  mov d3, XZR
+  mov d4, XZR
+  mov d5, XZR
+  mov d6, XZR
+  mov d7, XZR
+  mov d8, XZR
+  mov d9, XZR
+  mov d10, XZR
+  mov d11, XZR
+  mov d12, XZR
+  mov d13, XZR
+  mov d14, XZR
+  mov d15, XZR
+  mov d16, XZR
+  mov d17, XZR
+  mov d18, XZR
+  mov d19, XZR
+  mov d20, XZR
+  mov d21, XZR
+  mov d22, XZR
+  mov d23, XZR
+  mov d24, XZR
+  mov d25, XZR
+  mov d26, XZR
+  mov d27, XZR
+  mov d28, XZR
+  mov d29, XZR
+  mov d30, XZR
+  mov d31, XZR
 #endif /* BSP_START_NEEDS_REGISTER_INITIALIZATION */
 
 #endif /* AARCH64_MULTILIB_VFP */
 
-	/*
-	 * Invoke the start hook 0.
-	 *
-	 */
+  /*
+   * Invoke the start hook 0.
+   *
+   */
 
-	mov	x1, x5		/* machine type number or ~0 for DT boot */
-	bl	bsp_start_hook_0
+  mov x1, x5    /* machine type number or ~0 for DT boot */
+  bl bsp_start_hook_0
 
-	/* Branch to start hook 1 */
-	bl	bsp_start_hook_1
+  /* Branch to start hook 1 */
+  bl bsp_start_hook_1
 
-	/* Branch to boot card */
-	mov	x0, #0
-	bl	boot_card
+  /* Branch to boot card */
+  mov x0, #0
+  bl boot_card



More information about the vc mailing list