diff options
author | Achin Gupta <achin.gupta@arm.com> | 2013-10-25 09:08:21 +0100 |
---|---|---|
committer | James Morrissey <james.morrissey@arm.com> | 2013-10-25 09:37:16 +0100 |
commit | 4f6ad66ae9fcc8bcb3b0fcee10b7ab1ffcaf1a56 (patch) | |
tree | 475db5d74370cb62b02afab0900774955a59702f /bl31/aarch64/runtime_exceptions.S |
ARMv8 Trusted Firmware release v0.2
Diffstat (limited to 'bl31/aarch64/runtime_exceptions.S')
-rw-r--r-- | bl31/aarch64/runtime_exceptions.S | 248 |
1 files changed, 248 insertions, 0 deletions
diff --git a/bl31/aarch64/runtime_exceptions.S b/bl31/aarch64/runtime_exceptions.S new file mode 100644 index 00000000..21976add --- /dev/null +++ b/bl31/aarch64/runtime_exceptions.S @@ -0,0 +1,248 @@ +/* + * Copyright (c) 2013, ARM Limited. All rights reserved. + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions are met: + * + * Redistributions of source code must retain the above copyright notice, this + * list of conditions and the following disclaimer. + * + * Redistributions in binary form must reproduce the above copyright notice, + * this list of conditions and the following disclaimer in the documentation + * and/or other materials provided with the distribution. + * + * Neither the name of ARM nor the names of its contributors may be used + * to endorse or promote products derived from this software without specific + * prior written permission. + * + * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" + * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE + * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE + * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE + * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR + * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF + * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS + * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN + * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) + * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE + * POSSIBILITY OF SUCH DAMAGE. + */ + +#include <arch.h> +#include <runtime_svc.h> + + .globl runtime_exceptions + + +#include <asm_macros.S> + + + .section aarch64_code, "ax"; .align 11 + + .align 7 +runtime_exceptions: + /* ----------------------------------------------------- + * Current EL with _sp_el0 : 0x0 - 0x180 + * ----------------------------------------------------- + */ +sync_exception_sp_el0: + exception_entry save_regs + mov x0, #SYNC_EXCEPTION_SP_EL0 + mov x1, sp + bl sync_exception_handler + exception_exit restore_regs + eret + + .align 7 +irq_sp_el0: + exception_entry save_regs + mov x0, #IRQ_SP_EL0 + mov x1, sp + bl async_exception_handler + exception_exit restore_regs + eret + + .align 7 +fiq_sp_el0: + exception_entry save_regs + mov x0, #FIQ_SP_EL0 + mov x1, sp + bl async_exception_handler + exception_exit restore_regs + eret + + .align 7 +serror_sp_el0: + exception_entry save_regs + mov x0, #SERROR_SP_EL0 + mov x1, sp + bl async_exception_handler + exception_exit restore_regs + eret + + /* ----------------------------------------------------- + * Current EL with SPx: 0x200 - 0x380 + * ----------------------------------------------------- + */ + .align 7 +sync_exception_sp_elx: + exception_entry save_regs + mov x0, #SYNC_EXCEPTION_SP_ELX + mov x1, sp + bl sync_exception_handler + exception_exit restore_regs + eret + + .align 7 +irq_sp_elx: + exception_entry save_regs + mov x0, #IRQ_SP_ELX + mov x1, sp + bl async_exception_handler + exception_exit restore_regs + eret + + .align 7 +fiq_sp_elx: + exception_entry save_regs + mov x0, #FIQ_SP_ELX + mov x1, sp + bl async_exception_handler + exception_exit restore_regs + eret + + .align 7 +serror_sp_elx: + exception_entry save_regs + mov x0, #SERROR_SP_ELX + mov x1, sp + bl async_exception_handler + exception_exit restore_regs + eret + + /* ----------------------------------------------------- + * Lower EL using AArch64 : 0x400 - 0x580 + * ----------------------------------------------------- + */ + .align 7 +sync_exception_aarch64: + exception_entry save_regs + mov x0, #SYNC_EXCEPTION_AARCH64 + mov x1, sp + bl sync_exception_handler + exception_exit restore_regs + eret + + .align 7 +irq_aarch64: + exception_entry save_regs + mov x0, #IRQ_AARCH64 + mov x1, sp + bl async_exception_handler + exception_exit restore_regs + eret + + .align 7 +fiq_aarch64: + exception_entry save_regs + mov x0, #FIQ_AARCH64 + mov x1, sp + bl async_exception_handler + exception_exit restore_regs + eret + + .align 7 +serror_aarch64: + exception_entry save_regs + mov x0, #IRQ_AARCH32 + mov x1, sp + bl async_exception_handler + exception_exit restore_regs + eret + + /* ----------------------------------------------------- + * Lower EL using AArch32 : 0x600 - 0x780 + * ----------------------------------------------------- + */ + .align 7 +sync_exception_aarch32: + exception_entry save_regs + mov x0, #SYNC_EXCEPTION_AARCH32 + mov x1, sp + bl sync_exception_handler + exception_exit restore_regs + eret + + .align 7 +irq_aarch32: + exception_entry save_regs + mov x0, #IRQ_AARCH32 + mov x1, sp + bl async_exception_handler + exception_exit restore_regs + eret + + .align 7 +fiq_aarch32: + exception_entry save_regs + mov x0, #FIQ_AARCH32 + mov x1, sp + bl async_exception_handler + exception_exit restore_regs + eret + + .align 7 +serror_aarch32: + exception_entry save_regs + mov x0, #SERROR_AARCH32 + mov x1, sp + bl async_exception_handler + exception_exit restore_regs + eret + + .align 7 + +save_regs:; .type save_regs, %function + sub sp, sp, #0x100 + stp x0, x1, [sp, #0x0] + stp x2, x3, [sp, #0x10] + stp x4, x5, [sp, #0x20] + stp x6, x7, [sp, #0x30] + stp x8, x9, [sp, #0x40] + stp x10, x11, [sp, #0x50] + stp x12, x13, [sp, #0x60] + stp x14, x15, [sp, #0x70] + stp x16, x17, [sp, #0x80] + stp x18, x19, [sp, #0x90] + stp x20, x21, [sp, #0xa0] + stp x22, x23, [sp, #0xb0] + stp x24, x25, [sp, #0xc0] + stp x26, x27, [sp, #0xd0] + mrs x0, sp_el0 + stp x28, x0, [sp, #0xe0] + mrs x0, spsr_el3 + str x0, [sp, #0xf0] + ret + + +restore_regs:; .type restore_regs, %function + ldr x9, [sp, #0xf0] + msr spsr_el3, x9 + ldp x28, x9, [sp, #0xe0] + msr sp_el0, x9 + ldp x26, x27, [sp, #0xd0] + ldp x24, x25, [sp, #0xc0] + ldp x22, x23, [sp, #0xb0] + ldp x20, x21, [sp, #0xa0] + ldp x18, x19, [sp, #0x90] + ldp x16, x17, [sp, #0x80] + ldp x14, x15, [sp, #0x70] + ldp x12, x13, [sp, #0x60] + ldp x10, x11, [sp, #0x50] + ldp x8, x9, [sp, #0x40] + ldp x6, x7, [sp, #0x30] + ldp x4, x5, [sp, #0x20] + ldp x2, x3, [sp, #0x10] + ldp x0, x1, [sp, #0x0] + add sp, sp, #0x100 + ret |