blob: 378bf7ebb3b5aa8475021cfed56eaa63b8305b3b [file] [log] [blame]
/*
* Ripped off from arch/arm64/include/asm/sysreg.h
*
* Copyright (C) 2016, Red Hat Inc, Andrew Jones <drjones@redhat.com>
*
* This work is licensed under the terms of the GNU LGPL, version 2.
*/
#ifndef _ASMARM64_SYSREG_H_
#define _ASMARM64_SYSREG_H_
#define sys_reg(op0, op1, crn, crm, op2) \
((((op0)&3)<<19)|((op1)<<16)|((crn)<<12)|((crm)<<8)|((op2)<<5))
#ifdef __ASSEMBLY__
.irp num,0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30
.equ .L__reg_num_x\num, \num
.endr
.equ .L__reg_num_xzr, 31
.macro mrs_s, rt, sreg
.inst 0xd5200000|(\sreg)|(.L__reg_num_\rt)
.endm
.macro msr_s, sreg, rt
.inst 0xd5000000|(\sreg)|(.L__reg_num_\rt)
.endm
#else
#include <libcflat.h>
#define read_sysreg(r) ({ \
u64 __val; \
asm volatile("mrs %0, " xstr(r) : "=r" (__val)); \
__val; \
})
#define write_sysreg(v, r) do { \
u64 __val = (u64)v; \
asm volatile("msr " xstr(r) ", %x0" : : "rZ" (__val)); \
} while (0)
#define read_sysreg_s(r) ({ \
u64 __val; \
asm volatile("mrs_s %0, " xstr(r) : "=r" (__val)); \
__val; \
})
#define write_sysreg_s(v, r) do { \
u64 __val = (u64)v; \
asm volatile("msr_s " xstr(r) ", %x0" : : "rZ" (__val));\
} while (0)
#define write_regn_el0(__reg, __n, __val) \
write_sysreg((__val), __reg ## __n ## _el0)
#define read_regn_el0(__reg, __n) \
read_sysreg(__reg ## __n ## _el0)
asm(
" .irp num,0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30\n"
" .equ .L__reg_num_x\\num, \\num\n"
" .endr\n"
" .equ .L__reg_num_xzr, 31\n"
"\n"
" .macro mrs_s, rt, sreg\n"
" .inst 0xd5200000|(\\sreg)|(.L__reg_num_\\rt)\n"
" .endm\n"
"\n"
" .macro msr_s, sreg, rt\n"
" .inst 0xd5000000|(\\sreg)|(.L__reg_num_\\rt)\n"
" .endm\n"
);
#endif /* __ASSEMBLY__ */
#endif /* _ASMARM64_SYSREG_H_ */