1//===-- sanitizer_syscall_linux_aarch64.inc --------------------*- C++ -*-===//
2//
3//                     The LLVM Compiler Infrastructure
4//
5// This file is distributed under the University of Illinois Open Source
6// License. See LICENSE.TXT for details.
7//
8//===----------------------------------------------------------------------===//
9//
10// Implementations of internal_syscall and internal_iserror for Linux/aarch64.
11//
12//===----------------------------------------------------------------------===//
13
14#define SYSCALL(name) __NR_ ## name
15
16static uptr __internal_syscall(u64 nr) {
17  register u64 x8 asm("x8") = nr;
18  register u64 x0 asm("x0");
19  asm volatile("svc 0"
20               : "=r"(x0)
21               : "r"(x8)
22               : "memory", "cc");
23  return x0;
24}
25#define __internal_syscall0(n) \
26  (__internal_syscall)(n)
27
28static uptr __internal_syscall(u64 nr, u64 arg1) {
29  register u64 x8 asm("x8") = nr;
30  register u64 x0 asm("x0") = arg1;
31  asm volatile("svc 0"
32               : "=r"(x0)
33               : "r"(x8), "0"(x0)
34               : "memory", "cc");
35  return x0;
36}
37#define __internal_syscall1(n, a1) \
38  (__internal_syscall)(n, (u64)(a1))
39
40static uptr __internal_syscall(u64 nr, u64 arg1, long arg2) {
41  register u64 x8 asm("x8") = nr;
42  register u64 x0 asm("x0") = arg1;
43  register u64 x1 asm("x1") = arg2;
44  asm volatile("svc 0"
45               : "=r"(x0)
46               : "r"(x8), "0"(x0), "r"(x1)
47               : "memory", "cc");
48  return x0;
49}
50#define __internal_syscall2(n, a1, a2) \
51  (__internal_syscall)(n, (u64)(a1), (long)(a2))
52
53static uptr __internal_syscall(u64 nr, u64 arg1, long arg2, long arg3) {
54  register u64 x8 asm("x8") = nr;
55  register u64 x0 asm("x0") = arg1;
56  register u64 x1 asm("x1") = arg2;
57  register u64 x2 asm("x2") = arg3;
58  asm volatile("svc 0"
59               : "=r"(x0)
60               : "r"(x8), "0"(x0), "r"(x1), "r"(x2)
61               : "memory", "cc");
62  return x0;
63}
64#define __internal_syscall3(n, a1, a2, a3) \
65  (__internal_syscall)(n, (u64)(a1), (long)(a2), (long)(a3))
66
67static uptr __internal_syscall(u64 nr, u64 arg1, long arg2, long arg3,
68                               u64 arg4) {
69  register u64 x8 asm("x8") = nr;
70  register u64 x0 asm("x0") = arg1;
71  register u64 x1 asm("x1") = arg2;
72  register u64 x2 asm("x2") = arg3;
73  register u64 x3 asm("x3") = arg4;
74  asm volatile("svc 0"
75               : "=r"(x0)
76               : "r"(x8), "0"(x0), "r"(x1), "r"(x2), "r"(x3)
77               : "memory", "cc");
78  return x0;
79}
80#define __internal_syscall4(n, a1, a2, a3, a4) \
81  (__internal_syscall)(n, (u64)(a1), (long)(a2), (long)(a3), (long)(a4))
82
83static uptr __internal_syscall(u64 nr, u64 arg1, long arg2, long arg3,
84                               u64 arg4, long arg5) {
85  register u64 x8 asm("x8") = nr;
86  register u64 x0 asm("x0") = arg1;
87  register u64 x1 asm("x1") = arg2;
88  register u64 x2 asm("x2") = arg3;
89  register u64 x3 asm("x3") = arg4;
90  register u64 x4 asm("x4") = arg5;
91  asm volatile("svc 0"
92               : "=r"(x0)
93               : "r"(x8), "0"(x0), "r"(x1), "r"(x2), "r"(x3), "r"(x4)
94               : "memory", "cc");
95  return x0;
96}
97#define __internal_syscall5(n, a1, a2, a3, a4, a5) \
98  (__internal_syscall)(n, (u64)(a1), (long)(a2), (long)(a3), (long)(a4), \
99                       (u64)(a5))
100
101static uptr __internal_syscall(u64 nr, u64 arg1, long arg2, long arg3,
102                               u64 arg4, long arg5, long arg6) {
103  register u64 x8 asm("x8") = nr;
104  register u64 x0 asm("x0") = arg1;
105  register u64 x1 asm("x1") = arg2;
106  register u64 x2 asm("x2") = arg3;
107  register u64 x3 asm("x3") = arg4;
108  register u64 x4 asm("x4") = arg5;
109  register u64 x5 asm("x5") = arg6;
110  asm volatile("svc 0"
111               : "=r"(x0)
112               : "r"(x8), "0"(x0), "r"(x1), "r"(x2), "r"(x3), "r"(x4), "r"(x5)
113               : "memory", "cc");
114  return x0;
115}
116#define __internal_syscall6(n, a1, a2, a3, a4, a5, a6) \
117  (__internal_syscall)(n, (u64)(a1), (long)(a2), (long)(a3), (long)(a4), \
118                       (u64)(a5), (long)(a6))
119
120#define __SYSCALL_NARGS_X(a1, a2, a3, a4, a5, a6, a7, a8, n, ...) n
121#define __SYSCALL_NARGS(...) \
122  __SYSCALL_NARGS_X(__VA_ARGS__, 7, 6, 5, 4, 3, 2, 1, 0, )
123#define __SYSCALL_CONCAT_X(a, b) a##b
124#define __SYSCALL_CONCAT(a, b) __SYSCALL_CONCAT_X(a, b)
125#define __SYSCALL_DISP(b, ...) \
126  __SYSCALL_CONCAT(b, __SYSCALL_NARGS(__VA_ARGS__))(__VA_ARGS__)
127
128#define internal_syscall(...) __SYSCALL_DISP(__internal_syscall, __VA_ARGS__)
129
130// Helper function used to avoid cobbler errno.
131bool internal_iserror(uptr retval, int *rverrno) {
132  if (retval >= (uptr)-4095) {
133    if (rverrno)
134      *rverrno = -retval;
135    return true;
136  }
137  return false;
138}
139