-#define __SYSCALL_LL(x) \
+#define __SYSCALL_LL_E(x) \
((union { long long ll; long l[2]; }){ .ll = x }).l[0], \
((union { long long ll; long l[2]; }){ .ll = x }).l[1]
+#define __SYSCALL_LL_O(x) __SYSCALL_LL_E((x))
static inline long __syscall0(long __n)
{
{
unsigned long __ret;
__asm__ __volatile__ ("xchg %2,%%ebx ; int $128 ; xchg %2,%%ebx"
- : "=a"(__ret) : "a"(__n), "r"(__a1) : "memory");
+ : "=a"(__ret) : "a"(__n), "d"(__a1) : "memory");
return __ret;
}
{
unsigned long __ret;
__asm__ __volatile__ ("xchg %2,%%ebx ; int $128 ; xchg %2,%%ebx"
- : "=a"(__ret) : "a"(__n), "r"(__a1), "c"(__a2) : "memory");
+ : "=a"(__ret) : "a"(__n), "d"(__a1), "c"(__a2) : "memory");
return __ret;
}
{
unsigned long __ret;
__asm__ __volatile__ ("xchg %2,%%ebx ; int $128 ; xchg %2,%%ebx"
- : "=a"(__ret) : "a"(__n), "r"(__a1), "c"(__a2), "d"(__a3) : "memory");
+ : "=a"(__ret) : "a"(__n), "S"(__a1), "c"(__a2), "d"(__a3) : "memory");
return __ret;
}
{
unsigned long __ret;
__asm__ __volatile__ ("xchg %2,%%ebx ; int $128 ; xchg %2,%%ebx"
- : "=a"(__ret) : "a"(__n), "r"(__a1), "c"(__a2), "d"(__a3), "S"(__a4) : "memory");
+ : "=a"(__ret) : "a"(__n), "D"(__a1), "c"(__a2), "d"(__a3), "S"(__a4) : "memory");
return __ret;
}
+#if 0
static inline long __syscall5(long __n, long __a1, long __a2, long __a3, long __a4, long __a5)
{
unsigned long __ret;
: "=a"(__ret) : "a"(__n), "g"(__a1), "c"(__a2), "d"(__a3), "S"(__a4), "D"(__a5) : "memory");
return __ret;
}
+#else
+static inline long __syscall5(long __n, long __a1, long __a2, long __a3, long __a4, long __a5)
+{
+ return (__syscall)(__n, __a1, __a2, __a3, __a4, __a5);
+}
+#endif
static inline long __syscall6(long __n, long __a1, long __a2, long __a3, long __a4, long __a5, long __a6)
{
#define __NR_inotify_init1 332
#define __NR_preadv 333
#define __NR_pwritev 334
+#define __NR_prlimit64 340
/* fixup legacy 16-bit junk */
#undef __NR_lchown
#define SYS_inotify_init1 332
#define SYS_preadv 333
#define SYS_pwritev 334
+#define SYS_prlimit64 340
/* fixup legacy 16-bit junk */
#undef SYS_lchown