[planner] Resolves: rhbz#1141479 use system valgrind headers
Caolán McNamara
caolanm at fedoraproject.org
Tue Sep 16 10:43:18 UTC 2014
commit 8b657c4c02831872e93abac86407e1d0ce413fea
Author: Caolán McNamara <caolanm at redhat.com>
Date: Tue Sep 16 11:37:53 2014 +0100
Resolves: rhbz#1141479 use system valgrind headers
planner-rhbz1141479-paste.tasks.patch | 4062 +++++++++++++++++++++++++++++++++
planner.spec | 9 +-
2 files changed, 4069 insertions(+), 2 deletions(-)
---
diff --git a/planner-rhbz1141479-paste.tasks.patch b/planner-rhbz1141479-paste.tasks.patch
new file mode 100644
index 0000000..346d916
--- /dev/null
+++ b/planner-rhbz1141479-paste.tasks.patch
@@ -0,0 +1,4062 @@
+diff -ruN planner-0.14.6.orig/src/Makefile.am planner-0.14.6/src/Makefile.am
+--- planner-0.14.6.orig/src/Makefile.am 2014-09-16 10:33:17.393876669 +0100
++++ planner-0.14.6/src/Makefile.am 2014-09-16 10:43:08.811065116 +0100
+@@ -119,7 +119,6 @@
+ planner-window.h \
+ planner-working-time-dialog.c \
+ planner-working-time-dialog.h \
+- valgrind.h \
+ $(gantt_task_source) \
+ $(task_source) \
+ $(resource_source) \
+diff -ruN planner-0.14.6.orig/src/Makefile.in planner-0.14.6/src/Makefile.in
+--- planner-0.14.6.orig/src/Makefile.in 2014-09-16 10:33:17.392876660 +0100
++++ planner-0.14.6/src/Makefile.in 2014-09-16 10:43:06.947048833 +0100
+@@ -508,7 +508,6 @@
+ planner-window.h \
+ planner-working-time-dialog.c \
+ planner-working-time-dialog.h \
+- valgrind.h \
+ $(gantt_task_source) \
+ $(task_source) \
+ $(resource_source) \
+diff -ruN planner-0.14.6.orig/src/planner-plugin.c planner-0.14.6/src/planner-plugin.c
+--- planner-0.14.6.orig/src/planner-plugin.c 2014-09-16 10:33:17.387876616 +0100
++++ planner-0.14.6/src/planner-plugin.c 2014-09-16 10:43:01.870004484 +0100
+@@ -24,7 +24,7 @@
+ #include <glib.h>
+ #include "planner-window.h"
+ #include "planner-plugin.h"
+-#include "valgrind.h"
++#include "valgrind/valgrind.h"
+
+ static void mv_init (PlannerPlugin *plugin);
+ static void mv_class_init (PlannerPluginClass *class);
+diff -ruN planner-0.14.6.orig/src/valgrind.h planner-0.14.6/src/valgrind.h
+--- planner-0.14.6.orig/src/valgrind.h 2014-09-16 10:33:17.390876642 +0100
++++ planner-0.14.6/src/valgrind.h 1970-01-01 01:00:00.000000000 +0100
+@@ -1,4024 +0,0 @@
+-/* Aside from this comment and stripping of trailing white space
+- * this file is an unmodified copy of include/valgrind.h from
+- * valgrind 3.5.0
+- */
+-/* -*- c -*-
+- ----------------------------------------------------------------
+-
+- Notice that the following BSD-style license applies to this one
+- file (valgrind.h) only. The rest of Valgrind is licensed under the
+- terms of the GNU General Public License, version 2, unless
+- otherwise indicated. See the COPYING file in the source
+- distribution for details.
+-
+- ----------------------------------------------------------------
+-
+- This file is part of Valgrind, a dynamic binary instrumentation
+- framework.
+-
+- Copyright (C) 2000-2009 Julian Seward. All rights reserved.
+-
+- Redistribution and use in source and binary forms, with or without
+- modification, are permitted provided that the following conditions
+- are met:
+-
+- 1. Redistributions of source code must retain the above copyright
+- notice, this list of conditions and the following disclaimer.
+-
+- 2. The origin of this software must not be misrepresented; you must
+- not claim that you wrote the original software. If you use this
+- software in a product, an acknowledgment in the product
+- documentation would be appreciated but is not required.
+-
+- 3. Altered source versions must be plainly marked as such, and must
+- not be misrepresented as being the original software.
+-
+- 4. The name of the author may not be used to endorse or promote
+- products derived from this software without specific prior written
+- permission.
+-
+- THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS
+- OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
+- WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
+- ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY
+- DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
+- DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE
+- GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
+- INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY,
+- WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
+- NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
+- SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
+-
+- ----------------------------------------------------------------
+-
+- Notice that the above BSD-style license applies to this one file
+- (valgrind.h) only. The entire rest of Valgrind is licensed under
+- the terms of the GNU General Public License, version 2. See the
+- COPYING file in the source distribution for details.
+-
+- ----------------------------------------------------------------
+-*/
+-
+-
+-/* This file is for inclusion into client (your!) code.
+-
+- You can use these macros to manipulate and query Valgrind's
+- execution inside your own programs.
+-
+- The resulting executables will still run without Valgrind, just a
+- little bit more slowly than they otherwise would, but otherwise
+- unchanged. When not running on valgrind, each client request
+- consumes very few (eg. 7) instructions, so the resulting performance
+- loss is negligible unless you plan to execute client requests
+- millions of times per second. Nevertheless, if that is still a
+- problem, you can compile with the NVALGRIND symbol defined (gcc
+- -DNVALGRIND) so that client requests are not even compiled in. */
+-
+-#ifndef __VALGRIND_H
+-#define __VALGRIND_H
+-
+-#include <stdarg.h>
+-
+-/* Nb: this file might be included in a file compiled with -ansi. So
+- we can't use C++ style "//" comments nor the "asm" keyword (instead
+- use "__asm__"). */
+-
+-/* Derive some tags indicating what the target platform is. Note
+- that in this file we're using the compiler's CPP symbols for
+- identifying architectures, which are different to the ones we use
+- within the rest of Valgrind. Note, __powerpc__ is active for both
+- 32 and 64-bit PPC, whereas __powerpc64__ is only active for the
+- latter (on Linux, that is). */
+-#undef PLAT_x86_linux
+-#undef PLAT_amd64_linux
+-#undef PLAT_ppc32_linux
+-#undef PLAT_ppc64_linux
+-#undef PLAT_ppc32_aix5
+-#undef PLAT_ppc64_aix5
+-
+-
+-#if defined(_AIX) && defined(__64BIT__)
+-# define PLAT_ppc64_aix5 1
+-#elif defined(_AIX) && !defined(__64BIT__)
+-# define PLAT_ppc32_aix5 1
+-#elif defined(__APPLE__) && defined(__i386__)
+-# define PLAT_x86_darwin 1
+-#elif defined(__APPLE__) && defined(__x86_64__)
+-# define PLAT_amd64_darwin 1
+-#elif defined(__i386__)
+-# define PLAT_x86_linux 1
+-#elif defined(__x86_64__)
+-# define PLAT_amd64_linux 1
+-#elif defined(__powerpc__) && !defined(__powerpc64__)
+-# define PLAT_ppc32_linux 1
+-#elif defined(__powerpc__) && defined(__powerpc64__)
+-# define PLAT_ppc64_linux 1
+-#else
+-/* If we're not compiling for our target platform, don't generate
+- any inline asms. */
+-# if !defined(NVALGRIND)
+-# define NVALGRIND 1
+-# endif
+-#endif
+-
+-
+-/* ------------------------------------------------------------------ */
+-/* ARCHITECTURE SPECIFICS for SPECIAL INSTRUCTIONS. There is nothing */
+-/* in here of use to end-users -- skip to the next section. */
+-/* ------------------------------------------------------------------ */
+-
+-#if defined(NVALGRIND)
+-
+-/* Define NVALGRIND to completely remove the Valgrind magic sequence
+- from the compiled code (analogous to NDEBUG's effects on
+- assert()) */
+-#define VALGRIND_DO_CLIENT_REQUEST( \
+- _zzq_rlval, _zzq_default, _zzq_request, \
+- _zzq_arg1, _zzq_arg2, _zzq_arg3, _zzq_arg4, _zzq_arg5) \
+- { \
+- (_zzq_rlval) = (_zzq_default); \
+- }
+-
+-#else /* ! NVALGRIND */
+-
+-/* The following defines the magic code sequences which the JITter
+- spots and handles magically. Don't look too closely at them as
+- they will rot your brain.
+-
+- The assembly code sequences for all architectures is in this one
+- file. This is because this file must be stand-alone, and we don't
+- want to have multiple files.
+-
+- For VALGRIND_DO_CLIENT_REQUEST, we must ensure that the default
+- value gets put in the return slot, so that everything works when
+- this is executed not under Valgrind. Args are passed in a memory
+- block, and so there's no intrinsic limit to the number that could
+- be passed, but it's currently five.
+-
+- The macro args are:
+- _zzq_rlval result lvalue
+- _zzq_default default value (result returned when running on real CPU)
+- _zzq_request request code
+- _zzq_arg1..5 request params
+-
+- The other two macros are used to support function wrapping, and are
+- a lot simpler. VALGRIND_GET_NR_CONTEXT returns the value of the
+- guest's NRADDR pseudo-register and whatever other information is
+- needed to safely run the call original from the wrapper: on
+- ppc64-linux, the R2 value at the divert point is also needed. This
+- information is abstracted into a user-visible type, OrigFn.
+-
+- VALGRIND_CALL_NOREDIR_* behaves the same as the following on the
+- guest, but guarantees that the branch instruction will not be
+- redirected: x86: call *%eax, amd64: call *%rax, ppc32/ppc64:
+- branch-and-link-to-r11. VALGRIND_CALL_NOREDIR is just text, not a
+- complete inline asm, since it needs to be combined with more magic
+- inline asm stuff to be useful.
+-*/
+-
+-/* ------------------------- x86-{linux,darwin} ---------------- */
+-
+-#if defined(PLAT_x86_linux) || defined(PLAT_x86_darwin)
+-
+-typedef
+- struct {
+- unsigned int nraddr; /* where's the code? */
+- }
+- OrigFn;
+-
+-#define __SPECIAL_INSTRUCTION_PREAMBLE \
+- "roll $3, %%edi ; roll $13, %%edi\n\t" \
+- "roll $29, %%edi ; roll $19, %%edi\n\t"
+-
+-#define VALGRIND_DO_CLIENT_REQUEST( \
+- _zzq_rlval, _zzq_default, _zzq_request, \
+- _zzq_arg1, _zzq_arg2, _zzq_arg3, _zzq_arg4, _zzq_arg5) \
+- { volatile unsigned int _zzq_args[6]; \
+- volatile unsigned int _zzq_result; \
+- _zzq_args[0] = (unsigned int)(_zzq_request); \
+- _zzq_args[1] = (unsigned int)(_zzq_arg1); \
+- _zzq_args[2] = (unsigned int)(_zzq_arg2); \
+- _zzq_args[3] = (unsigned int)(_zzq_arg3); \
+- _zzq_args[4] = (unsigned int)(_zzq_arg4); \
+- _zzq_args[5] = (unsigned int)(_zzq_arg5); \
+- __asm__ volatile(__SPECIAL_INSTRUCTION_PREAMBLE \
+- /* %EDX = client_request ( %EAX ) */ \
+- "xchgl %%ebx,%%ebx" \
+- : "=d" (_zzq_result) \
+- : "a" (&_zzq_args[0]), "0" (_zzq_default) \
+- : "cc", "memory" \
+- ); \
+- _zzq_rlval = _zzq_result; \
+- }
+-
+-#define VALGRIND_GET_NR_CONTEXT(_zzq_rlval) \
+- { volatile OrigFn* _zzq_orig = &(_zzq_rlval); \
+- volatile unsigned int __addr; \
+- __asm__ volatile(__SPECIAL_INSTRUCTION_PREAMBLE \
+- /* %EAX = guest_NRADDR */ \
+- "xchgl %%ecx,%%ecx" \
+- : "=a" (__addr) \
+- : \
+- : "cc", "memory" \
+- ); \
+- _zzq_orig->nraddr = __addr; \
+- }
+-
+-#define VALGRIND_CALL_NOREDIR_EAX \
+- __SPECIAL_INSTRUCTION_PREAMBLE \
+- /* call-noredir *%EAX */ \
+- "xchgl %%edx,%%edx\n\t"
+-#endif /* PLAT_x86_linux || PLAT_x86_darwin */
+-
+-/* ------------------------ amd64-{linux,darwin} --------------- */
+-
+-#if defined(PLAT_amd64_linux) || defined(PLAT_amd64_darwin)
+-
+-typedef
+- struct {
+- unsigned long long int nraddr; /* where's the code? */
+- }
+- OrigFn;
+-
+-#define __SPECIAL_INSTRUCTION_PREAMBLE \
+- "rolq $3, %%rdi ; rolq $13, %%rdi\n\t" \
+- "rolq $61, %%rdi ; rolq $51, %%rdi\n\t"
+-
+-#define VALGRIND_DO_CLIENT_REQUEST( \
+- _zzq_rlval, _zzq_default, _zzq_request, \
+- _zzq_arg1, _zzq_arg2, _zzq_arg3, _zzq_arg4, _zzq_arg5) \
+- { volatile unsigned long long int _zzq_args[6]; \
+- volatile unsigned long long int _zzq_result; \
+- _zzq_args[0] = (unsigned long long int)(_zzq_request); \
+- _zzq_args[1] = (unsigned long long int)(_zzq_arg1); \
+- _zzq_args[2] = (unsigned long long int)(_zzq_arg2); \
+- _zzq_args[3] = (unsigned long long int)(_zzq_arg3); \
+- _zzq_args[4] = (unsigned long long int)(_zzq_arg4); \
+- _zzq_args[5] = (unsigned long long int)(_zzq_arg5); \
+- __asm__ volatile(__SPECIAL_INSTRUCTION_PREAMBLE \
+- /* %RDX = client_request ( %RAX ) */ \
+- "xchgq %%rbx,%%rbx" \
+- : "=d" (_zzq_result) \
+- : "a" (&_zzq_args[0]), "0" (_zzq_default) \
+- : "cc", "memory" \
+- ); \
+- _zzq_rlval = _zzq_result; \
+- }
+-
+-#define VALGRIND_GET_NR_CONTEXT(_zzq_rlval) \
+- { volatile OrigFn* _zzq_orig = &(_zzq_rlval); \
+- volatile unsigned long long int __addr; \
+- __asm__ volatile(__SPECIAL_INSTRUCTION_PREAMBLE \
+- /* %RAX = guest_NRADDR */ \
+- "xchgq %%rcx,%%rcx" \
+- : "=a" (__addr) \
+- : \
+- : "cc", "memory" \
+- ); \
+- _zzq_orig->nraddr = __addr; \
+- }
+-
+-#define VALGRIND_CALL_NOREDIR_RAX \
+- __SPECIAL_INSTRUCTION_PREAMBLE \
+- /* call-noredir *%RAX */ \
+- "xchgq %%rdx,%%rdx\n\t"
+-#endif /* PLAT_amd64_linux || PLAT_amd64_darwin */
+-
+-/* ------------------------ ppc32-linux ------------------------ */
+-
+-#if defined(PLAT_ppc32_linux)
+-
+-typedef
+- struct {
+- unsigned int nraddr; /* where's the code? */
+- }
+- OrigFn;
+-
+-#define __SPECIAL_INSTRUCTION_PREAMBLE \
+- "rlwinm 0,0,3,0,0 ; rlwinm 0,0,13,0,0\n\t" \
+- "rlwinm 0,0,29,0,0 ; rlwinm 0,0,19,0,0\n\t"
+-
+-#define VALGRIND_DO_CLIENT_REQUEST( \
+- _zzq_rlval, _zzq_default, _zzq_request, \
+- _zzq_arg1, _zzq_arg2, _zzq_arg3, _zzq_arg4, _zzq_arg5) \
+- \
+- { unsigned int _zzq_args[6]; \
+- unsigned int _zzq_result; \
+- unsigned int* _zzq_ptr; \
+- _zzq_args[0] = (unsigned int)(_zzq_request); \
+- _zzq_args[1] = (unsigned int)(_zzq_arg1); \
+- _zzq_args[2] = (unsigned int)(_zzq_arg2); \
+- _zzq_args[3] = (unsigned int)(_zzq_arg3); \
+- _zzq_args[4] = (unsigned int)(_zzq_arg4); \
+- _zzq_args[5] = (unsigned int)(_zzq_arg5); \
+- _zzq_ptr = _zzq_args; \
+- __asm__ volatile("mr 3,%1\n\t" /*default*/ \
+- "mr 4,%2\n\t" /*ptr*/ \
+- __SPECIAL_INSTRUCTION_PREAMBLE \
+- /* %R3 = client_request ( %R4 ) */ \
+- "or 1,1,1\n\t" \
+- "mr %0,3" /*result*/ \
+- : "=b" (_zzq_result) \
+- : "b" (_zzq_default), "b" (_zzq_ptr) \
+- : "cc", "memory", "r3", "r4"); \
+- _zzq_rlval = _zzq_result; \
+- }
+-
+-#define VALGRIND_GET_NR_CONTEXT(_zzq_rlval) \
+- { volatile OrigFn* _zzq_orig = &(_zzq_rlval); \
+- unsigned int __addr; \
+- __asm__ volatile(__SPECIAL_INSTRUCTION_PREAMBLE \
+- /* %R3 = guest_NRADDR */ \
+- "or 2,2,2\n\t" \
+- "mr %0,3" \
+- : "=b" (__addr) \
+- : \
+- : "cc", "memory", "r3" \
+- ); \
+- _zzq_orig->nraddr = __addr; \
+- }
+-
+-#define VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- __SPECIAL_INSTRUCTION_PREAMBLE \
+- /* branch-and-link-to-noredir *%R11 */ \
+- "or 3,3,3\n\t"
+-#endif /* PLAT_ppc32_linux */
+-
+-/* ------------------------ ppc64-linux ------------------------ */
+-
+-#if defined(PLAT_ppc64_linux)
+-
+-typedef
+- struct {
+- unsigned long long int nraddr; /* where's the code? */
+- unsigned long long int r2; /* what tocptr do we need? */
+- }
+- OrigFn;
+-
+-#define __SPECIAL_INSTRUCTION_PREAMBLE \
+- "rotldi 0,0,3 ; rotldi 0,0,13\n\t" \
+- "rotldi 0,0,61 ; rotldi 0,0,51\n\t"
+-
+-#define VALGRIND_DO_CLIENT_REQUEST( \
+- _zzq_rlval, _zzq_default, _zzq_request, \
+- _zzq_arg1, _zzq_arg2, _zzq_arg3, _zzq_arg4, _zzq_arg5) \
+- \
+- { unsigned long long int _zzq_args[6]; \
+- register unsigned long long int _zzq_result __asm__("r3"); \
+- register unsigned long long int* _zzq_ptr __asm__("r4"); \
+- _zzq_args[0] = (unsigned long long int)(_zzq_request); \
+- _zzq_args[1] = (unsigned long long int)(_zzq_arg1); \
+- _zzq_args[2] = (unsigned long long int)(_zzq_arg2); \
+- _zzq_args[3] = (unsigned long long int)(_zzq_arg3); \
+- _zzq_args[4] = (unsigned long long int)(_zzq_arg4); \
+- _zzq_args[5] = (unsigned long long int)(_zzq_arg5); \
+- _zzq_ptr = _zzq_args; \
+- __asm__ volatile(__SPECIAL_INSTRUCTION_PREAMBLE \
+- /* %R3 = client_request ( %R4 ) */ \
+- "or 1,1,1" \
+- : "=r" (_zzq_result) \
+- : "0" (_zzq_default), "r" (_zzq_ptr) \
+- : "cc", "memory"); \
+- _zzq_rlval = _zzq_result; \
+- }
+-
+-#define VALGRIND_GET_NR_CONTEXT(_zzq_rlval) \
+- { volatile OrigFn* _zzq_orig = &(_zzq_rlval); \
+- register unsigned long long int __addr __asm__("r3"); \
+- __asm__ volatile(__SPECIAL_INSTRUCTION_PREAMBLE \
+- /* %R3 = guest_NRADDR */ \
+- "or 2,2,2" \
+- : "=r" (__addr) \
+- : \
+- : "cc", "memory" \
+- ); \
+- _zzq_orig->nraddr = __addr; \
+- __asm__ volatile(__SPECIAL_INSTRUCTION_PREAMBLE \
+- /* %R3 = guest_NRADDR_GPR2 */ \
+- "or 4,4,4" \
+- : "=r" (__addr) \
+- : \
+- : "cc", "memory" \
+- ); \
+- _zzq_orig->r2 = __addr; \
+- }
+-
+-#define VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- __SPECIAL_INSTRUCTION_PREAMBLE \
+- /* branch-and-link-to-noredir *%R11 */ \
+- "or 3,3,3\n\t"
+-
+-#endif /* PLAT_ppc64_linux */
+-
+-/* ------------------------ ppc32-aix5 ------------------------- */
+-
+-#if defined(PLAT_ppc32_aix5)
+-
+-typedef
+- struct {
+- unsigned int nraddr; /* where's the code? */
+- unsigned int r2; /* what tocptr do we need? */
+- }
+- OrigFn;
+-
+-#define __SPECIAL_INSTRUCTION_PREAMBLE \
+- "rlwinm 0,0,3,0,0 ; rlwinm 0,0,13,0,0\n\t" \
+- "rlwinm 0,0,29,0,0 ; rlwinm 0,0,19,0,0\n\t"
+-
+-#define VALGRIND_DO_CLIENT_REQUEST( \
+- _zzq_rlval, _zzq_default, _zzq_request, \
+- _zzq_arg1, _zzq_arg2, _zzq_arg3, _zzq_arg4, _zzq_arg5) \
+- \
+- { unsigned int _zzq_args[7]; \
+- register unsigned int _zzq_result; \
+- register unsigned int* _zzq_ptr; \
+- _zzq_args[0] = (unsigned int)(_zzq_request); \
+- _zzq_args[1] = (unsigned int)(_zzq_arg1); \
+- _zzq_args[2] = (unsigned int)(_zzq_arg2); \
+- _zzq_args[3] = (unsigned int)(_zzq_arg3); \
+- _zzq_args[4] = (unsigned int)(_zzq_arg4); \
+- _zzq_args[5] = (unsigned int)(_zzq_arg5); \
+- _zzq_args[6] = (unsigned int)(_zzq_default); \
+- _zzq_ptr = _zzq_args; \
+- __asm__ volatile("mr 4,%1\n\t" \
+- "lwz 3, 24(4)\n\t" \
+- __SPECIAL_INSTRUCTION_PREAMBLE \
+- /* %R3 = client_request ( %R4 ) */ \
+- "or 1,1,1\n\t" \
+- "mr %0,3" \
+- : "=b" (_zzq_result) \
+- : "b" (_zzq_ptr) \
+- : "r3", "r4", "cc", "memory"); \
+- _zzq_rlval = _zzq_result; \
+- }
+-
+-#define VALGRIND_GET_NR_CONTEXT(_zzq_rlval) \
+- { volatile OrigFn* _zzq_orig = &(_zzq_rlval); \
+- register unsigned int __addr; \
+- __asm__ volatile(__SPECIAL_INSTRUCTION_PREAMBLE \
+- /* %R3 = guest_NRADDR */ \
+- "or 2,2,2\n\t" \
+- "mr %0,3" \
+- : "=b" (__addr) \
+- : \
+- : "r3", "cc", "memory" \
+- ); \
+- _zzq_orig->nraddr = __addr; \
+- __asm__ volatile(__SPECIAL_INSTRUCTION_PREAMBLE \
+- /* %R3 = guest_NRADDR_GPR2 */ \
+- "or 4,4,4\n\t" \
+- "mr %0,3" \
+- : "=b" (__addr) \
+- : \
+- : "r3", "cc", "memory" \
+- ); \
+- _zzq_orig->r2 = __addr; \
+- }
+-
+-#define VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- __SPECIAL_INSTRUCTION_PREAMBLE \
+- /* branch-and-link-to-noredir *%R11 */ \
+- "or 3,3,3\n\t"
+-
+-#endif /* PLAT_ppc32_aix5 */
+-
+-/* ------------------------ ppc64-aix5 ------------------------- */
+-
+-#if defined(PLAT_ppc64_aix5)
+-
+-typedef
+- struct {
+- unsigned long long int nraddr; /* where's the code? */
+- unsigned long long int r2; /* what tocptr do we need? */
+- }
+- OrigFn;
+-
+-#define __SPECIAL_INSTRUCTION_PREAMBLE \
+- "rotldi 0,0,3 ; rotldi 0,0,13\n\t" \
+- "rotldi 0,0,61 ; rotldi 0,0,51\n\t"
+-
+-#define VALGRIND_DO_CLIENT_REQUEST( \
+- _zzq_rlval, _zzq_default, _zzq_request, \
+- _zzq_arg1, _zzq_arg2, _zzq_arg3, _zzq_arg4, _zzq_arg5) \
+- \
+- { unsigned long long int _zzq_args[7]; \
+- register unsigned long long int _zzq_result; \
+- register unsigned long long int* _zzq_ptr; \
+- _zzq_args[0] = (unsigned int long long)(_zzq_request); \
+- _zzq_args[1] = (unsigned int long long)(_zzq_arg1); \
+- _zzq_args[2] = (unsigned int long long)(_zzq_arg2); \
+- _zzq_args[3] = (unsigned int long long)(_zzq_arg3); \
+- _zzq_args[4] = (unsigned int long long)(_zzq_arg4); \
+- _zzq_args[5] = (unsigned int long long)(_zzq_arg5); \
+- _zzq_args[6] = (unsigned int long long)(_zzq_default); \
+- _zzq_ptr = _zzq_args; \
+- __asm__ volatile("mr 4,%1\n\t" \
+- "ld 3, 48(4)\n\t" \
+- __SPECIAL_INSTRUCTION_PREAMBLE \
+- /* %R3 = client_request ( %R4 ) */ \
+- "or 1,1,1\n\t" \
+- "mr %0,3" \
+- : "=b" (_zzq_result) \
+- : "b" (_zzq_ptr) \
+- : "r3", "r4", "cc", "memory"); \
+- _zzq_rlval = _zzq_result; \
+- }
+-
+-#define VALGRIND_GET_NR_CONTEXT(_zzq_rlval) \
+- { volatile OrigFn* _zzq_orig = &(_zzq_rlval); \
+- register unsigned long long int __addr; \
+- __asm__ volatile(__SPECIAL_INSTRUCTION_PREAMBLE \
+- /* %R3 = guest_NRADDR */ \
+- "or 2,2,2\n\t" \
+- "mr %0,3" \
+- : "=b" (__addr) \
+- : \
+- : "r3", "cc", "memory" \
+- ); \
+- _zzq_orig->nraddr = __addr; \
+- __asm__ volatile(__SPECIAL_INSTRUCTION_PREAMBLE \
+- /* %R3 = guest_NRADDR_GPR2 */ \
+- "or 4,4,4\n\t" \
+- "mr %0,3" \
+- : "=b" (__addr) \
+- : \
+- : "r3", "cc", "memory" \
+- ); \
+- _zzq_orig->r2 = __addr; \
+- }
+-
+-#define VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- __SPECIAL_INSTRUCTION_PREAMBLE \
+- /* branch-and-link-to-noredir *%R11 */ \
+- "or 3,3,3\n\t"
+-
+-#endif /* PLAT_ppc64_aix5 */
+-
+-/* Insert assembly code for other platforms here... */
+-
+-#endif /* NVALGRIND */
+-
+-
+-/* ------------------------------------------------------------------ */
+-/* PLATFORM SPECIFICS for FUNCTION WRAPPING. This is all very */
+-/* ugly. It's the least-worst tradeoff I can think of. */
+-/* ------------------------------------------------------------------ */
+-
+-/* This section defines magic (a.k.a appalling-hack) macros for doing
+- guaranteed-no-redirection macros, so as to get from function
+- wrappers to the functions they are wrapping. The whole point is to
+- construct standard call sequences, but to do the call itself with a
+- special no-redirect call pseudo-instruction that the JIT
+- understands and handles specially. This section is long and
+- repetitious, and I can't see a way to make it shorter.
+-
+- The naming scheme is as follows:
+-
+- CALL_FN_{W,v}_{v,W,WW,WWW,WWWW,5W,6W,7W,etc}
+-
+- 'W' stands for "word" and 'v' for "void". Hence there are
+- different macros for calling arity 0, 1, 2, 3, 4, etc, functions,
+- and for each, the possibility of returning a word-typed result, or
+- no result.
+-*/
+-
+-/* Use these to write the name of your wrapper. NOTE: duplicates
+- VG_WRAP_FUNCTION_Z{U,Z} in pub_tool_redir.h. */
+-
+-/* Use an extra level of macroisation so as to ensure the soname/fnname
+- args are fully macro-expanded before pasting them together. */
+-#define VG_CONCAT4(_aa,_bb,_cc,_dd) _aa##_bb##_cc##_dd
+-
+-#define I_WRAP_SONAME_FNNAME_ZU(soname,fnname) \
+- VG_CONCAT4(_vgwZU_,soname,_,fnname)
+-
+-#define I_WRAP_SONAME_FNNAME_ZZ(soname,fnname) \
+- VG_CONCAT4(_vgwZZ_,soname,_,fnname)
+-
+-/* Use this macro from within a wrapper function to collect the
+- context (address and possibly other info) of the original function.
+- Once you have that you can then use it in one of the CALL_FN_
+- macros. The type of the argument _lval is OrigFn. */
+-#define VALGRIND_GET_ORIG_FN(_lval) VALGRIND_GET_NR_CONTEXT(_lval)
+-
+-/* Derivatives of the main macros below, for calling functions
+- returning void. */
+-
+-#define CALL_FN_v_v(fnptr) \
+- do { volatile unsigned long _junk; \
+- CALL_FN_W_v(_junk,fnptr); } while (0)
+-
+-#define CALL_FN_v_W(fnptr, arg1) \
+- do { volatile unsigned long _junk; \
+- CALL_FN_W_W(_junk,fnptr,arg1); } while (0)
+-
+-#define CALL_FN_v_WW(fnptr, arg1,arg2) \
+- do { volatile unsigned long _junk; \
+- CALL_FN_W_WW(_junk,fnptr,arg1,arg2); } while (0)
+-
+-#define CALL_FN_v_WWW(fnptr, arg1,arg2,arg3) \
+- do { volatile unsigned long _junk; \
+- CALL_FN_W_WWW(_junk,fnptr,arg1,arg2,arg3); } while (0)
+-
+-#define CALL_FN_v_WWWW(fnptr, arg1,arg2,arg3,arg4) \
+- do { volatile unsigned long _junk; \
+- CALL_FN_W_WWWW(_junk,fnptr,arg1,arg2,arg3,arg4); } while (0)
+-
+-#define CALL_FN_v_5W(fnptr, arg1,arg2,arg3,arg4,arg5) \
+- do { volatile unsigned long _junk; \
+- CALL_FN_W_5W(_junk,fnptr,arg1,arg2,arg3,arg4,arg5); } while (0)
+-
+-#define CALL_FN_v_6W(fnptr, arg1,arg2,arg3,arg4,arg5,arg6) \
+- do { volatile unsigned long _junk; \
+- CALL_FN_W_6W(_junk,fnptr,arg1,arg2,arg3,arg4,arg5,arg6); } while (0)
+-
+-#define CALL_FN_v_7W(fnptr, arg1,arg2,arg3,arg4,arg5,arg6,arg7) \
+- do { volatile unsigned long _junk; \
+- CALL_FN_W_7W(_junk,fnptr,arg1,arg2,arg3,arg4,arg5,arg6,arg7); } while (0)
+-
+-/* ------------------------- x86-{linux,darwin} ---------------- */
+-
+-#if defined(PLAT_x86_linux) || defined(PLAT_x86_darwin)
+-
+-/* These regs are trashed by the hidden call. No need to mention eax
+- as gcc can already see that, plus causes gcc to bomb. */
+-#define __CALLER_SAVED_REGS /*"eax"*/ "ecx", "edx"
+-
+-/* These CALL_FN_ macros assume that on x86-linux, sizeof(unsigned
+- long) == 4. */
+-
+-#define CALL_FN_W_v(lval, orig) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[1]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- __asm__ volatile( \
+- "movl (%%eax), %%eax\n\t" /* target->%eax */ \
+- VALGRIND_CALL_NOREDIR_EAX \
+- : /*out*/ "=a" (_res) \
+- : /*in*/ "a" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_W(lval, orig, arg1) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[2]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)(arg1); \
+- __asm__ volatile( \
+- "pushl 4(%%eax)\n\t" \
+- "movl (%%eax), %%eax\n\t" /* target->%eax */ \
+- VALGRIND_CALL_NOREDIR_EAX \
+- "addl $4, %%esp\n" \
+- : /*out*/ "=a" (_res) \
+- : /*in*/ "a" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_WW(lval, orig, arg1,arg2) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)(arg1); \
+- _argvec[2] = (unsigned long)(arg2); \
+- __asm__ volatile( \
+- "pushl 8(%%eax)\n\t" \
+- "pushl 4(%%eax)\n\t" \
+- "movl (%%eax), %%eax\n\t" /* target->%eax */ \
+- VALGRIND_CALL_NOREDIR_EAX \
+- "addl $8, %%esp\n" \
+- : /*out*/ "=a" (_res) \
+- : /*in*/ "a" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_WWW(lval, orig, arg1,arg2,arg3) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[4]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)(arg1); \
+- _argvec[2] = (unsigned long)(arg2); \
+- _argvec[3] = (unsigned long)(arg3); \
+- __asm__ volatile( \
+- "pushl 12(%%eax)\n\t" \
+- "pushl 8(%%eax)\n\t" \
+- "pushl 4(%%eax)\n\t" \
+- "movl (%%eax), %%eax\n\t" /* target->%eax */ \
+- VALGRIND_CALL_NOREDIR_EAX \
+- "addl $12, %%esp\n" \
+- : /*out*/ "=a" (_res) \
+- : /*in*/ "a" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_WWWW(lval, orig, arg1,arg2,arg3,arg4) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[5]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)(arg1); \
+- _argvec[2] = (unsigned long)(arg2); \
+- _argvec[3] = (unsigned long)(arg3); \
+- _argvec[4] = (unsigned long)(arg4); \
+- __asm__ volatile( \
+- "pushl 16(%%eax)\n\t" \
+- "pushl 12(%%eax)\n\t" \
+- "pushl 8(%%eax)\n\t" \
+- "pushl 4(%%eax)\n\t" \
+- "movl (%%eax), %%eax\n\t" /* target->%eax */ \
+- VALGRIND_CALL_NOREDIR_EAX \
+- "addl $16, %%esp\n" \
+- : /*out*/ "=a" (_res) \
+- : /*in*/ "a" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_5W(lval, orig, arg1,arg2,arg3,arg4,arg5) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[6]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)(arg1); \
+- _argvec[2] = (unsigned long)(arg2); \
+- _argvec[3] = (unsigned long)(arg3); \
+- _argvec[4] = (unsigned long)(arg4); \
+- _argvec[5] = (unsigned long)(arg5); \
+- __asm__ volatile( \
+- "pushl 20(%%eax)\n\t" \
+- "pushl 16(%%eax)\n\t" \
+- "pushl 12(%%eax)\n\t" \
+- "pushl 8(%%eax)\n\t" \
+- "pushl 4(%%eax)\n\t" \
+- "movl (%%eax), %%eax\n\t" /* target->%eax */ \
+- VALGRIND_CALL_NOREDIR_EAX \
+- "addl $20, %%esp\n" \
+- : /*out*/ "=a" (_res) \
+- : /*in*/ "a" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_6W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[7]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)(arg1); \
+- _argvec[2] = (unsigned long)(arg2); \
+- _argvec[3] = (unsigned long)(arg3); \
+- _argvec[4] = (unsigned long)(arg4); \
+- _argvec[5] = (unsigned long)(arg5); \
+- _argvec[6] = (unsigned long)(arg6); \
+- __asm__ volatile( \
+- "pushl 24(%%eax)\n\t" \
+- "pushl 20(%%eax)\n\t" \
+- "pushl 16(%%eax)\n\t" \
+- "pushl 12(%%eax)\n\t" \
+- "pushl 8(%%eax)\n\t" \
+- "pushl 4(%%eax)\n\t" \
+- "movl (%%eax), %%eax\n\t" /* target->%eax */ \
+- VALGRIND_CALL_NOREDIR_EAX \
+- "addl $24, %%esp\n" \
+- : /*out*/ "=a" (_res) \
+- : /*in*/ "a" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_7W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[8]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)(arg1); \
+- _argvec[2] = (unsigned long)(arg2); \
+- _argvec[3] = (unsigned long)(arg3); \
+- _argvec[4] = (unsigned long)(arg4); \
+- _argvec[5] = (unsigned long)(arg5); \
+- _argvec[6] = (unsigned long)(arg6); \
+- _argvec[7] = (unsigned long)(arg7); \
+- __asm__ volatile( \
+- "pushl 28(%%eax)\n\t" \
+- "pushl 24(%%eax)\n\t" \
+- "pushl 20(%%eax)\n\t" \
+- "pushl 16(%%eax)\n\t" \
+- "pushl 12(%%eax)\n\t" \
+- "pushl 8(%%eax)\n\t" \
+- "pushl 4(%%eax)\n\t" \
+- "movl (%%eax), %%eax\n\t" /* target->%eax */ \
+- VALGRIND_CALL_NOREDIR_EAX \
+- "addl $28, %%esp\n" \
+- : /*out*/ "=a" (_res) \
+- : /*in*/ "a" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_8W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7,arg8) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[9]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)(arg1); \
+- _argvec[2] = (unsigned long)(arg2); \
+- _argvec[3] = (unsigned long)(arg3); \
+- _argvec[4] = (unsigned long)(arg4); \
+- _argvec[5] = (unsigned long)(arg5); \
+- _argvec[6] = (unsigned long)(arg6); \
+- _argvec[7] = (unsigned long)(arg7); \
+- _argvec[8] = (unsigned long)(arg8); \
+- __asm__ volatile( \
+- "pushl 32(%%eax)\n\t" \
+- "pushl 28(%%eax)\n\t" \
+- "pushl 24(%%eax)\n\t" \
+- "pushl 20(%%eax)\n\t" \
+- "pushl 16(%%eax)\n\t" \
+- "pushl 12(%%eax)\n\t" \
+- "pushl 8(%%eax)\n\t" \
+- "pushl 4(%%eax)\n\t" \
+- "movl (%%eax), %%eax\n\t" /* target->%eax */ \
+- VALGRIND_CALL_NOREDIR_EAX \
+- "addl $32, %%esp\n" \
+- : /*out*/ "=a" (_res) \
+- : /*in*/ "a" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_9W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7,arg8,arg9) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[10]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)(arg1); \
+- _argvec[2] = (unsigned long)(arg2); \
+- _argvec[3] = (unsigned long)(arg3); \
+- _argvec[4] = (unsigned long)(arg4); \
+- _argvec[5] = (unsigned long)(arg5); \
+- _argvec[6] = (unsigned long)(arg6); \
+- _argvec[7] = (unsigned long)(arg7); \
+- _argvec[8] = (unsigned long)(arg8); \
+- _argvec[9] = (unsigned long)(arg9); \
+- __asm__ volatile( \
+- "pushl 36(%%eax)\n\t" \
+- "pushl 32(%%eax)\n\t" \
+- "pushl 28(%%eax)\n\t" \
+- "pushl 24(%%eax)\n\t" \
+- "pushl 20(%%eax)\n\t" \
+- "pushl 16(%%eax)\n\t" \
+- "pushl 12(%%eax)\n\t" \
+- "pushl 8(%%eax)\n\t" \
+- "pushl 4(%%eax)\n\t" \
+- "movl (%%eax), %%eax\n\t" /* target->%eax */ \
+- VALGRIND_CALL_NOREDIR_EAX \
+- "addl $36, %%esp\n" \
+- : /*out*/ "=a" (_res) \
+- : /*in*/ "a" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_10W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7,arg8,arg9,arg10) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[11]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)(arg1); \
+- _argvec[2] = (unsigned long)(arg2); \
+- _argvec[3] = (unsigned long)(arg3); \
+- _argvec[4] = (unsigned long)(arg4); \
+- _argvec[5] = (unsigned long)(arg5); \
+- _argvec[6] = (unsigned long)(arg6); \
+- _argvec[7] = (unsigned long)(arg7); \
+- _argvec[8] = (unsigned long)(arg8); \
+- _argvec[9] = (unsigned long)(arg9); \
+- _argvec[10] = (unsigned long)(arg10); \
+- __asm__ volatile( \
+- "pushl 40(%%eax)\n\t" \
+- "pushl 36(%%eax)\n\t" \
+- "pushl 32(%%eax)\n\t" \
+- "pushl 28(%%eax)\n\t" \
+- "pushl 24(%%eax)\n\t" \
+- "pushl 20(%%eax)\n\t" \
+- "pushl 16(%%eax)\n\t" \
+- "pushl 12(%%eax)\n\t" \
+- "pushl 8(%%eax)\n\t" \
+- "pushl 4(%%eax)\n\t" \
+- "movl (%%eax), %%eax\n\t" /* target->%eax */ \
+- VALGRIND_CALL_NOREDIR_EAX \
+- "addl $40, %%esp\n" \
+- : /*out*/ "=a" (_res) \
+- : /*in*/ "a" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_11W(lval, orig, arg1,arg2,arg3,arg4,arg5, \
+- arg6,arg7,arg8,arg9,arg10, \
+- arg11) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[12]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)(arg1); \
+- _argvec[2] = (unsigned long)(arg2); \
+- _argvec[3] = (unsigned long)(arg3); \
+- _argvec[4] = (unsigned long)(arg4); \
+- _argvec[5] = (unsigned long)(arg5); \
+- _argvec[6] = (unsigned long)(arg6); \
+- _argvec[7] = (unsigned long)(arg7); \
+- _argvec[8] = (unsigned long)(arg8); \
+- _argvec[9] = (unsigned long)(arg9); \
+- _argvec[10] = (unsigned long)(arg10); \
+- _argvec[11] = (unsigned long)(arg11); \
+- __asm__ volatile( \
+- "pushl 44(%%eax)\n\t" \
+- "pushl 40(%%eax)\n\t" \
+- "pushl 36(%%eax)\n\t" \
+- "pushl 32(%%eax)\n\t" \
+- "pushl 28(%%eax)\n\t" \
+- "pushl 24(%%eax)\n\t" \
+- "pushl 20(%%eax)\n\t" \
+- "pushl 16(%%eax)\n\t" \
+- "pushl 12(%%eax)\n\t" \
+- "pushl 8(%%eax)\n\t" \
+- "pushl 4(%%eax)\n\t" \
+- "movl (%%eax), %%eax\n\t" /* target->%eax */ \
+- VALGRIND_CALL_NOREDIR_EAX \
+- "addl $44, %%esp\n" \
+- : /*out*/ "=a" (_res) \
+- : /*in*/ "a" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_12W(lval, orig, arg1,arg2,arg3,arg4,arg5, \
+- arg6,arg7,arg8,arg9,arg10, \
+- arg11,arg12) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[13]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)(arg1); \
+- _argvec[2] = (unsigned long)(arg2); \
+- _argvec[3] = (unsigned long)(arg3); \
+- _argvec[4] = (unsigned long)(arg4); \
+- _argvec[5] = (unsigned long)(arg5); \
+- _argvec[6] = (unsigned long)(arg6); \
+- _argvec[7] = (unsigned long)(arg7); \
+- _argvec[8] = (unsigned long)(arg8); \
+- _argvec[9] = (unsigned long)(arg9); \
+- _argvec[10] = (unsigned long)(arg10); \
+- _argvec[11] = (unsigned long)(arg11); \
+- _argvec[12] = (unsigned long)(arg12); \
+- __asm__ volatile( \
+- "pushl 48(%%eax)\n\t" \
+- "pushl 44(%%eax)\n\t" \
+- "pushl 40(%%eax)\n\t" \
+- "pushl 36(%%eax)\n\t" \
+- "pushl 32(%%eax)\n\t" \
+- "pushl 28(%%eax)\n\t" \
+- "pushl 24(%%eax)\n\t" \
+- "pushl 20(%%eax)\n\t" \
+- "pushl 16(%%eax)\n\t" \
+- "pushl 12(%%eax)\n\t" \
+- "pushl 8(%%eax)\n\t" \
+- "pushl 4(%%eax)\n\t" \
+- "movl (%%eax), %%eax\n\t" /* target->%eax */ \
+- VALGRIND_CALL_NOREDIR_EAX \
+- "addl $48, %%esp\n" \
+- : /*out*/ "=a" (_res) \
+- : /*in*/ "a" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#endif /* PLAT_x86_linux || PLAT_x86_darwin */
+-
+-/* ------------------------ amd64-{linux,darwin} --------------- */
+-
+-#if defined(PLAT_amd64_linux) || defined(PLAT_amd64_darwin)
+-
+-/* ARGREGS: rdi rsi rdx rcx r8 r9 (the rest on stack in R-to-L order) */
+-
+-/* These regs are trashed by the hidden call. */
+-#define __CALLER_SAVED_REGS /*"rax",*/ "rcx", "rdx", "rsi", \
+- "rdi", "r8", "r9", "r10", "r11"
+-
+-/* These CALL_FN_ macros assume that on amd64-linux, sizeof(unsigned
+- long) == 8. */
+-
+-/* NB 9 Sept 07. There is a nasty kludge here in all these CALL_FN_
+- macros. In order not to trash the stack redzone, we need to drop
+- %rsp by 128 before the hidden call, and restore afterwards. The
+- nastyness is that it is only by luck that the stack still appears
+- to be unwindable during the hidden call - since then the behaviour
+- of any routine using this macro does not match what the CFI data
+- says. Sigh.
+-
+- Why is this important? Imagine that a wrapper has a stack
+- allocated local, and passes to the hidden call, a pointer to it.
+- Because gcc does not know about the hidden call, it may allocate
+- that local in the redzone. Unfortunately the hidden call may then
+- trash it before it comes to use it. So we must step clear of the
+- redzone, for the duration of the hidden call, to make it safe.
+-
+- Probably the same problem afflicts the other redzone-style ABIs too
+- (ppc64-linux, ppc32-aix5, ppc64-aix5); but for those, the stack is
+- self describing (none of this CFI nonsense) so at least messing
+- with the stack pointer doesn't give a danger of non-unwindable
+- stack. */
+-
+-#define CALL_FN_W_v(lval, orig) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[1]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- __asm__ volatile( \
+- "subq $128,%%rsp\n\t" \
+- "movq (%%rax), %%rax\n\t" /* target->%rax */ \
+- VALGRIND_CALL_NOREDIR_RAX \
+- "addq $128,%%rsp\n\t" \
+- : /*out*/ "=a" (_res) \
+- : /*in*/ "a" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_W(lval, orig, arg1) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[2]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)(arg1); \
+- __asm__ volatile( \
+- "subq $128,%%rsp\n\t" \
+- "movq 8(%%rax), %%rdi\n\t" \
+- "movq (%%rax), %%rax\n\t" /* target->%rax */ \
+- VALGRIND_CALL_NOREDIR_RAX \
+- "addq $128,%%rsp\n\t" \
+- : /*out*/ "=a" (_res) \
+- : /*in*/ "a" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_WW(lval, orig, arg1,arg2) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)(arg1); \
+- _argvec[2] = (unsigned long)(arg2); \
+- __asm__ volatile( \
+- "subq $128,%%rsp\n\t" \
+- "movq 16(%%rax), %%rsi\n\t" \
+- "movq 8(%%rax), %%rdi\n\t" \
+- "movq (%%rax), %%rax\n\t" /* target->%rax */ \
+- VALGRIND_CALL_NOREDIR_RAX \
+- "addq $128,%%rsp\n\t" \
+- : /*out*/ "=a" (_res) \
+- : /*in*/ "a" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_WWW(lval, orig, arg1,arg2,arg3) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[4]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)(arg1); \
+- _argvec[2] = (unsigned long)(arg2); \
+- _argvec[3] = (unsigned long)(arg3); \
+- __asm__ volatile( \
+- "subq $128,%%rsp\n\t" \
+- "movq 24(%%rax), %%rdx\n\t" \
+- "movq 16(%%rax), %%rsi\n\t" \
+- "movq 8(%%rax), %%rdi\n\t" \
+- "movq (%%rax), %%rax\n\t" /* target->%rax */ \
+- VALGRIND_CALL_NOREDIR_RAX \
+- "addq $128,%%rsp\n\t" \
+- : /*out*/ "=a" (_res) \
+- : /*in*/ "a" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_WWWW(lval, orig, arg1,arg2,arg3,arg4) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[5]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)(arg1); \
+- _argvec[2] = (unsigned long)(arg2); \
+- _argvec[3] = (unsigned long)(arg3); \
+- _argvec[4] = (unsigned long)(arg4); \
+- __asm__ volatile( \
+- "subq $128,%%rsp\n\t" \
+- "movq 32(%%rax), %%rcx\n\t" \
+- "movq 24(%%rax), %%rdx\n\t" \
+- "movq 16(%%rax), %%rsi\n\t" \
+- "movq 8(%%rax), %%rdi\n\t" \
+- "movq (%%rax), %%rax\n\t" /* target->%rax */ \
+- VALGRIND_CALL_NOREDIR_RAX \
+- "addq $128,%%rsp\n\t" \
+- : /*out*/ "=a" (_res) \
+- : /*in*/ "a" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_5W(lval, orig, arg1,arg2,arg3,arg4,arg5) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[6]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)(arg1); \
+- _argvec[2] = (unsigned long)(arg2); \
+- _argvec[3] = (unsigned long)(arg3); \
+- _argvec[4] = (unsigned long)(arg4); \
+- _argvec[5] = (unsigned long)(arg5); \
+- __asm__ volatile( \
+- "subq $128,%%rsp\n\t" \
+- "movq 40(%%rax), %%r8\n\t" \
+- "movq 32(%%rax), %%rcx\n\t" \
+- "movq 24(%%rax), %%rdx\n\t" \
+- "movq 16(%%rax), %%rsi\n\t" \
+- "movq 8(%%rax), %%rdi\n\t" \
+- "movq (%%rax), %%rax\n\t" /* target->%rax */ \
+- VALGRIND_CALL_NOREDIR_RAX \
+- "addq $128,%%rsp\n\t" \
+- : /*out*/ "=a" (_res) \
+- : /*in*/ "a" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_6W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[7]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)(arg1); \
+- _argvec[2] = (unsigned long)(arg2); \
+- _argvec[3] = (unsigned long)(arg3); \
+- _argvec[4] = (unsigned long)(arg4); \
+- _argvec[5] = (unsigned long)(arg5); \
+- _argvec[6] = (unsigned long)(arg6); \
+- __asm__ volatile( \
+- "subq $128,%%rsp\n\t" \
+- "movq 48(%%rax), %%r9\n\t" \
+- "movq 40(%%rax), %%r8\n\t" \
+- "movq 32(%%rax), %%rcx\n\t" \
+- "movq 24(%%rax), %%rdx\n\t" \
+- "movq 16(%%rax), %%rsi\n\t" \
+- "movq 8(%%rax), %%rdi\n\t" \
+- "movq (%%rax), %%rax\n\t" /* target->%rax */ \
+- "addq $128,%%rsp\n\t" \
+- VALGRIND_CALL_NOREDIR_RAX \
+- : /*out*/ "=a" (_res) \
+- : /*in*/ "a" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_7W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[8]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)(arg1); \
+- _argvec[2] = (unsigned long)(arg2); \
+- _argvec[3] = (unsigned long)(arg3); \
+- _argvec[4] = (unsigned long)(arg4); \
+- _argvec[5] = (unsigned long)(arg5); \
+- _argvec[6] = (unsigned long)(arg6); \
+- _argvec[7] = (unsigned long)(arg7); \
+- __asm__ volatile( \
+- "subq $128,%%rsp\n\t" \
+- "pushq 56(%%rax)\n\t" \
+- "movq 48(%%rax), %%r9\n\t" \
+- "movq 40(%%rax), %%r8\n\t" \
+- "movq 32(%%rax), %%rcx\n\t" \
+- "movq 24(%%rax), %%rdx\n\t" \
+- "movq 16(%%rax), %%rsi\n\t" \
+- "movq 8(%%rax), %%rdi\n\t" \
+- "movq (%%rax), %%rax\n\t" /* target->%rax */ \
+- VALGRIND_CALL_NOREDIR_RAX \
+- "addq $8, %%rsp\n" \
+- "addq $128,%%rsp\n\t" \
+- : /*out*/ "=a" (_res) \
+- : /*in*/ "a" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_8W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7,arg8) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[9]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)(arg1); \
+- _argvec[2] = (unsigned long)(arg2); \
+- _argvec[3] = (unsigned long)(arg3); \
+- _argvec[4] = (unsigned long)(arg4); \
+- _argvec[5] = (unsigned long)(arg5); \
+- _argvec[6] = (unsigned long)(arg6); \
+- _argvec[7] = (unsigned long)(arg7); \
+- _argvec[8] = (unsigned long)(arg8); \
+- __asm__ volatile( \
+- "subq $128,%%rsp\n\t" \
+- "pushq 64(%%rax)\n\t" \
+- "pushq 56(%%rax)\n\t" \
+- "movq 48(%%rax), %%r9\n\t" \
+- "movq 40(%%rax), %%r8\n\t" \
+- "movq 32(%%rax), %%rcx\n\t" \
+- "movq 24(%%rax), %%rdx\n\t" \
+- "movq 16(%%rax), %%rsi\n\t" \
+- "movq 8(%%rax), %%rdi\n\t" \
+- "movq (%%rax), %%rax\n\t" /* target->%rax */ \
+- VALGRIND_CALL_NOREDIR_RAX \
+- "addq $16, %%rsp\n" \
+- "addq $128,%%rsp\n\t" \
+- : /*out*/ "=a" (_res) \
+- : /*in*/ "a" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_9W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7,arg8,arg9) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[10]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)(arg1); \
+- _argvec[2] = (unsigned long)(arg2); \
+- _argvec[3] = (unsigned long)(arg3); \
+- _argvec[4] = (unsigned long)(arg4); \
+- _argvec[5] = (unsigned long)(arg5); \
+- _argvec[6] = (unsigned long)(arg6); \
+- _argvec[7] = (unsigned long)(arg7); \
+- _argvec[8] = (unsigned long)(arg8); \
+- _argvec[9] = (unsigned long)(arg9); \
+- __asm__ volatile( \
+- "subq $128,%%rsp\n\t" \
+- "pushq 72(%%rax)\n\t" \
+- "pushq 64(%%rax)\n\t" \
+- "pushq 56(%%rax)\n\t" \
+- "movq 48(%%rax), %%r9\n\t" \
+- "movq 40(%%rax), %%r8\n\t" \
+- "movq 32(%%rax), %%rcx\n\t" \
+- "movq 24(%%rax), %%rdx\n\t" \
+- "movq 16(%%rax), %%rsi\n\t" \
+- "movq 8(%%rax), %%rdi\n\t" \
+- "movq (%%rax), %%rax\n\t" /* target->%rax */ \
+- VALGRIND_CALL_NOREDIR_RAX \
+- "addq $24, %%rsp\n" \
+- "addq $128,%%rsp\n\t" \
+- : /*out*/ "=a" (_res) \
+- : /*in*/ "a" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_10W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7,arg8,arg9,arg10) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[11]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)(arg1); \
+- _argvec[2] = (unsigned long)(arg2); \
+- _argvec[3] = (unsigned long)(arg3); \
+- _argvec[4] = (unsigned long)(arg4); \
+- _argvec[5] = (unsigned long)(arg5); \
+- _argvec[6] = (unsigned long)(arg6); \
+- _argvec[7] = (unsigned long)(arg7); \
+- _argvec[8] = (unsigned long)(arg8); \
+- _argvec[9] = (unsigned long)(arg9); \
+- _argvec[10] = (unsigned long)(arg10); \
+- __asm__ volatile( \
+- "subq $128,%%rsp\n\t" \
+- "pushq 80(%%rax)\n\t" \
+- "pushq 72(%%rax)\n\t" \
+- "pushq 64(%%rax)\n\t" \
+- "pushq 56(%%rax)\n\t" \
+- "movq 48(%%rax), %%r9\n\t" \
+- "movq 40(%%rax), %%r8\n\t" \
+- "movq 32(%%rax), %%rcx\n\t" \
+- "movq 24(%%rax), %%rdx\n\t" \
+- "movq 16(%%rax), %%rsi\n\t" \
+- "movq 8(%%rax), %%rdi\n\t" \
+- "movq (%%rax), %%rax\n\t" /* target->%rax */ \
+- VALGRIND_CALL_NOREDIR_RAX \
+- "addq $32, %%rsp\n" \
+- "addq $128,%%rsp\n\t" \
+- : /*out*/ "=a" (_res) \
+- : /*in*/ "a" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_11W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7,arg8,arg9,arg10,arg11) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[12]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)(arg1); \
+- _argvec[2] = (unsigned long)(arg2); \
+- _argvec[3] = (unsigned long)(arg3); \
+- _argvec[4] = (unsigned long)(arg4); \
+- _argvec[5] = (unsigned long)(arg5); \
+- _argvec[6] = (unsigned long)(arg6); \
+- _argvec[7] = (unsigned long)(arg7); \
+- _argvec[8] = (unsigned long)(arg8); \
+- _argvec[9] = (unsigned long)(arg9); \
+- _argvec[10] = (unsigned long)(arg10); \
+- _argvec[11] = (unsigned long)(arg11); \
+- __asm__ volatile( \
+- "subq $128,%%rsp\n\t" \
+- "pushq 88(%%rax)\n\t" \
+- "pushq 80(%%rax)\n\t" \
+- "pushq 72(%%rax)\n\t" \
+- "pushq 64(%%rax)\n\t" \
+- "pushq 56(%%rax)\n\t" \
+- "movq 48(%%rax), %%r9\n\t" \
+- "movq 40(%%rax), %%r8\n\t" \
+- "movq 32(%%rax), %%rcx\n\t" \
+- "movq 24(%%rax), %%rdx\n\t" \
+- "movq 16(%%rax), %%rsi\n\t" \
+- "movq 8(%%rax), %%rdi\n\t" \
+- "movq (%%rax), %%rax\n\t" /* target->%rax */ \
+- VALGRIND_CALL_NOREDIR_RAX \
+- "addq $40, %%rsp\n" \
+- "addq $128,%%rsp\n\t" \
+- : /*out*/ "=a" (_res) \
+- : /*in*/ "a" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_12W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7,arg8,arg9,arg10,arg11,arg12) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[13]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)(arg1); \
+- _argvec[2] = (unsigned long)(arg2); \
+- _argvec[3] = (unsigned long)(arg3); \
+- _argvec[4] = (unsigned long)(arg4); \
+- _argvec[5] = (unsigned long)(arg5); \
+- _argvec[6] = (unsigned long)(arg6); \
+- _argvec[7] = (unsigned long)(arg7); \
+- _argvec[8] = (unsigned long)(arg8); \
+- _argvec[9] = (unsigned long)(arg9); \
+- _argvec[10] = (unsigned long)(arg10); \
+- _argvec[11] = (unsigned long)(arg11); \
+- _argvec[12] = (unsigned long)(arg12); \
+- __asm__ volatile( \
+- "subq $128,%%rsp\n\t" \
+- "pushq 96(%%rax)\n\t" \
+- "pushq 88(%%rax)\n\t" \
+- "pushq 80(%%rax)\n\t" \
+- "pushq 72(%%rax)\n\t" \
+- "pushq 64(%%rax)\n\t" \
+- "pushq 56(%%rax)\n\t" \
+- "movq 48(%%rax), %%r9\n\t" \
+- "movq 40(%%rax), %%r8\n\t" \
+- "movq 32(%%rax), %%rcx\n\t" \
+- "movq 24(%%rax), %%rdx\n\t" \
+- "movq 16(%%rax), %%rsi\n\t" \
+- "movq 8(%%rax), %%rdi\n\t" \
+- "movq (%%rax), %%rax\n\t" /* target->%rax */ \
+- VALGRIND_CALL_NOREDIR_RAX \
+- "addq $48, %%rsp\n" \
+- "addq $128,%%rsp\n\t" \
+- : /*out*/ "=a" (_res) \
+- : /*in*/ "a" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#endif /* PLAT_amd64_linux || PLAT_amd64_darwin */
+-
+-/* ------------------------ ppc32-linux ------------------------ */
+-
+-#if defined(PLAT_ppc32_linux)
+-
+-/* This is useful for finding out about the on-stack stuff:
+-
+- extern int f9 ( int,int,int,int,int,int,int,int,int );
+- extern int f10 ( int,int,int,int,int,int,int,int,int,int );
+- extern int f11 ( int,int,int,int,int,int,int,int,int,int,int );
+- extern int f12 ( int,int,int,int,int,int,int,int,int,int,int,int );
+-
+- int g9 ( void ) {
+- return f9(11,22,33,44,55,66,77,88,99);
+- }
+- int g10 ( void ) {
+- return f10(11,22,33,44,55,66,77,88,99,110);
+- }
+- int g11 ( void ) {
+- return f11(11,22,33,44,55,66,77,88,99,110,121);
+- }
+- int g12 ( void ) {
+- return f12(11,22,33,44,55,66,77,88,99,110,121,132);
+- }
+-*/
+-
+-/* ARGREGS: r3 r4 r5 r6 r7 r8 r9 r10 (the rest on stack somewhere) */
+-
+-/* These regs are trashed by the hidden call. */
+-#define __CALLER_SAVED_REGS \
+- "lr", "ctr", "xer", \
+- "cr0", "cr1", "cr2", "cr3", "cr4", "cr5", "cr6", "cr7", \
+- "r0", "r2", "r3", "r4", "r5", "r6", "r7", "r8", "r9", "r10", \
+- "r11", "r12", "r13"
+-
+-/* These CALL_FN_ macros assume that on ppc32-linux,
+- sizeof(unsigned long) == 4. */
+-
+-#define CALL_FN_W_v(lval, orig) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[1]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- "lwz 11,0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr %0,3" \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_W(lval, orig, arg1) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[2]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)arg1; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- "lwz 3,4(11)\n\t" /* arg1->r3 */ \
+- "lwz 11,0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr %0,3" \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_WW(lval, orig, arg1,arg2) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)arg1; \
+- _argvec[2] = (unsigned long)arg2; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- "lwz 3,4(11)\n\t" /* arg1->r3 */ \
+- "lwz 4,8(11)\n\t" \
+- "lwz 11,0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr %0,3" \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_WWW(lval, orig, arg1,arg2,arg3) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[4]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)arg1; \
+- _argvec[2] = (unsigned long)arg2; \
+- _argvec[3] = (unsigned long)arg3; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- "lwz 3,4(11)\n\t" /* arg1->r3 */ \
+- "lwz 4,8(11)\n\t" \
+- "lwz 5,12(11)\n\t" \
+- "lwz 11,0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr %0,3" \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_WWWW(lval, orig, arg1,arg2,arg3,arg4) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[5]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)arg1; \
+- _argvec[2] = (unsigned long)arg2; \
+- _argvec[3] = (unsigned long)arg3; \
+- _argvec[4] = (unsigned long)arg4; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- "lwz 3,4(11)\n\t" /* arg1->r3 */ \
+- "lwz 4,8(11)\n\t" \
+- "lwz 5,12(11)\n\t" \
+- "lwz 6,16(11)\n\t" /* arg4->r6 */ \
+- "lwz 11,0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr %0,3" \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_5W(lval, orig, arg1,arg2,arg3,arg4,arg5) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[6]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)arg1; \
+- _argvec[2] = (unsigned long)arg2; \
+- _argvec[3] = (unsigned long)arg3; \
+- _argvec[4] = (unsigned long)arg4; \
+- _argvec[5] = (unsigned long)arg5; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- "lwz 3,4(11)\n\t" /* arg1->r3 */ \
+- "lwz 4,8(11)\n\t" \
+- "lwz 5,12(11)\n\t" \
+- "lwz 6,16(11)\n\t" /* arg4->r6 */ \
+- "lwz 7,20(11)\n\t" \
+- "lwz 11,0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr %0,3" \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_6W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[7]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)arg1; \
+- _argvec[2] = (unsigned long)arg2; \
+- _argvec[3] = (unsigned long)arg3; \
+- _argvec[4] = (unsigned long)arg4; \
+- _argvec[5] = (unsigned long)arg5; \
+- _argvec[6] = (unsigned long)arg6; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- "lwz 3,4(11)\n\t" /* arg1->r3 */ \
+- "lwz 4,8(11)\n\t" \
+- "lwz 5,12(11)\n\t" \
+- "lwz 6,16(11)\n\t" /* arg4->r6 */ \
+- "lwz 7,20(11)\n\t" \
+- "lwz 8,24(11)\n\t" \
+- "lwz 11,0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr %0,3" \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_7W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[8]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)arg1; \
+- _argvec[2] = (unsigned long)arg2; \
+- _argvec[3] = (unsigned long)arg3; \
+- _argvec[4] = (unsigned long)arg4; \
+- _argvec[5] = (unsigned long)arg5; \
+- _argvec[6] = (unsigned long)arg6; \
+- _argvec[7] = (unsigned long)arg7; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- "lwz 3,4(11)\n\t" /* arg1->r3 */ \
+- "lwz 4,8(11)\n\t" \
+- "lwz 5,12(11)\n\t" \
+- "lwz 6,16(11)\n\t" /* arg4->r6 */ \
+- "lwz 7,20(11)\n\t" \
+- "lwz 8,24(11)\n\t" \
+- "lwz 9,28(11)\n\t" \
+- "lwz 11,0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr %0,3" \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_8W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7,arg8) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[9]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)arg1; \
+- _argvec[2] = (unsigned long)arg2; \
+- _argvec[3] = (unsigned long)arg3; \
+- _argvec[4] = (unsigned long)arg4; \
+- _argvec[5] = (unsigned long)arg5; \
+- _argvec[6] = (unsigned long)arg6; \
+- _argvec[7] = (unsigned long)arg7; \
+- _argvec[8] = (unsigned long)arg8; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- "lwz 3,4(11)\n\t" /* arg1->r3 */ \
+- "lwz 4,8(11)\n\t" \
+- "lwz 5,12(11)\n\t" \
+- "lwz 6,16(11)\n\t" /* arg4->r6 */ \
+- "lwz 7,20(11)\n\t" \
+- "lwz 8,24(11)\n\t" \
+- "lwz 9,28(11)\n\t" \
+- "lwz 10,32(11)\n\t" /* arg8->r10 */ \
+- "lwz 11,0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr %0,3" \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_9W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7,arg8,arg9) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[10]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)arg1; \
+- _argvec[2] = (unsigned long)arg2; \
+- _argvec[3] = (unsigned long)arg3; \
+- _argvec[4] = (unsigned long)arg4; \
+- _argvec[5] = (unsigned long)arg5; \
+- _argvec[6] = (unsigned long)arg6; \
+- _argvec[7] = (unsigned long)arg7; \
+- _argvec[8] = (unsigned long)arg8; \
+- _argvec[9] = (unsigned long)arg9; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- "addi 1,1,-16\n\t" \
+- /* arg9 */ \
+- "lwz 3,36(11)\n\t" \
+- "stw 3,8(1)\n\t" \
+- /* args1-8 */ \
+- "lwz 3,4(11)\n\t" /* arg1->r3 */ \
+- "lwz 4,8(11)\n\t" \
+- "lwz 5,12(11)\n\t" \
+- "lwz 6,16(11)\n\t" /* arg4->r6 */ \
+- "lwz 7,20(11)\n\t" \
+- "lwz 8,24(11)\n\t" \
+- "lwz 9,28(11)\n\t" \
+- "lwz 10,32(11)\n\t" /* arg8->r10 */ \
+- "lwz 11,0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "addi 1,1,16\n\t" \
+- "mr %0,3" \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_10W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7,arg8,arg9,arg10) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[11]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)arg1; \
+- _argvec[2] = (unsigned long)arg2; \
+- _argvec[3] = (unsigned long)arg3; \
+- _argvec[4] = (unsigned long)arg4; \
+- _argvec[5] = (unsigned long)arg5; \
+- _argvec[6] = (unsigned long)arg6; \
+- _argvec[7] = (unsigned long)arg7; \
+- _argvec[8] = (unsigned long)arg8; \
+- _argvec[9] = (unsigned long)arg9; \
+- _argvec[10] = (unsigned long)arg10; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- "addi 1,1,-16\n\t" \
+- /* arg10 */ \
+- "lwz 3,40(11)\n\t" \
+- "stw 3,12(1)\n\t" \
+- /* arg9 */ \
+- "lwz 3,36(11)\n\t" \
+- "stw 3,8(1)\n\t" \
+- /* args1-8 */ \
+- "lwz 3,4(11)\n\t" /* arg1->r3 */ \
+- "lwz 4,8(11)\n\t" \
+- "lwz 5,12(11)\n\t" \
+- "lwz 6,16(11)\n\t" /* arg4->r6 */ \
+- "lwz 7,20(11)\n\t" \
+- "lwz 8,24(11)\n\t" \
+- "lwz 9,28(11)\n\t" \
+- "lwz 10,32(11)\n\t" /* arg8->r10 */ \
+- "lwz 11,0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "addi 1,1,16\n\t" \
+- "mr %0,3" \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_11W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7,arg8,arg9,arg10,arg11) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[12]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)arg1; \
+- _argvec[2] = (unsigned long)arg2; \
+- _argvec[3] = (unsigned long)arg3; \
+- _argvec[4] = (unsigned long)arg4; \
+- _argvec[5] = (unsigned long)arg5; \
+- _argvec[6] = (unsigned long)arg6; \
+- _argvec[7] = (unsigned long)arg7; \
+- _argvec[8] = (unsigned long)arg8; \
+- _argvec[9] = (unsigned long)arg9; \
+- _argvec[10] = (unsigned long)arg10; \
+- _argvec[11] = (unsigned long)arg11; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- "addi 1,1,-32\n\t" \
+- /* arg11 */ \
+- "lwz 3,44(11)\n\t" \
+- "stw 3,16(1)\n\t" \
+- /* arg10 */ \
+- "lwz 3,40(11)\n\t" \
+- "stw 3,12(1)\n\t" \
+- /* arg9 */ \
+- "lwz 3,36(11)\n\t" \
+- "stw 3,8(1)\n\t" \
+- /* args1-8 */ \
+- "lwz 3,4(11)\n\t" /* arg1->r3 */ \
+- "lwz 4,8(11)\n\t" \
+- "lwz 5,12(11)\n\t" \
+- "lwz 6,16(11)\n\t" /* arg4->r6 */ \
+- "lwz 7,20(11)\n\t" \
+- "lwz 8,24(11)\n\t" \
+- "lwz 9,28(11)\n\t" \
+- "lwz 10,32(11)\n\t" /* arg8->r10 */ \
+- "lwz 11,0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "addi 1,1,32\n\t" \
+- "mr %0,3" \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_12W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7,arg8,arg9,arg10,arg11,arg12) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[13]; \
+- volatile unsigned long _res; \
+- _argvec[0] = (unsigned long)_orig.nraddr; \
+- _argvec[1] = (unsigned long)arg1; \
+- _argvec[2] = (unsigned long)arg2; \
+- _argvec[3] = (unsigned long)arg3; \
+- _argvec[4] = (unsigned long)arg4; \
+- _argvec[5] = (unsigned long)arg5; \
+- _argvec[6] = (unsigned long)arg6; \
+- _argvec[7] = (unsigned long)arg7; \
+- _argvec[8] = (unsigned long)arg8; \
+- _argvec[9] = (unsigned long)arg9; \
+- _argvec[10] = (unsigned long)arg10; \
+- _argvec[11] = (unsigned long)arg11; \
+- _argvec[12] = (unsigned long)arg12; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- "addi 1,1,-32\n\t" \
+- /* arg12 */ \
+- "lwz 3,48(11)\n\t" \
+- "stw 3,20(1)\n\t" \
+- /* arg11 */ \
+- "lwz 3,44(11)\n\t" \
+- "stw 3,16(1)\n\t" \
+- /* arg10 */ \
+- "lwz 3,40(11)\n\t" \
+- "stw 3,12(1)\n\t" \
+- /* arg9 */ \
+- "lwz 3,36(11)\n\t" \
+- "stw 3,8(1)\n\t" \
+- /* args1-8 */ \
+- "lwz 3,4(11)\n\t" /* arg1->r3 */ \
+- "lwz 4,8(11)\n\t" \
+- "lwz 5,12(11)\n\t" \
+- "lwz 6,16(11)\n\t" /* arg4->r6 */ \
+- "lwz 7,20(11)\n\t" \
+- "lwz 8,24(11)\n\t" \
+- "lwz 9,28(11)\n\t" \
+- "lwz 10,32(11)\n\t" /* arg8->r10 */ \
+- "lwz 11,0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "addi 1,1,32\n\t" \
+- "mr %0,3" \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[0]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#endif /* PLAT_ppc32_linux */
+-
+-/* ------------------------ ppc64-linux ------------------------ */
+-
+-#if defined(PLAT_ppc64_linux)
+-
+-/* ARGREGS: r3 r4 r5 r6 r7 r8 r9 r10 (the rest on stack somewhere) */
+-
+-/* These regs are trashed by the hidden call. */
+-#define __CALLER_SAVED_REGS \
+- "lr", "ctr", "xer", \
+- "cr0", "cr1", "cr2", "cr3", "cr4", "cr5", "cr6", "cr7", \
+- "r0", "r2", "r3", "r4", "r5", "r6", "r7", "r8", "r9", "r10", \
+- "r11", "r12", "r13"
+-
+-/* These CALL_FN_ macros assume that on ppc64-linux, sizeof(unsigned
+- long) == 8. */
+-
+-#define CALL_FN_W_v(lval, orig) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+0]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- "std 2,-16(11)\n\t" /* save tocptr */ \
+- "ld 2,-8(11)\n\t" /* use nraddr's tocptr */ \
+- "ld 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "ld 2,-16(11)" /* restore tocptr */ \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_W(lval, orig, arg1) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+1]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- "std 2,-16(11)\n\t" /* save tocptr */ \
+- "ld 2,-8(11)\n\t" /* use nraddr's tocptr */ \
+- "ld 3, 8(11)\n\t" /* arg1->r3 */ \
+- "ld 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "ld 2,-16(11)" /* restore tocptr */ \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_WW(lval, orig, arg1,arg2) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+2]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- "std 2,-16(11)\n\t" /* save tocptr */ \
+- "ld 2,-8(11)\n\t" /* use nraddr's tocptr */ \
+- "ld 3, 8(11)\n\t" /* arg1->r3 */ \
+- "ld 4, 16(11)\n\t" /* arg2->r4 */ \
+- "ld 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "ld 2,-16(11)" /* restore tocptr */ \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_WWW(lval, orig, arg1,arg2,arg3) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+3]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- "std 2,-16(11)\n\t" /* save tocptr */ \
+- "ld 2,-8(11)\n\t" /* use nraddr's tocptr */ \
+- "ld 3, 8(11)\n\t" /* arg1->r3 */ \
+- "ld 4, 16(11)\n\t" /* arg2->r4 */ \
+- "ld 5, 24(11)\n\t" /* arg3->r5 */ \
+- "ld 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "ld 2,-16(11)" /* restore tocptr */ \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_WWWW(lval, orig, arg1,arg2,arg3,arg4) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+4]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- _argvec[2+4] = (unsigned long)arg4; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- "std 2,-16(11)\n\t" /* save tocptr */ \
+- "ld 2,-8(11)\n\t" /* use nraddr's tocptr */ \
+- "ld 3, 8(11)\n\t" /* arg1->r3 */ \
+- "ld 4, 16(11)\n\t" /* arg2->r4 */ \
+- "ld 5, 24(11)\n\t" /* arg3->r5 */ \
+- "ld 6, 32(11)\n\t" /* arg4->r6 */ \
+- "ld 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "ld 2,-16(11)" /* restore tocptr */ \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_5W(lval, orig, arg1,arg2,arg3,arg4,arg5) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+5]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- _argvec[2+4] = (unsigned long)arg4; \
+- _argvec[2+5] = (unsigned long)arg5; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- "std 2,-16(11)\n\t" /* save tocptr */ \
+- "ld 2,-8(11)\n\t" /* use nraddr's tocptr */ \
+- "ld 3, 8(11)\n\t" /* arg1->r3 */ \
+- "ld 4, 16(11)\n\t" /* arg2->r4 */ \
+- "ld 5, 24(11)\n\t" /* arg3->r5 */ \
+- "ld 6, 32(11)\n\t" /* arg4->r6 */ \
+- "ld 7, 40(11)\n\t" /* arg5->r7 */ \
+- "ld 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "ld 2,-16(11)" /* restore tocptr */ \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_6W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+6]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- _argvec[2+4] = (unsigned long)arg4; \
+- _argvec[2+5] = (unsigned long)arg5; \
+- _argvec[2+6] = (unsigned long)arg6; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- "std 2,-16(11)\n\t" /* save tocptr */ \
+- "ld 2,-8(11)\n\t" /* use nraddr's tocptr */ \
+- "ld 3, 8(11)\n\t" /* arg1->r3 */ \
+- "ld 4, 16(11)\n\t" /* arg2->r4 */ \
+- "ld 5, 24(11)\n\t" /* arg3->r5 */ \
+- "ld 6, 32(11)\n\t" /* arg4->r6 */ \
+- "ld 7, 40(11)\n\t" /* arg5->r7 */ \
+- "ld 8, 48(11)\n\t" /* arg6->r8 */ \
+- "ld 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "ld 2,-16(11)" /* restore tocptr */ \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_7W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+7]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- _argvec[2+4] = (unsigned long)arg4; \
+- _argvec[2+5] = (unsigned long)arg5; \
+- _argvec[2+6] = (unsigned long)arg6; \
+- _argvec[2+7] = (unsigned long)arg7; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- "std 2,-16(11)\n\t" /* save tocptr */ \
+- "ld 2,-8(11)\n\t" /* use nraddr's tocptr */ \
+- "ld 3, 8(11)\n\t" /* arg1->r3 */ \
+- "ld 4, 16(11)\n\t" /* arg2->r4 */ \
+- "ld 5, 24(11)\n\t" /* arg3->r5 */ \
+- "ld 6, 32(11)\n\t" /* arg4->r6 */ \
+- "ld 7, 40(11)\n\t" /* arg5->r7 */ \
+- "ld 8, 48(11)\n\t" /* arg6->r8 */ \
+- "ld 9, 56(11)\n\t" /* arg7->r9 */ \
+- "ld 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "ld 2,-16(11)" /* restore tocptr */ \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_8W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7,arg8) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+8]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- _argvec[2+4] = (unsigned long)arg4; \
+- _argvec[2+5] = (unsigned long)arg5; \
+- _argvec[2+6] = (unsigned long)arg6; \
+- _argvec[2+7] = (unsigned long)arg7; \
+- _argvec[2+8] = (unsigned long)arg8; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- "std 2,-16(11)\n\t" /* save tocptr */ \
+- "ld 2,-8(11)\n\t" /* use nraddr's tocptr */ \
+- "ld 3, 8(11)\n\t" /* arg1->r3 */ \
+- "ld 4, 16(11)\n\t" /* arg2->r4 */ \
+- "ld 5, 24(11)\n\t" /* arg3->r5 */ \
+- "ld 6, 32(11)\n\t" /* arg4->r6 */ \
+- "ld 7, 40(11)\n\t" /* arg5->r7 */ \
+- "ld 8, 48(11)\n\t" /* arg6->r8 */ \
+- "ld 9, 56(11)\n\t" /* arg7->r9 */ \
+- "ld 10, 64(11)\n\t" /* arg8->r10 */ \
+- "ld 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "ld 2,-16(11)" /* restore tocptr */ \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_9W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7,arg8,arg9) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+9]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- _argvec[2+4] = (unsigned long)arg4; \
+- _argvec[2+5] = (unsigned long)arg5; \
+- _argvec[2+6] = (unsigned long)arg6; \
+- _argvec[2+7] = (unsigned long)arg7; \
+- _argvec[2+8] = (unsigned long)arg8; \
+- _argvec[2+9] = (unsigned long)arg9; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- "std 2,-16(11)\n\t" /* save tocptr */ \
+- "ld 2,-8(11)\n\t" /* use nraddr's tocptr */ \
+- "addi 1,1,-128\n\t" /* expand stack frame */ \
+- /* arg9 */ \
+- "ld 3,72(11)\n\t" \
+- "std 3,112(1)\n\t" \
+- /* args1-8 */ \
+- "ld 3, 8(11)\n\t" /* arg1->r3 */ \
+- "ld 4, 16(11)\n\t" /* arg2->r4 */ \
+- "ld 5, 24(11)\n\t" /* arg3->r5 */ \
+- "ld 6, 32(11)\n\t" /* arg4->r6 */ \
+- "ld 7, 40(11)\n\t" /* arg5->r7 */ \
+- "ld 8, 48(11)\n\t" /* arg6->r8 */ \
+- "ld 9, 56(11)\n\t" /* arg7->r9 */ \
+- "ld 10, 64(11)\n\t" /* arg8->r10 */ \
+- "ld 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "ld 2,-16(11)\n\t" /* restore tocptr */ \
+- "addi 1,1,128" /* restore frame */ \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_10W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7,arg8,arg9,arg10) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+10]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- _argvec[2+4] = (unsigned long)arg4; \
+- _argvec[2+5] = (unsigned long)arg5; \
+- _argvec[2+6] = (unsigned long)arg6; \
+- _argvec[2+7] = (unsigned long)arg7; \
+- _argvec[2+8] = (unsigned long)arg8; \
+- _argvec[2+9] = (unsigned long)arg9; \
+- _argvec[2+10] = (unsigned long)arg10; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- "std 2,-16(11)\n\t" /* save tocptr */ \
+- "ld 2,-8(11)\n\t" /* use nraddr's tocptr */ \
+- "addi 1,1,-128\n\t" /* expand stack frame */ \
+- /* arg10 */ \
+- "ld 3,80(11)\n\t" \
+- "std 3,120(1)\n\t" \
+- /* arg9 */ \
+- "ld 3,72(11)\n\t" \
+- "std 3,112(1)\n\t" \
+- /* args1-8 */ \
+- "ld 3, 8(11)\n\t" /* arg1->r3 */ \
+- "ld 4, 16(11)\n\t" /* arg2->r4 */ \
+- "ld 5, 24(11)\n\t" /* arg3->r5 */ \
+- "ld 6, 32(11)\n\t" /* arg4->r6 */ \
+- "ld 7, 40(11)\n\t" /* arg5->r7 */ \
+- "ld 8, 48(11)\n\t" /* arg6->r8 */ \
+- "ld 9, 56(11)\n\t" /* arg7->r9 */ \
+- "ld 10, 64(11)\n\t" /* arg8->r10 */ \
+- "ld 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "ld 2,-16(11)\n\t" /* restore tocptr */ \
+- "addi 1,1,128" /* restore frame */ \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_11W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7,arg8,arg9,arg10,arg11) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+11]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- _argvec[2+4] = (unsigned long)arg4; \
+- _argvec[2+5] = (unsigned long)arg5; \
+- _argvec[2+6] = (unsigned long)arg6; \
+- _argvec[2+7] = (unsigned long)arg7; \
+- _argvec[2+8] = (unsigned long)arg8; \
+- _argvec[2+9] = (unsigned long)arg9; \
+- _argvec[2+10] = (unsigned long)arg10; \
+- _argvec[2+11] = (unsigned long)arg11; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- "std 2,-16(11)\n\t" /* save tocptr */ \
+- "ld 2,-8(11)\n\t" /* use nraddr's tocptr */ \
+- "addi 1,1,-144\n\t" /* expand stack frame */ \
+- /* arg11 */ \
+- "ld 3,88(11)\n\t" \
+- "std 3,128(1)\n\t" \
+- /* arg10 */ \
+- "ld 3,80(11)\n\t" \
+- "std 3,120(1)\n\t" \
+- /* arg9 */ \
+- "ld 3,72(11)\n\t" \
+- "std 3,112(1)\n\t" \
+- /* args1-8 */ \
+- "ld 3, 8(11)\n\t" /* arg1->r3 */ \
+- "ld 4, 16(11)\n\t" /* arg2->r4 */ \
+- "ld 5, 24(11)\n\t" /* arg3->r5 */ \
+- "ld 6, 32(11)\n\t" /* arg4->r6 */ \
+- "ld 7, 40(11)\n\t" /* arg5->r7 */ \
+- "ld 8, 48(11)\n\t" /* arg6->r8 */ \
+- "ld 9, 56(11)\n\t" /* arg7->r9 */ \
+- "ld 10, 64(11)\n\t" /* arg8->r10 */ \
+- "ld 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "ld 2,-16(11)\n\t" /* restore tocptr */ \
+- "addi 1,1,144" /* restore frame */ \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_12W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7,arg8,arg9,arg10,arg11,arg12) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+12]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- _argvec[2+4] = (unsigned long)arg4; \
+- _argvec[2+5] = (unsigned long)arg5; \
+- _argvec[2+6] = (unsigned long)arg6; \
+- _argvec[2+7] = (unsigned long)arg7; \
+- _argvec[2+8] = (unsigned long)arg8; \
+- _argvec[2+9] = (unsigned long)arg9; \
+- _argvec[2+10] = (unsigned long)arg10; \
+- _argvec[2+11] = (unsigned long)arg11; \
+- _argvec[2+12] = (unsigned long)arg12; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- "std 2,-16(11)\n\t" /* save tocptr */ \
+- "ld 2,-8(11)\n\t" /* use nraddr's tocptr */ \
+- "addi 1,1,-144\n\t" /* expand stack frame */ \
+- /* arg12 */ \
+- "ld 3,96(11)\n\t" \
+- "std 3,136(1)\n\t" \
+- /* arg11 */ \
+- "ld 3,88(11)\n\t" \
+- "std 3,128(1)\n\t" \
+- /* arg10 */ \
+- "ld 3,80(11)\n\t" \
+- "std 3,120(1)\n\t" \
+- /* arg9 */ \
+- "ld 3,72(11)\n\t" \
+- "std 3,112(1)\n\t" \
+- /* args1-8 */ \
+- "ld 3, 8(11)\n\t" /* arg1->r3 */ \
+- "ld 4, 16(11)\n\t" /* arg2->r4 */ \
+- "ld 5, 24(11)\n\t" /* arg3->r5 */ \
+- "ld 6, 32(11)\n\t" /* arg4->r6 */ \
+- "ld 7, 40(11)\n\t" /* arg5->r7 */ \
+- "ld 8, 48(11)\n\t" /* arg6->r8 */ \
+- "ld 9, 56(11)\n\t" /* arg7->r9 */ \
+- "ld 10, 64(11)\n\t" /* arg8->r10 */ \
+- "ld 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "ld 2,-16(11)\n\t" /* restore tocptr */ \
+- "addi 1,1,144" /* restore frame */ \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#endif /* PLAT_ppc64_linux */
+-
+-/* ------------------------ ppc32-aix5 ------------------------- */
+-
+-#if defined(PLAT_ppc32_aix5)
+-
+-/* ARGREGS: r3 r4 r5 r6 r7 r8 r9 r10 (the rest on stack somewhere) */
+-
+-/* These regs are trashed by the hidden call. */
+-#define __CALLER_SAVED_REGS \
+- "lr", "ctr", "xer", \
+- "cr0", "cr1", "cr2", "cr3", "cr4", "cr5", "cr6", "cr7", \
+- "r0", "r2", "r3", "r4", "r5", "r6", "r7", "r8", "r9", "r10", \
+- "r11", "r12", "r13"
+-
+-/* Expand the stack frame, copying enough info that unwinding
+- still works. Trashes r3. */
+-
+-#define VG_EXPAND_FRAME_BY_trashes_r3(_n_fr) \
+- "addi 1,1,-" #_n_fr "\n\t" \
+- "lwz 3," #_n_fr "(1)\n\t" \
+- "stw 3,0(1)\n\t"
+-
+-#define VG_CONTRACT_FRAME_BY(_n_fr) \
+- "addi 1,1," #_n_fr "\n\t"
+-
+-/* These CALL_FN_ macros assume that on ppc32-aix5, sizeof(unsigned
+- long) == 4. */
+-
+-#define CALL_FN_W_v(lval, orig) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+0]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- VG_EXPAND_FRAME_BY_trashes_r3(512) \
+- "stw 2,-8(11)\n\t" /* save tocptr */ \
+- "lwz 2,-4(11)\n\t" /* use nraddr's tocptr */ \
+- "lwz 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "lwz 2,-8(11)\n\t" /* restore tocptr */ \
+- VG_CONTRACT_FRAME_BY(512) \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_W(lval, orig, arg1) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+1]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- VG_EXPAND_FRAME_BY_trashes_r3(512) \
+- "stw 2,-8(11)\n\t" /* save tocptr */ \
+- "lwz 2,-4(11)\n\t" /* use nraddr's tocptr */ \
+- "lwz 3, 4(11)\n\t" /* arg1->r3 */ \
+- "lwz 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "lwz 2,-8(11)\n\t" /* restore tocptr */ \
+- VG_CONTRACT_FRAME_BY(512) \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_WW(lval, orig, arg1,arg2) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+2]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- VG_EXPAND_FRAME_BY_trashes_r3(512) \
+- "stw 2,-8(11)\n\t" /* save tocptr */ \
+- "lwz 2,-4(11)\n\t" /* use nraddr's tocptr */ \
+- "lwz 3, 4(11)\n\t" /* arg1->r3 */ \
+- "lwz 4, 8(11)\n\t" /* arg2->r4 */ \
+- "lwz 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "lwz 2,-8(11)\n\t" /* restore tocptr */ \
+- VG_CONTRACT_FRAME_BY(512) \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_WWW(lval, orig, arg1,arg2,arg3) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+3]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- VG_EXPAND_FRAME_BY_trashes_r3(512) \
+- "stw 2,-8(11)\n\t" /* save tocptr */ \
+- "lwz 2,-4(11)\n\t" /* use nraddr's tocptr */ \
+- "lwz 3, 4(11)\n\t" /* arg1->r3 */ \
+- "lwz 4, 8(11)\n\t" /* arg2->r4 */ \
+- "lwz 5, 12(11)\n\t" /* arg3->r5 */ \
+- "lwz 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "lwz 2,-8(11)\n\t" /* restore tocptr */ \
+- VG_CONTRACT_FRAME_BY(512) \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_WWWW(lval, orig, arg1,arg2,arg3,arg4) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+4]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- _argvec[2+4] = (unsigned long)arg4; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- VG_EXPAND_FRAME_BY_trashes_r3(512) \
+- "stw 2,-8(11)\n\t" /* save tocptr */ \
+- "lwz 2,-4(11)\n\t" /* use nraddr's tocptr */ \
+- "lwz 3, 4(11)\n\t" /* arg1->r3 */ \
+- "lwz 4, 8(11)\n\t" /* arg2->r4 */ \
+- "lwz 5, 12(11)\n\t" /* arg3->r5 */ \
+- "lwz 6, 16(11)\n\t" /* arg4->r6 */ \
+- "lwz 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "lwz 2,-8(11)\n\t" /* restore tocptr */ \
+- VG_CONTRACT_FRAME_BY(512) \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_5W(lval, orig, arg1,arg2,arg3,arg4,arg5) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+5]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- _argvec[2+4] = (unsigned long)arg4; \
+- _argvec[2+5] = (unsigned long)arg5; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- VG_EXPAND_FRAME_BY_trashes_r3(512) \
+- "stw 2,-8(11)\n\t" /* save tocptr */ \
+- "lwz 2,-4(11)\n\t" /* use nraddr's tocptr */ \
+- "lwz 3, 4(11)\n\t" /* arg1->r3 */ \
+- "lwz 4, 8(11)\n\t" /* arg2->r4 */ \
+- "lwz 5, 12(11)\n\t" /* arg3->r5 */ \
+- "lwz 6, 16(11)\n\t" /* arg4->r6 */ \
+- "lwz 7, 20(11)\n\t" /* arg5->r7 */ \
+- "lwz 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "lwz 2,-8(11)\n\t" /* restore tocptr */ \
+- VG_CONTRACT_FRAME_BY(512) \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_6W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+6]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- _argvec[2+4] = (unsigned long)arg4; \
+- _argvec[2+5] = (unsigned long)arg5; \
+- _argvec[2+6] = (unsigned long)arg6; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- VG_EXPAND_FRAME_BY_trashes_r3(512) \
+- "stw 2,-8(11)\n\t" /* save tocptr */ \
+- "lwz 2,-4(11)\n\t" /* use nraddr's tocptr */ \
+- "lwz 3, 4(11)\n\t" /* arg1->r3 */ \
+- "lwz 4, 8(11)\n\t" /* arg2->r4 */ \
+- "lwz 5, 12(11)\n\t" /* arg3->r5 */ \
+- "lwz 6, 16(11)\n\t" /* arg4->r6 */ \
+- "lwz 7, 20(11)\n\t" /* arg5->r7 */ \
+- "lwz 8, 24(11)\n\t" /* arg6->r8 */ \
+- "lwz 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "lwz 2,-8(11)\n\t" /* restore tocptr */ \
+- VG_CONTRACT_FRAME_BY(512) \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_7W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+7]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- _argvec[2+4] = (unsigned long)arg4; \
+- _argvec[2+5] = (unsigned long)arg5; \
+- _argvec[2+6] = (unsigned long)arg6; \
+- _argvec[2+7] = (unsigned long)arg7; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- VG_EXPAND_FRAME_BY_trashes_r3(512) \
+- "stw 2,-8(11)\n\t" /* save tocptr */ \
+- "lwz 2,-4(11)\n\t" /* use nraddr's tocptr */ \
+- "lwz 3, 4(11)\n\t" /* arg1->r3 */ \
+- "lwz 4, 8(11)\n\t" /* arg2->r4 */ \
+- "lwz 5, 12(11)\n\t" /* arg3->r5 */ \
+- "lwz 6, 16(11)\n\t" /* arg4->r6 */ \
+- "lwz 7, 20(11)\n\t" /* arg5->r7 */ \
+- "lwz 8, 24(11)\n\t" /* arg6->r8 */ \
+- "lwz 9, 28(11)\n\t" /* arg7->r9 */ \
+- "lwz 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "lwz 2,-8(11)\n\t" /* restore tocptr */ \
+- VG_CONTRACT_FRAME_BY(512) \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_8W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7,arg8) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+8]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- _argvec[2+4] = (unsigned long)arg4; \
+- _argvec[2+5] = (unsigned long)arg5; \
+- _argvec[2+6] = (unsigned long)arg6; \
+- _argvec[2+7] = (unsigned long)arg7; \
+- _argvec[2+8] = (unsigned long)arg8; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- VG_EXPAND_FRAME_BY_trashes_r3(512) \
+- "stw 2,-8(11)\n\t" /* save tocptr */ \
+- "lwz 2,-4(11)\n\t" /* use nraddr's tocptr */ \
+- "lwz 3, 4(11)\n\t" /* arg1->r3 */ \
+- "lwz 4, 8(11)\n\t" /* arg2->r4 */ \
+- "lwz 5, 12(11)\n\t" /* arg3->r5 */ \
+- "lwz 6, 16(11)\n\t" /* arg4->r6 */ \
+- "lwz 7, 20(11)\n\t" /* arg5->r7 */ \
+- "lwz 8, 24(11)\n\t" /* arg6->r8 */ \
+- "lwz 9, 28(11)\n\t" /* arg7->r9 */ \
+- "lwz 10, 32(11)\n\t" /* arg8->r10 */ \
+- "lwz 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "lwz 2,-8(11)\n\t" /* restore tocptr */ \
+- VG_CONTRACT_FRAME_BY(512) \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_9W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7,arg8,arg9) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+9]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- _argvec[2+4] = (unsigned long)arg4; \
+- _argvec[2+5] = (unsigned long)arg5; \
+- _argvec[2+6] = (unsigned long)arg6; \
+- _argvec[2+7] = (unsigned long)arg7; \
+- _argvec[2+8] = (unsigned long)arg8; \
+- _argvec[2+9] = (unsigned long)arg9; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- VG_EXPAND_FRAME_BY_trashes_r3(512) \
+- "stw 2,-8(11)\n\t" /* save tocptr */ \
+- "lwz 2,-4(11)\n\t" /* use nraddr's tocptr */ \
+- VG_EXPAND_FRAME_BY_trashes_r3(64) \
+- /* arg9 */ \
+- "lwz 3,36(11)\n\t" \
+- "stw 3,56(1)\n\t" \
+- /* args1-8 */ \
+- "lwz 3, 4(11)\n\t" /* arg1->r3 */ \
+- "lwz 4, 8(11)\n\t" /* arg2->r4 */ \
+- "lwz 5, 12(11)\n\t" /* arg3->r5 */ \
+- "lwz 6, 16(11)\n\t" /* arg4->r6 */ \
+- "lwz 7, 20(11)\n\t" /* arg5->r7 */ \
+- "lwz 8, 24(11)\n\t" /* arg6->r8 */ \
+- "lwz 9, 28(11)\n\t" /* arg7->r9 */ \
+- "lwz 10, 32(11)\n\t" /* arg8->r10 */ \
+- "lwz 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "lwz 2,-8(11)\n\t" /* restore tocptr */ \
+- VG_CONTRACT_FRAME_BY(64) \
+- VG_CONTRACT_FRAME_BY(512) \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_10W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7,arg8,arg9,arg10) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+10]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- _argvec[2+4] = (unsigned long)arg4; \
+- _argvec[2+5] = (unsigned long)arg5; \
+- _argvec[2+6] = (unsigned long)arg6; \
+- _argvec[2+7] = (unsigned long)arg7; \
+- _argvec[2+8] = (unsigned long)arg8; \
+- _argvec[2+9] = (unsigned long)arg9; \
+- _argvec[2+10] = (unsigned long)arg10; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- VG_EXPAND_FRAME_BY_trashes_r3(512) \
+- "stw 2,-8(11)\n\t" /* save tocptr */ \
+- "lwz 2,-4(11)\n\t" /* use nraddr's tocptr */ \
+- VG_EXPAND_FRAME_BY_trashes_r3(64) \
+- /* arg10 */ \
+- "lwz 3,40(11)\n\t" \
+- "stw 3,60(1)\n\t" \
+- /* arg9 */ \
+- "lwz 3,36(11)\n\t" \
+- "stw 3,56(1)\n\t" \
+- /* args1-8 */ \
+- "lwz 3, 4(11)\n\t" /* arg1->r3 */ \
+- "lwz 4, 8(11)\n\t" /* arg2->r4 */ \
+- "lwz 5, 12(11)\n\t" /* arg3->r5 */ \
+- "lwz 6, 16(11)\n\t" /* arg4->r6 */ \
+- "lwz 7, 20(11)\n\t" /* arg5->r7 */ \
+- "lwz 8, 24(11)\n\t" /* arg6->r8 */ \
+- "lwz 9, 28(11)\n\t" /* arg7->r9 */ \
+- "lwz 10, 32(11)\n\t" /* arg8->r10 */ \
+- "lwz 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "lwz 2,-8(11)\n\t" /* restore tocptr */ \
+- VG_CONTRACT_FRAME_BY(64) \
+- VG_CONTRACT_FRAME_BY(512) \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_11W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7,arg8,arg9,arg10,arg11) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+11]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- _argvec[2+4] = (unsigned long)arg4; \
+- _argvec[2+5] = (unsigned long)arg5; \
+- _argvec[2+6] = (unsigned long)arg6; \
+- _argvec[2+7] = (unsigned long)arg7; \
+- _argvec[2+8] = (unsigned long)arg8; \
+- _argvec[2+9] = (unsigned long)arg9; \
+- _argvec[2+10] = (unsigned long)arg10; \
+- _argvec[2+11] = (unsigned long)arg11; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- VG_EXPAND_FRAME_BY_trashes_r3(512) \
+- "stw 2,-8(11)\n\t" /* save tocptr */ \
+- "lwz 2,-4(11)\n\t" /* use nraddr's tocptr */ \
+- VG_EXPAND_FRAME_BY_trashes_r3(72) \
+- /* arg11 */ \
+- "lwz 3,44(11)\n\t" \
+- "stw 3,64(1)\n\t" \
+- /* arg10 */ \
+- "lwz 3,40(11)\n\t" \
+- "stw 3,60(1)\n\t" \
+- /* arg9 */ \
+- "lwz 3,36(11)\n\t" \
+- "stw 3,56(1)\n\t" \
+- /* args1-8 */ \
+- "lwz 3, 4(11)\n\t" /* arg1->r3 */ \
+- "lwz 4, 8(11)\n\t" /* arg2->r4 */ \
+- "lwz 5, 12(11)\n\t" /* arg3->r5 */ \
+- "lwz 6, 16(11)\n\t" /* arg4->r6 */ \
+- "lwz 7, 20(11)\n\t" /* arg5->r7 */ \
+- "lwz 8, 24(11)\n\t" /* arg6->r8 */ \
+- "lwz 9, 28(11)\n\t" /* arg7->r9 */ \
+- "lwz 10, 32(11)\n\t" /* arg8->r10 */ \
+- "lwz 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "lwz 2,-8(11)\n\t" /* restore tocptr */ \
+- VG_CONTRACT_FRAME_BY(72) \
+- VG_CONTRACT_FRAME_BY(512) \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_12W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7,arg8,arg9,arg10,arg11,arg12) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+12]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- _argvec[2+4] = (unsigned long)arg4; \
+- _argvec[2+5] = (unsigned long)arg5; \
+- _argvec[2+6] = (unsigned long)arg6; \
+- _argvec[2+7] = (unsigned long)arg7; \
+- _argvec[2+8] = (unsigned long)arg8; \
+- _argvec[2+9] = (unsigned long)arg9; \
+- _argvec[2+10] = (unsigned long)arg10; \
+- _argvec[2+11] = (unsigned long)arg11; \
+- _argvec[2+12] = (unsigned long)arg12; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- VG_EXPAND_FRAME_BY_trashes_r3(512) \
+- "stw 2,-8(11)\n\t" /* save tocptr */ \
+- "lwz 2,-4(11)\n\t" /* use nraddr's tocptr */ \
+- VG_EXPAND_FRAME_BY_trashes_r3(72) \
+- /* arg12 */ \
+- "lwz 3,48(11)\n\t" \
+- "stw 3,68(1)\n\t" \
+- /* arg11 */ \
+- "lwz 3,44(11)\n\t" \
+- "stw 3,64(1)\n\t" \
+- /* arg10 */ \
+- "lwz 3,40(11)\n\t" \
+- "stw 3,60(1)\n\t" \
+- /* arg9 */ \
+- "lwz 3,36(11)\n\t" \
+- "stw 3,56(1)\n\t" \
+- /* args1-8 */ \
+- "lwz 3, 4(11)\n\t" /* arg1->r3 */ \
+- "lwz 4, 8(11)\n\t" /* arg2->r4 */ \
+- "lwz 5, 12(11)\n\t" /* arg3->r5 */ \
+- "lwz 6, 16(11)\n\t" /* arg4->r6 */ \
+- "lwz 7, 20(11)\n\t" /* arg5->r7 */ \
+- "lwz 8, 24(11)\n\t" /* arg6->r8 */ \
+- "lwz 9, 28(11)\n\t" /* arg7->r9 */ \
+- "lwz 10, 32(11)\n\t" /* arg8->r10 */ \
+- "lwz 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "lwz 2,-8(11)\n\t" /* restore tocptr */ \
+- VG_CONTRACT_FRAME_BY(72) \
+- VG_CONTRACT_FRAME_BY(512) \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#endif /* PLAT_ppc32_aix5 */
+-
+-/* ------------------------ ppc64-aix5 ------------------------- */
+-
+-#if defined(PLAT_ppc64_aix5)
+-
+-/* ARGREGS: r3 r4 r5 r6 r7 r8 r9 r10 (the rest on stack somewhere) */
+-
+-/* These regs are trashed by the hidden call. */
+-#define __CALLER_SAVED_REGS \
+- "lr", "ctr", "xer", \
+- "cr0", "cr1", "cr2", "cr3", "cr4", "cr5", "cr6", "cr7", \
+- "r0", "r2", "r3", "r4", "r5", "r6", "r7", "r8", "r9", "r10", \
+- "r11", "r12", "r13"
+-
+-/* Expand the stack frame, copying enough info that unwinding
+- still works. Trashes r3. */
+-
+-#define VG_EXPAND_FRAME_BY_trashes_r3(_n_fr) \
+- "addi 1,1,-" #_n_fr "\n\t" \
+- "ld 3," #_n_fr "(1)\n\t" \
+- "std 3,0(1)\n\t"
+-
+-#define VG_CONTRACT_FRAME_BY(_n_fr) \
+- "addi 1,1," #_n_fr "\n\t"
+-
+-/* These CALL_FN_ macros assume that on ppc64-aix5, sizeof(unsigned
+- long) == 8. */
+-
+-#define CALL_FN_W_v(lval, orig) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+0]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- VG_EXPAND_FRAME_BY_trashes_r3(512) \
+- "std 2,-16(11)\n\t" /* save tocptr */ \
+- "ld 2,-8(11)\n\t" /* use nraddr's tocptr */ \
+- "ld 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "ld 2,-16(11)\n\t" /* restore tocptr */ \
+- VG_CONTRACT_FRAME_BY(512) \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_W(lval, orig, arg1) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+1]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- VG_EXPAND_FRAME_BY_trashes_r3(512) \
+- "std 2,-16(11)\n\t" /* save tocptr */ \
+- "ld 2,-8(11)\n\t" /* use nraddr's tocptr */ \
+- "ld 3, 8(11)\n\t" /* arg1->r3 */ \
+- "ld 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "ld 2,-16(11)\n\t" /* restore tocptr */ \
+- VG_CONTRACT_FRAME_BY(512) \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_WW(lval, orig, arg1,arg2) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+2]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- VG_EXPAND_FRAME_BY_trashes_r3(512) \
+- "std 2,-16(11)\n\t" /* save tocptr */ \
+- "ld 2,-8(11)\n\t" /* use nraddr's tocptr */ \
+- "ld 3, 8(11)\n\t" /* arg1->r3 */ \
+- "ld 4, 16(11)\n\t" /* arg2->r4 */ \
+- "ld 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "ld 2,-16(11)\n\t" /* restore tocptr */ \
+- VG_CONTRACT_FRAME_BY(512) \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_WWW(lval, orig, arg1,arg2,arg3) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+3]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- VG_EXPAND_FRAME_BY_trashes_r3(512) \
+- "std 2,-16(11)\n\t" /* save tocptr */ \
+- "ld 2,-8(11)\n\t" /* use nraddr's tocptr */ \
+- "ld 3, 8(11)\n\t" /* arg1->r3 */ \
+- "ld 4, 16(11)\n\t" /* arg2->r4 */ \
+- "ld 5, 24(11)\n\t" /* arg3->r5 */ \
+- "ld 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "ld 2,-16(11)\n\t" /* restore tocptr */ \
+- VG_CONTRACT_FRAME_BY(512) \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_WWWW(lval, orig, arg1,arg2,arg3,arg4) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+4]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- _argvec[2+4] = (unsigned long)arg4; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- VG_EXPAND_FRAME_BY_trashes_r3(512) \
+- "std 2,-16(11)\n\t" /* save tocptr */ \
+- "ld 2,-8(11)\n\t" /* use nraddr's tocptr */ \
+- "ld 3, 8(11)\n\t" /* arg1->r3 */ \
+- "ld 4, 16(11)\n\t" /* arg2->r4 */ \
+- "ld 5, 24(11)\n\t" /* arg3->r5 */ \
+- "ld 6, 32(11)\n\t" /* arg4->r6 */ \
+- "ld 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "ld 2,-16(11)\n\t" /* restore tocptr */ \
+- VG_CONTRACT_FRAME_BY(512) \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_5W(lval, orig, arg1,arg2,arg3,arg4,arg5) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+5]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- _argvec[2+4] = (unsigned long)arg4; \
+- _argvec[2+5] = (unsigned long)arg5; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- VG_EXPAND_FRAME_BY_trashes_r3(512) \
+- "std 2,-16(11)\n\t" /* save tocptr */ \
+- "ld 2,-8(11)\n\t" /* use nraddr's tocptr */ \
+- "ld 3, 8(11)\n\t" /* arg1->r3 */ \
+- "ld 4, 16(11)\n\t" /* arg2->r4 */ \
+- "ld 5, 24(11)\n\t" /* arg3->r5 */ \
+- "ld 6, 32(11)\n\t" /* arg4->r6 */ \
+- "ld 7, 40(11)\n\t" /* arg5->r7 */ \
+- "ld 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "ld 2,-16(11)\n\t" /* restore tocptr */ \
+- VG_CONTRACT_FRAME_BY(512) \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_6W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+6]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- _argvec[2+4] = (unsigned long)arg4; \
+- _argvec[2+5] = (unsigned long)arg5; \
+- _argvec[2+6] = (unsigned long)arg6; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- VG_EXPAND_FRAME_BY_trashes_r3(512) \
+- "std 2,-16(11)\n\t" /* save tocptr */ \
+- "ld 2,-8(11)\n\t" /* use nraddr's tocptr */ \
+- "ld 3, 8(11)\n\t" /* arg1->r3 */ \
+- "ld 4, 16(11)\n\t" /* arg2->r4 */ \
+- "ld 5, 24(11)\n\t" /* arg3->r5 */ \
+- "ld 6, 32(11)\n\t" /* arg4->r6 */ \
+- "ld 7, 40(11)\n\t" /* arg5->r7 */ \
+- "ld 8, 48(11)\n\t" /* arg6->r8 */ \
+- "ld 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "ld 2,-16(11)\n\t" /* restore tocptr */ \
+- VG_CONTRACT_FRAME_BY(512) \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_7W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+7]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- _argvec[2+4] = (unsigned long)arg4; \
+- _argvec[2+5] = (unsigned long)arg5; \
+- _argvec[2+6] = (unsigned long)arg6; \
+- _argvec[2+7] = (unsigned long)arg7; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- VG_EXPAND_FRAME_BY_trashes_r3(512) \
+- "std 2,-16(11)\n\t" /* save tocptr */ \
+- "ld 2,-8(11)\n\t" /* use nraddr's tocptr */ \
+- "ld 3, 8(11)\n\t" /* arg1->r3 */ \
+- "ld 4, 16(11)\n\t" /* arg2->r4 */ \
+- "ld 5, 24(11)\n\t" /* arg3->r5 */ \
+- "ld 6, 32(11)\n\t" /* arg4->r6 */ \
+- "ld 7, 40(11)\n\t" /* arg5->r7 */ \
+- "ld 8, 48(11)\n\t" /* arg6->r8 */ \
+- "ld 9, 56(11)\n\t" /* arg7->r9 */ \
+- "ld 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "ld 2,-16(11)\n\t" /* restore tocptr */ \
+- VG_CONTRACT_FRAME_BY(512) \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_8W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7,arg8) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+8]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- _argvec[2+4] = (unsigned long)arg4; \
+- _argvec[2+5] = (unsigned long)arg5; \
+- _argvec[2+6] = (unsigned long)arg6; \
+- _argvec[2+7] = (unsigned long)arg7; \
+- _argvec[2+8] = (unsigned long)arg8; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- VG_EXPAND_FRAME_BY_trashes_r3(512) \
+- "std 2,-16(11)\n\t" /* save tocptr */ \
+- "ld 2,-8(11)\n\t" /* use nraddr's tocptr */ \
+- "ld 3, 8(11)\n\t" /* arg1->r3 */ \
+- "ld 4, 16(11)\n\t" /* arg2->r4 */ \
+- "ld 5, 24(11)\n\t" /* arg3->r5 */ \
+- "ld 6, 32(11)\n\t" /* arg4->r6 */ \
+- "ld 7, 40(11)\n\t" /* arg5->r7 */ \
+- "ld 8, 48(11)\n\t" /* arg6->r8 */ \
+- "ld 9, 56(11)\n\t" /* arg7->r9 */ \
+- "ld 10, 64(11)\n\t" /* arg8->r10 */ \
+- "ld 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "ld 2,-16(11)\n\t" /* restore tocptr */ \
+- VG_CONTRACT_FRAME_BY(512) \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_9W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7,arg8,arg9) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+9]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- _argvec[2+4] = (unsigned long)arg4; \
+- _argvec[2+5] = (unsigned long)arg5; \
+- _argvec[2+6] = (unsigned long)arg6; \
+- _argvec[2+7] = (unsigned long)arg7; \
+- _argvec[2+8] = (unsigned long)arg8; \
+- _argvec[2+9] = (unsigned long)arg9; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- VG_EXPAND_FRAME_BY_trashes_r3(512) \
+- "std 2,-16(11)\n\t" /* save tocptr */ \
+- "ld 2,-8(11)\n\t" /* use nraddr's tocptr */ \
+- VG_EXPAND_FRAME_BY_trashes_r3(128) \
+- /* arg9 */ \
+- "ld 3,72(11)\n\t" \
+- "std 3,112(1)\n\t" \
+- /* args1-8 */ \
+- "ld 3, 8(11)\n\t" /* arg1->r3 */ \
+- "ld 4, 16(11)\n\t" /* arg2->r4 */ \
+- "ld 5, 24(11)\n\t" /* arg3->r5 */ \
+- "ld 6, 32(11)\n\t" /* arg4->r6 */ \
+- "ld 7, 40(11)\n\t" /* arg5->r7 */ \
+- "ld 8, 48(11)\n\t" /* arg6->r8 */ \
+- "ld 9, 56(11)\n\t" /* arg7->r9 */ \
+- "ld 10, 64(11)\n\t" /* arg8->r10 */ \
+- "ld 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "ld 2,-16(11)\n\t" /* restore tocptr */ \
+- VG_CONTRACT_FRAME_BY(128) \
+- VG_CONTRACT_FRAME_BY(512) \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_10W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7,arg8,arg9,arg10) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+10]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- _argvec[2+4] = (unsigned long)arg4; \
+- _argvec[2+5] = (unsigned long)arg5; \
+- _argvec[2+6] = (unsigned long)arg6; \
+- _argvec[2+7] = (unsigned long)arg7; \
+- _argvec[2+8] = (unsigned long)arg8; \
+- _argvec[2+9] = (unsigned long)arg9; \
+- _argvec[2+10] = (unsigned long)arg10; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- VG_EXPAND_FRAME_BY_trashes_r3(512) \
+- "std 2,-16(11)\n\t" /* save tocptr */ \
+- "ld 2,-8(11)\n\t" /* use nraddr's tocptr */ \
+- VG_EXPAND_FRAME_BY_trashes_r3(128) \
+- /* arg10 */ \
+- "ld 3,80(11)\n\t" \
+- "std 3,120(1)\n\t" \
+- /* arg9 */ \
+- "ld 3,72(11)\n\t" \
+- "std 3,112(1)\n\t" \
+- /* args1-8 */ \
+- "ld 3, 8(11)\n\t" /* arg1->r3 */ \
+- "ld 4, 16(11)\n\t" /* arg2->r4 */ \
+- "ld 5, 24(11)\n\t" /* arg3->r5 */ \
+- "ld 6, 32(11)\n\t" /* arg4->r6 */ \
+- "ld 7, 40(11)\n\t" /* arg5->r7 */ \
+- "ld 8, 48(11)\n\t" /* arg6->r8 */ \
+- "ld 9, 56(11)\n\t" /* arg7->r9 */ \
+- "ld 10, 64(11)\n\t" /* arg8->r10 */ \
+- "ld 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "ld 2,-16(11)\n\t" /* restore tocptr */ \
+- VG_CONTRACT_FRAME_BY(128) \
+- VG_CONTRACT_FRAME_BY(512) \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_11W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7,arg8,arg9,arg10,arg11) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+11]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- _argvec[2+4] = (unsigned long)arg4; \
+- _argvec[2+5] = (unsigned long)arg5; \
+- _argvec[2+6] = (unsigned long)arg6; \
+- _argvec[2+7] = (unsigned long)arg7; \
+- _argvec[2+8] = (unsigned long)arg8; \
+- _argvec[2+9] = (unsigned long)arg9; \
+- _argvec[2+10] = (unsigned long)arg10; \
+- _argvec[2+11] = (unsigned long)arg11; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- VG_EXPAND_FRAME_BY_trashes_r3(512) \
+- "std 2,-16(11)\n\t" /* save tocptr */ \
+- "ld 2,-8(11)\n\t" /* use nraddr's tocptr */ \
+- VG_EXPAND_FRAME_BY_trashes_r3(144) \
+- /* arg11 */ \
+- "ld 3,88(11)\n\t" \
+- "std 3,128(1)\n\t" \
+- /* arg10 */ \
+- "ld 3,80(11)\n\t" \
+- "std 3,120(1)\n\t" \
+- /* arg9 */ \
+- "ld 3,72(11)\n\t" \
+- "std 3,112(1)\n\t" \
+- /* args1-8 */ \
+- "ld 3, 8(11)\n\t" /* arg1->r3 */ \
+- "ld 4, 16(11)\n\t" /* arg2->r4 */ \
+- "ld 5, 24(11)\n\t" /* arg3->r5 */ \
+- "ld 6, 32(11)\n\t" /* arg4->r6 */ \
+- "ld 7, 40(11)\n\t" /* arg5->r7 */ \
+- "ld 8, 48(11)\n\t" /* arg6->r8 */ \
+- "ld 9, 56(11)\n\t" /* arg7->r9 */ \
+- "ld 10, 64(11)\n\t" /* arg8->r10 */ \
+- "ld 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "ld 2,-16(11)\n\t" /* restore tocptr */ \
+- VG_CONTRACT_FRAME_BY(144) \
+- VG_CONTRACT_FRAME_BY(512) \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#define CALL_FN_W_12W(lval, orig, arg1,arg2,arg3,arg4,arg5,arg6, \
+- arg7,arg8,arg9,arg10,arg11,arg12) \
+- do { \
+- volatile OrigFn _orig = (orig); \
+- volatile unsigned long _argvec[3+12]; \
+- volatile unsigned long _res; \
+- /* _argvec[0] holds current r2 across the call */ \
+- _argvec[1] = (unsigned long)_orig.r2; \
+- _argvec[2] = (unsigned long)_orig.nraddr; \
+- _argvec[2+1] = (unsigned long)arg1; \
+- _argvec[2+2] = (unsigned long)arg2; \
+- _argvec[2+3] = (unsigned long)arg3; \
+- _argvec[2+4] = (unsigned long)arg4; \
+- _argvec[2+5] = (unsigned long)arg5; \
+- _argvec[2+6] = (unsigned long)arg6; \
+- _argvec[2+7] = (unsigned long)arg7; \
+- _argvec[2+8] = (unsigned long)arg8; \
+- _argvec[2+9] = (unsigned long)arg9; \
+- _argvec[2+10] = (unsigned long)arg10; \
+- _argvec[2+11] = (unsigned long)arg11; \
+- _argvec[2+12] = (unsigned long)arg12; \
+- __asm__ volatile( \
+- "mr 11,%1\n\t" \
+- VG_EXPAND_FRAME_BY_trashes_r3(512) \
+- "std 2,-16(11)\n\t" /* save tocptr */ \
+- "ld 2,-8(11)\n\t" /* use nraddr's tocptr */ \
+- VG_EXPAND_FRAME_BY_trashes_r3(144) \
+- /* arg12 */ \
+- "ld 3,96(11)\n\t" \
+- "std 3,136(1)\n\t" \
+- /* arg11 */ \
+- "ld 3,88(11)\n\t" \
+- "std 3,128(1)\n\t" \
+- /* arg10 */ \
+- "ld 3,80(11)\n\t" \
+- "std 3,120(1)\n\t" \
+- /* arg9 */ \
+- "ld 3,72(11)\n\t" \
+- "std 3,112(1)\n\t" \
+- /* args1-8 */ \
+- "ld 3, 8(11)\n\t" /* arg1->r3 */ \
+- "ld 4, 16(11)\n\t" /* arg2->r4 */ \
+- "ld 5, 24(11)\n\t" /* arg3->r5 */ \
+- "ld 6, 32(11)\n\t" /* arg4->r6 */ \
+- "ld 7, 40(11)\n\t" /* arg5->r7 */ \
+- "ld 8, 48(11)\n\t" /* arg6->r8 */ \
+- "ld 9, 56(11)\n\t" /* arg7->r9 */ \
+- "ld 10, 64(11)\n\t" /* arg8->r10 */ \
+- "ld 11, 0(11)\n\t" /* target->r11 */ \
+- VALGRIND_BRANCH_AND_LINK_TO_NOREDIR_R11 \
+- "mr 11,%1\n\t" \
+- "mr %0,3\n\t" \
+- "ld 2,-16(11)\n\t" /* restore tocptr */ \
+- VG_CONTRACT_FRAME_BY(144) \
+- VG_CONTRACT_FRAME_BY(512) \
+- : /*out*/ "=r" (_res) \
+- : /*in*/ "r" (&_argvec[2]) \
+- : /*trash*/ "cc", "memory", __CALLER_SAVED_REGS \
+- ); \
+- lval = (__typeof__(lval)) _res; \
+- } while (0)
+-
+-#endif /* PLAT_ppc64_aix5 */
+-
+-
+-/* ------------------------------------------------------------------ */
+-/* ARCHITECTURE INDEPENDENT MACROS for CLIENT REQUESTS. */
+-/* */
+-/* ------------------------------------------------------------------ */
+-
+-/* Some request codes. There are many more of these, but most are not
+- exposed to end-user view. These are the public ones, all of the
+- form 0x1000 + small_number.
+-
+- Core ones are in the range 0x00000000--0x0000ffff. The non-public
+- ones start at 0x2000.
+-*/
+-
+-/* These macros are used by tools -- they must be public, but don't
+- embed them into other programs. */
+-#define VG_USERREQ_TOOL_BASE(a,b) \
+- ((unsigned int)(((a)&0xff) << 24 | ((b)&0xff) << 16))
+-#define VG_IS_TOOL_USERREQ(a, b, v) \
+- (VG_USERREQ_TOOL_BASE(a,b) == ((v) & 0xffff0000))
+-
+-/* !! ABIWARNING !! ABIWARNING !! ABIWARNING !! ABIWARNING !!
+- This enum comprises an ABI exported by Valgrind to programs
+- which use client requests. DO NOT CHANGE THE ORDER OF THESE
+- ENTRIES, NOR DELETE ANY -- add new ones at the end. */
+-typedef
+- enum { VG_USERREQ__RUNNING_ON_VALGRIND = 0x1001,
+- VG_USERREQ__DISCARD_TRANSLATIONS = 0x1002,
+-
+- /* These allow any function to be called from the simulated
+- CPU but run on the real CPU. Nb: the first arg passed to
+- the function is always the ThreadId of the running
+- thread! So CLIENT_CALL0 actually requires a 1 arg
+- function, etc. */
+- VG_USERREQ__CLIENT_CALL0 = 0x1101,
+- VG_USERREQ__CLIENT_CALL1 = 0x1102,
+- VG_USERREQ__CLIENT_CALL2 = 0x1103,
+- VG_USERREQ__CLIENT_CALL3 = 0x1104,
+-
+- /* Can be useful in regression testing suites -- eg. can
+- send Valgrind's output to /dev/null and still count
+- errors. */
+- VG_USERREQ__COUNT_ERRORS = 0x1201,
+-
+- /* These are useful and can be interpreted by any tool that
+- tracks malloc() et al, by using vg_replace_malloc.c. */
+- VG_USERREQ__MALLOCLIKE_BLOCK = 0x1301,
+- VG_USERREQ__FREELIKE_BLOCK = 0x1302,
+- /* Memory pool support. */
+- VG_USERREQ__CREATE_MEMPOOL = 0x1303,
+- VG_USERREQ__DESTROY_MEMPOOL = 0x1304,
+- VG_USERREQ__MEMPOOL_ALLOC = 0x1305,
+- VG_USERREQ__MEMPOOL_FREE = 0x1306,
+- VG_USERREQ__MEMPOOL_TRIM = 0x1307,
+- VG_USERREQ__MOVE_MEMPOOL = 0x1308,
+- VG_USERREQ__MEMPOOL_CHANGE = 0x1309,
+- VG_USERREQ__MEMPOOL_EXISTS = 0x130a,
+-
+- /* Allow printfs to valgrind log. */
+- VG_USERREQ__PRINTF = 0x1401,
+- VG_USERREQ__PRINTF_BACKTRACE = 0x1402,
+-
+- /* Stack support. */
+- VG_USERREQ__STACK_REGISTER = 0x1501,
+- VG_USERREQ__STACK_DEREGISTER = 0x1502,
+- VG_USERREQ__STACK_CHANGE = 0x1503,
+-
+- /* Wine support */
+- VG_USERREQ__LOAD_PDB_DEBUGINFO = 0x1601
+- } Vg_ClientRequest;
+-
+-#if !defined(__GNUC__)
+-# define __extension__ /* */
+-#endif
+-
+-/* Returns the number of Valgrinds this code is running under. That
+- is, 0 if running natively, 1 if running under Valgrind, 2 if
+- running under Valgrind which is running under another Valgrind,
+- etc. */
+-#define RUNNING_ON_VALGRIND __extension__ \
+- ({unsigned int _qzz_res; \
+- VALGRIND_DO_CLIENT_REQUEST(_qzz_res, 0 /* if not */, \
+- VG_USERREQ__RUNNING_ON_VALGRIND, \
+- 0, 0, 0, 0, 0); \
+- _qzz_res; \
+- })
+-
+-
+-/* Discard translation of code in the range [_qzz_addr .. _qzz_addr +
+- _qzz_len - 1]. Useful if you are debugging a JITter or some such,
+- since it provides a way to make sure valgrind will retranslate the
+- invalidated area. Returns no value. */
+-#define VALGRIND_DISCARD_TRANSLATIONS(_qzz_addr,_qzz_len) \
+- {unsigned int _qzz_res; \
+- VALGRIND_DO_CLIENT_REQUEST(_qzz_res, 0, \
+- VG_USERREQ__DISCARD_TRANSLATIONS, \
+- _qzz_addr, _qzz_len, 0, 0, 0); \
+- }
+-
+-
+-/* These requests are for getting Valgrind itself to print something.
+- Possibly with a backtrace. This is a really ugly hack. The return value
+- is the number of characters printed, excluding the "**<pid>** " part at the
+- start and the backtrace (if present). */
+-
+-#if defined(NVALGRIND)
+-
+-# define VALGRIND_PRINTF(...)
+-# define VALGRIND_PRINTF_BACKTRACE(...)
+-
+-#else /* NVALGRIND */
+-
+-/* Modern GCC will optimize the static routine out if unused,
+- and unused attribute will shut down warnings about it. */
+-static int VALGRIND_PRINTF(const char *format, ...)
+- __attribute__((format(__printf__, 1, 2), __unused__));
+-static int
+-VALGRIND_PRINTF(const char *format, ...)
+-{
+- unsigned long _qzz_res;
+- va_list vargs;
+- va_start(vargs, format);
+- VALGRIND_DO_CLIENT_REQUEST(_qzz_res, 0, VG_USERREQ__PRINTF,
+- (unsigned long)format, (unsigned long)vargs,
+- 0, 0, 0);
+- va_end(vargs);
+- return (int)_qzz_res;
+-}
+-
+-static int VALGRIND_PRINTF_BACKTRACE(const char *format, ...)
+- __attribute__((format(__printf__, 1, 2), __unused__));
+-static int
+-VALGRIND_PRINTF_BACKTRACE(const char *format, ...)
+-{
+- unsigned long _qzz_res;
+- va_list vargs;
+- va_start(vargs, format);
+- VALGRIND_DO_CLIENT_REQUEST(_qzz_res, 0, VG_USERREQ__PRINTF_BACKTRACE,
+- (unsigned long)format, (unsigned long)vargs,
+- 0, 0, 0);
+- va_end(vargs);
+- return (int)_qzz_res;
+-}
+-
+-#endif /* NVALGRIND */
+-
+-
+-/* These requests allow control to move from the simulated CPU to the
+- real CPU, calling an arbitary function.
+-
+- Note that the current ThreadId is inserted as the first argument.
+- So this call:
+-
+- VALGRIND_NON_SIMD_CALL2(f, arg1, arg2)
+-
+- requires f to have this signature:
+-
+- Word f(Word tid, Word arg1, Word arg2)
+-
+- where "Word" is a word-sized type.
+-
+- Note that these client requests are not entirely reliable. For example,
+- if you call a function with them that subsequently calls printf(),
+- there's a high chance Valgrind will crash. Generally, your prospects of
+- these working are made higher if the called function does not refer to
+- any global variables, and does not refer to any libc or other functions
+- (printf et al). Any kind of entanglement with libc or dynamic linking is
+- likely to have a bad outcome, for tricky reasons which we've grappled
+- with a lot in the past.
+-*/
+-#define VALGRIND_NON_SIMD_CALL0(_qyy_fn) \
+- __extension__ \
+- ({unsigned long _qyy_res; \
+- VALGRIND_DO_CLIENT_REQUEST(_qyy_res, 0 /* default return */, \
+- VG_USERREQ__CLIENT_CALL0, \
+- _qyy_fn, \
+- 0, 0, 0, 0); \
+- _qyy_res; \
+- })
+-
+-#define VALGRIND_NON_SIMD_CALL1(_qyy_fn, _qyy_arg1) \
+- __extension__ \
+- ({unsigned long _qyy_res; \
+- VALGRIND_DO_CLIENT_REQUEST(_qyy_res, 0 /* default return */, \
+- VG_USERREQ__CLIENT_CALL1, \
+- _qyy_fn, \
+- _qyy_arg1, 0, 0, 0); \
+- _qyy_res; \
+- })
+-
+-#define VALGRIND_NON_SIMD_CALL2(_qyy_fn, _qyy_arg1, _qyy_arg2) \
+- __extension__ \
+- ({unsigned long _qyy_res; \
+- VALGRIND_DO_CLIENT_REQUEST(_qyy_res, 0 /* default return */, \
+- VG_USERREQ__CLIENT_CALL2, \
+- _qyy_fn, \
+- _qyy_arg1, _qyy_arg2, 0, 0); \
+- _qyy_res; \
+- })
+-
+-#define VALGRIND_NON_SIMD_CALL3(_qyy_fn, _qyy_arg1, _qyy_arg2, _qyy_arg3) \
+- __extension__ \
+- ({unsigned long _qyy_res; \
+- VALGRIND_DO_CLIENT_REQUEST(_qyy_res, 0 /* default return */, \
+- VG_USERREQ__CLIENT_CALL3, \
+- _qyy_fn, \
+- _qyy_arg1, _qyy_arg2, \
+- _qyy_arg3, 0); \
+- _qyy_res; \
+- })
+-
+-
+-/* Counts the number of errors that have been recorded by a tool. Nb:
+- the tool must record the errors with VG_(maybe_record_error)() or
+- VG_(unique_error)() for them to be counted. */
+-#define VALGRIND_COUNT_ERRORS \
+- __extension__ \
+- ({unsigned int _qyy_res; \
+- VALGRIND_DO_CLIENT_REQUEST(_qyy_res, 0 /* default return */, \
+- VG_USERREQ__COUNT_ERRORS, \
+- 0, 0, 0, 0, 0); \
+- _qyy_res; \
+- })
+-
+-/* Several Valgrind tools (Memcheck, Massif, Helgrind, DRD) rely on knowing
+- when heap blocks are allocated in order to give accurate results. This
+- happens automatically for the standard allocator functions such as
+- malloc(), calloc(), realloc(), memalign(), new, new[], free(), delete,
+- delete[], etc.
+-
+- But if your program uses a custom allocator, this doesn't automatically
+- happen, and Valgrind will not do as well. For example, if you allocate
+- superblocks with mmap() and then allocates chunks of the superblocks, all
+- Valgrind's observations will be at the mmap() level and it won't know that
+- the chunks should be considered separate entities. In Memcheck's case,
+- that means you probably won't get heap block overrun detection (because
+- there won't be redzones marked as unaddressable) and you definitely won't
+- get any leak detection.
+-
+- The following client requests allow a custom allocator to be annotated so
+- that it can be handled accurately by Valgrind.
+-
+- VALGRIND_MALLOCLIKE_BLOCK marks a region of memory as having been allocated
+- by a malloc()-like function. For Memcheck (an illustrative case), this
+- does two things:
+-
+- - It records that the block has been allocated. This means any addresses
+- within the block mentioned in error messages will be
+- identified as belonging to the block. It also means that if the block
+- isn't freed it will be detected by the leak checker.
+-
+- - It marks the block as being addressable and undefined (if 'is_zeroed' is
+- not set), or addressable and defined (if 'is_zeroed' is set). This
+- controls how accesses to the block by the program are handled.
+-
+- 'addr' is the start of the usable block (ie. after any
+- redzone), 'sizeB' is its size. 'rzB' is the redzone size if the allocator
+- can apply redzones -- these are blocks of padding at the start and end of
+- each block. Adding redzones is recommended as it makes it much more likely
+- Valgrind will spot block overruns. `is_zeroed' indicates if the memory is
+- zeroed (or filled with another predictable value), as is the case for
+- calloc().
+-
+- VALGRIND_MALLOCLIKE_BLOCK should be put immediately after the point where a
+- heap block -- that will be used by the client program -- is allocated.
+- It's best to put it at the outermost level of the allocator if possible;
+- for example, if you have a function my_alloc() which calls
+- internal_alloc(), and the client request is put inside internal_alloc(),
+- stack traces relating to the heap block will contain entries for both
+- my_alloc() and internal_alloc(), which is probably not what you want.
+-
+- For Memcheck users: if you use VALGRIND_MALLOCLIKE_BLOCK to carve out
+- custom blocks from within a heap block, B, that has been allocated with
+- malloc/calloc/new/etc, then block B will be *ignored* during leak-checking
+- -- the custom blocks will take precedence.
+-
+- VALGRIND_FREELIKE_BLOCK is the partner to VALGRIND_MALLOCLIKE_BLOCK. For
+- Memcheck, it does two things:
+-
+- - It records that the block has been deallocated. This assumes that the
+- block was annotated as having been allocated via
+- VALGRIND_MALLOCLIKE_BLOCK. Otherwise, an error will be issued.
+-
+- - It marks the block as being unaddressable.
+-
+- VALGRIND_FREELIKE_BLOCK should be put immediately after the point where a
+- heap block is deallocated.
+-
+- In many cases, these two client requests will not be enough to get your
+- allocator working well with Memcheck. More specifically, if your allocator
+- writes to freed blocks in any way then a VALGRIND_MAKE_MEM_UNDEFINED call
+- will be necessary to mark the memory as addressable just before the zeroing
+- occurs, otherwise you'll get a lot of invalid write errors. For example,
+- you'll need to do this if your allocator recycles freed blocks, but it
+- zeroes them before handing them back out (via VALGRIND_MALLOCLIKE_BLOCK).
+- Alternatively, if your allocator reuses freed blocks for allocator-internal
+- data structures, VALGRIND_MAKE_MEM_UNDEFINED calls will also be necessary.
+-
+- Really, what's happening is a blurring of the lines between the client
+- program and the allocator... after VALGRIND_FREELIKE_BLOCK is called, the
+- memory should be considered unaddressable to the client program, but the
+- allocator knows more than the rest of the client program and so may be able
+- to safely access it. Extra client requests are necessary for Valgrind to
+- understand the distinction between the allocator and the rest of the
+- program.
+-
+- Note: there is currently no VALGRIND_REALLOCLIKE_BLOCK client request; it
+- has to be emulated with MALLOCLIKE/FREELIKE and memory copying.
+-
+- Ignored if addr == 0.
+-*/
+-#define VALGRIND_MALLOCLIKE_BLOCK(addr, sizeB, rzB, is_zeroed) \
+- {unsigned int _qzz_res; \
+- VALGRIND_DO_CLIENT_REQUEST(_qzz_res, 0, \
+- VG_USERREQ__MALLOCLIKE_BLOCK, \
+- addr, sizeB, rzB, is_zeroed, 0); \
+- }
+-
+-/* See the comment for VALGRIND_MALLOCLIKE_BLOCK for details.
+- Ignored if addr == 0.
+-*/
+-#define VALGRIND_FREELIKE_BLOCK(addr, rzB) \
+- {unsigned int _qzz_res; \
+- VALGRIND_DO_CLIENT_REQUEST(_qzz_res, 0, \
+- VG_USERREQ__FREELIKE_BLOCK, \
+- addr, rzB, 0, 0, 0); \
+- }
+-
+-/* Create a memory pool. */
+-#define VALGRIND_CREATE_MEMPOOL(pool, rzB, is_zeroed) \
+- {unsigned int _qzz_res; \
+- VALGRIND_DO_CLIENT_REQUEST(_qzz_res, 0, \
+- VG_USERREQ__CREATE_MEMPOOL, \
+- pool, rzB, is_zeroed, 0, 0); \
+- }
+-
+-/* Destroy a memory pool. */
+-#define VALGRIND_DESTROY_MEMPOOL(pool) \
+- {unsigned int _qzz_res; \
+- VALGRIND_DO_CLIENT_REQUEST(_qzz_res, 0, \
+- VG_USERREQ__DESTROY_MEMPOOL, \
+- pool, 0, 0, 0, 0); \
+- }
+-
+-/* Associate a piece of memory with a memory pool. */
+-#define VALGRIND_MEMPOOL_ALLOC(pool, addr, size) \
+- {unsigned int _qzz_res; \
+- VALGRIND_DO_CLIENT_REQUEST(_qzz_res, 0, \
+- VG_USERREQ__MEMPOOL_ALLOC, \
+- pool, addr, size, 0, 0); \
+- }
+-
+-/* Disassociate a piece of memory from a memory pool. */
+-#define VALGRIND_MEMPOOL_FREE(pool, addr) \
+- {unsigned int _qzz_res; \
+- VALGRIND_DO_CLIENT_REQUEST(_qzz_res, 0, \
+- VG_USERREQ__MEMPOOL_FREE, \
+- pool, addr, 0, 0, 0); \
+- }
+-
+-/* Disassociate any pieces outside a particular range. */
+-#define VALGRIND_MEMPOOL_TRIM(pool, addr, size) \
+- {unsigned int _qzz_res; \
+- VALGRIND_DO_CLIENT_REQUEST(_qzz_res, 0, \
+- VG_USERREQ__MEMPOOL_TRIM, \
+- pool, addr, size, 0, 0); \
+- }
+-
+-/* Resize and/or move a piece associated with a memory pool. */
+-#define VALGRIND_MOVE_MEMPOOL(poolA, poolB) \
+- {unsigned int _qzz_res; \
+- VALGRIND_DO_CLIENT_REQUEST(_qzz_res, 0, \
+- VG_USERREQ__MOVE_MEMPOOL, \
+- poolA, poolB, 0, 0, 0); \
+- }
+-
+-/* Resize and/or move a piece associated with a memory pool. */
+-#define VALGRIND_MEMPOOL_CHANGE(pool, addrA, addrB, size) \
+- {unsigned int _qzz_res; \
+- VALGRIND_DO_CLIENT_REQUEST(_qzz_res, 0, \
+- VG_USERREQ__MEMPOOL_CHANGE, \
+- pool, addrA, addrB, size, 0); \
+- }
+-
+-/* Return 1 if a mempool exists, else 0. */
+-#define VALGRIND_MEMPOOL_EXISTS(pool) \
+- __extension__ \
+- ({unsigned int _qzz_res; \
+- VALGRIND_DO_CLIENT_REQUEST(_qzz_res, 0, \
+- VG_USERREQ__MEMPOOL_EXISTS, \
+- pool, 0, 0, 0, 0); \
+- _qzz_res; \
+- })
+-
+-/* Mark a piece of memory as being a stack. Returns a stack id. */
+-#define VALGRIND_STACK_REGISTER(start, end) \
+- __extension__ \
+- ({unsigned int _qzz_res; \
+- VALGRIND_DO_CLIENT_REQUEST(_qzz_res, 0, \
+- VG_USERREQ__STACK_REGISTER, \
+- start, end, 0, 0, 0); \
+- _qzz_res; \
+- })
+-
+-/* Unmark the piece of memory associated with a stack id as being a
+- stack. */
+-#define VALGRIND_STACK_DEREGISTER(id) \
+- {unsigned int _qzz_res; \
+- VALGRIND_DO_CLIENT_REQUEST(_qzz_res, 0, \
+- VG_USERREQ__STACK_DEREGISTER, \
+- id, 0, 0, 0, 0); \
+- }
+-
+-/* Change the start and end address of the stack id. */
+-#define VALGRIND_STACK_CHANGE(id, start, end) \
+- {unsigned int _qzz_res; \
+- VALGRIND_DO_CLIENT_REQUEST(_qzz_res, 0, \
+- VG_USERREQ__STACK_CHANGE, \
+- id, start, end, 0, 0); \
+- }
+-
+-/* Load PDB debug info for Wine PE image_map. */
+-#define VALGRIND_LOAD_PDB_DEBUGINFO(fd, ptr, total_size, delta) \
+- {unsigned int _qzz_res; \
+- VALGRIND_DO_CLIENT_REQUEST(_qzz_res, 0, \
+- VG_USERREQ__LOAD_PDB_DEBUGINFO, \
+- fd, ptr, total_size, delta, 0); \
+- }
+-
+-
+-#undef PLAT_x86_linux
+-#undef PLAT_amd64_linux
+-#undef PLAT_ppc32_linux
+-#undef PLAT_ppc64_linux
+-#undef PLAT_ppc32_aix5
+-#undef PLAT_ppc64_aix5
+-
+-#endif /* __VALGRIND_H */
diff --git a/planner.spec b/planner.spec
index 709543d..cabce30 100644
--- a/planner.spec
+++ b/planner.spec
@@ -1,7 +1,7 @@
Summary: A graphical project management tool
Name: planner
Version: 0.14.6
-Release: 17%{?dist}
+Release: 18%{?dist}
License: GPLv2+
Group: Applications/Productivity
URL: http://live.gnome.org/Planner
@@ -12,7 +12,7 @@ BuildRequires: libglade2-devel >= 2.0.0, gnome-vfs2-devel >= 2.0.2
BuildRequires: libxml2-devel >= 2.5.4
BuildRequires: libxslt-devel >= 1.0.27, libgsf-devel, gtk-doc, pygtk2-devel
BuildRequires: scrollkeeper, glib2-devel, python-devel, intltool, autoconf
-BuildRequires: automake, libtool
+BuildRequires: automake, libtool, valgrind-devel
Obsoletes: planner-eds < 0.14.6-6
Patch1: planner-gnome596173-xdg.patch
@@ -29,6 +29,7 @@ Patch11: planner-gnome615641-multiselect.patch
Patch12: planner-gnome655803-bad-LC_TIME.patch
Patch13: planner-desktop-warnings.patch
Patch14: planner-aarch64.patch
+Patch15: planner-rhbz1141479-paste.tasks.patch
%description
Planner is a visual project management application which allows users to
@@ -64,6 +65,7 @@ files to allow you to develop with planner.
%patch12 -p1 -b .gnome655803
%patch13 -p1 -b .desktop-warnings
%patch14 -p1 -b .aarch64
+%patch15 -p1 -b .systemvalgrind
%build
rm -rf libegg
@@ -149,6 +151,9 @@ rm -rf $RPM_BUILD_ROOT
%{_libdir}/pkgconfig/*
%changelog
+* Tue Sep 16 2014 Caolán McNamara <caolanm at redhat.com> - 0.14.6-18
+- Resolves: rhbz#1141479 use system valgrind headers
+
* Mon Sep 08 2014 Rex Dieter <rdieter at fedoraproject.org> 0.14.6-17
- fix/update scriptlets
More information about the scm-commits
mailing list