mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-11-30 19:26:38 +07:00
0444e8f6d7
This fixes a problem where the code that attempts to repair/convert an object of incorrect type is only executed on the first time the predefined method is called. The mechanism that disables warnings on subsequent calls was interfering with the repair mechanism. ACPICA BZ 781. http://acpica.org/bugzilla/show_bug.cgi?id=781 Signed-off-by: Bob Moore <robert.moore@intel.com> Signed-off-by: Lin Ming <ming.m.lin@intel.com> Signed-off-by: Len Brown <len.brown@intel.com>
576 lines
22 KiB
C
576 lines
22 KiB
C
/******************************************************************************
|
|
*
|
|
* Name: acmacros.h - C macros for the entire subsystem.
|
|
*
|
|
*****************************************************************************/
|
|
|
|
/*
|
|
* Copyright (C) 2000 - 2008, Intel Corp.
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions, and the following disclaimer,
|
|
* without modification.
|
|
* 2. Redistributions in binary form must reproduce at minimum a disclaimer
|
|
* substantially similar to the "NO WARRANTY" disclaimer below
|
|
* ("Disclaimer") and any redistribution must be conditioned upon
|
|
* including a substantially similar Disclaimer requirement for further
|
|
* binary redistribution.
|
|
* 3. Neither the names of the above-listed copyright holders nor the names
|
|
* of any contributors may be used to endorse or promote products derived
|
|
* from this software without specific prior written permission.
|
|
*
|
|
* Alternatively, this software may be distributed under the terms of the
|
|
* GNU General Public License ("GPL") version 2 as published by the Free
|
|
* Software Foundation.
|
|
*
|
|
* NO WARRANTY
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTIBILITY AND FITNESS FOR
|
|
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
* HOLDERS OR CONTRIBUTORS BE LIABLE FOR SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,
|
|
* STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING
|
|
* IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
|
* POSSIBILITY OF SUCH DAMAGES.
|
|
*/
|
|
|
|
#ifndef __ACMACROS_H__
|
|
#define __ACMACROS_H__
|
|
|
|
/*
|
|
* Extract data using a pointer. Any more than a byte and we
|
|
* get into potential aligment issues -- see the STORE macros below.
|
|
* Use with care.
|
|
*/
|
|
#define ACPI_GET8(ptr) *ACPI_CAST_PTR (u8, ptr)
|
|
#define ACPI_GET16(ptr) *ACPI_CAST_PTR (u16, ptr)
|
|
#define ACPI_GET32(ptr) *ACPI_CAST_PTR (u32, ptr)
|
|
#define ACPI_GET64(ptr) *ACPI_CAST_PTR (u64, ptr)
|
|
#define ACPI_SET8(ptr) *ACPI_CAST_PTR (u8, ptr)
|
|
#define ACPI_SET16(ptr) *ACPI_CAST_PTR (u16, ptr)
|
|
#define ACPI_SET32(ptr) *ACPI_CAST_PTR (u32, ptr)
|
|
#define ACPI_SET64(ptr) *ACPI_CAST_PTR (u64, ptr)
|
|
|
|
/*
|
|
* printf() format helpers
|
|
*/
|
|
|
|
/* Split 64-bit integer into two 32-bit values. Use with %8.8_x%8.8_x */
|
|
|
|
#define ACPI_FORMAT_UINT64(i) ACPI_HIDWORD(i), ACPI_LODWORD(i)
|
|
|
|
#if ACPI_MACHINE_WIDTH == 64
|
|
#define ACPI_FORMAT_NATIVE_UINT(i) ACPI_FORMAT_UINT64(i)
|
|
#else
|
|
#define ACPI_FORMAT_NATIVE_UINT(i) 0, (i)
|
|
#endif
|
|
|
|
/*
|
|
* Macros for moving data around to/from buffers that are possibly unaligned.
|
|
* If the hardware supports the transfer of unaligned data, just do the store.
|
|
* Otherwise, we have to move one byte at a time.
|
|
*/
|
|
#ifdef ACPI_BIG_ENDIAN
|
|
/*
|
|
* Macros for big-endian machines
|
|
*/
|
|
|
|
/* These macros reverse the bytes during the move, converting little-endian to big endian */
|
|
|
|
/* Big Endian <== Little Endian */
|
|
/* Hi...Lo Lo...Hi */
|
|
/* 16-bit source, 16/32/64 destination */
|
|
|
|
#define ACPI_MOVE_16_TO_16(d, s) {(( u8 *)(void *)(d))[0] = ((u8 *)(void *)(s))[1];\
|
|
(( u8 *)(void *)(d))[1] = ((u8 *)(void *)(s))[0];}
|
|
|
|
#define ACPI_MOVE_16_TO_32(d, s) {(*(u32 *)(void *)(d))=0;\
|
|
((u8 *)(void *)(d))[2] = ((u8 *)(void *)(s))[1];\
|
|
((u8 *)(void *)(d))[3] = ((u8 *)(void *)(s))[0];}
|
|
|
|
#define ACPI_MOVE_16_TO_64(d, s) {(*(u64 *)(void *)(d))=0;\
|
|
((u8 *)(void *)(d))[6] = ((u8 *)(void *)(s))[1];\
|
|
((u8 *)(void *)(d))[7] = ((u8 *)(void *)(s))[0];}
|
|
|
|
/* 32-bit source, 16/32/64 destination */
|
|
|
|
#define ACPI_MOVE_32_TO_16(d, s) ACPI_MOVE_16_TO_16(d, s) /* Truncate to 16 */
|
|
|
|
#define ACPI_MOVE_32_TO_32(d, s) {(( u8 *)(void *)(d))[0] = ((u8 *)(void *)(s))[3];\
|
|
(( u8 *)(void *)(d))[1] = ((u8 *)(void *)(s))[2];\
|
|
(( u8 *)(void *)(d))[2] = ((u8 *)(void *)(s))[1];\
|
|
(( u8 *)(void *)(d))[3] = ((u8 *)(void *)(s))[0];}
|
|
|
|
#define ACPI_MOVE_32_TO_64(d, s) {(*(u64 *)(void *)(d))=0;\
|
|
((u8 *)(void *)(d))[4] = ((u8 *)(void *)(s))[3];\
|
|
((u8 *)(void *)(d))[5] = ((u8 *)(void *)(s))[2];\
|
|
((u8 *)(void *)(d))[6] = ((u8 *)(void *)(s))[1];\
|
|
((u8 *)(void *)(d))[7] = ((u8 *)(void *)(s))[0];}
|
|
|
|
/* 64-bit source, 16/32/64 destination */
|
|
|
|
#define ACPI_MOVE_64_TO_16(d, s) ACPI_MOVE_16_TO_16(d, s) /* Truncate to 16 */
|
|
|
|
#define ACPI_MOVE_64_TO_32(d, s) ACPI_MOVE_32_TO_32(d, s) /* Truncate to 32 */
|
|
|
|
#define ACPI_MOVE_64_TO_64(d, s) {(( u8 *)(void *)(d))[0] = ((u8 *)(void *)(s))[7];\
|
|
(( u8 *)(void *)(d))[1] = ((u8 *)(void *)(s))[6];\
|
|
(( u8 *)(void *)(d))[2] = ((u8 *)(void *)(s))[5];\
|
|
(( u8 *)(void *)(d))[3] = ((u8 *)(void *)(s))[4];\
|
|
(( u8 *)(void *)(d))[4] = ((u8 *)(void *)(s))[3];\
|
|
(( u8 *)(void *)(d))[5] = ((u8 *)(void *)(s))[2];\
|
|
(( u8 *)(void *)(d))[6] = ((u8 *)(void *)(s))[1];\
|
|
(( u8 *)(void *)(d))[7] = ((u8 *)(void *)(s))[0];}
|
|
#else
|
|
/*
|
|
* Macros for little-endian machines
|
|
*/
|
|
|
|
#ifndef ACPI_MISALIGNMENT_NOT_SUPPORTED
|
|
|
|
/* The hardware supports unaligned transfers, just do the little-endian move */
|
|
|
|
/* 16-bit source, 16/32/64 destination */
|
|
|
|
#define ACPI_MOVE_16_TO_16(d, s) *(u16 *)(void *)(d) = *(u16 *)(void *)(s)
|
|
#define ACPI_MOVE_16_TO_32(d, s) *(u32 *)(void *)(d) = *(u16 *)(void *)(s)
|
|
#define ACPI_MOVE_16_TO_64(d, s) *(u64 *)(void *)(d) = *(u16 *)(void *)(s)
|
|
|
|
/* 32-bit source, 16/32/64 destination */
|
|
|
|
#define ACPI_MOVE_32_TO_16(d, s) ACPI_MOVE_16_TO_16(d, s) /* Truncate to 16 */
|
|
#define ACPI_MOVE_32_TO_32(d, s) *(u32 *)(void *)(d) = *(u32 *)(void *)(s)
|
|
#define ACPI_MOVE_32_TO_64(d, s) *(u64 *)(void *)(d) = *(u32 *)(void *)(s)
|
|
|
|
/* 64-bit source, 16/32/64 destination */
|
|
|
|
#define ACPI_MOVE_64_TO_16(d, s) ACPI_MOVE_16_TO_16(d, s) /* Truncate to 16 */
|
|
#define ACPI_MOVE_64_TO_32(d, s) ACPI_MOVE_32_TO_32(d, s) /* Truncate to 32 */
|
|
#define ACPI_MOVE_64_TO_64(d, s) *(u64 *)(void *)(d) = *(u64 *)(void *)(s)
|
|
|
|
#else
|
|
/*
|
|
* The hardware does not support unaligned transfers. We must move the
|
|
* data one byte at a time. These macros work whether the source or
|
|
* the destination (or both) is/are unaligned. (Little-endian move)
|
|
*/
|
|
|
|
/* 16-bit source, 16/32/64 destination */
|
|
|
|
#define ACPI_MOVE_16_TO_16(d, s) {(( u8 *)(void *)(d))[0] = ((u8 *)(void *)(s))[0];\
|
|
(( u8 *)(void *)(d))[1] = ((u8 *)(void *)(s))[1];}
|
|
|
|
#define ACPI_MOVE_16_TO_32(d, s) {(*(u32 *)(void *)(d)) = 0; ACPI_MOVE_16_TO_16(d, s);}
|
|
#define ACPI_MOVE_16_TO_64(d, s) {(*(u64 *)(void *)(d)) = 0; ACPI_MOVE_16_TO_16(d, s);}
|
|
|
|
/* 32-bit source, 16/32/64 destination */
|
|
|
|
#define ACPI_MOVE_32_TO_16(d, s) ACPI_MOVE_16_TO_16(d, s) /* Truncate to 16 */
|
|
|
|
#define ACPI_MOVE_32_TO_32(d, s) {(( u8 *)(void *)(d))[0] = ((u8 *)(void *)(s))[0];\
|
|
(( u8 *)(void *)(d))[1] = ((u8 *)(void *)(s))[1];\
|
|
(( u8 *)(void *)(d))[2] = ((u8 *)(void *)(s))[2];\
|
|
(( u8 *)(void *)(d))[3] = ((u8 *)(void *)(s))[3];}
|
|
|
|
#define ACPI_MOVE_32_TO_64(d, s) {(*(u64 *)(void *)(d)) = 0; ACPI_MOVE_32_TO_32(d, s);}
|
|
|
|
/* 64-bit source, 16/32/64 destination */
|
|
|
|
#define ACPI_MOVE_64_TO_16(d, s) ACPI_MOVE_16_TO_16(d, s) /* Truncate to 16 */
|
|
#define ACPI_MOVE_64_TO_32(d, s) ACPI_MOVE_32_TO_32(d, s) /* Truncate to 32 */
|
|
#define ACPI_MOVE_64_TO_64(d, s) {(( u8 *)(void *)(d))[0] = ((u8 *)(void *)(s))[0];\
|
|
(( u8 *)(void *)(d))[1] = ((u8 *)(void *)(s))[1];\
|
|
(( u8 *)(void *)(d))[2] = ((u8 *)(void *)(s))[2];\
|
|
(( u8 *)(void *)(d))[3] = ((u8 *)(void *)(s))[3];\
|
|
(( u8 *)(void *)(d))[4] = ((u8 *)(void *)(s))[4];\
|
|
(( u8 *)(void *)(d))[5] = ((u8 *)(void *)(s))[5];\
|
|
(( u8 *)(void *)(d))[6] = ((u8 *)(void *)(s))[6];\
|
|
(( u8 *)(void *)(d))[7] = ((u8 *)(void *)(s))[7];}
|
|
#endif
|
|
#endif
|
|
|
|
/* Macros based on machine integer width */
|
|
|
|
#if ACPI_MACHINE_WIDTH == 32
|
|
#define ACPI_MOVE_SIZE_TO_16(d, s) ACPI_MOVE_32_TO_16(d, s)
|
|
|
|
#elif ACPI_MACHINE_WIDTH == 64
|
|
#define ACPI_MOVE_SIZE_TO_16(d, s) ACPI_MOVE_64_TO_16(d, s)
|
|
|
|
#else
|
|
#error unknown ACPI_MACHINE_WIDTH
|
|
#endif
|
|
|
|
/*
|
|
* Fast power-of-two math macros for non-optimized compilers
|
|
*/
|
|
#define _ACPI_DIV(value, power_of2) ((u32) ((value) >> (power_of2)))
|
|
#define _ACPI_MUL(value, power_of2) ((u32) ((value) << (power_of2)))
|
|
#define _ACPI_MOD(value, divisor) ((u32) ((value) & ((divisor) -1)))
|
|
|
|
#define ACPI_DIV_2(a) _ACPI_DIV(a, 1)
|
|
#define ACPI_MUL_2(a) _ACPI_MUL(a, 1)
|
|
#define ACPI_MOD_2(a) _ACPI_MOD(a, 2)
|
|
|
|
#define ACPI_DIV_4(a) _ACPI_DIV(a, 2)
|
|
#define ACPI_MUL_4(a) _ACPI_MUL(a, 2)
|
|
#define ACPI_MOD_4(a) _ACPI_MOD(a, 4)
|
|
|
|
#define ACPI_DIV_8(a) _ACPI_DIV(a, 3)
|
|
#define ACPI_MUL_8(a) _ACPI_MUL(a, 3)
|
|
#define ACPI_MOD_8(a) _ACPI_MOD(a, 8)
|
|
|
|
#define ACPI_DIV_16(a) _ACPI_DIV(a, 4)
|
|
#define ACPI_MUL_16(a) _ACPI_MUL(a, 4)
|
|
#define ACPI_MOD_16(a) _ACPI_MOD(a, 16)
|
|
|
|
#define ACPI_DIV_32(a) _ACPI_DIV(a, 5)
|
|
#define ACPI_MUL_32(a) _ACPI_MUL(a, 5)
|
|
#define ACPI_MOD_32(a) _ACPI_MOD(a, 32)
|
|
|
|
/*
|
|
* Rounding macros (Power of two boundaries only)
|
|
*/
|
|
#define ACPI_ROUND_DOWN(value, boundary) (((acpi_size)(value)) & \
|
|
(~(((acpi_size) boundary)-1)))
|
|
|
|
#define ACPI_ROUND_UP(value, boundary) ((((acpi_size)(value)) + \
|
|
(((acpi_size) boundary)-1)) & \
|
|
(~(((acpi_size) boundary)-1)))
|
|
|
|
/* Note: sizeof(acpi_size) evaluates to either 4 or 8 (32- vs 64-bit mode) */
|
|
|
|
#define ACPI_ROUND_DOWN_TO_32BIT(a) ACPI_ROUND_DOWN(a, 4)
|
|
#define ACPI_ROUND_DOWN_TO_64BIT(a) ACPI_ROUND_DOWN(a, 8)
|
|
#define ACPI_ROUND_DOWN_TO_NATIVE_WORD(a) ACPI_ROUND_DOWN(a, sizeof(acpi_size))
|
|
|
|
#define ACPI_ROUND_UP_TO_32BIT(a) ACPI_ROUND_UP(a, 4)
|
|
#define ACPI_ROUND_UP_TO_64BIT(a) ACPI_ROUND_UP(a, 8)
|
|
#define ACPI_ROUND_UP_TO_NATIVE_WORD(a) ACPI_ROUND_UP(a, sizeof(acpi_size))
|
|
|
|
#define ACPI_ROUND_BITS_UP_TO_BYTES(a) ACPI_DIV_8((a) + 7)
|
|
#define ACPI_ROUND_BITS_DOWN_TO_BYTES(a) ACPI_DIV_8((a))
|
|
|
|
#define ACPI_ROUND_UP_TO_1K(a) (((a) + 1023) >> 10)
|
|
|
|
/* Generic (non-power-of-two) rounding */
|
|
|
|
#define ACPI_ROUND_UP_TO(value, boundary) (((value) + ((boundary)-1)) / (boundary))
|
|
|
|
#define ACPI_IS_MISALIGNED(value) (((acpi_size) value) & (sizeof(acpi_size)-1))
|
|
|
|
/*
|
|
* Bitmask creation
|
|
* Bit positions start at zero.
|
|
* MASK_BITS_ABOVE creates a mask starting AT the position and above
|
|
* MASK_BITS_BELOW creates a mask starting one bit BELOW the position
|
|
*/
|
|
#define ACPI_MASK_BITS_ABOVE(position) (~((ACPI_INTEGER_MAX) << ((u32) (position))))
|
|
#define ACPI_MASK_BITS_BELOW(position) ((ACPI_INTEGER_MAX) << ((u32) (position)))
|
|
|
|
/* Bitfields within ACPI registers */
|
|
|
|
#define ACPI_REGISTER_PREPARE_BITS(val, pos, mask) ((val << pos) & mask)
|
|
#define ACPI_REGISTER_INSERT_VALUE(reg, pos, mask, val) reg = (reg & (~(mask))) | ACPI_REGISTER_PREPARE_BITS(val, pos, mask)
|
|
|
|
#define ACPI_INSERT_BITS(target, mask, source) target = ((target & (~(mask))) | (source & mask))
|
|
|
|
/*
|
|
* A struct acpi_namespace_node can appear in some contexts
|
|
* where a pointer to a union acpi_operand_object can also
|
|
* appear. This macro is used to distinguish them.
|
|
*
|
|
* The "Descriptor" field is the first field in both structures.
|
|
*/
|
|
#define ACPI_GET_DESCRIPTOR_TYPE(d) (((union acpi_descriptor *)(void *)(d))->common.descriptor_type)
|
|
#define ACPI_SET_DESCRIPTOR_TYPE(d, t) (((union acpi_descriptor *)(void *)(d))->common.descriptor_type = t)
|
|
|
|
/*
|
|
* Macros for the master AML opcode table
|
|
*/
|
|
#if defined (ACPI_DISASSEMBLER) || defined (ACPI_DEBUG_OUTPUT)
|
|
#define ACPI_OP(name, Pargs, Iargs, obj_type, class, type, flags) \
|
|
{name, (u32)(Pargs), (u32)(Iargs), (u32)(flags), obj_type, class, type}
|
|
#else
|
|
#define ACPI_OP(name, Pargs, Iargs, obj_type, class, type, flags) \
|
|
{(u32)(Pargs), (u32)(Iargs), (u32)(flags), obj_type, class, type}
|
|
#endif
|
|
|
|
#define ARG_TYPE_WIDTH 5
|
|
#define ARG_1(x) ((u32)(x))
|
|
#define ARG_2(x) ((u32)(x) << (1 * ARG_TYPE_WIDTH))
|
|
#define ARG_3(x) ((u32)(x) << (2 * ARG_TYPE_WIDTH))
|
|
#define ARG_4(x) ((u32)(x) << (3 * ARG_TYPE_WIDTH))
|
|
#define ARG_5(x) ((u32)(x) << (4 * ARG_TYPE_WIDTH))
|
|
#define ARG_6(x) ((u32)(x) << (5 * ARG_TYPE_WIDTH))
|
|
|
|
#define ARGI_LIST1(a) (ARG_1(a))
|
|
#define ARGI_LIST2(a, b) (ARG_1(b)|ARG_2(a))
|
|
#define ARGI_LIST3(a, b, c) (ARG_1(c)|ARG_2(b)|ARG_3(a))
|
|
#define ARGI_LIST4(a, b, c, d) (ARG_1(d)|ARG_2(c)|ARG_3(b)|ARG_4(a))
|
|
#define ARGI_LIST5(a, b, c, d, e) (ARG_1(e)|ARG_2(d)|ARG_3(c)|ARG_4(b)|ARG_5(a))
|
|
#define ARGI_LIST6(a, b, c, d, e, f) (ARG_1(f)|ARG_2(e)|ARG_3(d)|ARG_4(c)|ARG_5(b)|ARG_6(a))
|
|
|
|
#define ARGP_LIST1(a) (ARG_1(a))
|
|
#define ARGP_LIST2(a, b) (ARG_1(a)|ARG_2(b))
|
|
#define ARGP_LIST3(a, b, c) (ARG_1(a)|ARG_2(b)|ARG_3(c))
|
|
#define ARGP_LIST4(a, b, c, d) (ARG_1(a)|ARG_2(b)|ARG_3(c)|ARG_4(d))
|
|
#define ARGP_LIST5(a, b, c, d, e) (ARG_1(a)|ARG_2(b)|ARG_3(c)|ARG_4(d)|ARG_5(e))
|
|
#define ARGP_LIST6(a, b, c, d, e, f) (ARG_1(a)|ARG_2(b)|ARG_3(c)|ARG_4(d)|ARG_5(e)|ARG_6(f))
|
|
|
|
#define GET_CURRENT_ARG_TYPE(list) (list & ((u32) 0x1F))
|
|
#define INCREMENT_ARG_LIST(list) (list >>= ((u32) ARG_TYPE_WIDTH))
|
|
|
|
/*
|
|
* Ascii error messages can be configured out
|
|
*/
|
|
#ifndef ACPI_NO_ERROR_MESSAGES
|
|
|
|
/*
|
|
* Error reporting. Callers module and line number are inserted by AE_INFO,
|
|
* the plist contains a set of parens to allow variable-length lists.
|
|
* These macros are used for both the debug and non-debug versions of the code.
|
|
*/
|
|
#define ACPI_ERROR_NAMESPACE(s, e) acpi_ns_report_error (AE_INFO, s, e);
|
|
#define ACPI_ERROR_METHOD(s, n, p, e) acpi_ns_report_method_error (AE_INFO, s, n, p, e);
|
|
#define ACPI_WARN_PREDEFINED(plist) acpi_ut_predefined_warning plist
|
|
|
|
#else
|
|
|
|
/* No error messages */
|
|
|
|
#define ACPI_ERROR_NAMESPACE(s, e)
|
|
#define ACPI_ERROR_METHOD(s, n, p, e)
|
|
#define ACPI_WARN_PREDEFINED(plist)
|
|
#endif /* ACPI_NO_ERROR_MESSAGES */
|
|
|
|
/*
|
|
* Debug macros that are conditionally compiled
|
|
*/
|
|
#ifdef ACPI_DEBUG_OUTPUT
|
|
|
|
/*
|
|
* Function entry tracing
|
|
*/
|
|
#ifdef CONFIG_ACPI_DEBUG_FUNC_TRACE
|
|
|
|
#define ACPI_FUNCTION_TRACE(a) ACPI_FUNCTION_NAME(a) \
|
|
acpi_ut_trace(ACPI_DEBUG_PARAMETERS)
|
|
#define ACPI_FUNCTION_TRACE_PTR(a, b) ACPI_FUNCTION_NAME(a) \
|
|
acpi_ut_trace_ptr(ACPI_DEBUG_PARAMETERS, (void *)b)
|
|
#define ACPI_FUNCTION_TRACE_U32(a, b) ACPI_FUNCTION_NAME(a) \
|
|
acpi_ut_trace_u32(ACPI_DEBUG_PARAMETERS, (u32)b)
|
|
#define ACPI_FUNCTION_TRACE_STR(a, b) ACPI_FUNCTION_NAME(a) \
|
|
acpi_ut_trace_str(ACPI_DEBUG_PARAMETERS, (char *)b)
|
|
|
|
#define ACPI_FUNCTION_ENTRY() acpi_ut_track_stack_ptr()
|
|
|
|
/*
|
|
* Function exit tracing.
|
|
* WARNING: These macros include a return statement. This is usually considered
|
|
* bad form, but having a separate exit macro is very ugly and difficult to maintain.
|
|
* One of the FUNCTION_TRACE macros above must be used in conjunction with these macros
|
|
* so that "_AcpiFunctionName" is defined.
|
|
*
|
|
* Note: the DO_WHILE0 macro is used to prevent some compilers from complaining
|
|
* about these constructs.
|
|
*/
|
|
#ifdef ACPI_USE_DO_WHILE_0
|
|
#define ACPI_DO_WHILE0(a) do a while(0)
|
|
#else
|
|
#define ACPI_DO_WHILE0(a) a
|
|
#endif
|
|
|
|
#define return_VOID ACPI_DO_WHILE0 ({ \
|
|
acpi_ut_exit (ACPI_DEBUG_PARAMETERS); \
|
|
return;})
|
|
/*
|
|
* There are two versions of most of the return macros. The default version is
|
|
* safer, since it avoids side-effects by guaranteeing that the argument will
|
|
* not be evaluated twice.
|
|
*
|
|
* A less-safe version of the macros is provided for optional use if the
|
|
* compiler uses excessive CPU stack (for example, this may happen in the
|
|
* debug case if code optimzation is disabled.)
|
|
*/
|
|
#ifndef ACPI_SIMPLE_RETURN_MACROS
|
|
|
|
#define return_ACPI_STATUS(s) ACPI_DO_WHILE0 ({ \
|
|
register acpi_status _s = (s); \
|
|
acpi_ut_status_exit (ACPI_DEBUG_PARAMETERS, _s); \
|
|
return (_s); })
|
|
#define return_PTR(s) ACPI_DO_WHILE0 ({ \
|
|
register void *_s = (void *) (s); \
|
|
acpi_ut_ptr_exit (ACPI_DEBUG_PARAMETERS, (u8 *) _s); \
|
|
return (_s); })
|
|
#define return_VALUE(s) ACPI_DO_WHILE0 ({ \
|
|
register acpi_integer _s = (s); \
|
|
acpi_ut_value_exit (ACPI_DEBUG_PARAMETERS, _s); \
|
|
return (_s); })
|
|
#define return_UINT8(s) ACPI_DO_WHILE0 ({ \
|
|
register u8 _s = (u8) (s); \
|
|
acpi_ut_value_exit (ACPI_DEBUG_PARAMETERS, (acpi_integer) _s); \
|
|
return (_s); })
|
|
#define return_UINT32(s) ACPI_DO_WHILE0 ({ \
|
|
register u32 _s = (u32) (s); \
|
|
acpi_ut_value_exit (ACPI_DEBUG_PARAMETERS, (acpi_integer) _s); \
|
|
return (_s); })
|
|
#else /* Use original less-safe macros */
|
|
|
|
#define return_ACPI_STATUS(s) ACPI_DO_WHILE0 ({ \
|
|
acpi_ut_status_exit (ACPI_DEBUG_PARAMETERS, (s)); \
|
|
return((s)); })
|
|
#define return_PTR(s) ACPI_DO_WHILE0 ({ \
|
|
acpi_ut_ptr_exit (ACPI_DEBUG_PARAMETERS, (u8 *) (s)); \
|
|
return((s)); })
|
|
#define return_VALUE(s) ACPI_DO_WHILE0 ({ \
|
|
acpi_ut_value_exit (ACPI_DEBUG_PARAMETERS, (acpi_integer) (s)); \
|
|
return((s)); })
|
|
#define return_UINT8(s) return_VALUE(s)
|
|
#define return_UINT32(s) return_VALUE(s)
|
|
|
|
#endif /* ACPI_SIMPLE_RETURN_MACROS */
|
|
|
|
#else /* !CONFIG_ACPI_DEBUG_FUNC_TRACE */
|
|
|
|
#define ACPI_FUNCTION_TRACE(a)
|
|
#define ACPI_FUNCTION_TRACE_PTR(a,b)
|
|
#define ACPI_FUNCTION_TRACE_U32(a,b)
|
|
#define ACPI_FUNCTION_TRACE_STR(a,b)
|
|
#define ACPI_FUNCTION_EXIT
|
|
#define ACPI_FUNCTION_STATUS_EXIT(s)
|
|
#define ACPI_FUNCTION_VALUE_EXIT(s)
|
|
#define ACPI_FUNCTION_TRACE(a)
|
|
#define ACPI_FUNCTION_ENTRY()
|
|
|
|
#define return_VOID return
|
|
#define return_ACPI_STATUS(s) return(s)
|
|
#define return_VALUE(s) return(s)
|
|
#define return_UINT8(s) return(s)
|
|
#define return_UINT32(s) return(s)
|
|
#define return_PTR(s) return(s)
|
|
|
|
#endif /* CONFIG_ACPI_DEBUG_FUNC_TRACE */
|
|
|
|
/* Conditional execution */
|
|
|
|
#define ACPI_DEBUG_EXEC(a) a
|
|
#define ACPI_NORMAL_EXEC(a)
|
|
|
|
#define ACPI_DEBUG_DEFINE(a) a;
|
|
#define ACPI_DEBUG_ONLY_MEMBERS(a) a;
|
|
#define _VERBOSE_STRUCTURES
|
|
|
|
/* Stack and buffer dumping */
|
|
|
|
#define ACPI_DUMP_STACK_ENTRY(a) acpi_ex_dump_operand((a), 0)
|
|
#define ACPI_DUMP_OPERANDS(a, b, c) acpi_ex_dump_operands(a, b, c)
|
|
|
|
#define ACPI_DUMP_ENTRY(a, b) acpi_ns_dump_entry (a, b)
|
|
#define ACPI_DUMP_PATHNAME(a, b, c, d) acpi_ns_dump_pathname(a, b, c, d)
|
|
#define ACPI_DUMP_RESOURCE_LIST(a) acpi_rs_dump_resource_list(a)
|
|
#define ACPI_DUMP_BUFFER(a, b) acpi_ut_dump_buffer((u8 *) a, b, DB_BYTE_DISPLAY, _COMPONENT)
|
|
|
|
#else
|
|
/*
|
|
* This is the non-debug case -- make everything go away,
|
|
* leaving no executable debug code!
|
|
*/
|
|
#define ACPI_DEBUG_EXEC(a)
|
|
#define ACPI_NORMAL_EXEC(a) a;
|
|
|
|
#define ACPI_DEBUG_DEFINE(a) do { } while(0)
|
|
#define ACPI_DEBUG_ONLY_MEMBERS(a) do { } while(0)
|
|
#define ACPI_FUNCTION_TRACE(a) do { } while(0)
|
|
#define ACPI_FUNCTION_TRACE_PTR(a, b) do { } while(0)
|
|
#define ACPI_FUNCTION_TRACE_U32(a, b) do { } while(0)
|
|
#define ACPI_FUNCTION_TRACE_STR(a, b) do { } while(0)
|
|
#define ACPI_FUNCTION_EXIT do { } while(0)
|
|
#define ACPI_FUNCTION_STATUS_EXIT(s) do { } while(0)
|
|
#define ACPI_FUNCTION_VALUE_EXIT(s) do { } while(0)
|
|
#define ACPI_FUNCTION_ENTRY() do { } while(0)
|
|
#define ACPI_DUMP_STACK_ENTRY(a) do { } while(0)
|
|
#define ACPI_DUMP_OPERANDS(a, b, c) do { } while(0)
|
|
#define ACPI_DUMP_ENTRY(a, b) do { } while(0)
|
|
#define ACPI_DUMP_TABLES(a, b) do { } while(0)
|
|
#define ACPI_DUMP_PATHNAME(a, b, c, d) do { } while(0)
|
|
#define ACPI_DUMP_RESOURCE_LIST(a) do { } while(0)
|
|
#define ACPI_DUMP_BUFFER(a, b) do { } while(0)
|
|
|
|
#define return_VOID return
|
|
#define return_ACPI_STATUS(s) return(s)
|
|
#define return_VALUE(s) return(s)
|
|
#define return_UINT8(s) return(s)
|
|
#define return_UINT32(s) return(s)
|
|
#define return_PTR(s) return(s)
|
|
|
|
#endif /* ACPI_DEBUG_OUTPUT */
|
|
|
|
/*
|
|
* Some code only gets executed when the debugger is built in.
|
|
* Note that this is entirely independent of whether the
|
|
* DEBUG_PRINT stuff (set by ACPI_DEBUG_OUTPUT) is on, or not.
|
|
*/
|
|
#ifdef ACPI_DEBUGGER
|
|
#define ACPI_DEBUGGER_EXEC(a) a
|
|
#else
|
|
#define ACPI_DEBUGGER_EXEC(a)
|
|
#endif
|
|
|
|
#ifdef ACPI_DEBUG_OUTPUT
|
|
/*
|
|
* 1) Set name to blanks
|
|
* 2) Copy the object name
|
|
*/
|
|
#define ACPI_ADD_OBJECT_NAME(a,b) ACPI_MEMSET (a->common.name, ' ', sizeof (a->common.name));\
|
|
ACPI_STRNCPY (a->common.name, acpi_gbl_ns_type_names[b], sizeof (a->common.name))
|
|
#else
|
|
|
|
#define ACPI_ADD_OBJECT_NAME(a,b)
|
|
#endif
|
|
|
|
/*
|
|
* Memory allocation tracking (DEBUG ONLY)
|
|
*/
|
|
#define ACPI_MEM_PARAMETERS _COMPONENT, _acpi_module_name, __LINE__
|
|
|
|
#ifndef ACPI_DBG_TRACK_ALLOCATIONS
|
|
|
|
/* Memory allocation */
|
|
|
|
#ifndef ACPI_ALLOCATE
|
|
#define ACPI_ALLOCATE(a) acpi_ut_allocate((acpi_size)(a), ACPI_MEM_PARAMETERS)
|
|
#endif
|
|
#ifndef ACPI_ALLOCATE_ZEROED
|
|
#define ACPI_ALLOCATE_ZEROED(a) acpi_ut_allocate_zeroed((acpi_size)(a), ACPI_MEM_PARAMETERS)
|
|
#endif
|
|
#ifndef ACPI_FREE
|
|
#define ACPI_FREE(a) acpio_os_free(a)
|
|
#endif
|
|
#define ACPI_MEM_TRACKING(a)
|
|
|
|
#else
|
|
|
|
/* Memory allocation */
|
|
|
|
#define ACPI_ALLOCATE(a) acpi_ut_allocate_and_track((acpi_size)(a), ACPI_MEM_PARAMETERS)
|
|
#define ACPI_ALLOCATE_ZEROED(a) acpi_ut_allocate_zeroed_and_track((acpi_size)(a), ACPI_MEM_PARAMETERS)
|
|
#define ACPI_FREE(a) acpi_ut_free_and_track(a, ACPI_MEM_PARAMETERS)
|
|
#define ACPI_MEM_TRACKING(a) a
|
|
|
|
#endif /* ACPI_DBG_TRACK_ALLOCATIONS */
|
|
|
|
/* Preemption point */
|
|
#ifndef ACPI_PREEMPTION_POINT
|
|
#define ACPI_PREEMPTION_POINT() /* no preemption */
|
|
#endif
|
|
|
|
#endif /* ACMACROS_H */
|