/* * Copyright (c) 2016-2024, Arm Limited and Contributors. All rights reserved. * Copyright (c) 2020, NVIDIA Corporation. All rights reserved. * * SPDX-License-Identifier: BSD-3-Clause */ #ifndef UTILS_DEF_H #define UTILS_DEF_H #include /* Compute the number of elements in the given array */ #define ARRAY_SIZE(a) \ (sizeof(a) / sizeof((a)[0])) #define IS_POWER_OF_TWO(x) \ (((x) & ((x) - 1)) == 0) #define SIZE_FROM_LOG2_WORDS(n) (U(4) << (n)) #define BIT_32(nr) (U(1) << (nr)) #define BIT_64(nr) (ULL(1) << (nr)) #ifdef __aarch64__ #define BIT BIT_64 #else #define BIT BIT_32 #endif /* * Create a contiguous bitmask starting at bit position @l and ending at * position @h. For example * GENMASK_64(39, 21) gives us the 64bit vector 0x000000ffffe00000. */ #if defined(__LINKER__) || defined(__ASSEMBLER__) #define GENMASK_32(h, l) \ (((0xFFFFFFFF) << (l)) & (0xFFFFFFFF >> (32 - 1 - (h)))) #define GENMASK_64(h, l) \ ((~0 << (l)) & (~0 >> (64 - 1 - (h)))) #else #define GENMASK_32(h, l) \ (((~UINT32_C(0)) << (l)) & (~UINT32_C(0) >> (32 - 1 - (h)))) #define GENMASK_64(h, l) \ (((~UINT64_C(0)) << (l)) & (~UINT64_C(0) >> (64 - 1 - (h)))) #endif #ifdef __aarch64__ #define GENMASK GENMASK_64 #else #define GENMASK GENMASK_32 #endif #define HI(addr) (addr >> 32) #define LO(addr) (addr & 0xffffffff) /* * This variant of div_round_up can be used in macro definition but should not * be used in C code as the `div` parameter is evaluated twice. */ #define DIV_ROUND_UP_2EVAL(n, d) (((n) + (d) - 1) / (d)) #define div_round_up(val, div) __extension__ ({ \ __typeof__(div) _div = (div); \ ((val) + _div - (__typeof__(div)) 1) / _div; \ }) #define MIN(x, y) __extension__ ({ \ __typeof__(x) _x = (x); \ __typeof__(y) _y = (y); \ (void)(&_x == &_y); \ (_x < _y) ? _x : _y; \ }) #define MAX(x, y) __extension__ ({ \ __typeof__(x) _x = (x); \ __typeof__(y) _y = (y); \ (void)(&_x == &_y); \ (_x > _y) ? _x : _y; \ }) #define CLAMP(x, min, max) __extension__ ({ \ __typeof__(x) _x = (x); \ __typeof__(min) _min = (min); \ __typeof__(max) _max = (max); \ (void)(&_x == &_min); \ (void)(&_x == &_max); \ ((_x > _max) ? _max : ((_x < _min) ? _min : _x)); \ }) /* * The round_up() macro rounds up a value to the given boundary in a * type-agnostic yet type-safe manner. The boundary must be a power of two. * In other words, it computes the smallest multiple of boundary which is * greater than or equal to value. * * round_down() is similar but rounds the value down instead. */ #define round_boundary(value, boundary) \ ((__typeof__(value))((boundary) - 1)) #define round_up(value, boundary) \ ((((value) - 1) | round_boundary(value, boundary)) + 1) #define round_down(value, boundary) \ ((value) & ~round_boundary(value, boundary)) /* add operation together with checking whether the operation overflowed * The result is '*res', * return 0 on success and 1 on overflow */ #define add_overflow(a, b, res) __builtin_add_overflow((a), (b), (res)) /* * Round up a value to align with a given size and * check whether overflow happens. * The rounduped value is '*res', * return 0 on success and 1 on overflow */ #define round_up_overflow(v, size, res) (__extension__({ \ typeof(res) __res = res; \ typeof(*(__res)) __roundup_tmp = 0; \ typeof(v) __roundup_mask = (typeof(v))(size) - 1; \ \ add_overflow((v), __roundup_mask, &__roundup_tmp) ? 1 : \ (void)(*(__res) = __roundup_tmp & ~__roundup_mask), 0; \ })) /* * Add a with b, then round up the result to align with a given size and * check whether overflow happens. * The rounduped value is '*res', * return 0 on success and 1 on overflow */ #define add_with_round_up_overflow(a, b, size, res) (__extension__({ \ typeof(a) __a = (a); \ typeof(__a) __add_res = 0; \ \ add_overflow((__a), (b), &__add_res) ? 1 : \ round_up_overflow(__add_res, (size), (res)) ? 1 : 0; \ })) /** * Helper macro to ensure a value lies on a given boundary. */ #define is_aligned(value, boundary) \ (round_up((uintptr_t) value, boundary) == \ round_down((uintptr_t) value, boundary)) /* * Evaluates to 1 if (ptr + inc) overflows, 0 otherwise. * Both arguments must be unsigned pointer values (i.e. uintptr_t). */ #define check_uptr_overflow(_ptr, _inc) \ ((_ptr) > (UINTPTR_MAX - (_inc))) /* * Evaluates to 1 if (u32 + inc) overflows, 0 otherwise. * Both arguments must be 32-bit unsigned integers (i.e. effectively uint32_t). */ #define check_u32_overflow(_u32, _inc) \ ((_u32) > (UINT32_MAX - (_inc))) /* Register size of the current architecture. */ #ifdef __aarch64__ #define REGSZ U(8) #else #define REGSZ U(4) #endif /* * Test for the current architecture version to be at least the version * expected. */ #define ARM_ARCH_AT_LEAST(_maj, _min) \ ((ARM_ARCH_MAJOR > (_maj)) || \ ((ARM_ARCH_MAJOR == (_maj)) && (ARM_ARCH_MINOR >= (_min)))) /* * Import an assembly or linker symbol as a C expression with the specified * type */ #define IMPORT_SYM(type, sym, name) \ extern char sym[];\ static const __attribute__((unused)) type name = (type) sym; /* * When the symbol is used to hold a pointer, its alignment can be asserted * with this macro. For example, if there is a linker symbol that is going to * be used as a 64-bit pointer, the value of the linker symbol must also be * aligned to 64 bit. This macro makes sure this is the case. */ #define ASSERT_SYM_PTR_ALIGN(sym) assert(((size_t)(sym) % __alignof__(*(sym))) == 0) #define COMPILER_BARRIER() __asm__ volatile ("" ::: "memory") /* Compiler builtin of GCC >= 9 and planned in llvm */ #ifdef __HAVE_SPECULATION_SAFE_VALUE # define SPECULATION_SAFE_VALUE(var) __builtin_speculation_safe_value(var) #else # define SPECULATION_SAFE_VALUE(var) var #endif /* * Ticks elapsed in one second with a signal of 1 MHz */ #define MHZ_TICKS_PER_SEC U(1000000) /* * Ticks elapsed in one second with a signal of 1 KHz */ #define KHZ_TICKS_PER_SEC U(1000) #endif /* UTILS_DEF_H */