/* Routines required for instrumenting a program. */ /* Compile this one with gcc. */ /* Copyright (C) 1989-2020 Free Software Foundation, Inc. This file is part of GCC. GCC is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 3, or (at your option) any later version. GCC is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. Under Section 7 of GPL version 3, you are granted additional permissions described in the GCC Runtime Library Exception, version 3.1, as published by the Free Software Foundation. You should have received a copy of the GNU General Public License and a copy of the GCC Runtime Library Exception along with this program; see the files COPYING3 and COPYING.RUNTIME respectively. If not, see . */ #include "libgcov.h" #if !defined(inhibit_libc) /* Detect whether target can support atomic update of profilers. */ #if __SIZEOF_LONG_LONG__ == 4 && __GCC_HAVE_SYNC_COMPARE_AND_SWAP_4 #define GCOV_SUPPORTS_ATOMIC 1 #else #if __SIZEOF_LONG_LONG__ == 8 && __GCC_HAVE_SYNC_COMPARE_AND_SWAP_8 #define GCOV_SUPPORTS_ATOMIC 1 #else #define GCOV_SUPPORTS_ATOMIC 0 #endif #endif #ifdef L_gcov_interval_profiler /* If VALUE is in interval , then increases the corresponding counter in COUNTERS. If the VALUE is above or below the interval, COUNTERS[STEPS] or COUNTERS[STEPS + 1] is increased instead. */ void __gcov_interval_profiler (gcov_type *counters, gcov_type value, int start, unsigned steps) { gcov_type delta = value - start; if (delta < 0) counters[steps + 1]++; else if (delta >= steps) counters[steps]++; else counters[delta]++; } #endif #if defined(L_gcov_interval_profiler_atomic) && GCOV_SUPPORTS_ATOMIC /* If VALUE is in interval , then increases the corresponding counter in COUNTERS. If the VALUE is above or below the interval, COUNTERS[STEPS] or COUNTERS[STEPS + 1] is increased instead. Function is thread-safe. */ void __gcov_interval_profiler_atomic (gcov_type *counters, gcov_type value, int start, unsigned steps) { gcov_type delta = value - start; if (delta < 0) __atomic_fetch_add (&counters[steps + 1], 1, __ATOMIC_RELAXED); else if (delta >= steps) __atomic_fetch_add (&counters[steps], 1, __ATOMIC_RELAXED); else __atomic_fetch_add (&counters[delta], 1, __ATOMIC_RELAXED); } #endif #ifdef L_gcov_pow2_profiler /* If VALUE is a power of two, COUNTERS[1] is incremented. Otherwise COUNTERS[0] is incremented. */ void __gcov_pow2_profiler (gcov_type *counters, gcov_type value) { if (value == 0 || (value & (value - 1))) counters[0]++; else counters[1]++; } #endif #if defined(L_gcov_pow2_profiler_atomic) && GCOV_SUPPORTS_ATOMIC /* If VALUE is a power of two, COUNTERS[1] is incremented. Otherwise COUNTERS[0] is incremented. Function is thread-safe. */ void __gcov_pow2_profiler_atomic (gcov_type *counters, gcov_type value) { if (value == 0 || (value & (value - 1))) __atomic_fetch_add (&counters[0], 1, __ATOMIC_RELAXED); else __atomic_fetch_add (&counters[1], 1, __ATOMIC_RELAXED); } #endif /* Tries to determine N most commons value among its inputs. */ static inline void __gcov_topn_values_profiler_body (gcov_type *counters, gcov_type value, int use_atomic) { if (use_atomic) __atomic_fetch_add (&counters[0], 1, __ATOMIC_RELAXED); else counters[0]++; ++counters; /* First try to find an existing value. */ int empty_counter = -1; for (unsigned i = 0; i < GCOV_TOPN_VALUES; i++) if (value == counters[2 * i]) { if (use_atomic) __atomic_fetch_add (&counters[2 * i + 1], GCOV_TOPN_VALUES, __ATOMIC_RELAXED); else counters[2 * i + 1] += GCOV_TOPN_VALUES; return; } else if (counters[2 * i + 1] <= 0) empty_counter = i; /* Find an empty slot for a new value. */ if (empty_counter != -1) { counters[2 * empty_counter] = value; counters[2 * empty_counter + 1] = GCOV_TOPN_VALUES; return; } /* We haven't found an empty slot, then decrement all counter values by one. */ for (unsigned i = 0; i < GCOV_TOPN_VALUES; i++) if (use_atomic) __atomic_fetch_sub (&counters[2 * i + 1], 1, __ATOMIC_RELAXED); else counters[2 * i + 1]--; } #ifdef L_gcov_topn_values_profiler void __gcov_topn_values_profiler (gcov_type *counters, gcov_type value) { __gcov_topn_values_profiler_body (counters, value, 0); } #endif #if defined(L_gcov_topn_values_profiler_atomic) && GCOV_SUPPORTS_ATOMIC /* Update one value profilers (COUNTERS) for a given VALUE. CAVEAT: Following function is not thread-safe, only total number of executions (COUNTERS[2]) is update with an atomic instruction. Problem is that one cannot atomically update two counters (COUNTERS[0] and COUNTERS[1]), for more information please read following email thread: https://gcc.gnu.org/ml/gcc-patches/2016-08/msg00024.html. */ void __gcov_topn_values_profiler_atomic (gcov_type *counters, gcov_type value) { __gcov_topn_values_profiler_body (counters, value, 1); } #endif #ifdef L_gcov_indirect_call_profiler_v4 /* These two variables are used to actually track caller and callee. Keep them in TLS memory so races are not common (they are written to often). The variables are set directly by GCC instrumented code, so declaration here must match one in tree-profile.c */ #if defined(HAVE_CC_TLS) && !defined (USE_EMUTLS) __thread #endif struct indirect_call_tuple __gcov_indirect_call; /* By default, the C++ compiler will use function addresses in the vtable entries. Setting TARGET_VTABLE_USES_DESCRIPTORS to nonzero tells the compiler to use function descriptors instead. The value of this macro says how many words wide the descriptor is (normally 2). It is assumed that the address of a function descriptor may be treated as a pointer to a function. */ /* Tries to determine the most common value among its inputs. */ static inline void __gcov_indirect_call_profiler_body (gcov_type value, void *cur_func, int use_atomic) { /* If the C++ virtual tables contain function descriptors then one function may have multiple descriptors and we need to dereference the descriptors to see if they point to the same function. */ if (cur_func == __gcov_indirect_call.callee || (__LIBGCC_VTABLE_USES_DESCRIPTORS__ && *(void **) cur_func == *(void **) __gcov_indirect_call.callee)) __gcov_topn_values_profiler_body (__gcov_indirect_call.counters, value, use_atomic); __gcov_indirect_call.callee = NULL; } void __gcov_indirect_call_profiler_v4 (gcov_type value, void *cur_func) { __gcov_indirect_call_profiler_body (value, cur_func, 0); } #if GCOV_SUPPORTS_ATOMIC void __gcov_indirect_call_profiler_v4_atomic (gcov_type value, void *cur_func) { __gcov_indirect_call_profiler_body (value, cur_func, 1); } #endif #endif #ifdef L_gcov_time_profiler /* Counter for first visit of each function. */ gcov_type __gcov_time_profiler_counter ATTRIBUTE_HIDDEN; #endif #ifdef L_gcov_average_profiler /* Increase corresponding COUNTER by VALUE. FIXME: Perhaps we want to saturate up. */ void __gcov_average_profiler (gcov_type *counters, gcov_type value) { counters[0] += value; counters[1] ++; } #endif #if defined(L_gcov_average_profiler_atomic) && GCOV_SUPPORTS_ATOMIC /* Increase corresponding COUNTER by VALUE. FIXME: Perhaps we want to saturate up. Function is thread-safe. */ void __gcov_average_profiler_atomic (gcov_type *counters, gcov_type value) { __atomic_fetch_add (&counters[0], value, __ATOMIC_RELAXED); __atomic_fetch_add (&counters[1], 1, __ATOMIC_RELAXED); } #endif #ifdef L_gcov_ior_profiler /* Bitwise-OR VALUE into COUNTER. */ void __gcov_ior_profiler (gcov_type *counters, gcov_type value) { *counters |= value; } #endif #if defined(L_gcov_ior_profiler_atomic) && GCOV_SUPPORTS_ATOMIC /* Bitwise-OR VALUE into COUNTER. Function is thread-safe. */ void __gcov_ior_profiler_atomic (gcov_type *counters, gcov_type value) { __atomic_fetch_or (&counters[0], value, __ATOMIC_RELAXED); } #endif #endif /* inhibit_libc */