+ *length = 0;
+
+ if (*length == 0)
+ error (_("Could not find the frame base for \"%s\"."),
+ SYMBOL_NATURAL_NAME (framefunc));
+}
+
+/* Helper function for dwarf2_evaluate_loc_desc. Computes the CFA for
+ the frame in BATON. */
+
+static CORE_ADDR
+dwarf_expr_frame_cfa (void *baton)
+{
+ struct dwarf_expr_baton *debaton = (struct dwarf_expr_baton *) baton;
+
+ return dwarf2_frame_cfa (debaton->frame);
+}
+
+/* Helper function for dwarf2_evaluate_loc_desc. Computes the PC for
+ the frame in BATON. */
+
+static CORE_ADDR
+dwarf_expr_frame_pc (void *baton)
+{
+ struct dwarf_expr_baton *debaton = (struct dwarf_expr_baton *) baton;
+
+ return get_frame_address_in_block (debaton->frame);
+}
+
+/* Using the objfile specified in BATON, find the address for the
+ current thread's thread-local storage with offset OFFSET. */
+static CORE_ADDR
+dwarf_expr_tls_address (void *baton, CORE_ADDR offset)
+{
+ struct dwarf_expr_baton *debaton = (struct dwarf_expr_baton *) baton;
+ struct objfile *objfile = dwarf2_per_cu_objfile (debaton->per_cu);
+
+ return target_translate_tls_address (objfile, offset);
+}
+
+/* Call DWARF subroutine from DW_AT_location of DIE at DIE_OFFSET in
+ current CU (as is PER_CU). State of the CTX is not affected by the
+ call and return. */
+
+static void
+per_cu_dwarf_call (struct dwarf_expr_context *ctx, cu_offset die_offset,
+ struct dwarf2_per_cu_data *per_cu,
+ CORE_ADDR (*get_frame_pc) (void *baton),
+ void *baton)
+{
+ struct dwarf2_locexpr_baton block;
+
+ block = dwarf2_fetch_die_loc_cu_off (die_offset, per_cu, get_frame_pc, baton);
+
+ /* DW_OP_call_ref is currently not supported. */
+ gdb_assert (block.per_cu == per_cu);
+
+ dwarf_expr_eval (ctx, block.data, block.size);
+}
+
+/* Helper interface of per_cu_dwarf_call for dwarf2_evaluate_loc_desc. */
+
+static void
+dwarf_expr_dwarf_call (struct dwarf_expr_context *ctx, cu_offset die_offset)
+{
+ struct dwarf_expr_baton *debaton = ctx->baton;
+
+ per_cu_dwarf_call (ctx, die_offset, debaton->per_cu,
+ ctx->funcs->get_frame_pc, ctx->baton);
+}
+
+/* Callback function for dwarf2_evaluate_loc_desc. */
+
+static struct type *
+dwarf_expr_get_base_type (struct dwarf_expr_context *ctx,
+ cu_offset die_offset)
+{
+ struct dwarf_expr_baton *debaton = ctx->baton;
+
+ return dwarf2_get_die_type (die_offset, debaton->per_cu);
+}
+
+/* See dwarf2loc.h. */
+
+unsigned int entry_values_debug = 0;
+
+/* Helper to set entry_values_debug. */
+
+static void
+show_entry_values_debug (struct ui_file *file, int from_tty,
+ struct cmd_list_element *c, const char *value)
+{
+ fprintf_filtered (file,
+ _("Entry values and tail call frames debugging is %s.\n"),
+ value);
+}
+
+/* Find DW_TAG_GNU_call_site's DW_AT_GNU_call_site_target address.
+ CALLER_FRAME (for registers) can be NULL if it is not known. This function
+ always returns valid address or it throws NO_ENTRY_VALUE_ERROR. */
+
+static CORE_ADDR
+call_site_to_target_addr (struct gdbarch *call_site_gdbarch,
+ struct call_site *call_site,
+ struct frame_info *caller_frame)
+{
+ switch (FIELD_LOC_KIND (call_site->target))
+ {
+ case FIELD_LOC_KIND_DWARF_BLOCK:
+ {
+ struct dwarf2_locexpr_baton *dwarf_block;
+ struct value *val;
+ struct type *caller_core_addr_type;
+ struct gdbarch *caller_arch;
+
+ dwarf_block = FIELD_DWARF_BLOCK (call_site->target);
+ if (dwarf_block == NULL)
+ {
+ struct bound_minimal_symbol msym;
+
+ msym = lookup_minimal_symbol_by_pc (call_site->pc - 1);
+ throw_error (NO_ENTRY_VALUE_ERROR,
+ _("DW_AT_GNU_call_site_target is not specified "
+ "at %s in %s"),
+ paddress (call_site_gdbarch, call_site->pc),
+ (msym.minsym == NULL ? "???"
+ : SYMBOL_PRINT_NAME (msym.minsym)));
+
+ }
+ if (caller_frame == NULL)
+ {
+ struct bound_minimal_symbol msym;
+
+ msym = lookup_minimal_symbol_by_pc (call_site->pc - 1);
+ throw_error (NO_ENTRY_VALUE_ERROR,
+ _("DW_AT_GNU_call_site_target DWARF block resolving "
+ "requires known frame which is currently not "
+ "available at %s in %s"),
+ paddress (call_site_gdbarch, call_site->pc),
+ (msym.minsym == NULL ? "???"
+ : SYMBOL_PRINT_NAME (msym.minsym)));
+
+ }
+ caller_arch = get_frame_arch (caller_frame);
+ caller_core_addr_type = builtin_type (caller_arch)->builtin_func_ptr;
+ val = dwarf2_evaluate_loc_desc (caller_core_addr_type, caller_frame,
+ dwarf_block->data, dwarf_block->size,
+ dwarf_block->per_cu);
+ /* DW_AT_GNU_call_site_target is a DWARF expression, not a DWARF
+ location. */
+ if (VALUE_LVAL (val) == lval_memory)
+ return value_address (val);
+ else
+ return value_as_address (val);
+ }
+
+ case FIELD_LOC_KIND_PHYSNAME:
+ {
+ const char *physname;
+ struct minimal_symbol *msym;
+
+ physname = FIELD_STATIC_PHYSNAME (call_site->target);
+
+ /* Handle both the mangled and demangled PHYSNAME. */
+ msym = lookup_minimal_symbol (physname, NULL, NULL);
+ if (msym == NULL)
+ {
+ msym = lookup_minimal_symbol_by_pc (call_site->pc - 1).minsym;
+ throw_error (NO_ENTRY_VALUE_ERROR,
+ _("Cannot find function \"%s\" for a call site target "
+ "at %s in %s"),
+ physname, paddress (call_site_gdbarch, call_site->pc),
+ msym == NULL ? "???" : SYMBOL_PRINT_NAME (msym));
+
+ }
+ return SYMBOL_VALUE_ADDRESS (msym);
+ }
+
+ case FIELD_LOC_KIND_PHYSADDR:
+ return FIELD_STATIC_PHYSADDR (call_site->target);
+
+ default:
+ internal_error (__FILE__, __LINE__, _("invalid call site target kind"));
+ }
+}
+
+/* Convert function entry point exact address ADDR to the function which is
+ compliant with TAIL_CALL_LIST_COMPLETE condition. Throw
+ NO_ENTRY_VALUE_ERROR otherwise. */
+
+static struct symbol *
+func_addr_to_tail_call_list (struct gdbarch *gdbarch, CORE_ADDR addr)
+{
+ struct symbol *sym = find_pc_function (addr);
+ struct type *type;
+
+ if (sym == NULL || BLOCK_START (SYMBOL_BLOCK_VALUE (sym)) != addr)
+ throw_error (NO_ENTRY_VALUE_ERROR,
+ _("DW_TAG_GNU_call_site resolving failed to find function "
+ "name for address %s"),
+ paddress (gdbarch, addr));
+
+ type = SYMBOL_TYPE (sym);
+ gdb_assert (TYPE_CODE (type) == TYPE_CODE_FUNC);
+ gdb_assert (TYPE_SPECIFIC_FIELD (type) == TYPE_SPECIFIC_FUNC);
+
+ return sym;
+}
+
+/* Verify function with entry point exact address ADDR can never call itself
+ via its tail calls (incl. transitively). Throw NO_ENTRY_VALUE_ERROR if it
+ can call itself via tail calls.
+
+ If a funtion can tail call itself its entry value based parameters are
+ unreliable. There is no verification whether the value of some/all
+ parameters is unchanged through the self tail call, we expect if there is
+ a self tail call all the parameters can be modified. */
+
+static void
+func_verify_no_selftailcall (struct gdbarch *gdbarch, CORE_ADDR verify_addr)
+{
+ struct obstack addr_obstack;
+ struct cleanup *old_chain;
+ CORE_ADDR addr;
+
+ /* Track here CORE_ADDRs which were already visited. */
+ htab_t addr_hash;
+
+ /* The verification is completely unordered. Track here function addresses
+ which still need to be iterated. */
+ VEC (CORE_ADDR) *todo = NULL;
+
+ obstack_init (&addr_obstack);
+ old_chain = make_cleanup_obstack_free (&addr_obstack);
+ addr_hash = htab_create_alloc_ex (64, core_addr_hash, core_addr_eq, NULL,
+ &addr_obstack, hashtab_obstack_allocate,
+ NULL);
+ make_cleanup_htab_delete (addr_hash);
+
+ make_cleanup (VEC_cleanup (CORE_ADDR), &todo);
+
+ VEC_safe_push (CORE_ADDR, todo, verify_addr);
+ while (!VEC_empty (CORE_ADDR, todo))
+ {
+ struct symbol *func_sym;
+ struct call_site *call_site;
+
+ addr = VEC_pop (CORE_ADDR, todo);
+
+ func_sym = func_addr_to_tail_call_list (gdbarch, addr);
+
+ for (call_site = TYPE_TAIL_CALL_LIST (SYMBOL_TYPE (func_sym));
+ call_site; call_site = call_site->tail_call_next)
+ {
+ CORE_ADDR target_addr;
+ void **slot;
+
+ /* CALLER_FRAME with registers is not available for tail-call jumped
+ frames. */
+ target_addr = call_site_to_target_addr (gdbarch, call_site, NULL);
+
+ if (target_addr == verify_addr)
+ {
+ struct bound_minimal_symbol msym;
+
+ msym = lookup_minimal_symbol_by_pc (verify_addr);
+ throw_error (NO_ENTRY_VALUE_ERROR,
+ _("DW_OP_GNU_entry_value resolving has found "
+ "function \"%s\" at %s can call itself via tail "
+ "calls"),
+ (msym.minsym == NULL ? "???"
+ : SYMBOL_PRINT_NAME (msym.minsym)),
+ paddress (gdbarch, verify_addr));
+ }
+
+ slot = htab_find_slot (addr_hash, &target_addr, INSERT);
+ if (*slot == NULL)
+ {
+ *slot = obstack_copy (&addr_obstack, &target_addr,
+ sizeof (target_addr));
+ VEC_safe_push (CORE_ADDR, todo, target_addr);
+ }
+ }
+ }
+
+ do_cleanups (old_chain);
+}
+
+/* Print user readable form of CALL_SITE->PC to gdb_stdlog. Used only for
+ ENTRY_VALUES_DEBUG. */
+
+static void
+tailcall_dump (struct gdbarch *gdbarch, const struct call_site *call_site)
+{
+ CORE_ADDR addr = call_site->pc;
+ struct bound_minimal_symbol msym = lookup_minimal_symbol_by_pc (addr - 1);
+
+ fprintf_unfiltered (gdb_stdlog, " %s(%s)", paddress (gdbarch, addr),
+ (msym.minsym == NULL ? "???"
+ : SYMBOL_PRINT_NAME (msym.minsym)));
+
+}
+
+/* vec.h needs single word type name, typedef it. */
+typedef struct call_site *call_sitep;
+
+/* Define VEC (call_sitep) functions. */
+DEF_VEC_P (call_sitep);
+
+/* Intersect RESULTP with CHAIN to keep RESULTP unambiguous, keep in RESULTP
+ only top callers and bottom callees which are present in both. GDBARCH is
+ used only for ENTRY_VALUES_DEBUG. RESULTP is NULL after return if there are
+ no remaining possibilities to provide unambiguous non-trivial result.
+ RESULTP should point to NULL on the first (initialization) call. Caller is
+ responsible for xfree of any RESULTP data. */
+
+static void
+chain_candidate (struct gdbarch *gdbarch, struct call_site_chain **resultp,
+ VEC (call_sitep) *chain)
+{
+ struct call_site_chain *result = *resultp;
+ long length = VEC_length (call_sitep, chain);
+ int callers, callees, idx;
+
+ if (result == NULL)
+ {
+ /* Create the initial chain containing all the passed PCs. */
+
+ result = xmalloc (sizeof (*result) + sizeof (*result->call_site)
+ * (length - 1));
+ result->length = length;
+ result->callers = result->callees = length;
+ if (!VEC_empty (call_sitep, chain))
+ memcpy (result->call_site, VEC_address (call_sitep, chain),
+ sizeof (*result->call_site) * length);
+ *resultp = result;
+
+ if (entry_values_debug)
+ {
+ fprintf_unfiltered (gdb_stdlog, "tailcall: initial:");
+ for (idx = 0; idx < length; idx++)
+ tailcall_dump (gdbarch, result->call_site[idx]);
+ fputc_unfiltered ('\n', gdb_stdlog);
+ }
+
+ return;
+ }
+
+ if (entry_values_debug)
+ {
+ fprintf_unfiltered (gdb_stdlog, "tailcall: compare:");
+ for (idx = 0; idx < length; idx++)
+ tailcall_dump (gdbarch, VEC_index (call_sitep, chain, idx));
+ fputc_unfiltered ('\n', gdb_stdlog);
+ }
+
+ /* Intersect callers. */
+
+ callers = min (result->callers, length);
+ for (idx = 0; idx < callers; idx++)
+ if (result->call_site[idx] != VEC_index (call_sitep, chain, idx))
+ {
+ result->callers = idx;
+ break;
+ }
+
+ /* Intersect callees. */
+
+ callees = min (result->callees, length);
+ for (idx = 0; idx < callees; idx++)
+ if (result->call_site[result->length - 1 - idx]
+ != VEC_index (call_sitep, chain, length - 1 - idx))
+ {
+ result->callees = idx;
+ break;
+ }
+
+ if (entry_values_debug)
+ {
+ fprintf_unfiltered (gdb_stdlog, "tailcall: reduced:");
+ for (idx = 0; idx < result->callers; idx++)
+ tailcall_dump (gdbarch, result->call_site[idx]);
+ fputs_unfiltered (" |", gdb_stdlog);
+ for (idx = 0; idx < result->callees; idx++)
+ tailcall_dump (gdbarch, result->call_site[result->length
+ - result->callees + idx]);
+ fputc_unfiltered ('\n', gdb_stdlog);
+ }
+
+ if (result->callers == 0 && result->callees == 0)
+ {
+ /* There are no common callers or callees. It could be also a direct
+ call (which has length 0) with ambiguous possibility of an indirect
+ call - CALLERS == CALLEES == 0 is valid during the first allocation
+ but any subsequence processing of such entry means ambiguity. */
+ xfree (result);
+ *resultp = NULL;
+ return;
+ }
+
+ /* See call_site_find_chain_1 why there is no way to reach the bottom callee
+ PC again. In such case there must be two different code paths to reach
+ it, therefore some of the former determined intermediate PCs must differ
+ and the unambiguous chain gets shortened. */
+ gdb_assert (result->callers + result->callees < result->length);
+}
+
+/* Create and return call_site_chain for CALLER_PC and CALLEE_PC. All the
+ assumed frames between them use GDBARCH. Use depth first search so we can
+ keep single CHAIN of call_site's back to CALLER_PC. Function recursion
+ would have needless GDB stack overhead. Caller is responsible for xfree of
+ the returned result. Any unreliability results in thrown
+ NO_ENTRY_VALUE_ERROR. */
+
+static struct call_site_chain *
+call_site_find_chain_1 (struct gdbarch *gdbarch, CORE_ADDR caller_pc,
+ CORE_ADDR callee_pc)
+{
+ CORE_ADDR save_callee_pc = callee_pc;
+ struct obstack addr_obstack;
+ struct cleanup *back_to_retval, *back_to_workdata;
+ struct call_site_chain *retval = NULL;
+ struct call_site *call_site;
+
+ /* Mark CALL_SITEs so we do not visit the same ones twice. */
+ htab_t addr_hash;
+
+ /* CHAIN contains only the intermediate CALL_SITEs. Neither CALLER_PC's
+ call_site nor any possible call_site at CALLEE_PC's function is there.
+ Any CALL_SITE in CHAIN will be iterated to its siblings - via
+ TAIL_CALL_NEXT. This is inappropriate for CALLER_PC's call_site. */
+ VEC (call_sitep) *chain = NULL;
+
+ /* We are not interested in the specific PC inside the callee function. */
+ callee_pc = get_pc_function_start (callee_pc);
+ if (callee_pc == 0)
+ throw_error (NO_ENTRY_VALUE_ERROR, _("Unable to find function for PC %s"),
+ paddress (gdbarch, save_callee_pc));
+
+ back_to_retval = make_cleanup (free_current_contents, &retval);
+
+ obstack_init (&addr_obstack);
+ back_to_workdata = make_cleanup_obstack_free (&addr_obstack);
+ addr_hash = htab_create_alloc_ex (64, core_addr_hash, core_addr_eq, NULL,
+ &addr_obstack, hashtab_obstack_allocate,
+ NULL);
+ make_cleanup_htab_delete (addr_hash);
+
+ make_cleanup (VEC_cleanup (call_sitep), &chain);
+
+ /* Do not push CALL_SITE to CHAIN. Push there only the first tail call site
+ at the target's function. All the possible tail call sites in the
+ target's function will get iterated as already pushed into CHAIN via their
+ TAIL_CALL_NEXT. */
+ call_site = call_site_for_pc (gdbarch, caller_pc);
+
+ while (call_site)
+ {
+ CORE_ADDR target_func_addr;
+ struct call_site *target_call_site;
+
+ /* CALLER_FRAME with registers is not available for tail-call jumped
+ frames. */
+ target_func_addr = call_site_to_target_addr (gdbarch, call_site, NULL);
+
+ if (target_func_addr == callee_pc)
+ {
+ chain_candidate (gdbarch, &retval, chain);
+ if (retval == NULL)
+ break;
+
+ /* There is no way to reach CALLEE_PC again as we would prevent
+ entering it twice as being already marked in ADDR_HASH. */
+ target_call_site = NULL;
+ }
+ else
+ {
+ struct symbol *target_func;
+
+ target_func = func_addr_to_tail_call_list (gdbarch, target_func_addr);
+ target_call_site = TYPE_TAIL_CALL_LIST (SYMBOL_TYPE (target_func));
+ }
+
+ do
+ {
+ /* Attempt to visit TARGET_CALL_SITE. */
+
+ if (target_call_site)
+ {
+ void **slot;
+
+ slot = htab_find_slot (addr_hash, &target_call_site->pc, INSERT);
+ if (*slot == NULL)
+ {
+ /* Successfully entered TARGET_CALL_SITE. */
+
+ *slot = &target_call_site->pc;
+ VEC_safe_push (call_sitep, chain, target_call_site);
+ break;
+ }
+ }
+
+ /* Backtrack (without revisiting the originating call_site). Try the
+ callers's sibling; if there isn't any try the callers's callers's
+ sibling etc. */
+
+ target_call_site = NULL;
+ while (!VEC_empty (call_sitep, chain))
+ {
+ call_site = VEC_pop (call_sitep, chain);
+
+ gdb_assert (htab_find_slot (addr_hash, &call_site->pc,
+ NO_INSERT) != NULL);
+ htab_remove_elt (addr_hash, &call_site->pc);
+
+ target_call_site = call_site->tail_call_next;
+ if (target_call_site)
+ break;
+ }
+ }
+ while (target_call_site);
+
+ if (VEC_empty (call_sitep, chain))
+ call_site = NULL;
+ else
+ call_site = VEC_last (call_sitep, chain);
+ }
+
+ if (retval == NULL)