convert to_set_circular_trace_buffer
[deliverable/binutils-gdb.git] / gdb / target.c
... / ...
CommitLineData
1/* Select target systems and architectures at runtime for GDB.
2
3 Copyright (C) 1990-2014 Free Software Foundation, Inc.
4
5 Contributed by Cygnus Support.
6
7 This file is part of GDB.
8
9 This program is free software; you can redistribute it and/or modify
10 it under the terms of the GNU General Public License as published by
11 the Free Software Foundation; either version 3 of the License, or
12 (at your option) any later version.
13
14 This program is distributed in the hope that it will be useful,
15 but WITHOUT ANY WARRANTY; without even the implied warranty of
16 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
17 GNU General Public License for more details.
18
19 You should have received a copy of the GNU General Public License
20 along with this program. If not, see <http://www.gnu.org/licenses/>. */
21
22#include "defs.h"
23#include <errno.h>
24#include <string.h>
25#include "target.h"
26#include "target-dcache.h"
27#include "gdbcmd.h"
28#include "symtab.h"
29#include "inferior.h"
30#include "bfd.h"
31#include "symfile.h"
32#include "objfiles.h"
33#include "dcache.h"
34#include <signal.h>
35#include "regcache.h"
36#include "gdb_assert.h"
37#include "gdbcore.h"
38#include "exceptions.h"
39#include "target-descriptions.h"
40#include "gdbthread.h"
41#include "solib.h"
42#include "exec.h"
43#include "inline-frame.h"
44#include "tracepoint.h"
45#include "gdb/fileio.h"
46#include "agent.h"
47
48static void target_info (char *, int);
49
50static void default_terminal_info (struct target_ops *, const char *, int);
51
52static int default_watchpoint_addr_within_range (struct target_ops *,
53 CORE_ADDR, CORE_ADDR, int);
54
55static int default_region_ok_for_hw_watchpoint (struct target_ops *,
56 CORE_ADDR, int);
57
58static void default_rcmd (struct target_ops *, char *, struct ui_file *);
59
60static ptid_t default_get_ada_task_ptid (struct target_ops *self,
61 long lwp, long tid);
62
63static void tcomplain (void) ATTRIBUTE_NORETURN;
64
65static int nomemory (CORE_ADDR, char *, int, int, struct target_ops *);
66
67static int return_zero (void);
68
69static void *return_null (void);
70
71void target_ignore (void);
72
73static void target_command (char *, int);
74
75static struct target_ops *find_default_run_target (char *);
76
77static target_xfer_partial_ftype default_xfer_partial;
78
79static struct gdbarch *default_thread_architecture (struct target_ops *ops,
80 ptid_t ptid);
81
82static int dummy_find_memory_regions (struct target_ops *self,
83 find_memory_region_ftype ignore1,
84 void *ignore2);
85
86static char *dummy_make_corefile_notes (struct target_ops *self,
87 bfd *ignore1, int *ignore2);
88
89static int find_default_can_async_p (struct target_ops *ignore);
90
91static int find_default_is_async_p (struct target_ops *ignore);
92
93static enum exec_direction_kind default_execution_direction
94 (struct target_ops *self);
95
96#include "target-delegates.c"
97
98static void init_dummy_target (void);
99
100static struct target_ops debug_target;
101
102static void debug_to_open (char *, int);
103
104static void debug_to_prepare_to_store (struct target_ops *self,
105 struct regcache *);
106
107static void debug_to_files_info (struct target_ops *);
108
109static int debug_to_insert_breakpoint (struct target_ops *, struct gdbarch *,
110 struct bp_target_info *);
111
112static int debug_to_remove_breakpoint (struct target_ops *, struct gdbarch *,
113 struct bp_target_info *);
114
115static int debug_to_can_use_hw_breakpoint (struct target_ops *self,
116 int, int, int);
117
118static int debug_to_insert_hw_breakpoint (struct target_ops *self,
119 struct gdbarch *,
120 struct bp_target_info *);
121
122static int debug_to_remove_hw_breakpoint (struct target_ops *self,
123 struct gdbarch *,
124 struct bp_target_info *);
125
126static int debug_to_insert_watchpoint (struct target_ops *self,
127 CORE_ADDR, int, int,
128 struct expression *);
129
130static int debug_to_remove_watchpoint (struct target_ops *self,
131 CORE_ADDR, int, int,
132 struct expression *);
133
134static int debug_to_stopped_data_address (struct target_ops *, CORE_ADDR *);
135
136static int debug_to_watchpoint_addr_within_range (struct target_ops *,
137 CORE_ADDR, CORE_ADDR, int);
138
139static int debug_to_region_ok_for_hw_watchpoint (struct target_ops *self,
140 CORE_ADDR, int);
141
142static int debug_to_can_accel_watchpoint_condition (struct target_ops *self,
143 CORE_ADDR, int, int,
144 struct expression *);
145
146static void debug_to_terminal_init (struct target_ops *self);
147
148static void debug_to_terminal_inferior (struct target_ops *self);
149
150static void debug_to_terminal_ours_for_output (struct target_ops *self);
151
152static void debug_to_terminal_save_ours (struct target_ops *self);
153
154static void debug_to_terminal_ours (struct target_ops *self);
155
156static void debug_to_load (struct target_ops *self, char *, int);
157
158static int debug_to_can_run (struct target_ops *self);
159
160static void debug_to_stop (struct target_ops *self, ptid_t);
161
162/* Pointer to array of target architecture structures; the size of the
163 array; the current index into the array; the allocated size of the
164 array. */
165struct target_ops **target_structs;
166unsigned target_struct_size;
167unsigned target_struct_allocsize;
168#define DEFAULT_ALLOCSIZE 10
169
170/* The initial current target, so that there is always a semi-valid
171 current target. */
172
173static struct target_ops dummy_target;
174
175/* Top of target stack. */
176
177static struct target_ops *target_stack;
178
179/* The target structure we are currently using to talk to a process
180 or file or whatever "inferior" we have. */
181
182struct target_ops current_target;
183
184/* Command list for target. */
185
186static struct cmd_list_element *targetlist = NULL;
187
188/* Nonzero if we should trust readonly sections from the
189 executable when reading memory. */
190
191static int trust_readonly = 0;
192
193/* Nonzero if we should show true memory content including
194 memory breakpoint inserted by gdb. */
195
196static int show_memory_breakpoints = 0;
197
198/* These globals control whether GDB attempts to perform these
199 operations; they are useful for targets that need to prevent
200 inadvertant disruption, such as in non-stop mode. */
201
202int may_write_registers = 1;
203
204int may_write_memory = 1;
205
206int may_insert_breakpoints = 1;
207
208int may_insert_tracepoints = 1;
209
210int may_insert_fast_tracepoints = 1;
211
212int may_stop = 1;
213
214/* Non-zero if we want to see trace of target level stuff. */
215
216static unsigned int targetdebug = 0;
217static void
218show_targetdebug (struct ui_file *file, int from_tty,
219 struct cmd_list_element *c, const char *value)
220{
221 fprintf_filtered (file, _("Target debugging is %s.\n"), value);
222}
223
224static void setup_target_debug (void);
225
226/* The user just typed 'target' without the name of a target. */
227
228static void
229target_command (char *arg, int from_tty)
230{
231 fputs_filtered ("Argument required (target name). Try `help target'\n",
232 gdb_stdout);
233}
234
235/* Default target_has_* methods for process_stratum targets. */
236
237int
238default_child_has_all_memory (struct target_ops *ops)
239{
240 /* If no inferior selected, then we can't read memory here. */
241 if (ptid_equal (inferior_ptid, null_ptid))
242 return 0;
243
244 return 1;
245}
246
247int
248default_child_has_memory (struct target_ops *ops)
249{
250 /* If no inferior selected, then we can't read memory here. */
251 if (ptid_equal (inferior_ptid, null_ptid))
252 return 0;
253
254 return 1;
255}
256
257int
258default_child_has_stack (struct target_ops *ops)
259{
260 /* If no inferior selected, there's no stack. */
261 if (ptid_equal (inferior_ptid, null_ptid))
262 return 0;
263
264 return 1;
265}
266
267int
268default_child_has_registers (struct target_ops *ops)
269{
270 /* Can't read registers from no inferior. */
271 if (ptid_equal (inferior_ptid, null_ptid))
272 return 0;
273
274 return 1;
275}
276
277int
278default_child_has_execution (struct target_ops *ops, ptid_t the_ptid)
279{
280 /* If there's no thread selected, then we can't make it run through
281 hoops. */
282 if (ptid_equal (the_ptid, null_ptid))
283 return 0;
284
285 return 1;
286}
287
288
289int
290target_has_all_memory_1 (void)
291{
292 struct target_ops *t;
293
294 for (t = current_target.beneath; t != NULL; t = t->beneath)
295 if (t->to_has_all_memory (t))
296 return 1;
297
298 return 0;
299}
300
301int
302target_has_memory_1 (void)
303{
304 struct target_ops *t;
305
306 for (t = current_target.beneath; t != NULL; t = t->beneath)
307 if (t->to_has_memory (t))
308 return 1;
309
310 return 0;
311}
312
313int
314target_has_stack_1 (void)
315{
316 struct target_ops *t;
317
318 for (t = current_target.beneath; t != NULL; t = t->beneath)
319 if (t->to_has_stack (t))
320 return 1;
321
322 return 0;
323}
324
325int
326target_has_registers_1 (void)
327{
328 struct target_ops *t;
329
330 for (t = current_target.beneath; t != NULL; t = t->beneath)
331 if (t->to_has_registers (t))
332 return 1;
333
334 return 0;
335}
336
337int
338target_has_execution_1 (ptid_t the_ptid)
339{
340 struct target_ops *t;
341
342 for (t = current_target.beneath; t != NULL; t = t->beneath)
343 if (t->to_has_execution (t, the_ptid))
344 return 1;
345
346 return 0;
347}
348
349int
350target_has_execution_current (void)
351{
352 return target_has_execution_1 (inferior_ptid);
353}
354
355/* Complete initialization of T. This ensures that various fields in
356 T are set, if needed by the target implementation. */
357
358void
359complete_target_initialization (struct target_ops *t)
360{
361 /* Provide default values for all "must have" methods. */
362 if (t->to_xfer_partial == NULL)
363 t->to_xfer_partial = default_xfer_partial;
364
365 if (t->to_has_all_memory == NULL)
366 t->to_has_all_memory = (int (*) (struct target_ops *)) return_zero;
367
368 if (t->to_has_memory == NULL)
369 t->to_has_memory = (int (*) (struct target_ops *)) return_zero;
370
371 if (t->to_has_stack == NULL)
372 t->to_has_stack = (int (*) (struct target_ops *)) return_zero;
373
374 if (t->to_has_registers == NULL)
375 t->to_has_registers = (int (*) (struct target_ops *)) return_zero;
376
377 if (t->to_has_execution == NULL)
378 t->to_has_execution = (int (*) (struct target_ops *, ptid_t)) return_zero;
379
380 install_delegators (t);
381}
382
383/* Add possible target architecture T to the list and add a new
384 command 'target T->to_shortname'. Set COMPLETER as the command's
385 completer if not NULL. */
386
387void
388add_target_with_completer (struct target_ops *t,
389 completer_ftype *completer)
390{
391 struct cmd_list_element *c;
392
393 complete_target_initialization (t);
394
395 if (!target_structs)
396 {
397 target_struct_allocsize = DEFAULT_ALLOCSIZE;
398 target_structs = (struct target_ops **) xmalloc
399 (target_struct_allocsize * sizeof (*target_structs));
400 }
401 if (target_struct_size >= target_struct_allocsize)
402 {
403 target_struct_allocsize *= 2;
404 target_structs = (struct target_ops **)
405 xrealloc ((char *) target_structs,
406 target_struct_allocsize * sizeof (*target_structs));
407 }
408 target_structs[target_struct_size++] = t;
409
410 if (targetlist == NULL)
411 add_prefix_cmd ("target", class_run, target_command, _("\
412Connect to a target machine or process.\n\
413The first argument is the type or protocol of the target machine.\n\
414Remaining arguments are interpreted by the target protocol. For more\n\
415information on the arguments for a particular protocol, type\n\
416`help target ' followed by the protocol name."),
417 &targetlist, "target ", 0, &cmdlist);
418 c = add_cmd (t->to_shortname, no_class, t->to_open, t->to_doc,
419 &targetlist);
420 if (completer != NULL)
421 set_cmd_completer (c, completer);
422}
423
424/* Add a possible target architecture to the list. */
425
426void
427add_target (struct target_ops *t)
428{
429 add_target_with_completer (t, NULL);
430}
431
432/* See target.h. */
433
434void
435add_deprecated_target_alias (struct target_ops *t, char *alias)
436{
437 struct cmd_list_element *c;
438 char *alt;
439
440 /* If we use add_alias_cmd, here, we do not get the deprecated warning,
441 see PR cli/15104. */
442 c = add_cmd (alias, no_class, t->to_open, t->to_doc, &targetlist);
443 alt = xstrprintf ("target %s", t->to_shortname);
444 deprecate_cmd (c, alt);
445}
446
447/* Stub functions */
448
449void
450target_ignore (void)
451{
452}
453
454void
455target_kill (void)
456{
457 struct target_ops *t;
458
459 for (t = current_target.beneath; t != NULL; t = t->beneath)
460 if (t->to_kill != NULL)
461 {
462 if (targetdebug)
463 fprintf_unfiltered (gdb_stdlog, "target_kill ()\n");
464
465 t->to_kill (t);
466 return;
467 }
468
469 noprocess ();
470}
471
472void
473target_load (char *arg, int from_tty)
474{
475 target_dcache_invalidate ();
476 (*current_target.to_load) (&current_target, arg, from_tty);
477}
478
479void
480target_create_inferior (char *exec_file, char *args,
481 char **env, int from_tty)
482{
483 struct target_ops *t;
484
485 for (t = current_target.beneath; t != NULL; t = t->beneath)
486 {
487 if (t->to_create_inferior != NULL)
488 {
489 t->to_create_inferior (t, exec_file, args, env, from_tty);
490 if (targetdebug)
491 fprintf_unfiltered (gdb_stdlog,
492 "target_create_inferior (%s, %s, xxx, %d)\n",
493 exec_file, args, from_tty);
494 return;
495 }
496 }
497
498 internal_error (__FILE__, __LINE__,
499 _("could not find a target to create inferior"));
500}
501
502void
503target_terminal_inferior (void)
504{
505 /* A background resume (``run&'') should leave GDB in control of the
506 terminal. Use target_can_async_p, not target_is_async_p, since at
507 this point the target is not async yet. However, if sync_execution
508 is not set, we know it will become async prior to resume. */
509 if (target_can_async_p () && !sync_execution)
510 return;
511
512 /* If GDB is resuming the inferior in the foreground, install
513 inferior's terminal modes. */
514 (*current_target.to_terminal_inferior) (&current_target);
515}
516
517static int
518nomemory (CORE_ADDR memaddr, char *myaddr, int len, int write,
519 struct target_ops *t)
520{
521 errno = EIO; /* Can't read/write this location. */
522 return 0; /* No bytes handled. */
523}
524
525static void
526tcomplain (void)
527{
528 error (_("You can't do that when your target is `%s'"),
529 current_target.to_shortname);
530}
531
532void
533noprocess (void)
534{
535 error (_("You can't do that without a process to debug."));
536}
537
538static void
539default_terminal_info (struct target_ops *self, const char *args, int from_tty)
540{
541 printf_unfiltered (_("No saved terminal information.\n"));
542}
543
544/* A default implementation for the to_get_ada_task_ptid target method.
545
546 This function builds the PTID by using both LWP and TID as part of
547 the PTID lwp and tid elements. The pid used is the pid of the
548 inferior_ptid. */
549
550static ptid_t
551default_get_ada_task_ptid (struct target_ops *self, long lwp, long tid)
552{
553 return ptid_build (ptid_get_pid (inferior_ptid), lwp, tid);
554}
555
556static enum exec_direction_kind
557default_execution_direction (struct target_ops *self)
558{
559 if (!target_can_execute_reverse)
560 return EXEC_FORWARD;
561 else if (!target_can_async_p ())
562 return EXEC_FORWARD;
563 else
564 gdb_assert_not_reached ("\
565to_execution_direction must be implemented for reverse async");
566}
567
568/* Go through the target stack from top to bottom, copying over zero
569 entries in current_target, then filling in still empty entries. In
570 effect, we are doing class inheritance through the pushed target
571 vectors.
572
573 NOTE: cagney/2003-10-17: The problem with this inheritance, as it
574 is currently implemented, is that it discards any knowledge of
575 which target an inherited method originally belonged to.
576 Consequently, new new target methods should instead explicitly and
577 locally search the target stack for the target that can handle the
578 request. */
579
580static void
581update_current_target (void)
582{
583 struct target_ops *t;
584
585 /* First, reset current's contents. */
586 memset (&current_target, 0, sizeof (current_target));
587
588 /* Install the delegators. */
589 install_delegators (&current_target);
590
591#define INHERIT(FIELD, TARGET) \
592 if (!current_target.FIELD) \
593 current_target.FIELD = (TARGET)->FIELD
594
595 for (t = target_stack; t; t = t->beneath)
596 {
597 INHERIT (to_shortname, t);
598 INHERIT (to_longname, t);
599 INHERIT (to_doc, t);
600 /* Do not inherit to_open. */
601 /* Do not inherit to_close. */
602 /* Do not inherit to_attach. */
603 /* Do not inherit to_post_attach. */
604 INHERIT (to_attach_no_wait, t);
605 /* Do not inherit to_detach. */
606 /* Do not inherit to_disconnect. */
607 /* Do not inherit to_resume. */
608 /* Do not inherit to_wait. */
609 /* Do not inherit to_fetch_registers. */
610 /* Do not inherit to_store_registers. */
611 /* Do not inherit to_prepare_to_store. */
612 INHERIT (deprecated_xfer_memory, t);
613 /* Do not inherit to_files_info. */
614 /* Do not inherit to_insert_breakpoint. */
615 /* Do not inherit to_remove_breakpoint. */
616 /* Do not inherit to_can_use_hw_breakpoint. */
617 /* Do not inherit to_insert_hw_breakpoint. */
618 /* Do not inherit to_remove_hw_breakpoint. */
619 /* Do not inherit to_ranged_break_num_registers. */
620 /* Do not inherit to_insert_watchpoint. */
621 /* Do not inherit to_remove_watchpoint. */
622 /* Do not inherit to_insert_mask_watchpoint. */
623 /* Do not inherit to_remove_mask_watchpoint. */
624 /* Do not inherit to_stopped_data_address. */
625 INHERIT (to_have_steppable_watchpoint, t);
626 INHERIT (to_have_continuable_watchpoint, t);
627 /* Do not inherit to_stopped_by_watchpoint. */
628 /* Do not inherit to_watchpoint_addr_within_range. */
629 /* Do not inherit to_region_ok_for_hw_watchpoint. */
630 /* Do not inherit to_can_accel_watchpoint_condition. */
631 /* Do not inherit to_masked_watch_num_registers. */
632 /* Do not inherit to_terminal_init. */
633 /* Do not inherit to_terminal_inferior. */
634 /* Do not inherit to_terminal_ours_for_output. */
635 /* Do not inherit to_terminal_ours. */
636 /* Do not inherit to_terminal_save_ours. */
637 /* Do not inherit to_terminal_info. */
638 /* Do not inherit to_kill. */
639 /* Do not inherit to_load. */
640 /* Do no inherit to_create_inferior. */
641 /* Do not inherit to_post_startup_inferior. */
642 /* Do not inherit to_insert_fork_catchpoint. */
643 /* Do not inherit to_remove_fork_catchpoint. */
644 /* Do not inherit to_insert_vfork_catchpoint. */
645 /* Do not inherit to_remove_vfork_catchpoint. */
646 /* Do not inherit to_follow_fork. */
647 /* Do not inherit to_insert_exec_catchpoint. */
648 /* Do not inherit to_remove_exec_catchpoint. */
649 /* Do not inherit to_set_syscall_catchpoint. */
650 /* Do not inherit to_has_exited. */
651 /* Do not inherit to_mourn_inferior. */
652 INHERIT (to_can_run, t);
653 /* Do not inherit to_pass_signals. */
654 /* Do not inherit to_program_signals. */
655 /* Do not inherit to_thread_alive. */
656 /* Do not inherit to_find_new_threads. */
657 /* Do not inherit to_pid_to_str. */
658 /* Do not inherit to_extra_thread_info. */
659 /* Do not inherit to_thread_name. */
660 INHERIT (to_stop, t);
661 /* Do not inherit to_xfer_partial. */
662 /* Do not inherit to_rcmd. */
663 /* Do not inherit to_pid_to_exec_file. */
664 /* Do not inherit to_log_command. */
665 INHERIT (to_stratum, t);
666 /* Do not inherit to_has_all_memory. */
667 /* Do not inherit to_has_memory. */
668 /* Do not inherit to_has_stack. */
669 /* Do not inherit to_has_registers. */
670 /* Do not inherit to_has_execution. */
671 INHERIT (to_has_thread_control, t);
672 /* Do not inherit to_can_async_p. */
673 /* Do not inherit to_is_async_p. */
674 /* Do not inherit to_async. */
675 /* Do not inherit to_find_memory_regions. */
676 /* Do not inherit to_make_corefile_notes. */
677 /* Do not inherit to_get_bookmark. */
678 /* Do not inherit to_goto_bookmark. */
679 /* Do not inherit to_get_thread_local_address. */
680 /* Do not inherit to_can_execute_reverse. */
681 /* Do not inherit to_execution_direction. */
682 /* Do not inherit to_thread_architecture. */
683 /* Do not inherit to_read_description. */
684 /* Do not inherit to_get_ada_task_ptid. */
685 /* Do not inherit to_search_memory. */
686 /* Do not inherit to_supports_multi_process. */
687 /* Do not inherit to_supports_enable_disable_tracepoint. */
688 /* Do not inherit to_supports_string_tracing. */
689 /* Do not inherit to_trace_init. */
690 /* Do not inherit to_download_tracepoint. */
691 /* Do not inherit to_can_download_tracepoint. */
692 /* Do not inherit to_download_trace_state_variable. */
693 /* Do not inherit to_enable_tracepoint. */
694 /* Do not inherit to_disable_tracepoint. */
695 /* Do not inherit to_trace_set_readonly_regions. */
696 /* Do not inherit to_trace_start. */
697 /* Do not inherit to_get_trace_status. */
698 /* Do not inherit to_get_tracepoint_status. */
699 /* Do not inherit to_trace_stop. */
700 /* Do not inherit to_trace_find. */
701 /* Do not inherit to_get_trace_state_variable_value. */
702 /* Do not inherit to_save_trace_data. */
703 /* Do not inherit to_upload_tracepoints. */
704 /* Do not inherit to_upload_trace_state_variables. */
705 /* Do not inherit to_get_raw_trace_data. */
706 /* Do not inherit to_get_min_fast_tracepoint_insn_len. */
707 /* Do not inherit to_set_disconnected_tracing. */
708 /* Do not inherit to_set_circular_trace_buffer. */
709 INHERIT (to_set_trace_buffer_size, t);
710 INHERIT (to_set_trace_notes, t);
711 INHERIT (to_get_tib_address, t);
712 INHERIT (to_set_permissions, t);
713 INHERIT (to_static_tracepoint_marker_at, t);
714 INHERIT (to_static_tracepoint_markers_by_strid, t);
715 INHERIT (to_traceframe_info, t);
716 INHERIT (to_use_agent, t);
717 INHERIT (to_can_use_agent, t);
718 INHERIT (to_augmented_libraries_svr4_read, t);
719 INHERIT (to_magic, t);
720 INHERIT (to_supports_evaluation_of_breakpoint_conditions, t);
721 INHERIT (to_can_run_breakpoint_commands, t);
722 /* Do not inherit to_memory_map. */
723 /* Do not inherit to_flash_erase. */
724 /* Do not inherit to_flash_done. */
725 }
726#undef INHERIT
727
728 /* Clean up a target struct so it no longer has any zero pointers in
729 it. Some entries are defaulted to a method that print an error,
730 others are hard-wired to a standard recursive default. */
731
732#define de_fault(field, value) \
733 if (!current_target.field) \
734 current_target.field = value
735
736 de_fault (to_open,
737 (void (*) (char *, int))
738 tcomplain);
739 de_fault (to_close,
740 (void (*) (struct target_ops *))
741 target_ignore);
742 de_fault (deprecated_xfer_memory,
743 (int (*) (CORE_ADDR, gdb_byte *, int, int,
744 struct mem_attrib *, struct target_ops *))
745 nomemory);
746 de_fault (to_can_run,
747 (int (*) (struct target_ops *))
748 return_zero);
749 de_fault (to_stop,
750 (void (*) (struct target_ops *, ptid_t))
751 target_ignore);
752 current_target.to_read_description = NULL;
753 de_fault (to_set_trace_buffer_size,
754 (void (*) (struct target_ops *, LONGEST))
755 target_ignore);
756 de_fault (to_set_trace_notes,
757 (int (*) (struct target_ops *,
758 const char *, const char *, const char *))
759 return_zero);
760 de_fault (to_get_tib_address,
761 (int (*) (struct target_ops *, ptid_t, CORE_ADDR *))
762 tcomplain);
763 de_fault (to_set_permissions,
764 (void (*) (struct target_ops *))
765 target_ignore);
766 de_fault (to_static_tracepoint_marker_at,
767 (int (*) (struct target_ops *,
768 CORE_ADDR, struct static_tracepoint_marker *))
769 return_zero);
770 de_fault (to_static_tracepoint_markers_by_strid,
771 (VEC(static_tracepoint_marker_p) * (*) (struct target_ops *,
772 const char *))
773 tcomplain);
774 de_fault (to_traceframe_info,
775 (struct traceframe_info * (*) (struct target_ops *))
776 return_null);
777 de_fault (to_supports_evaluation_of_breakpoint_conditions,
778 (int (*) (struct target_ops *))
779 return_zero);
780 de_fault (to_can_run_breakpoint_commands,
781 (int (*) (struct target_ops *))
782 return_zero);
783 de_fault (to_use_agent,
784 (int (*) (struct target_ops *, int))
785 tcomplain);
786 de_fault (to_can_use_agent,
787 (int (*) (struct target_ops *))
788 return_zero);
789 de_fault (to_augmented_libraries_svr4_read,
790 (int (*) (struct target_ops *))
791 return_zero);
792
793#undef de_fault
794
795 /* Finally, position the target-stack beneath the squashed
796 "current_target". That way code looking for a non-inherited
797 target method can quickly and simply find it. */
798 current_target.beneath = target_stack;
799
800 if (targetdebug)
801 setup_target_debug ();
802}
803
804/* Push a new target type into the stack of the existing target accessors,
805 possibly superseding some of the existing accessors.
806
807 Rather than allow an empty stack, we always have the dummy target at
808 the bottom stratum, so we can call the function vectors without
809 checking them. */
810
811void
812push_target (struct target_ops *t)
813{
814 struct target_ops **cur;
815
816 /* Check magic number. If wrong, it probably means someone changed
817 the struct definition, but not all the places that initialize one. */
818 if (t->to_magic != OPS_MAGIC)
819 {
820 fprintf_unfiltered (gdb_stderr,
821 "Magic number of %s target struct wrong\n",
822 t->to_shortname);
823 internal_error (__FILE__, __LINE__,
824 _("failed internal consistency check"));
825 }
826
827 /* Find the proper stratum to install this target in. */
828 for (cur = &target_stack; (*cur) != NULL; cur = &(*cur)->beneath)
829 {
830 if ((int) (t->to_stratum) >= (int) (*cur)->to_stratum)
831 break;
832 }
833
834 /* If there's already targets at this stratum, remove them. */
835 /* FIXME: cagney/2003-10-15: I think this should be popping all
836 targets to CUR, and not just those at this stratum level. */
837 while ((*cur) != NULL && t->to_stratum == (*cur)->to_stratum)
838 {
839 /* There's already something at this stratum level. Close it,
840 and un-hook it from the stack. */
841 struct target_ops *tmp = (*cur);
842
843 (*cur) = (*cur)->beneath;
844 tmp->beneath = NULL;
845 target_close (tmp);
846 }
847
848 /* We have removed all targets in our stratum, now add the new one. */
849 t->beneath = (*cur);
850 (*cur) = t;
851
852 update_current_target ();
853}
854
855/* Remove a target_ops vector from the stack, wherever it may be.
856 Return how many times it was removed (0 or 1). */
857
858int
859unpush_target (struct target_ops *t)
860{
861 struct target_ops **cur;
862 struct target_ops *tmp;
863
864 if (t->to_stratum == dummy_stratum)
865 internal_error (__FILE__, __LINE__,
866 _("Attempt to unpush the dummy target"));
867
868 /* Look for the specified target. Note that we assume that a target
869 can only occur once in the target stack. */
870
871 for (cur = &target_stack; (*cur) != NULL; cur = &(*cur)->beneath)
872 {
873 if ((*cur) == t)
874 break;
875 }
876
877 /* If we don't find target_ops, quit. Only open targets should be
878 closed. */
879 if ((*cur) == NULL)
880 return 0;
881
882 /* Unchain the target. */
883 tmp = (*cur);
884 (*cur) = (*cur)->beneath;
885 tmp->beneath = NULL;
886
887 update_current_target ();
888
889 /* Finally close the target. Note we do this after unchaining, so
890 any target method calls from within the target_close
891 implementation don't end up in T anymore. */
892 target_close (t);
893
894 return 1;
895}
896
897void
898pop_all_targets_above (enum strata above_stratum)
899{
900 while ((int) (current_target.to_stratum) > (int) above_stratum)
901 {
902 if (!unpush_target (target_stack))
903 {
904 fprintf_unfiltered (gdb_stderr,
905 "pop_all_targets couldn't find target %s\n",
906 target_stack->to_shortname);
907 internal_error (__FILE__, __LINE__,
908 _("failed internal consistency check"));
909 break;
910 }
911 }
912}
913
914void
915pop_all_targets (void)
916{
917 pop_all_targets_above (dummy_stratum);
918}
919
920/* Return 1 if T is now pushed in the target stack. Return 0 otherwise. */
921
922int
923target_is_pushed (struct target_ops *t)
924{
925 struct target_ops **cur;
926
927 /* Check magic number. If wrong, it probably means someone changed
928 the struct definition, but not all the places that initialize one. */
929 if (t->to_magic != OPS_MAGIC)
930 {
931 fprintf_unfiltered (gdb_stderr,
932 "Magic number of %s target struct wrong\n",
933 t->to_shortname);
934 internal_error (__FILE__, __LINE__,
935 _("failed internal consistency check"));
936 }
937
938 for (cur = &target_stack; (*cur) != NULL; cur = &(*cur)->beneath)
939 if (*cur == t)
940 return 1;
941
942 return 0;
943}
944
945/* Using the objfile specified in OBJFILE, find the address for the
946 current thread's thread-local storage with offset OFFSET. */
947CORE_ADDR
948target_translate_tls_address (struct objfile *objfile, CORE_ADDR offset)
949{
950 volatile CORE_ADDR addr = 0;
951 struct target_ops *target;
952
953 for (target = current_target.beneath;
954 target != NULL;
955 target = target->beneath)
956 {
957 if (target->to_get_thread_local_address != NULL)
958 break;
959 }
960
961 if (target != NULL
962 && gdbarch_fetch_tls_load_module_address_p (target_gdbarch ()))
963 {
964 ptid_t ptid = inferior_ptid;
965 volatile struct gdb_exception ex;
966
967 TRY_CATCH (ex, RETURN_MASK_ALL)
968 {
969 CORE_ADDR lm_addr;
970
971 /* Fetch the load module address for this objfile. */
972 lm_addr = gdbarch_fetch_tls_load_module_address (target_gdbarch (),
973 objfile);
974 /* If it's 0, throw the appropriate exception. */
975 if (lm_addr == 0)
976 throw_error (TLS_LOAD_MODULE_NOT_FOUND_ERROR,
977 _("TLS load module not found"));
978
979 addr = target->to_get_thread_local_address (target, ptid,
980 lm_addr, offset);
981 }
982 /* If an error occurred, print TLS related messages here. Otherwise,
983 throw the error to some higher catcher. */
984 if (ex.reason < 0)
985 {
986 int objfile_is_library = (objfile->flags & OBJF_SHARED);
987
988 switch (ex.error)
989 {
990 case TLS_NO_LIBRARY_SUPPORT_ERROR:
991 error (_("Cannot find thread-local variables "
992 "in this thread library."));
993 break;
994 case TLS_LOAD_MODULE_NOT_FOUND_ERROR:
995 if (objfile_is_library)
996 error (_("Cannot find shared library `%s' in dynamic"
997 " linker's load module list"), objfile_name (objfile));
998 else
999 error (_("Cannot find executable file `%s' in dynamic"
1000 " linker's load module list"), objfile_name (objfile));
1001 break;
1002 case TLS_NOT_ALLOCATED_YET_ERROR:
1003 if (objfile_is_library)
1004 error (_("The inferior has not yet allocated storage for"
1005 " thread-local variables in\n"
1006 "the shared library `%s'\n"
1007 "for %s"),
1008 objfile_name (objfile), target_pid_to_str (ptid));
1009 else
1010 error (_("The inferior has not yet allocated storage for"
1011 " thread-local variables in\n"
1012 "the executable `%s'\n"
1013 "for %s"),
1014 objfile_name (objfile), target_pid_to_str (ptid));
1015 break;
1016 case TLS_GENERIC_ERROR:
1017 if (objfile_is_library)
1018 error (_("Cannot find thread-local storage for %s, "
1019 "shared library %s:\n%s"),
1020 target_pid_to_str (ptid),
1021 objfile_name (objfile), ex.message);
1022 else
1023 error (_("Cannot find thread-local storage for %s, "
1024 "executable file %s:\n%s"),
1025 target_pid_to_str (ptid),
1026 objfile_name (objfile), ex.message);
1027 break;
1028 default:
1029 throw_exception (ex);
1030 break;
1031 }
1032 }
1033 }
1034 /* It wouldn't be wrong here to try a gdbarch method, too; finding
1035 TLS is an ABI-specific thing. But we don't do that yet. */
1036 else
1037 error (_("Cannot find thread-local variables on this target"));
1038
1039 return addr;
1040}
1041
1042const char *
1043target_xfer_status_to_string (enum target_xfer_status err)
1044{
1045#define CASE(X) case X: return #X
1046 switch (err)
1047 {
1048 CASE(TARGET_XFER_E_IO);
1049 CASE(TARGET_XFER_E_UNAVAILABLE);
1050 default:
1051 return "<unknown>";
1052 }
1053#undef CASE
1054};
1055
1056
1057#undef MIN
1058#define MIN(A, B) (((A) <= (B)) ? (A) : (B))
1059
1060/* target_read_string -- read a null terminated string, up to LEN bytes,
1061 from MEMADDR in target. Set *ERRNOP to the errno code, or 0 if successful.
1062 Set *STRING to a pointer to malloc'd memory containing the data; the caller
1063 is responsible for freeing it. Return the number of bytes successfully
1064 read. */
1065
1066int
1067target_read_string (CORE_ADDR memaddr, char **string, int len, int *errnop)
1068{
1069 int tlen, offset, i;
1070 gdb_byte buf[4];
1071 int errcode = 0;
1072 char *buffer;
1073 int buffer_allocated;
1074 char *bufptr;
1075 unsigned int nbytes_read = 0;
1076
1077 gdb_assert (string);
1078
1079 /* Small for testing. */
1080 buffer_allocated = 4;
1081 buffer = xmalloc (buffer_allocated);
1082 bufptr = buffer;
1083
1084 while (len > 0)
1085 {
1086 tlen = MIN (len, 4 - (memaddr & 3));
1087 offset = memaddr & 3;
1088
1089 errcode = target_read_memory (memaddr & ~3, buf, sizeof buf);
1090 if (errcode != 0)
1091 {
1092 /* The transfer request might have crossed the boundary to an
1093 unallocated region of memory. Retry the transfer, requesting
1094 a single byte. */
1095 tlen = 1;
1096 offset = 0;
1097 errcode = target_read_memory (memaddr, buf, 1);
1098 if (errcode != 0)
1099 goto done;
1100 }
1101
1102 if (bufptr - buffer + tlen > buffer_allocated)
1103 {
1104 unsigned int bytes;
1105
1106 bytes = bufptr - buffer;
1107 buffer_allocated *= 2;
1108 buffer = xrealloc (buffer, buffer_allocated);
1109 bufptr = buffer + bytes;
1110 }
1111
1112 for (i = 0; i < tlen; i++)
1113 {
1114 *bufptr++ = buf[i + offset];
1115 if (buf[i + offset] == '\000')
1116 {
1117 nbytes_read += i + 1;
1118 goto done;
1119 }
1120 }
1121
1122 memaddr += tlen;
1123 len -= tlen;
1124 nbytes_read += tlen;
1125 }
1126done:
1127 *string = buffer;
1128 if (errnop != NULL)
1129 *errnop = errcode;
1130 return nbytes_read;
1131}
1132
1133struct target_section_table *
1134target_get_section_table (struct target_ops *target)
1135{
1136 struct target_ops *t;
1137
1138 if (targetdebug)
1139 fprintf_unfiltered (gdb_stdlog, "target_get_section_table ()\n");
1140
1141 for (t = target; t != NULL; t = t->beneath)
1142 if (t->to_get_section_table != NULL)
1143 return (*t->to_get_section_table) (t);
1144
1145 return NULL;
1146}
1147
1148/* Find a section containing ADDR. */
1149
1150struct target_section *
1151target_section_by_addr (struct target_ops *target, CORE_ADDR addr)
1152{
1153 struct target_section_table *table = target_get_section_table (target);
1154 struct target_section *secp;
1155
1156 if (table == NULL)
1157 return NULL;
1158
1159 for (secp = table->sections; secp < table->sections_end; secp++)
1160 {
1161 if (addr >= secp->addr && addr < secp->endaddr)
1162 return secp;
1163 }
1164 return NULL;
1165}
1166
1167/* Read memory from the live target, even if currently inspecting a
1168 traceframe. The return is the same as that of target_read. */
1169
1170static enum target_xfer_status
1171target_read_live_memory (enum target_object object,
1172 ULONGEST memaddr, gdb_byte *myaddr, ULONGEST len,
1173 ULONGEST *xfered_len)
1174{
1175 enum target_xfer_status ret;
1176 struct cleanup *cleanup;
1177
1178 /* Switch momentarily out of tfind mode so to access live memory.
1179 Note that this must not clear global state, such as the frame
1180 cache, which must still remain valid for the previous traceframe.
1181 We may be _building_ the frame cache at this point. */
1182 cleanup = make_cleanup_restore_traceframe_number ();
1183 set_traceframe_number (-1);
1184
1185 ret = target_xfer_partial (current_target.beneath, object, NULL,
1186 myaddr, NULL, memaddr, len, xfered_len);
1187
1188 do_cleanups (cleanup);
1189 return ret;
1190}
1191
1192/* Using the set of read-only target sections of OPS, read live
1193 read-only memory. Note that the actual reads start from the
1194 top-most target again.
1195
1196 For interface/parameters/return description see target.h,
1197 to_xfer_partial. */
1198
1199static enum target_xfer_status
1200memory_xfer_live_readonly_partial (struct target_ops *ops,
1201 enum target_object object,
1202 gdb_byte *readbuf, ULONGEST memaddr,
1203 ULONGEST len, ULONGEST *xfered_len)
1204{
1205 struct target_section *secp;
1206 struct target_section_table *table;
1207
1208 secp = target_section_by_addr (ops, memaddr);
1209 if (secp != NULL
1210 && (bfd_get_section_flags (secp->the_bfd_section->owner,
1211 secp->the_bfd_section)
1212 & SEC_READONLY))
1213 {
1214 struct target_section *p;
1215 ULONGEST memend = memaddr + len;
1216
1217 table = target_get_section_table (ops);
1218
1219 for (p = table->sections; p < table->sections_end; p++)
1220 {
1221 if (memaddr >= p->addr)
1222 {
1223 if (memend <= p->endaddr)
1224 {
1225 /* Entire transfer is within this section. */
1226 return target_read_live_memory (object, memaddr,
1227 readbuf, len, xfered_len);
1228 }
1229 else if (memaddr >= p->endaddr)
1230 {
1231 /* This section ends before the transfer starts. */
1232 continue;
1233 }
1234 else
1235 {
1236 /* This section overlaps the transfer. Just do half. */
1237 len = p->endaddr - memaddr;
1238 return target_read_live_memory (object, memaddr,
1239 readbuf, len, xfered_len);
1240 }
1241 }
1242 }
1243 }
1244
1245 return TARGET_XFER_EOF;
1246}
1247
1248/* Read memory from more than one valid target. A core file, for
1249 instance, could have some of memory but delegate other bits to
1250 the target below it. So, we must manually try all targets. */
1251
1252static enum target_xfer_status
1253raw_memory_xfer_partial (struct target_ops *ops, gdb_byte *readbuf,
1254 const gdb_byte *writebuf, ULONGEST memaddr, LONGEST len,
1255 ULONGEST *xfered_len)
1256{
1257 enum target_xfer_status res;
1258
1259 do
1260 {
1261 res = ops->to_xfer_partial (ops, TARGET_OBJECT_MEMORY, NULL,
1262 readbuf, writebuf, memaddr, len,
1263 xfered_len);
1264 if (res == TARGET_XFER_OK)
1265 break;
1266
1267 /* Stop if the target reports that the memory is not available. */
1268 if (res == TARGET_XFER_E_UNAVAILABLE)
1269 break;
1270
1271 /* We want to continue past core files to executables, but not
1272 past a running target's memory. */
1273 if (ops->to_has_all_memory (ops))
1274 break;
1275
1276 ops = ops->beneath;
1277 }
1278 while (ops != NULL);
1279
1280 return res;
1281}
1282
1283/* Perform a partial memory transfer.
1284 For docs see target.h, to_xfer_partial. */
1285
1286static enum target_xfer_status
1287memory_xfer_partial_1 (struct target_ops *ops, enum target_object object,
1288 gdb_byte *readbuf, const gdb_byte *writebuf, ULONGEST memaddr,
1289 ULONGEST len, ULONGEST *xfered_len)
1290{
1291 enum target_xfer_status res;
1292 int reg_len;
1293 struct mem_region *region;
1294 struct inferior *inf;
1295
1296 /* For accesses to unmapped overlay sections, read directly from
1297 files. Must do this first, as MEMADDR may need adjustment. */
1298 if (readbuf != NULL && overlay_debugging)
1299 {
1300 struct obj_section *section = find_pc_overlay (memaddr);
1301
1302 if (pc_in_unmapped_range (memaddr, section))
1303 {
1304 struct target_section_table *table
1305 = target_get_section_table (ops);
1306 const char *section_name = section->the_bfd_section->name;
1307
1308 memaddr = overlay_mapped_address (memaddr, section);
1309 return section_table_xfer_memory_partial (readbuf, writebuf,
1310 memaddr, len, xfered_len,
1311 table->sections,
1312 table->sections_end,
1313 section_name);
1314 }
1315 }
1316
1317 /* Try the executable files, if "trust-readonly-sections" is set. */
1318 if (readbuf != NULL && trust_readonly)
1319 {
1320 struct target_section *secp;
1321 struct target_section_table *table;
1322
1323 secp = target_section_by_addr (ops, memaddr);
1324 if (secp != NULL
1325 && (bfd_get_section_flags (secp->the_bfd_section->owner,
1326 secp->the_bfd_section)
1327 & SEC_READONLY))
1328 {
1329 table = target_get_section_table (ops);
1330 return section_table_xfer_memory_partial (readbuf, writebuf,
1331 memaddr, len, xfered_len,
1332 table->sections,
1333 table->sections_end,
1334 NULL);
1335 }
1336 }
1337
1338 /* If reading unavailable memory in the context of traceframes, and
1339 this address falls within a read-only section, fallback to
1340 reading from live memory. */
1341 if (readbuf != NULL && get_traceframe_number () != -1)
1342 {
1343 VEC(mem_range_s) *available;
1344
1345 /* If we fail to get the set of available memory, then the
1346 target does not support querying traceframe info, and so we
1347 attempt reading from the traceframe anyway (assuming the
1348 target implements the old QTro packet then). */
1349 if (traceframe_available_memory (&available, memaddr, len))
1350 {
1351 struct cleanup *old_chain;
1352
1353 old_chain = make_cleanup (VEC_cleanup(mem_range_s), &available);
1354
1355 if (VEC_empty (mem_range_s, available)
1356 || VEC_index (mem_range_s, available, 0)->start != memaddr)
1357 {
1358 /* Don't read into the traceframe's available
1359 memory. */
1360 if (!VEC_empty (mem_range_s, available))
1361 {
1362 LONGEST oldlen = len;
1363
1364 len = VEC_index (mem_range_s, available, 0)->start - memaddr;
1365 gdb_assert (len <= oldlen);
1366 }
1367
1368 do_cleanups (old_chain);
1369
1370 /* This goes through the topmost target again. */
1371 res = memory_xfer_live_readonly_partial (ops, object,
1372 readbuf, memaddr,
1373 len, xfered_len);
1374 if (res == TARGET_XFER_OK)
1375 return TARGET_XFER_OK;
1376 else
1377 {
1378 /* No use trying further, we know some memory starting
1379 at MEMADDR isn't available. */
1380 *xfered_len = len;
1381 return TARGET_XFER_E_UNAVAILABLE;
1382 }
1383 }
1384
1385 /* Don't try to read more than how much is available, in
1386 case the target implements the deprecated QTro packet to
1387 cater for older GDBs (the target's knowledge of read-only
1388 sections may be outdated by now). */
1389 len = VEC_index (mem_range_s, available, 0)->length;
1390
1391 do_cleanups (old_chain);
1392 }
1393 }
1394
1395 /* Try GDB's internal data cache. */
1396 region = lookup_mem_region (memaddr);
1397 /* region->hi == 0 means there's no upper bound. */
1398 if (memaddr + len < region->hi || region->hi == 0)
1399 reg_len = len;
1400 else
1401 reg_len = region->hi - memaddr;
1402
1403 switch (region->attrib.mode)
1404 {
1405 case MEM_RO:
1406 if (writebuf != NULL)
1407 return TARGET_XFER_E_IO;
1408 break;
1409
1410 case MEM_WO:
1411 if (readbuf != NULL)
1412 return TARGET_XFER_E_IO;
1413 break;
1414
1415 case MEM_FLASH:
1416 /* We only support writing to flash during "load" for now. */
1417 if (writebuf != NULL)
1418 error (_("Writing to flash memory forbidden in this context"));
1419 break;
1420
1421 case MEM_NONE:
1422 return TARGET_XFER_E_IO;
1423 }
1424
1425 if (!ptid_equal (inferior_ptid, null_ptid))
1426 inf = find_inferior_pid (ptid_get_pid (inferior_ptid));
1427 else
1428 inf = NULL;
1429
1430 if (inf != NULL
1431 /* The dcache reads whole cache lines; that doesn't play well
1432 with reading from a trace buffer, because reading outside of
1433 the collected memory range fails. */
1434 && get_traceframe_number () == -1
1435 && (region->attrib.cache
1436 || (stack_cache_enabled_p () && object == TARGET_OBJECT_STACK_MEMORY)
1437 || (code_cache_enabled_p () && object == TARGET_OBJECT_CODE_MEMORY)))
1438 {
1439 DCACHE *dcache = target_dcache_get_or_init ();
1440 int l;
1441
1442 if (readbuf != NULL)
1443 l = dcache_xfer_memory (ops, dcache, memaddr, readbuf, reg_len, 0);
1444 else
1445 /* FIXME drow/2006-08-09: If we're going to preserve const
1446 correctness dcache_xfer_memory should take readbuf and
1447 writebuf. */
1448 l = dcache_xfer_memory (ops, dcache, memaddr, (void *) writebuf,
1449 reg_len, 1);
1450 if (l <= 0)
1451 return TARGET_XFER_E_IO;
1452 else
1453 {
1454 *xfered_len = (ULONGEST) l;
1455 return TARGET_XFER_OK;
1456 }
1457 }
1458
1459 /* If none of those methods found the memory we wanted, fall back
1460 to a target partial transfer. Normally a single call to
1461 to_xfer_partial is enough; if it doesn't recognize an object
1462 it will call the to_xfer_partial of the next target down.
1463 But for memory this won't do. Memory is the only target
1464 object which can be read from more than one valid target.
1465 A core file, for instance, could have some of memory but
1466 delegate other bits to the target below it. So, we must
1467 manually try all targets. */
1468
1469 res = raw_memory_xfer_partial (ops, readbuf, writebuf, memaddr, reg_len,
1470 xfered_len);
1471
1472 /* Make sure the cache gets updated no matter what - if we are writing
1473 to the stack. Even if this write is not tagged as such, we still need
1474 to update the cache. */
1475
1476 if (res == TARGET_XFER_OK
1477 && inf != NULL
1478 && writebuf != NULL
1479 && target_dcache_init_p ()
1480 && !region->attrib.cache
1481 && ((stack_cache_enabled_p () && object != TARGET_OBJECT_STACK_MEMORY)
1482 || (code_cache_enabled_p () && object != TARGET_OBJECT_CODE_MEMORY)))
1483 {
1484 DCACHE *dcache = target_dcache_get ();
1485
1486 dcache_update (dcache, memaddr, (void *) writebuf, reg_len);
1487 }
1488
1489 /* If we still haven't got anything, return the last error. We
1490 give up. */
1491 return res;
1492}
1493
1494/* Perform a partial memory transfer. For docs see target.h,
1495 to_xfer_partial. */
1496
1497static enum target_xfer_status
1498memory_xfer_partial (struct target_ops *ops, enum target_object object,
1499 gdb_byte *readbuf, const gdb_byte *writebuf,
1500 ULONGEST memaddr, ULONGEST len, ULONGEST *xfered_len)
1501{
1502 enum target_xfer_status res;
1503
1504 /* Zero length requests are ok and require no work. */
1505 if (len == 0)
1506 return TARGET_XFER_EOF;
1507
1508 /* Fill in READBUF with breakpoint shadows, or WRITEBUF with
1509 breakpoint insns, thus hiding out from higher layers whether
1510 there are software breakpoints inserted in the code stream. */
1511 if (readbuf != NULL)
1512 {
1513 res = memory_xfer_partial_1 (ops, object, readbuf, NULL, memaddr, len,
1514 xfered_len);
1515
1516 if (res == TARGET_XFER_OK && !show_memory_breakpoints)
1517 breakpoint_xfer_memory (readbuf, NULL, NULL, memaddr, res);
1518 }
1519 else
1520 {
1521 void *buf;
1522 struct cleanup *old_chain;
1523
1524 /* A large write request is likely to be partially satisfied
1525 by memory_xfer_partial_1. We will continually malloc
1526 and free a copy of the entire write request for breakpoint
1527 shadow handling even though we only end up writing a small
1528 subset of it. Cap writes to 4KB to mitigate this. */
1529 len = min (4096, len);
1530
1531 buf = xmalloc (len);
1532 old_chain = make_cleanup (xfree, buf);
1533 memcpy (buf, writebuf, len);
1534
1535 breakpoint_xfer_memory (NULL, buf, writebuf, memaddr, len);
1536 res = memory_xfer_partial_1 (ops, object, NULL, buf, memaddr, len,
1537 xfered_len);
1538
1539 do_cleanups (old_chain);
1540 }
1541
1542 return res;
1543}
1544
1545static void
1546restore_show_memory_breakpoints (void *arg)
1547{
1548 show_memory_breakpoints = (uintptr_t) arg;
1549}
1550
1551struct cleanup *
1552make_show_memory_breakpoints_cleanup (int show)
1553{
1554 int current = show_memory_breakpoints;
1555
1556 show_memory_breakpoints = show;
1557 return make_cleanup (restore_show_memory_breakpoints,
1558 (void *) (uintptr_t) current);
1559}
1560
1561/* For docs see target.h, to_xfer_partial. */
1562
1563enum target_xfer_status
1564target_xfer_partial (struct target_ops *ops,
1565 enum target_object object, const char *annex,
1566 gdb_byte *readbuf, const gdb_byte *writebuf,
1567 ULONGEST offset, ULONGEST len,
1568 ULONGEST *xfered_len)
1569{
1570 enum target_xfer_status retval;
1571
1572 gdb_assert (ops->to_xfer_partial != NULL);
1573
1574 /* Transfer is done when LEN is zero. */
1575 if (len == 0)
1576 return TARGET_XFER_EOF;
1577
1578 if (writebuf && !may_write_memory)
1579 error (_("Writing to memory is not allowed (addr %s, len %s)"),
1580 core_addr_to_string_nz (offset), plongest (len));
1581
1582 *xfered_len = 0;
1583
1584 /* If this is a memory transfer, let the memory-specific code
1585 have a look at it instead. Memory transfers are more
1586 complicated. */
1587 if (object == TARGET_OBJECT_MEMORY || object == TARGET_OBJECT_STACK_MEMORY
1588 || object == TARGET_OBJECT_CODE_MEMORY)
1589 retval = memory_xfer_partial (ops, object, readbuf,
1590 writebuf, offset, len, xfered_len);
1591 else if (object == TARGET_OBJECT_RAW_MEMORY)
1592 {
1593 /* Request the normal memory object from other layers. */
1594 retval = raw_memory_xfer_partial (ops, readbuf, writebuf, offset, len,
1595 xfered_len);
1596 }
1597 else
1598 retval = ops->to_xfer_partial (ops, object, annex, readbuf,
1599 writebuf, offset, len, xfered_len);
1600
1601 if (targetdebug)
1602 {
1603 const unsigned char *myaddr = NULL;
1604
1605 fprintf_unfiltered (gdb_stdlog,
1606 "%s:target_xfer_partial "
1607 "(%d, %s, %s, %s, %s, %s) = %d, %s",
1608 ops->to_shortname,
1609 (int) object,
1610 (annex ? annex : "(null)"),
1611 host_address_to_string (readbuf),
1612 host_address_to_string (writebuf),
1613 core_addr_to_string_nz (offset),
1614 pulongest (len), retval,
1615 pulongest (*xfered_len));
1616
1617 if (readbuf)
1618 myaddr = readbuf;
1619 if (writebuf)
1620 myaddr = writebuf;
1621 if (retval == TARGET_XFER_OK && myaddr != NULL)
1622 {
1623 int i;
1624
1625 fputs_unfiltered (", bytes =", gdb_stdlog);
1626 for (i = 0; i < *xfered_len; i++)
1627 {
1628 if ((((intptr_t) &(myaddr[i])) & 0xf) == 0)
1629 {
1630 if (targetdebug < 2 && i > 0)
1631 {
1632 fprintf_unfiltered (gdb_stdlog, " ...");
1633 break;
1634 }
1635 fprintf_unfiltered (gdb_stdlog, "\n");
1636 }
1637
1638 fprintf_unfiltered (gdb_stdlog, " %02x", myaddr[i] & 0xff);
1639 }
1640 }
1641
1642 fputc_unfiltered ('\n', gdb_stdlog);
1643 }
1644
1645 /* Check implementations of to_xfer_partial update *XFERED_LEN
1646 properly. Do assertion after printing debug messages, so that we
1647 can find more clues on assertion failure from debugging messages. */
1648 if (retval == TARGET_XFER_OK || retval == TARGET_XFER_E_UNAVAILABLE)
1649 gdb_assert (*xfered_len > 0);
1650
1651 return retval;
1652}
1653
1654/* Read LEN bytes of target memory at address MEMADDR, placing the
1655 results in GDB's memory at MYADDR. Returns either 0 for success or
1656 TARGET_XFER_E_IO if any error occurs.
1657
1658 If an error occurs, no guarantee is made about the contents of the data at
1659 MYADDR. In particular, the caller should not depend upon partial reads
1660 filling the buffer with good data. There is no way for the caller to know
1661 how much good data might have been transfered anyway. Callers that can
1662 deal with partial reads should call target_read (which will retry until
1663 it makes no progress, and then return how much was transferred). */
1664
1665int
1666target_read_memory (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1667{
1668 /* Dispatch to the topmost target, not the flattened current_target.
1669 Memory accesses check target->to_has_(all_)memory, and the
1670 flattened target doesn't inherit those. */
1671 if (target_read (current_target.beneath, TARGET_OBJECT_MEMORY, NULL,
1672 myaddr, memaddr, len) == len)
1673 return 0;
1674 else
1675 return TARGET_XFER_E_IO;
1676}
1677
1678/* Like target_read_memory, but specify explicitly that this is a read
1679 from the target's raw memory. That is, this read bypasses the
1680 dcache, breakpoint shadowing, etc. */
1681
1682int
1683target_read_raw_memory (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1684{
1685 /* See comment in target_read_memory about why the request starts at
1686 current_target.beneath. */
1687 if (target_read (current_target.beneath, TARGET_OBJECT_RAW_MEMORY, NULL,
1688 myaddr, memaddr, len) == len)
1689 return 0;
1690 else
1691 return TARGET_XFER_E_IO;
1692}
1693
1694/* Like target_read_memory, but specify explicitly that this is a read from
1695 the target's stack. This may trigger different cache behavior. */
1696
1697int
1698target_read_stack (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1699{
1700 /* See comment in target_read_memory about why the request starts at
1701 current_target.beneath. */
1702 if (target_read (current_target.beneath, TARGET_OBJECT_STACK_MEMORY, NULL,
1703 myaddr, memaddr, len) == len)
1704 return 0;
1705 else
1706 return TARGET_XFER_E_IO;
1707}
1708
1709/* Like target_read_memory, but specify explicitly that this is a read from
1710 the target's code. This may trigger different cache behavior. */
1711
1712int
1713target_read_code (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1714{
1715 /* See comment in target_read_memory about why the request starts at
1716 current_target.beneath. */
1717 if (target_read (current_target.beneath, TARGET_OBJECT_CODE_MEMORY, NULL,
1718 myaddr, memaddr, len) == len)
1719 return 0;
1720 else
1721 return TARGET_XFER_E_IO;
1722}
1723
1724/* Write LEN bytes from MYADDR to target memory at address MEMADDR.
1725 Returns either 0 for success or TARGET_XFER_E_IO if any
1726 error occurs. If an error occurs, no guarantee is made about how
1727 much data got written. Callers that can deal with partial writes
1728 should call target_write. */
1729
1730int
1731target_write_memory (CORE_ADDR memaddr, const gdb_byte *myaddr, ssize_t len)
1732{
1733 /* See comment in target_read_memory about why the request starts at
1734 current_target.beneath. */
1735 if (target_write (current_target.beneath, TARGET_OBJECT_MEMORY, NULL,
1736 myaddr, memaddr, len) == len)
1737 return 0;
1738 else
1739 return TARGET_XFER_E_IO;
1740}
1741
1742/* Write LEN bytes from MYADDR to target raw memory at address
1743 MEMADDR. Returns either 0 for success or TARGET_XFER_E_IO
1744 if any error occurs. If an error occurs, no guarantee is made
1745 about how much data got written. Callers that can deal with
1746 partial writes should call target_write. */
1747
1748int
1749target_write_raw_memory (CORE_ADDR memaddr, const gdb_byte *myaddr, ssize_t len)
1750{
1751 /* See comment in target_read_memory about why the request starts at
1752 current_target.beneath. */
1753 if (target_write (current_target.beneath, TARGET_OBJECT_RAW_MEMORY, NULL,
1754 myaddr, memaddr, len) == len)
1755 return 0;
1756 else
1757 return TARGET_XFER_E_IO;
1758}
1759
1760/* Fetch the target's memory map. */
1761
1762VEC(mem_region_s) *
1763target_memory_map (void)
1764{
1765 VEC(mem_region_s) *result;
1766 struct mem_region *last_one, *this_one;
1767 int ix;
1768 struct target_ops *t;
1769
1770 if (targetdebug)
1771 fprintf_unfiltered (gdb_stdlog, "target_memory_map ()\n");
1772
1773 for (t = current_target.beneath; t != NULL; t = t->beneath)
1774 if (t->to_memory_map != NULL)
1775 break;
1776
1777 if (t == NULL)
1778 return NULL;
1779
1780 result = t->to_memory_map (t);
1781 if (result == NULL)
1782 return NULL;
1783
1784 qsort (VEC_address (mem_region_s, result),
1785 VEC_length (mem_region_s, result),
1786 sizeof (struct mem_region), mem_region_cmp);
1787
1788 /* Check that regions do not overlap. Simultaneously assign
1789 a numbering for the "mem" commands to use to refer to
1790 each region. */
1791 last_one = NULL;
1792 for (ix = 0; VEC_iterate (mem_region_s, result, ix, this_one); ix++)
1793 {
1794 this_one->number = ix;
1795
1796 if (last_one && last_one->hi > this_one->lo)
1797 {
1798 warning (_("Overlapping regions in memory map: ignoring"));
1799 VEC_free (mem_region_s, result);
1800 return NULL;
1801 }
1802 last_one = this_one;
1803 }
1804
1805 return result;
1806}
1807
1808void
1809target_flash_erase (ULONGEST address, LONGEST length)
1810{
1811 struct target_ops *t;
1812
1813 for (t = current_target.beneath; t != NULL; t = t->beneath)
1814 if (t->to_flash_erase != NULL)
1815 {
1816 if (targetdebug)
1817 fprintf_unfiltered (gdb_stdlog, "target_flash_erase (%s, %s)\n",
1818 hex_string (address), phex (length, 0));
1819 t->to_flash_erase (t, address, length);
1820 return;
1821 }
1822
1823 tcomplain ();
1824}
1825
1826void
1827target_flash_done (void)
1828{
1829 struct target_ops *t;
1830
1831 for (t = current_target.beneath; t != NULL; t = t->beneath)
1832 if (t->to_flash_done != NULL)
1833 {
1834 if (targetdebug)
1835 fprintf_unfiltered (gdb_stdlog, "target_flash_done\n");
1836 t->to_flash_done (t);
1837 return;
1838 }
1839
1840 tcomplain ();
1841}
1842
1843static void
1844show_trust_readonly (struct ui_file *file, int from_tty,
1845 struct cmd_list_element *c, const char *value)
1846{
1847 fprintf_filtered (file,
1848 _("Mode for reading from readonly sections is %s.\n"),
1849 value);
1850}
1851
1852/* More generic transfers. */
1853
1854static enum target_xfer_status
1855default_xfer_partial (struct target_ops *ops, enum target_object object,
1856 const char *annex, gdb_byte *readbuf,
1857 const gdb_byte *writebuf, ULONGEST offset, ULONGEST len,
1858 ULONGEST *xfered_len)
1859{
1860 if (object == TARGET_OBJECT_MEMORY
1861 && ops->deprecated_xfer_memory != NULL)
1862 /* If available, fall back to the target's
1863 "deprecated_xfer_memory" method. */
1864 {
1865 int xfered = -1;
1866
1867 errno = 0;
1868 if (writebuf != NULL)
1869 {
1870 void *buffer = xmalloc (len);
1871 struct cleanup *cleanup = make_cleanup (xfree, buffer);
1872
1873 memcpy (buffer, writebuf, len);
1874 xfered = ops->deprecated_xfer_memory (offset, buffer, len,
1875 1/*write*/, NULL, ops);
1876 do_cleanups (cleanup);
1877 }
1878 if (readbuf != NULL)
1879 xfered = ops->deprecated_xfer_memory (offset, readbuf, len,
1880 0/*read*/, NULL, ops);
1881 if (xfered > 0)
1882 {
1883 *xfered_len = (ULONGEST) xfered;
1884 return TARGET_XFER_E_IO;
1885 }
1886 else if (xfered == 0 && errno == 0)
1887 /* "deprecated_xfer_memory" uses 0, cross checked against
1888 ERRNO as one indication of an error. */
1889 return TARGET_XFER_EOF;
1890 else
1891 return TARGET_XFER_E_IO;
1892 }
1893 else
1894 {
1895 gdb_assert (ops->beneath != NULL);
1896 return ops->beneath->to_xfer_partial (ops->beneath, object, annex,
1897 readbuf, writebuf, offset, len,
1898 xfered_len);
1899 }
1900}
1901
1902/* Target vector read/write partial wrapper functions. */
1903
1904static enum target_xfer_status
1905target_read_partial (struct target_ops *ops,
1906 enum target_object object,
1907 const char *annex, gdb_byte *buf,
1908 ULONGEST offset, ULONGEST len,
1909 ULONGEST *xfered_len)
1910{
1911 return target_xfer_partial (ops, object, annex, buf, NULL, offset, len,
1912 xfered_len);
1913}
1914
1915static enum target_xfer_status
1916target_write_partial (struct target_ops *ops,
1917 enum target_object object,
1918 const char *annex, const gdb_byte *buf,
1919 ULONGEST offset, LONGEST len, ULONGEST *xfered_len)
1920{
1921 return target_xfer_partial (ops, object, annex, NULL, buf, offset, len,
1922 xfered_len);
1923}
1924
1925/* Wrappers to perform the full transfer. */
1926
1927/* For docs on target_read see target.h. */
1928
1929LONGEST
1930target_read (struct target_ops *ops,
1931 enum target_object object,
1932 const char *annex, gdb_byte *buf,
1933 ULONGEST offset, LONGEST len)
1934{
1935 LONGEST xfered = 0;
1936
1937 while (xfered < len)
1938 {
1939 ULONGEST xfered_len;
1940 enum target_xfer_status status;
1941
1942 status = target_read_partial (ops, object, annex,
1943 (gdb_byte *) buf + xfered,
1944 offset + xfered, len - xfered,
1945 &xfered_len);
1946
1947 /* Call an observer, notifying them of the xfer progress? */
1948 if (status == TARGET_XFER_EOF)
1949 return xfered;
1950 else if (status == TARGET_XFER_OK)
1951 {
1952 xfered += xfered_len;
1953 QUIT;
1954 }
1955 else
1956 return -1;
1957
1958 }
1959 return len;
1960}
1961
1962/* Assuming that the entire [begin, end) range of memory cannot be
1963 read, try to read whatever subrange is possible to read.
1964
1965 The function returns, in RESULT, either zero or one memory block.
1966 If there's a readable subrange at the beginning, it is completely
1967 read and returned. Any further readable subrange will not be read.
1968 Otherwise, if there's a readable subrange at the end, it will be
1969 completely read and returned. Any readable subranges before it
1970 (obviously, not starting at the beginning), will be ignored. In
1971 other cases -- either no readable subrange, or readable subrange(s)
1972 that is neither at the beginning, or end, nothing is returned.
1973
1974 The purpose of this function is to handle a read across a boundary
1975 of accessible memory in a case when memory map is not available.
1976 The above restrictions are fine for this case, but will give
1977 incorrect results if the memory is 'patchy'. However, supporting
1978 'patchy' memory would require trying to read every single byte,
1979 and it seems unacceptable solution. Explicit memory map is
1980 recommended for this case -- and target_read_memory_robust will
1981 take care of reading multiple ranges then. */
1982
1983static void
1984read_whatever_is_readable (struct target_ops *ops,
1985 ULONGEST begin, ULONGEST end,
1986 VEC(memory_read_result_s) **result)
1987{
1988 gdb_byte *buf = xmalloc (end - begin);
1989 ULONGEST current_begin = begin;
1990 ULONGEST current_end = end;
1991 int forward;
1992 memory_read_result_s r;
1993 ULONGEST xfered_len;
1994
1995 /* If we previously failed to read 1 byte, nothing can be done here. */
1996 if (end - begin <= 1)
1997 {
1998 xfree (buf);
1999 return;
2000 }
2001
2002 /* Check that either first or the last byte is readable, and give up
2003 if not. This heuristic is meant to permit reading accessible memory
2004 at the boundary of accessible region. */
2005 if (target_read_partial (ops, TARGET_OBJECT_MEMORY, NULL,
2006 buf, begin, 1, &xfered_len) == TARGET_XFER_OK)
2007 {
2008 forward = 1;
2009 ++current_begin;
2010 }
2011 else if (target_read_partial (ops, TARGET_OBJECT_MEMORY, NULL,
2012 buf + (end-begin) - 1, end - 1, 1,
2013 &xfered_len) == TARGET_XFER_OK)
2014 {
2015 forward = 0;
2016 --current_end;
2017 }
2018 else
2019 {
2020 xfree (buf);
2021 return;
2022 }
2023
2024 /* Loop invariant is that the [current_begin, current_end) was previously
2025 found to be not readable as a whole.
2026
2027 Note loop condition -- if the range has 1 byte, we can't divide the range
2028 so there's no point trying further. */
2029 while (current_end - current_begin > 1)
2030 {
2031 ULONGEST first_half_begin, first_half_end;
2032 ULONGEST second_half_begin, second_half_end;
2033 LONGEST xfer;
2034 ULONGEST middle = current_begin + (current_end - current_begin)/2;
2035
2036 if (forward)
2037 {
2038 first_half_begin = current_begin;
2039 first_half_end = middle;
2040 second_half_begin = middle;
2041 second_half_end = current_end;
2042 }
2043 else
2044 {
2045 first_half_begin = middle;
2046 first_half_end = current_end;
2047 second_half_begin = current_begin;
2048 second_half_end = middle;
2049 }
2050
2051 xfer = target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2052 buf + (first_half_begin - begin),
2053 first_half_begin,
2054 first_half_end - first_half_begin);
2055
2056 if (xfer == first_half_end - first_half_begin)
2057 {
2058 /* This half reads up fine. So, the error must be in the
2059 other half. */
2060 current_begin = second_half_begin;
2061 current_end = second_half_end;
2062 }
2063 else
2064 {
2065 /* This half is not readable. Because we've tried one byte, we
2066 know some part of this half if actually redable. Go to the next
2067 iteration to divide again and try to read.
2068
2069 We don't handle the other half, because this function only tries
2070 to read a single readable subrange. */
2071 current_begin = first_half_begin;
2072 current_end = first_half_end;
2073 }
2074 }
2075
2076 if (forward)
2077 {
2078 /* The [begin, current_begin) range has been read. */
2079 r.begin = begin;
2080 r.end = current_begin;
2081 r.data = buf;
2082 }
2083 else
2084 {
2085 /* The [current_end, end) range has been read. */
2086 LONGEST rlen = end - current_end;
2087
2088 r.data = xmalloc (rlen);
2089 memcpy (r.data, buf + current_end - begin, rlen);
2090 r.begin = current_end;
2091 r.end = end;
2092 xfree (buf);
2093 }
2094 VEC_safe_push(memory_read_result_s, (*result), &r);
2095}
2096
2097void
2098free_memory_read_result_vector (void *x)
2099{
2100 VEC(memory_read_result_s) *v = x;
2101 memory_read_result_s *current;
2102 int ix;
2103
2104 for (ix = 0; VEC_iterate (memory_read_result_s, v, ix, current); ++ix)
2105 {
2106 xfree (current->data);
2107 }
2108 VEC_free (memory_read_result_s, v);
2109}
2110
2111VEC(memory_read_result_s) *
2112read_memory_robust (struct target_ops *ops, ULONGEST offset, LONGEST len)
2113{
2114 VEC(memory_read_result_s) *result = 0;
2115
2116 LONGEST xfered = 0;
2117 while (xfered < len)
2118 {
2119 struct mem_region *region = lookup_mem_region (offset + xfered);
2120 LONGEST rlen;
2121
2122 /* If there is no explicit region, a fake one should be created. */
2123 gdb_assert (region);
2124
2125 if (region->hi == 0)
2126 rlen = len - xfered;
2127 else
2128 rlen = region->hi - offset;
2129
2130 if (region->attrib.mode == MEM_NONE || region->attrib.mode == MEM_WO)
2131 {
2132 /* Cannot read this region. Note that we can end up here only
2133 if the region is explicitly marked inaccessible, or
2134 'inaccessible-by-default' is in effect. */
2135 xfered += rlen;
2136 }
2137 else
2138 {
2139 LONGEST to_read = min (len - xfered, rlen);
2140 gdb_byte *buffer = (gdb_byte *)xmalloc (to_read);
2141
2142 LONGEST xfer = target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2143 (gdb_byte *) buffer,
2144 offset + xfered, to_read);
2145 /* Call an observer, notifying them of the xfer progress? */
2146 if (xfer <= 0)
2147 {
2148 /* Got an error reading full chunk. See if maybe we can read
2149 some subrange. */
2150 xfree (buffer);
2151 read_whatever_is_readable (ops, offset + xfered,
2152 offset + xfered + to_read, &result);
2153 xfered += to_read;
2154 }
2155 else
2156 {
2157 struct memory_read_result r;
2158 r.data = buffer;
2159 r.begin = offset + xfered;
2160 r.end = r.begin + xfer;
2161 VEC_safe_push (memory_read_result_s, result, &r);
2162 xfered += xfer;
2163 }
2164 QUIT;
2165 }
2166 }
2167 return result;
2168}
2169
2170
2171/* An alternative to target_write with progress callbacks. */
2172
2173LONGEST
2174target_write_with_progress (struct target_ops *ops,
2175 enum target_object object,
2176 const char *annex, const gdb_byte *buf,
2177 ULONGEST offset, LONGEST len,
2178 void (*progress) (ULONGEST, void *), void *baton)
2179{
2180 LONGEST xfered = 0;
2181
2182 /* Give the progress callback a chance to set up. */
2183 if (progress)
2184 (*progress) (0, baton);
2185
2186 while (xfered < len)
2187 {
2188 ULONGEST xfered_len;
2189 enum target_xfer_status status;
2190
2191 status = target_write_partial (ops, object, annex,
2192 (gdb_byte *) buf + xfered,
2193 offset + xfered, len - xfered,
2194 &xfered_len);
2195
2196 if (status == TARGET_XFER_EOF)
2197 return xfered;
2198 if (TARGET_XFER_STATUS_ERROR_P (status))
2199 return -1;
2200
2201 gdb_assert (status == TARGET_XFER_OK);
2202 if (progress)
2203 (*progress) (xfered_len, baton);
2204
2205 xfered += xfered_len;
2206 QUIT;
2207 }
2208 return len;
2209}
2210
2211/* For docs on target_write see target.h. */
2212
2213LONGEST
2214target_write (struct target_ops *ops,
2215 enum target_object object,
2216 const char *annex, const gdb_byte *buf,
2217 ULONGEST offset, LONGEST len)
2218{
2219 return target_write_with_progress (ops, object, annex, buf, offset, len,
2220 NULL, NULL);
2221}
2222
2223/* Read OBJECT/ANNEX using OPS. Store the result in *BUF_P and return
2224 the size of the transferred data. PADDING additional bytes are
2225 available in *BUF_P. This is a helper function for
2226 target_read_alloc; see the declaration of that function for more
2227 information. */
2228
2229static LONGEST
2230target_read_alloc_1 (struct target_ops *ops, enum target_object object,
2231 const char *annex, gdb_byte **buf_p, int padding)
2232{
2233 size_t buf_alloc, buf_pos;
2234 gdb_byte *buf;
2235
2236 /* This function does not have a length parameter; it reads the
2237 entire OBJECT). Also, it doesn't support objects fetched partly
2238 from one target and partly from another (in a different stratum,
2239 e.g. a core file and an executable). Both reasons make it
2240 unsuitable for reading memory. */
2241 gdb_assert (object != TARGET_OBJECT_MEMORY);
2242
2243 /* Start by reading up to 4K at a time. The target will throttle
2244 this number down if necessary. */
2245 buf_alloc = 4096;
2246 buf = xmalloc (buf_alloc);
2247 buf_pos = 0;
2248 while (1)
2249 {
2250 ULONGEST xfered_len;
2251 enum target_xfer_status status;
2252
2253 status = target_read_partial (ops, object, annex, &buf[buf_pos],
2254 buf_pos, buf_alloc - buf_pos - padding,
2255 &xfered_len);
2256
2257 if (status == TARGET_XFER_EOF)
2258 {
2259 /* Read all there was. */
2260 if (buf_pos == 0)
2261 xfree (buf);
2262 else
2263 *buf_p = buf;
2264 return buf_pos;
2265 }
2266 else if (status != TARGET_XFER_OK)
2267 {
2268 /* An error occurred. */
2269 xfree (buf);
2270 return TARGET_XFER_E_IO;
2271 }
2272
2273 buf_pos += xfered_len;
2274
2275 /* If the buffer is filling up, expand it. */
2276 if (buf_alloc < buf_pos * 2)
2277 {
2278 buf_alloc *= 2;
2279 buf = xrealloc (buf, buf_alloc);
2280 }
2281
2282 QUIT;
2283 }
2284}
2285
2286/* Read OBJECT/ANNEX using OPS. Store the result in *BUF_P and return
2287 the size of the transferred data. See the declaration in "target.h"
2288 function for more information about the return value. */
2289
2290LONGEST
2291target_read_alloc (struct target_ops *ops, enum target_object object,
2292 const char *annex, gdb_byte **buf_p)
2293{
2294 return target_read_alloc_1 (ops, object, annex, buf_p, 0);
2295}
2296
2297/* Read OBJECT/ANNEX using OPS. The result is NUL-terminated and
2298 returned as a string, allocated using xmalloc. If an error occurs
2299 or the transfer is unsupported, NULL is returned. Empty objects
2300 are returned as allocated but empty strings. A warning is issued
2301 if the result contains any embedded NUL bytes. */
2302
2303char *
2304target_read_stralloc (struct target_ops *ops, enum target_object object,
2305 const char *annex)
2306{
2307 gdb_byte *buffer;
2308 char *bufstr;
2309 LONGEST i, transferred;
2310
2311 transferred = target_read_alloc_1 (ops, object, annex, &buffer, 1);
2312 bufstr = (char *) buffer;
2313
2314 if (transferred < 0)
2315 return NULL;
2316
2317 if (transferred == 0)
2318 return xstrdup ("");
2319
2320 bufstr[transferred] = 0;
2321
2322 /* Check for embedded NUL bytes; but allow trailing NULs. */
2323 for (i = strlen (bufstr); i < transferred; i++)
2324 if (bufstr[i] != 0)
2325 {
2326 warning (_("target object %d, annex %s, "
2327 "contained unexpected null characters"),
2328 (int) object, annex ? annex : "(none)");
2329 break;
2330 }
2331
2332 return bufstr;
2333}
2334
2335/* Memory transfer methods. */
2336
2337void
2338get_target_memory (struct target_ops *ops, CORE_ADDR addr, gdb_byte *buf,
2339 LONGEST len)
2340{
2341 /* This method is used to read from an alternate, non-current
2342 target. This read must bypass the overlay support (as symbols
2343 don't match this target), and GDB's internal cache (wrong cache
2344 for this target). */
2345 if (target_read (ops, TARGET_OBJECT_RAW_MEMORY, NULL, buf, addr, len)
2346 != len)
2347 memory_error (TARGET_XFER_E_IO, addr);
2348}
2349
2350ULONGEST
2351get_target_memory_unsigned (struct target_ops *ops, CORE_ADDR addr,
2352 int len, enum bfd_endian byte_order)
2353{
2354 gdb_byte buf[sizeof (ULONGEST)];
2355
2356 gdb_assert (len <= sizeof (buf));
2357 get_target_memory (ops, addr, buf, len);
2358 return extract_unsigned_integer (buf, len, byte_order);
2359}
2360
2361/* See target.h. */
2362
2363int
2364target_insert_breakpoint (struct gdbarch *gdbarch,
2365 struct bp_target_info *bp_tgt)
2366{
2367 if (!may_insert_breakpoints)
2368 {
2369 warning (_("May not insert breakpoints"));
2370 return 1;
2371 }
2372
2373 return current_target.to_insert_breakpoint (&current_target,
2374 gdbarch, bp_tgt);
2375}
2376
2377/* See target.h. */
2378
2379int
2380target_remove_breakpoint (struct gdbarch *gdbarch,
2381 struct bp_target_info *bp_tgt)
2382{
2383 /* This is kind of a weird case to handle, but the permission might
2384 have been changed after breakpoints were inserted - in which case
2385 we should just take the user literally and assume that any
2386 breakpoints should be left in place. */
2387 if (!may_insert_breakpoints)
2388 {
2389 warning (_("May not remove breakpoints"));
2390 return 1;
2391 }
2392
2393 return current_target.to_remove_breakpoint (&current_target,
2394 gdbarch, bp_tgt);
2395}
2396
2397static void
2398target_info (char *args, int from_tty)
2399{
2400 struct target_ops *t;
2401 int has_all_mem = 0;
2402
2403 if (symfile_objfile != NULL)
2404 printf_unfiltered (_("Symbols from \"%s\".\n"),
2405 objfile_name (symfile_objfile));
2406
2407 for (t = target_stack; t != NULL; t = t->beneath)
2408 {
2409 if (!(*t->to_has_memory) (t))
2410 continue;
2411
2412 if ((int) (t->to_stratum) <= (int) dummy_stratum)
2413 continue;
2414 if (has_all_mem)
2415 printf_unfiltered (_("\tWhile running this, "
2416 "GDB does not access memory from...\n"));
2417 printf_unfiltered ("%s:\n", t->to_longname);
2418 (t->to_files_info) (t);
2419 has_all_mem = (*t->to_has_all_memory) (t);
2420 }
2421}
2422
2423/* This function is called before any new inferior is created, e.g.
2424 by running a program, attaching, or connecting to a target.
2425 It cleans up any state from previous invocations which might
2426 change between runs. This is a subset of what target_preopen
2427 resets (things which might change between targets). */
2428
2429void
2430target_pre_inferior (int from_tty)
2431{
2432 /* Clear out solib state. Otherwise the solib state of the previous
2433 inferior might have survived and is entirely wrong for the new
2434 target. This has been observed on GNU/Linux using glibc 2.3. How
2435 to reproduce:
2436
2437 bash$ ./foo&
2438 [1] 4711
2439 bash$ ./foo&
2440 [1] 4712
2441 bash$ gdb ./foo
2442 [...]
2443 (gdb) attach 4711
2444 (gdb) detach
2445 (gdb) attach 4712
2446 Cannot access memory at address 0xdeadbeef
2447 */
2448
2449 /* In some OSs, the shared library list is the same/global/shared
2450 across inferiors. If code is shared between processes, so are
2451 memory regions and features. */
2452 if (!gdbarch_has_global_solist (target_gdbarch ()))
2453 {
2454 no_shared_libraries (NULL, from_tty);
2455
2456 invalidate_target_mem_regions ();
2457
2458 target_clear_description ();
2459 }
2460
2461 agent_capability_invalidate ();
2462}
2463
2464/* Callback for iterate_over_inferiors. Gets rid of the given
2465 inferior. */
2466
2467static int
2468dispose_inferior (struct inferior *inf, void *args)
2469{
2470 struct thread_info *thread;
2471
2472 thread = any_thread_of_process (inf->pid);
2473 if (thread)
2474 {
2475 switch_to_thread (thread->ptid);
2476
2477 /* Core inferiors actually should be detached, not killed. */
2478 if (target_has_execution)
2479 target_kill ();
2480 else
2481 target_detach (NULL, 0);
2482 }
2483
2484 return 0;
2485}
2486
2487/* This is to be called by the open routine before it does
2488 anything. */
2489
2490void
2491target_preopen (int from_tty)
2492{
2493 dont_repeat ();
2494
2495 if (have_inferiors ())
2496 {
2497 if (!from_tty
2498 || !have_live_inferiors ()
2499 || query (_("A program is being debugged already. Kill it? ")))
2500 iterate_over_inferiors (dispose_inferior, NULL);
2501 else
2502 error (_("Program not killed."));
2503 }
2504
2505 /* Calling target_kill may remove the target from the stack. But if
2506 it doesn't (which seems like a win for UDI), remove it now. */
2507 /* Leave the exec target, though. The user may be switching from a
2508 live process to a core of the same program. */
2509 pop_all_targets_above (file_stratum);
2510
2511 target_pre_inferior (from_tty);
2512}
2513
2514/* Detach a target after doing deferred register stores. */
2515
2516void
2517target_detach (const char *args, int from_tty)
2518{
2519 struct target_ops* t;
2520
2521 if (gdbarch_has_global_breakpoints (target_gdbarch ()))
2522 /* Don't remove global breakpoints here. They're removed on
2523 disconnection from the target. */
2524 ;
2525 else
2526 /* If we're in breakpoints-always-inserted mode, have to remove
2527 them before detaching. */
2528 remove_breakpoints_pid (ptid_get_pid (inferior_ptid));
2529
2530 prepare_for_detach ();
2531
2532 current_target.to_detach (&current_target, args, from_tty);
2533 if (targetdebug)
2534 fprintf_unfiltered (gdb_stdlog, "target_detach (%s, %d)\n",
2535 args, from_tty);
2536}
2537
2538void
2539target_disconnect (char *args, int from_tty)
2540{
2541 struct target_ops *t;
2542
2543 /* If we're in breakpoints-always-inserted mode or if breakpoints
2544 are global across processes, we have to remove them before
2545 disconnecting. */
2546 remove_breakpoints ();
2547
2548 for (t = current_target.beneath; t != NULL; t = t->beneath)
2549 if (t->to_disconnect != NULL)
2550 {
2551 if (targetdebug)
2552 fprintf_unfiltered (gdb_stdlog, "target_disconnect (%s, %d)\n",
2553 args, from_tty);
2554 t->to_disconnect (t, args, from_tty);
2555 return;
2556 }
2557
2558 tcomplain ();
2559}
2560
2561ptid_t
2562target_wait (ptid_t ptid, struct target_waitstatus *status, int options)
2563{
2564 struct target_ops *t;
2565 ptid_t retval = (current_target.to_wait) (&current_target, ptid,
2566 status, options);
2567
2568 if (targetdebug)
2569 {
2570 char *status_string;
2571 char *options_string;
2572
2573 status_string = target_waitstatus_to_string (status);
2574 options_string = target_options_to_string (options);
2575 fprintf_unfiltered (gdb_stdlog,
2576 "target_wait (%d, status, options={%s})"
2577 " = %d, %s\n",
2578 ptid_get_pid (ptid), options_string,
2579 ptid_get_pid (retval), status_string);
2580 xfree (status_string);
2581 xfree (options_string);
2582 }
2583
2584 return retval;
2585}
2586
2587char *
2588target_pid_to_str (ptid_t ptid)
2589{
2590 struct target_ops *t;
2591
2592 for (t = current_target.beneath; t != NULL; t = t->beneath)
2593 {
2594 if (t->to_pid_to_str != NULL)
2595 return (*t->to_pid_to_str) (t, ptid);
2596 }
2597
2598 return normal_pid_to_str (ptid);
2599}
2600
2601char *
2602target_thread_name (struct thread_info *info)
2603{
2604 return current_target.to_thread_name (&current_target, info);
2605}
2606
2607void
2608target_resume (ptid_t ptid, int step, enum gdb_signal signal)
2609{
2610 struct target_ops *t;
2611
2612 target_dcache_invalidate ();
2613
2614 current_target.to_resume (&current_target, ptid, step, signal);
2615 if (targetdebug)
2616 fprintf_unfiltered (gdb_stdlog, "target_resume (%d, %s, %s)\n",
2617 ptid_get_pid (ptid),
2618 step ? "step" : "continue",
2619 gdb_signal_to_name (signal));
2620
2621 registers_changed_ptid (ptid);
2622 set_executing (ptid, 1);
2623 set_running (ptid, 1);
2624 clear_inline_frame_state (ptid);
2625}
2626
2627void
2628target_pass_signals (int numsigs, unsigned char *pass_signals)
2629{
2630 struct target_ops *t;
2631
2632 for (t = current_target.beneath; t != NULL; t = t->beneath)
2633 {
2634 if (t->to_pass_signals != NULL)
2635 {
2636 if (targetdebug)
2637 {
2638 int i;
2639
2640 fprintf_unfiltered (gdb_stdlog, "target_pass_signals (%d, {",
2641 numsigs);
2642
2643 for (i = 0; i < numsigs; i++)
2644 if (pass_signals[i])
2645 fprintf_unfiltered (gdb_stdlog, " %s",
2646 gdb_signal_to_name (i));
2647
2648 fprintf_unfiltered (gdb_stdlog, " })\n");
2649 }
2650
2651 (*t->to_pass_signals) (t, numsigs, pass_signals);
2652 return;
2653 }
2654 }
2655}
2656
2657void
2658target_program_signals (int numsigs, unsigned char *program_signals)
2659{
2660 struct target_ops *t;
2661
2662 for (t = current_target.beneath; t != NULL; t = t->beneath)
2663 {
2664 if (t->to_program_signals != NULL)
2665 {
2666 if (targetdebug)
2667 {
2668 int i;
2669
2670 fprintf_unfiltered (gdb_stdlog, "target_program_signals (%d, {",
2671 numsigs);
2672
2673 for (i = 0; i < numsigs; i++)
2674 if (program_signals[i])
2675 fprintf_unfiltered (gdb_stdlog, " %s",
2676 gdb_signal_to_name (i));
2677
2678 fprintf_unfiltered (gdb_stdlog, " })\n");
2679 }
2680
2681 (*t->to_program_signals) (t, numsigs, program_signals);
2682 return;
2683 }
2684 }
2685}
2686
2687/* Look through the list of possible targets for a target that can
2688 follow forks. */
2689
2690int
2691target_follow_fork (int follow_child, int detach_fork)
2692{
2693 struct target_ops *t;
2694
2695 for (t = current_target.beneath; t != NULL; t = t->beneath)
2696 {
2697 if (t->to_follow_fork != NULL)
2698 {
2699 int retval = t->to_follow_fork (t, follow_child, detach_fork);
2700
2701 if (targetdebug)
2702 fprintf_unfiltered (gdb_stdlog,
2703 "target_follow_fork (%d, %d) = %d\n",
2704 follow_child, detach_fork, retval);
2705 return retval;
2706 }
2707 }
2708
2709 /* Some target returned a fork event, but did not know how to follow it. */
2710 internal_error (__FILE__, __LINE__,
2711 _("could not find a target to follow fork"));
2712}
2713
2714void
2715target_mourn_inferior (void)
2716{
2717 struct target_ops *t;
2718
2719 for (t = current_target.beneath; t != NULL; t = t->beneath)
2720 {
2721 if (t->to_mourn_inferior != NULL)
2722 {
2723 t->to_mourn_inferior (t);
2724 if (targetdebug)
2725 fprintf_unfiltered (gdb_stdlog, "target_mourn_inferior ()\n");
2726
2727 /* We no longer need to keep handles on any of the object files.
2728 Make sure to release them to avoid unnecessarily locking any
2729 of them while we're not actually debugging. */
2730 bfd_cache_close_all ();
2731
2732 return;
2733 }
2734 }
2735
2736 internal_error (__FILE__, __LINE__,
2737 _("could not find a target to follow mourn inferior"));
2738}
2739
2740/* Look for a target which can describe architectural features, starting
2741 from TARGET. If we find one, return its description. */
2742
2743const struct target_desc *
2744target_read_description (struct target_ops *target)
2745{
2746 struct target_ops *t;
2747
2748 for (t = target; t != NULL; t = t->beneath)
2749 if (t->to_read_description != NULL)
2750 {
2751 const struct target_desc *tdesc;
2752
2753 tdesc = t->to_read_description (t);
2754 if (tdesc)
2755 return tdesc;
2756 }
2757
2758 return NULL;
2759}
2760
2761/* The default implementation of to_search_memory.
2762 This implements a basic search of memory, reading target memory and
2763 performing the search here (as opposed to performing the search in on the
2764 target side with, for example, gdbserver). */
2765
2766int
2767simple_search_memory (struct target_ops *ops,
2768 CORE_ADDR start_addr, ULONGEST search_space_len,
2769 const gdb_byte *pattern, ULONGEST pattern_len,
2770 CORE_ADDR *found_addrp)
2771{
2772 /* NOTE: also defined in find.c testcase. */
2773#define SEARCH_CHUNK_SIZE 16000
2774 const unsigned chunk_size = SEARCH_CHUNK_SIZE;
2775 /* Buffer to hold memory contents for searching. */
2776 gdb_byte *search_buf;
2777 unsigned search_buf_size;
2778 struct cleanup *old_cleanups;
2779
2780 search_buf_size = chunk_size + pattern_len - 1;
2781
2782 /* No point in trying to allocate a buffer larger than the search space. */
2783 if (search_space_len < search_buf_size)
2784 search_buf_size = search_space_len;
2785
2786 search_buf = malloc (search_buf_size);
2787 if (search_buf == NULL)
2788 error (_("Unable to allocate memory to perform the search."));
2789 old_cleanups = make_cleanup (free_current_contents, &search_buf);
2790
2791 /* Prime the search buffer. */
2792
2793 if (target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2794 search_buf, start_addr, search_buf_size) != search_buf_size)
2795 {
2796 warning (_("Unable to access %s bytes of target "
2797 "memory at %s, halting search."),
2798 pulongest (search_buf_size), hex_string (start_addr));
2799 do_cleanups (old_cleanups);
2800 return -1;
2801 }
2802
2803 /* Perform the search.
2804
2805 The loop is kept simple by allocating [N + pattern-length - 1] bytes.
2806 When we've scanned N bytes we copy the trailing bytes to the start and
2807 read in another N bytes. */
2808
2809 while (search_space_len >= pattern_len)
2810 {
2811 gdb_byte *found_ptr;
2812 unsigned nr_search_bytes = min (search_space_len, search_buf_size);
2813
2814 found_ptr = memmem (search_buf, nr_search_bytes,
2815 pattern, pattern_len);
2816
2817 if (found_ptr != NULL)
2818 {
2819 CORE_ADDR found_addr = start_addr + (found_ptr - search_buf);
2820
2821 *found_addrp = found_addr;
2822 do_cleanups (old_cleanups);
2823 return 1;
2824 }
2825
2826 /* Not found in this chunk, skip to next chunk. */
2827
2828 /* Don't let search_space_len wrap here, it's unsigned. */
2829 if (search_space_len >= chunk_size)
2830 search_space_len -= chunk_size;
2831 else
2832 search_space_len = 0;
2833
2834 if (search_space_len >= pattern_len)
2835 {
2836 unsigned keep_len = search_buf_size - chunk_size;
2837 CORE_ADDR read_addr = start_addr + chunk_size + keep_len;
2838 int nr_to_read;
2839
2840 /* Copy the trailing part of the previous iteration to the front
2841 of the buffer for the next iteration. */
2842 gdb_assert (keep_len == pattern_len - 1);
2843 memcpy (search_buf, search_buf + chunk_size, keep_len);
2844
2845 nr_to_read = min (search_space_len - keep_len, chunk_size);
2846
2847 if (target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2848 search_buf + keep_len, read_addr,
2849 nr_to_read) != nr_to_read)
2850 {
2851 warning (_("Unable to access %s bytes of target "
2852 "memory at %s, halting search."),
2853 plongest (nr_to_read),
2854 hex_string (read_addr));
2855 do_cleanups (old_cleanups);
2856 return -1;
2857 }
2858
2859 start_addr += chunk_size;
2860 }
2861 }
2862
2863 /* Not found. */
2864
2865 do_cleanups (old_cleanups);
2866 return 0;
2867}
2868
2869/* Search SEARCH_SPACE_LEN bytes beginning at START_ADDR for the
2870 sequence of bytes in PATTERN with length PATTERN_LEN.
2871
2872 The result is 1 if found, 0 if not found, and -1 if there was an error
2873 requiring halting of the search (e.g. memory read error).
2874 If the pattern is found the address is recorded in FOUND_ADDRP. */
2875
2876int
2877target_search_memory (CORE_ADDR start_addr, ULONGEST search_space_len,
2878 const gdb_byte *pattern, ULONGEST pattern_len,
2879 CORE_ADDR *found_addrp)
2880{
2881 struct target_ops *t;
2882 int found;
2883
2884 /* We don't use INHERIT to set current_target.to_search_memory,
2885 so we have to scan the target stack and handle targetdebug
2886 ourselves. */
2887
2888 if (targetdebug)
2889 fprintf_unfiltered (gdb_stdlog, "target_search_memory (%s, ...)\n",
2890 hex_string (start_addr));
2891
2892 for (t = current_target.beneath; t != NULL; t = t->beneath)
2893 if (t->to_search_memory != NULL)
2894 break;
2895
2896 if (t != NULL)
2897 {
2898 found = t->to_search_memory (t, start_addr, search_space_len,
2899 pattern, pattern_len, found_addrp);
2900 }
2901 else
2902 {
2903 /* If a special version of to_search_memory isn't available, use the
2904 simple version. */
2905 found = simple_search_memory (current_target.beneath,
2906 start_addr, search_space_len,
2907 pattern, pattern_len, found_addrp);
2908 }
2909
2910 if (targetdebug)
2911 fprintf_unfiltered (gdb_stdlog, " = %d\n", found);
2912
2913 return found;
2914}
2915
2916/* Look through the currently pushed targets. If none of them will
2917 be able to restart the currently running process, issue an error
2918 message. */
2919
2920void
2921target_require_runnable (void)
2922{
2923 struct target_ops *t;
2924
2925 for (t = target_stack; t != NULL; t = t->beneath)
2926 {
2927 /* If this target knows how to create a new program, then
2928 assume we will still be able to after killing the current
2929 one. Either killing and mourning will not pop T, or else
2930 find_default_run_target will find it again. */
2931 if (t->to_create_inferior != NULL)
2932 return;
2933
2934 /* Do not worry about thread_stratum targets that can not
2935 create inferiors. Assume they will be pushed again if
2936 necessary, and continue to the process_stratum. */
2937 if (t->to_stratum == thread_stratum
2938 || t->to_stratum == arch_stratum)
2939 continue;
2940
2941 error (_("The \"%s\" target does not support \"run\". "
2942 "Try \"help target\" or \"continue\"."),
2943 t->to_shortname);
2944 }
2945
2946 /* This function is only called if the target is running. In that
2947 case there should have been a process_stratum target and it
2948 should either know how to create inferiors, or not... */
2949 internal_error (__FILE__, __LINE__, _("No targets found"));
2950}
2951
2952/* Look through the list of possible targets for a target that can
2953 execute a run or attach command without any other data. This is
2954 used to locate the default process stratum.
2955
2956 If DO_MESG is not NULL, the result is always valid (error() is
2957 called for errors); else, return NULL on error. */
2958
2959static struct target_ops *
2960find_default_run_target (char *do_mesg)
2961{
2962 struct target_ops **t;
2963 struct target_ops *runable = NULL;
2964 int count;
2965
2966 count = 0;
2967
2968 for (t = target_structs; t < target_structs + target_struct_size;
2969 ++t)
2970 {
2971 if ((*t)->to_can_run && target_can_run (*t))
2972 {
2973 runable = *t;
2974 ++count;
2975 }
2976 }
2977
2978 if (count != 1)
2979 {
2980 if (do_mesg)
2981 error (_("Don't know how to %s. Try \"help target\"."), do_mesg);
2982 else
2983 return NULL;
2984 }
2985
2986 return runable;
2987}
2988
2989void
2990find_default_attach (struct target_ops *ops, char *args, int from_tty)
2991{
2992 struct target_ops *t;
2993
2994 t = find_default_run_target ("attach");
2995 (t->to_attach) (t, args, from_tty);
2996 return;
2997}
2998
2999void
3000find_default_create_inferior (struct target_ops *ops,
3001 char *exec_file, char *allargs, char **env,
3002 int from_tty)
3003{
3004 struct target_ops *t;
3005
3006 t = find_default_run_target ("run");
3007 (t->to_create_inferior) (t, exec_file, allargs, env, from_tty);
3008 return;
3009}
3010
3011static int
3012find_default_can_async_p (struct target_ops *ignore)
3013{
3014 struct target_ops *t;
3015
3016 /* This may be called before the target is pushed on the stack;
3017 look for the default process stratum. If there's none, gdb isn't
3018 configured with a native debugger, and target remote isn't
3019 connected yet. */
3020 t = find_default_run_target (NULL);
3021 if (t && t->to_can_async_p != delegate_can_async_p)
3022 return (t->to_can_async_p) (t);
3023 return 0;
3024}
3025
3026static int
3027find_default_is_async_p (struct target_ops *ignore)
3028{
3029 struct target_ops *t;
3030
3031 /* This may be called before the target is pushed on the stack;
3032 look for the default process stratum. If there's none, gdb isn't
3033 configured with a native debugger, and target remote isn't
3034 connected yet. */
3035 t = find_default_run_target (NULL);
3036 if (t && t->to_is_async_p != delegate_is_async_p)
3037 return (t->to_is_async_p) (t);
3038 return 0;
3039}
3040
3041static int
3042find_default_supports_non_stop (struct target_ops *self)
3043{
3044 struct target_ops *t;
3045
3046 t = find_default_run_target (NULL);
3047 if (t && t->to_supports_non_stop)
3048 return (t->to_supports_non_stop) (t);
3049 return 0;
3050}
3051
3052int
3053target_supports_non_stop (void)
3054{
3055 struct target_ops *t;
3056
3057 for (t = &current_target; t != NULL; t = t->beneath)
3058 if (t->to_supports_non_stop)
3059 return t->to_supports_non_stop (t);
3060
3061 return 0;
3062}
3063
3064/* Implement the "info proc" command. */
3065
3066int
3067target_info_proc (char *args, enum info_proc_what what)
3068{
3069 struct target_ops *t;
3070
3071 /* If we're already connected to something that can get us OS
3072 related data, use it. Otherwise, try using the native
3073 target. */
3074 if (current_target.to_stratum >= process_stratum)
3075 t = current_target.beneath;
3076 else
3077 t = find_default_run_target (NULL);
3078
3079 for (; t != NULL; t = t->beneath)
3080 {
3081 if (t->to_info_proc != NULL)
3082 {
3083 t->to_info_proc (t, args, what);
3084
3085 if (targetdebug)
3086 fprintf_unfiltered (gdb_stdlog,
3087 "target_info_proc (\"%s\", %d)\n", args, what);
3088
3089 return 1;
3090 }
3091 }
3092
3093 return 0;
3094}
3095
3096static int
3097find_default_supports_disable_randomization (struct target_ops *self)
3098{
3099 struct target_ops *t;
3100
3101 t = find_default_run_target (NULL);
3102 if (t && t->to_supports_disable_randomization)
3103 return (t->to_supports_disable_randomization) (t);
3104 return 0;
3105}
3106
3107int
3108target_supports_disable_randomization (void)
3109{
3110 struct target_ops *t;
3111
3112 for (t = &current_target; t != NULL; t = t->beneath)
3113 if (t->to_supports_disable_randomization)
3114 return t->to_supports_disable_randomization (t);
3115
3116 return 0;
3117}
3118
3119char *
3120target_get_osdata (const char *type)
3121{
3122 struct target_ops *t;
3123
3124 /* If we're already connected to something that can get us OS
3125 related data, use it. Otherwise, try using the native
3126 target. */
3127 if (current_target.to_stratum >= process_stratum)
3128 t = current_target.beneath;
3129 else
3130 t = find_default_run_target ("get OS data");
3131
3132 if (!t)
3133 return NULL;
3134
3135 return target_read_stralloc (t, TARGET_OBJECT_OSDATA, type);
3136}
3137
3138/* Determine the current address space of thread PTID. */
3139
3140struct address_space *
3141target_thread_address_space (ptid_t ptid)
3142{
3143 struct address_space *aspace;
3144 struct inferior *inf;
3145 struct target_ops *t;
3146
3147 for (t = current_target.beneath; t != NULL; t = t->beneath)
3148 {
3149 if (t->to_thread_address_space != NULL)
3150 {
3151 aspace = t->to_thread_address_space (t, ptid);
3152 gdb_assert (aspace);
3153
3154 if (targetdebug)
3155 fprintf_unfiltered (gdb_stdlog,
3156 "target_thread_address_space (%s) = %d\n",
3157 target_pid_to_str (ptid),
3158 address_space_num (aspace));
3159 return aspace;
3160 }
3161 }
3162
3163 /* Fall-back to the "main" address space of the inferior. */
3164 inf = find_inferior_pid (ptid_get_pid (ptid));
3165
3166 if (inf == NULL || inf->aspace == NULL)
3167 internal_error (__FILE__, __LINE__,
3168 _("Can't determine the current "
3169 "address space of thread %s\n"),
3170 target_pid_to_str (ptid));
3171
3172 return inf->aspace;
3173}
3174
3175
3176/* Target file operations. */
3177
3178static struct target_ops *
3179default_fileio_target (void)
3180{
3181 /* If we're already connected to something that can perform
3182 file I/O, use it. Otherwise, try using the native target. */
3183 if (current_target.to_stratum >= process_stratum)
3184 return current_target.beneath;
3185 else
3186 return find_default_run_target ("file I/O");
3187}
3188
3189/* Open FILENAME on the target, using FLAGS and MODE. Return a
3190 target file descriptor, or -1 if an error occurs (and set
3191 *TARGET_ERRNO). */
3192int
3193target_fileio_open (const char *filename, int flags, int mode,
3194 int *target_errno)
3195{
3196 struct target_ops *t;
3197
3198 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3199 {
3200 if (t->to_fileio_open != NULL)
3201 {
3202 int fd = t->to_fileio_open (t, filename, flags, mode, target_errno);
3203
3204 if (targetdebug)
3205 fprintf_unfiltered (gdb_stdlog,
3206 "target_fileio_open (%s,0x%x,0%o) = %d (%d)\n",
3207 filename, flags, mode,
3208 fd, fd != -1 ? 0 : *target_errno);
3209 return fd;
3210 }
3211 }
3212
3213 *target_errno = FILEIO_ENOSYS;
3214 return -1;
3215}
3216
3217/* Write up to LEN bytes from WRITE_BUF to FD on the target.
3218 Return the number of bytes written, or -1 if an error occurs
3219 (and set *TARGET_ERRNO). */
3220int
3221target_fileio_pwrite (int fd, const gdb_byte *write_buf, int len,
3222 ULONGEST offset, int *target_errno)
3223{
3224 struct target_ops *t;
3225
3226 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3227 {
3228 if (t->to_fileio_pwrite != NULL)
3229 {
3230 int ret = t->to_fileio_pwrite (t, fd, write_buf, len, offset,
3231 target_errno);
3232
3233 if (targetdebug)
3234 fprintf_unfiltered (gdb_stdlog,
3235 "target_fileio_pwrite (%d,...,%d,%s) "
3236 "= %d (%d)\n",
3237 fd, len, pulongest (offset),
3238 ret, ret != -1 ? 0 : *target_errno);
3239 return ret;
3240 }
3241 }
3242
3243 *target_errno = FILEIO_ENOSYS;
3244 return -1;
3245}
3246
3247/* Read up to LEN bytes FD on the target into READ_BUF.
3248 Return the number of bytes read, or -1 if an error occurs
3249 (and set *TARGET_ERRNO). */
3250int
3251target_fileio_pread (int fd, gdb_byte *read_buf, int len,
3252 ULONGEST offset, int *target_errno)
3253{
3254 struct target_ops *t;
3255
3256 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3257 {
3258 if (t->to_fileio_pread != NULL)
3259 {
3260 int ret = t->to_fileio_pread (t, fd, read_buf, len, offset,
3261 target_errno);
3262
3263 if (targetdebug)
3264 fprintf_unfiltered (gdb_stdlog,
3265 "target_fileio_pread (%d,...,%d,%s) "
3266 "= %d (%d)\n",
3267 fd, len, pulongest (offset),
3268 ret, ret != -1 ? 0 : *target_errno);
3269 return ret;
3270 }
3271 }
3272
3273 *target_errno = FILEIO_ENOSYS;
3274 return -1;
3275}
3276
3277/* Close FD on the target. Return 0, or -1 if an error occurs
3278 (and set *TARGET_ERRNO). */
3279int
3280target_fileio_close (int fd, int *target_errno)
3281{
3282 struct target_ops *t;
3283
3284 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3285 {
3286 if (t->to_fileio_close != NULL)
3287 {
3288 int ret = t->to_fileio_close (t, fd, target_errno);
3289
3290 if (targetdebug)
3291 fprintf_unfiltered (gdb_stdlog,
3292 "target_fileio_close (%d) = %d (%d)\n",
3293 fd, ret, ret != -1 ? 0 : *target_errno);
3294 return ret;
3295 }
3296 }
3297
3298 *target_errno = FILEIO_ENOSYS;
3299 return -1;
3300}
3301
3302/* Unlink FILENAME on the target. Return 0, or -1 if an error
3303 occurs (and set *TARGET_ERRNO). */
3304int
3305target_fileio_unlink (const char *filename, int *target_errno)
3306{
3307 struct target_ops *t;
3308
3309 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3310 {
3311 if (t->to_fileio_unlink != NULL)
3312 {
3313 int ret = t->to_fileio_unlink (t, filename, target_errno);
3314
3315 if (targetdebug)
3316 fprintf_unfiltered (gdb_stdlog,
3317 "target_fileio_unlink (%s) = %d (%d)\n",
3318 filename, ret, ret != -1 ? 0 : *target_errno);
3319 return ret;
3320 }
3321 }
3322
3323 *target_errno = FILEIO_ENOSYS;
3324 return -1;
3325}
3326
3327/* Read value of symbolic link FILENAME on the target. Return a
3328 null-terminated string allocated via xmalloc, or NULL if an error
3329 occurs (and set *TARGET_ERRNO). */
3330char *
3331target_fileio_readlink (const char *filename, int *target_errno)
3332{
3333 struct target_ops *t;
3334
3335 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3336 {
3337 if (t->to_fileio_readlink != NULL)
3338 {
3339 char *ret = t->to_fileio_readlink (t, filename, target_errno);
3340
3341 if (targetdebug)
3342 fprintf_unfiltered (gdb_stdlog,
3343 "target_fileio_readlink (%s) = %s (%d)\n",
3344 filename, ret? ret : "(nil)",
3345 ret? 0 : *target_errno);
3346 return ret;
3347 }
3348 }
3349
3350 *target_errno = FILEIO_ENOSYS;
3351 return NULL;
3352}
3353
3354static void
3355target_fileio_close_cleanup (void *opaque)
3356{
3357 int fd = *(int *) opaque;
3358 int target_errno;
3359
3360 target_fileio_close (fd, &target_errno);
3361}
3362
3363/* Read target file FILENAME. Store the result in *BUF_P and
3364 return the size of the transferred data. PADDING additional bytes are
3365 available in *BUF_P. This is a helper function for
3366 target_fileio_read_alloc; see the declaration of that function for more
3367 information. */
3368
3369static LONGEST
3370target_fileio_read_alloc_1 (const char *filename,
3371 gdb_byte **buf_p, int padding)
3372{
3373 struct cleanup *close_cleanup;
3374 size_t buf_alloc, buf_pos;
3375 gdb_byte *buf;
3376 LONGEST n;
3377 int fd;
3378 int target_errno;
3379
3380 fd = target_fileio_open (filename, FILEIO_O_RDONLY, 0700, &target_errno);
3381 if (fd == -1)
3382 return -1;
3383
3384 close_cleanup = make_cleanup (target_fileio_close_cleanup, &fd);
3385
3386 /* Start by reading up to 4K at a time. The target will throttle
3387 this number down if necessary. */
3388 buf_alloc = 4096;
3389 buf = xmalloc (buf_alloc);
3390 buf_pos = 0;
3391 while (1)
3392 {
3393 n = target_fileio_pread (fd, &buf[buf_pos],
3394 buf_alloc - buf_pos - padding, buf_pos,
3395 &target_errno);
3396 if (n < 0)
3397 {
3398 /* An error occurred. */
3399 do_cleanups (close_cleanup);
3400 xfree (buf);
3401 return -1;
3402 }
3403 else if (n == 0)
3404 {
3405 /* Read all there was. */
3406 do_cleanups (close_cleanup);
3407 if (buf_pos == 0)
3408 xfree (buf);
3409 else
3410 *buf_p = buf;
3411 return buf_pos;
3412 }
3413
3414 buf_pos += n;
3415
3416 /* If the buffer is filling up, expand it. */
3417 if (buf_alloc < buf_pos * 2)
3418 {
3419 buf_alloc *= 2;
3420 buf = xrealloc (buf, buf_alloc);
3421 }
3422
3423 QUIT;
3424 }
3425}
3426
3427/* Read target file FILENAME. Store the result in *BUF_P and return
3428 the size of the transferred data. See the declaration in "target.h"
3429 function for more information about the return value. */
3430
3431LONGEST
3432target_fileio_read_alloc (const char *filename, gdb_byte **buf_p)
3433{
3434 return target_fileio_read_alloc_1 (filename, buf_p, 0);
3435}
3436
3437/* Read target file FILENAME. The result is NUL-terminated and
3438 returned as a string, allocated using xmalloc. If an error occurs
3439 or the transfer is unsupported, NULL is returned. Empty objects
3440 are returned as allocated but empty strings. A warning is issued
3441 if the result contains any embedded NUL bytes. */
3442
3443char *
3444target_fileio_read_stralloc (const char *filename)
3445{
3446 gdb_byte *buffer;
3447 char *bufstr;
3448 LONGEST i, transferred;
3449
3450 transferred = target_fileio_read_alloc_1 (filename, &buffer, 1);
3451 bufstr = (char *) buffer;
3452
3453 if (transferred < 0)
3454 return NULL;
3455
3456 if (transferred == 0)
3457 return xstrdup ("");
3458
3459 bufstr[transferred] = 0;
3460
3461 /* Check for embedded NUL bytes; but allow trailing NULs. */
3462 for (i = strlen (bufstr); i < transferred; i++)
3463 if (bufstr[i] != 0)
3464 {
3465 warning (_("target file %s "
3466 "contained unexpected null characters"),
3467 filename);
3468 break;
3469 }
3470
3471 return bufstr;
3472}
3473
3474
3475static int
3476default_region_ok_for_hw_watchpoint (struct target_ops *self,
3477 CORE_ADDR addr, int len)
3478{
3479 return (len <= gdbarch_ptr_bit (target_gdbarch ()) / TARGET_CHAR_BIT);
3480}
3481
3482static int
3483default_watchpoint_addr_within_range (struct target_ops *target,
3484 CORE_ADDR addr,
3485 CORE_ADDR start, int length)
3486{
3487 return addr >= start && addr < start + length;
3488}
3489
3490static struct gdbarch *
3491default_thread_architecture (struct target_ops *ops, ptid_t ptid)
3492{
3493 return target_gdbarch ();
3494}
3495
3496static int
3497return_zero (void)
3498{
3499 return 0;
3500}
3501
3502static void *
3503return_null (void)
3504{
3505 return 0;
3506}
3507
3508/*
3509 * Find the next target down the stack from the specified target.
3510 */
3511
3512struct target_ops *
3513find_target_beneath (struct target_ops *t)
3514{
3515 return t->beneath;
3516}
3517
3518/* See target.h. */
3519
3520struct target_ops *
3521find_target_at (enum strata stratum)
3522{
3523 struct target_ops *t;
3524
3525 for (t = current_target.beneath; t != NULL; t = t->beneath)
3526 if (t->to_stratum == stratum)
3527 return t;
3528
3529 return NULL;
3530}
3531
3532\f
3533/* The inferior process has died. Long live the inferior! */
3534
3535void
3536generic_mourn_inferior (void)
3537{
3538 ptid_t ptid;
3539
3540 ptid = inferior_ptid;
3541 inferior_ptid = null_ptid;
3542
3543 /* Mark breakpoints uninserted in case something tries to delete a
3544 breakpoint while we delete the inferior's threads (which would
3545 fail, since the inferior is long gone). */
3546 mark_breakpoints_out ();
3547
3548 if (!ptid_equal (ptid, null_ptid))
3549 {
3550 int pid = ptid_get_pid (ptid);
3551 exit_inferior (pid);
3552 }
3553
3554 /* Note this wipes step-resume breakpoints, so needs to be done
3555 after exit_inferior, which ends up referencing the step-resume
3556 breakpoints through clear_thread_inferior_resources. */
3557 breakpoint_init_inferior (inf_exited);
3558
3559 registers_changed ();
3560
3561 reopen_exec_file ();
3562 reinit_frame_cache ();
3563
3564 if (deprecated_detach_hook)
3565 deprecated_detach_hook ();
3566}
3567\f
3568/* Convert a normal process ID to a string. Returns the string in a
3569 static buffer. */
3570
3571char *
3572normal_pid_to_str (ptid_t ptid)
3573{
3574 static char buf[32];
3575
3576 xsnprintf (buf, sizeof buf, "process %d", ptid_get_pid (ptid));
3577 return buf;
3578}
3579
3580static char *
3581dummy_pid_to_str (struct target_ops *ops, ptid_t ptid)
3582{
3583 return normal_pid_to_str (ptid);
3584}
3585
3586/* Error-catcher for target_find_memory_regions. */
3587static int
3588dummy_find_memory_regions (struct target_ops *self,
3589 find_memory_region_ftype ignore1, void *ignore2)
3590{
3591 error (_("Command not implemented for this target."));
3592 return 0;
3593}
3594
3595/* Error-catcher for target_make_corefile_notes. */
3596static char *
3597dummy_make_corefile_notes (struct target_ops *self,
3598 bfd *ignore1, int *ignore2)
3599{
3600 error (_("Command not implemented for this target."));
3601 return NULL;
3602}
3603
3604/* Set up the handful of non-empty slots needed by the dummy target
3605 vector. */
3606
3607static void
3608init_dummy_target (void)
3609{
3610 dummy_target.to_shortname = "None";
3611 dummy_target.to_longname = "None";
3612 dummy_target.to_doc = "";
3613 dummy_target.to_create_inferior = find_default_create_inferior;
3614 dummy_target.to_supports_non_stop = find_default_supports_non_stop;
3615 dummy_target.to_supports_disable_randomization
3616 = find_default_supports_disable_randomization;
3617 dummy_target.to_pid_to_str = dummy_pid_to_str;
3618 dummy_target.to_stratum = dummy_stratum;
3619 dummy_target.to_has_all_memory = (int (*) (struct target_ops *)) return_zero;
3620 dummy_target.to_has_memory = (int (*) (struct target_ops *)) return_zero;
3621 dummy_target.to_has_stack = (int (*) (struct target_ops *)) return_zero;
3622 dummy_target.to_has_registers = (int (*) (struct target_ops *)) return_zero;
3623 dummy_target.to_has_execution
3624 = (int (*) (struct target_ops *, ptid_t)) return_zero;
3625 dummy_target.to_magic = OPS_MAGIC;
3626
3627 install_dummy_methods (&dummy_target);
3628}
3629\f
3630static void
3631debug_to_open (char *args, int from_tty)
3632{
3633 debug_target.to_open (args, from_tty);
3634
3635 fprintf_unfiltered (gdb_stdlog, "target_open (%s, %d)\n", args, from_tty);
3636}
3637
3638void
3639target_close (struct target_ops *targ)
3640{
3641 gdb_assert (!target_is_pushed (targ));
3642
3643 if (targ->to_xclose != NULL)
3644 targ->to_xclose (targ);
3645 else if (targ->to_close != NULL)
3646 targ->to_close (targ);
3647
3648 if (targetdebug)
3649 fprintf_unfiltered (gdb_stdlog, "target_close ()\n");
3650}
3651
3652void
3653target_attach (char *args, int from_tty)
3654{
3655 current_target.to_attach (&current_target, args, from_tty);
3656 if (targetdebug)
3657 fprintf_unfiltered (gdb_stdlog, "target_attach (%s, %d)\n",
3658 args, from_tty);
3659}
3660
3661int
3662target_thread_alive (ptid_t ptid)
3663{
3664 struct target_ops *t;
3665
3666 for (t = current_target.beneath; t != NULL; t = t->beneath)
3667 {
3668 if (t->to_thread_alive != NULL)
3669 {
3670 int retval;
3671
3672 retval = t->to_thread_alive (t, ptid);
3673 if (targetdebug)
3674 fprintf_unfiltered (gdb_stdlog, "target_thread_alive (%d) = %d\n",
3675 ptid_get_pid (ptid), retval);
3676
3677 return retval;
3678 }
3679 }
3680
3681 return 0;
3682}
3683
3684void
3685target_find_new_threads (void)
3686{
3687 struct target_ops *t;
3688
3689 for (t = current_target.beneath; t != NULL; t = t->beneath)
3690 {
3691 if (t->to_find_new_threads != NULL)
3692 {
3693 t->to_find_new_threads (t);
3694 if (targetdebug)
3695 fprintf_unfiltered (gdb_stdlog, "target_find_new_threads ()\n");
3696
3697 return;
3698 }
3699 }
3700}
3701
3702void
3703target_stop (ptid_t ptid)
3704{
3705 if (!may_stop)
3706 {
3707 warning (_("May not interrupt or stop the target, ignoring attempt"));
3708 return;
3709 }
3710
3711 (*current_target.to_stop) (&current_target, ptid);
3712}
3713
3714static void
3715debug_to_post_attach (struct target_ops *self, int pid)
3716{
3717 debug_target.to_post_attach (&debug_target, pid);
3718
3719 fprintf_unfiltered (gdb_stdlog, "target_post_attach (%d)\n", pid);
3720}
3721
3722/* Concatenate ELEM to LIST, a comma separate list, and return the
3723 result. The LIST incoming argument is released. */
3724
3725static char *
3726str_comma_list_concat_elem (char *list, const char *elem)
3727{
3728 if (list == NULL)
3729 return xstrdup (elem);
3730 else
3731 return reconcat (list, list, ", ", elem, (char *) NULL);
3732}
3733
3734/* Helper for target_options_to_string. If OPT is present in
3735 TARGET_OPTIONS, append the OPT_STR (string version of OPT) in RET.
3736 Returns the new resulting string. OPT is removed from
3737 TARGET_OPTIONS. */
3738
3739static char *
3740do_option (int *target_options, char *ret,
3741 int opt, char *opt_str)
3742{
3743 if ((*target_options & opt) != 0)
3744 {
3745 ret = str_comma_list_concat_elem (ret, opt_str);
3746 *target_options &= ~opt;
3747 }
3748
3749 return ret;
3750}
3751
3752char *
3753target_options_to_string (int target_options)
3754{
3755 char *ret = NULL;
3756
3757#define DO_TARG_OPTION(OPT) \
3758 ret = do_option (&target_options, ret, OPT, #OPT)
3759
3760 DO_TARG_OPTION (TARGET_WNOHANG);
3761
3762 if (target_options != 0)
3763 ret = str_comma_list_concat_elem (ret, "unknown???");
3764
3765 if (ret == NULL)
3766 ret = xstrdup ("");
3767 return ret;
3768}
3769
3770static void
3771debug_print_register (const char * func,
3772 struct regcache *regcache, int regno)
3773{
3774 struct gdbarch *gdbarch = get_regcache_arch (regcache);
3775
3776 fprintf_unfiltered (gdb_stdlog, "%s ", func);
3777 if (regno >= 0 && regno < gdbarch_num_regs (gdbarch)
3778 && gdbarch_register_name (gdbarch, regno) != NULL
3779 && gdbarch_register_name (gdbarch, regno)[0] != '\0')
3780 fprintf_unfiltered (gdb_stdlog, "(%s)",
3781 gdbarch_register_name (gdbarch, regno));
3782 else
3783 fprintf_unfiltered (gdb_stdlog, "(%d)", regno);
3784 if (regno >= 0 && regno < gdbarch_num_regs (gdbarch))
3785 {
3786 enum bfd_endian byte_order = gdbarch_byte_order (gdbarch);
3787 int i, size = register_size (gdbarch, regno);
3788 gdb_byte buf[MAX_REGISTER_SIZE];
3789
3790 regcache_raw_collect (regcache, regno, buf);
3791 fprintf_unfiltered (gdb_stdlog, " = ");
3792 for (i = 0; i < size; i++)
3793 {
3794 fprintf_unfiltered (gdb_stdlog, "%02x", buf[i]);
3795 }
3796 if (size <= sizeof (LONGEST))
3797 {
3798 ULONGEST val = extract_unsigned_integer (buf, size, byte_order);
3799
3800 fprintf_unfiltered (gdb_stdlog, " %s %s",
3801 core_addr_to_string_nz (val), plongest (val));
3802 }
3803 }
3804 fprintf_unfiltered (gdb_stdlog, "\n");
3805}
3806
3807void
3808target_fetch_registers (struct regcache *regcache, int regno)
3809{
3810 struct target_ops *t;
3811
3812 for (t = current_target.beneath; t != NULL; t = t->beneath)
3813 {
3814 if (t->to_fetch_registers != NULL)
3815 {
3816 t->to_fetch_registers (t, regcache, regno);
3817 if (targetdebug)
3818 debug_print_register ("target_fetch_registers", regcache, regno);
3819 return;
3820 }
3821 }
3822}
3823
3824void
3825target_store_registers (struct regcache *regcache, int regno)
3826{
3827 struct target_ops *t;
3828
3829 if (!may_write_registers)
3830 error (_("Writing to registers is not allowed (regno %d)"), regno);
3831
3832 current_target.to_store_registers (&current_target, regcache, regno);
3833 if (targetdebug)
3834 {
3835 debug_print_register ("target_store_registers", regcache, regno);
3836 }
3837}
3838
3839int
3840target_core_of_thread (ptid_t ptid)
3841{
3842 struct target_ops *t;
3843
3844 for (t = current_target.beneath; t != NULL; t = t->beneath)
3845 {
3846 if (t->to_core_of_thread != NULL)
3847 {
3848 int retval = t->to_core_of_thread (t, ptid);
3849
3850 if (targetdebug)
3851 fprintf_unfiltered (gdb_stdlog,
3852 "target_core_of_thread (%d) = %d\n",
3853 ptid_get_pid (ptid), retval);
3854 return retval;
3855 }
3856 }
3857
3858 return -1;
3859}
3860
3861int
3862target_verify_memory (const gdb_byte *data, CORE_ADDR memaddr, ULONGEST size)
3863{
3864 struct target_ops *t;
3865
3866 for (t = current_target.beneath; t != NULL; t = t->beneath)
3867 {
3868 if (t->to_verify_memory != NULL)
3869 {
3870 int retval = t->to_verify_memory (t, data, memaddr, size);
3871
3872 if (targetdebug)
3873 fprintf_unfiltered (gdb_stdlog,
3874 "target_verify_memory (%s, %s) = %d\n",
3875 paddress (target_gdbarch (), memaddr),
3876 pulongest (size),
3877 retval);
3878 return retval;
3879 }
3880 }
3881
3882 tcomplain ();
3883}
3884
3885/* The documentation for this function is in its prototype declaration in
3886 target.h. */
3887
3888int
3889target_insert_mask_watchpoint (CORE_ADDR addr, CORE_ADDR mask, int rw)
3890{
3891 struct target_ops *t;
3892
3893 for (t = current_target.beneath; t != NULL; t = t->beneath)
3894 if (t->to_insert_mask_watchpoint != NULL)
3895 {
3896 int ret;
3897
3898 ret = t->to_insert_mask_watchpoint (t, addr, mask, rw);
3899
3900 if (targetdebug)
3901 fprintf_unfiltered (gdb_stdlog, "\
3902target_insert_mask_watchpoint (%s, %s, %d) = %d\n",
3903 core_addr_to_string (addr),
3904 core_addr_to_string (mask), rw, ret);
3905
3906 return ret;
3907 }
3908
3909 return 1;
3910}
3911
3912/* The documentation for this function is in its prototype declaration in
3913 target.h. */
3914
3915int
3916target_remove_mask_watchpoint (CORE_ADDR addr, CORE_ADDR mask, int rw)
3917{
3918 struct target_ops *t;
3919
3920 for (t = current_target.beneath; t != NULL; t = t->beneath)
3921 if (t->to_remove_mask_watchpoint != NULL)
3922 {
3923 int ret;
3924
3925 ret = t->to_remove_mask_watchpoint (t, addr, mask, rw);
3926
3927 if (targetdebug)
3928 fprintf_unfiltered (gdb_stdlog, "\
3929target_remove_mask_watchpoint (%s, %s, %d) = %d\n",
3930 core_addr_to_string (addr),
3931 core_addr_to_string (mask), rw, ret);
3932
3933 return ret;
3934 }
3935
3936 return 1;
3937}
3938
3939/* The documentation for this function is in its prototype declaration
3940 in target.h. */
3941
3942int
3943target_masked_watch_num_registers (CORE_ADDR addr, CORE_ADDR mask)
3944{
3945 struct target_ops *t;
3946
3947 for (t = current_target.beneath; t != NULL; t = t->beneath)
3948 if (t->to_masked_watch_num_registers != NULL)
3949 return t->to_masked_watch_num_registers (t, addr, mask);
3950
3951 return -1;
3952}
3953
3954/* The documentation for this function is in its prototype declaration
3955 in target.h. */
3956
3957int
3958target_ranged_break_num_registers (void)
3959{
3960 struct target_ops *t;
3961
3962 for (t = current_target.beneath; t != NULL; t = t->beneath)
3963 if (t->to_ranged_break_num_registers != NULL)
3964 return t->to_ranged_break_num_registers (t);
3965
3966 return -1;
3967}
3968
3969/* See target.h. */
3970
3971struct btrace_target_info *
3972target_enable_btrace (ptid_t ptid)
3973{
3974 struct target_ops *t;
3975
3976 for (t = current_target.beneath; t != NULL; t = t->beneath)
3977 if (t->to_enable_btrace != NULL)
3978 return t->to_enable_btrace (t, ptid);
3979
3980 tcomplain ();
3981 return NULL;
3982}
3983
3984/* See target.h. */
3985
3986void
3987target_disable_btrace (struct btrace_target_info *btinfo)
3988{
3989 struct target_ops *t;
3990
3991 for (t = current_target.beneath; t != NULL; t = t->beneath)
3992 if (t->to_disable_btrace != NULL)
3993 {
3994 t->to_disable_btrace (t, btinfo);
3995 return;
3996 }
3997
3998 tcomplain ();
3999}
4000
4001/* See target.h. */
4002
4003void
4004target_teardown_btrace (struct btrace_target_info *btinfo)
4005{
4006 struct target_ops *t;
4007
4008 for (t = current_target.beneath; t != NULL; t = t->beneath)
4009 if (t->to_teardown_btrace != NULL)
4010 {
4011 t->to_teardown_btrace (t, btinfo);
4012 return;
4013 }
4014
4015 tcomplain ();
4016}
4017
4018/* See target.h. */
4019
4020enum btrace_error
4021target_read_btrace (VEC (btrace_block_s) **btrace,
4022 struct btrace_target_info *btinfo,
4023 enum btrace_read_type type)
4024{
4025 struct target_ops *t;
4026
4027 for (t = current_target.beneath; t != NULL; t = t->beneath)
4028 if (t->to_read_btrace != NULL)
4029 return t->to_read_btrace (t, btrace, btinfo, type);
4030
4031 tcomplain ();
4032 return BTRACE_ERR_NOT_SUPPORTED;
4033}
4034
4035/* See target.h. */
4036
4037void
4038target_stop_recording (void)
4039{
4040 struct target_ops *t;
4041
4042 for (t = current_target.beneath; t != NULL; t = t->beneath)
4043 if (t->to_stop_recording != NULL)
4044 {
4045 t->to_stop_recording (t);
4046 return;
4047 }
4048
4049 /* This is optional. */
4050}
4051
4052/* See target.h. */
4053
4054void
4055target_info_record (void)
4056{
4057 struct target_ops *t;
4058
4059 for (t = current_target.beneath; t != NULL; t = t->beneath)
4060 if (t->to_info_record != NULL)
4061 {
4062 t->to_info_record (t);
4063 return;
4064 }
4065
4066 tcomplain ();
4067}
4068
4069/* See target.h. */
4070
4071void
4072target_save_record (const char *filename)
4073{
4074 struct target_ops *t;
4075
4076 for (t = current_target.beneath; t != NULL; t = t->beneath)
4077 if (t->to_save_record != NULL)
4078 {
4079 t->to_save_record (t, filename);
4080 return;
4081 }
4082
4083 tcomplain ();
4084}
4085
4086/* See target.h. */
4087
4088int
4089target_supports_delete_record (void)
4090{
4091 struct target_ops *t;
4092
4093 for (t = current_target.beneath; t != NULL; t = t->beneath)
4094 if (t->to_delete_record != NULL)
4095 return 1;
4096
4097 return 0;
4098}
4099
4100/* See target.h. */
4101
4102void
4103target_delete_record (void)
4104{
4105 struct target_ops *t;
4106
4107 for (t = current_target.beneath; t != NULL; t = t->beneath)
4108 if (t->to_delete_record != NULL)
4109 {
4110 t->to_delete_record (t);
4111 return;
4112 }
4113
4114 tcomplain ();
4115}
4116
4117/* See target.h. */
4118
4119int
4120target_record_is_replaying (void)
4121{
4122 struct target_ops *t;
4123
4124 for (t = current_target.beneath; t != NULL; t = t->beneath)
4125 if (t->to_record_is_replaying != NULL)
4126 return t->to_record_is_replaying (t);
4127
4128 return 0;
4129}
4130
4131/* See target.h. */
4132
4133void
4134target_goto_record_begin (void)
4135{
4136 struct target_ops *t;
4137
4138 for (t = current_target.beneath; t != NULL; t = t->beneath)
4139 if (t->to_goto_record_begin != NULL)
4140 {
4141 t->to_goto_record_begin (t);
4142 return;
4143 }
4144
4145 tcomplain ();
4146}
4147
4148/* See target.h. */
4149
4150void
4151target_goto_record_end (void)
4152{
4153 struct target_ops *t;
4154
4155 for (t = current_target.beneath; t != NULL; t = t->beneath)
4156 if (t->to_goto_record_end != NULL)
4157 {
4158 t->to_goto_record_end (t);
4159 return;
4160 }
4161
4162 tcomplain ();
4163}
4164
4165/* See target.h. */
4166
4167void
4168target_goto_record (ULONGEST insn)
4169{
4170 struct target_ops *t;
4171
4172 for (t = current_target.beneath; t != NULL; t = t->beneath)
4173 if (t->to_goto_record != NULL)
4174 {
4175 t->to_goto_record (t, insn);
4176 return;
4177 }
4178
4179 tcomplain ();
4180}
4181
4182/* See target.h. */
4183
4184void
4185target_insn_history (int size, int flags)
4186{
4187 struct target_ops *t;
4188
4189 for (t = current_target.beneath; t != NULL; t = t->beneath)
4190 if (t->to_insn_history != NULL)
4191 {
4192 t->to_insn_history (t, size, flags);
4193 return;
4194 }
4195
4196 tcomplain ();
4197}
4198
4199/* See target.h. */
4200
4201void
4202target_insn_history_from (ULONGEST from, int size, int flags)
4203{
4204 struct target_ops *t;
4205
4206 for (t = current_target.beneath; t != NULL; t = t->beneath)
4207 if (t->to_insn_history_from != NULL)
4208 {
4209 t->to_insn_history_from (t, from, size, flags);
4210 return;
4211 }
4212
4213 tcomplain ();
4214}
4215
4216/* See target.h. */
4217
4218void
4219target_insn_history_range (ULONGEST begin, ULONGEST end, int flags)
4220{
4221 struct target_ops *t;
4222
4223 for (t = current_target.beneath; t != NULL; t = t->beneath)
4224 if (t->to_insn_history_range != NULL)
4225 {
4226 t->to_insn_history_range (t, begin, end, flags);
4227 return;
4228 }
4229
4230 tcomplain ();
4231}
4232
4233/* See target.h. */
4234
4235void
4236target_call_history (int size, int flags)
4237{
4238 struct target_ops *t;
4239
4240 for (t = current_target.beneath; t != NULL; t = t->beneath)
4241 if (t->to_call_history != NULL)
4242 {
4243 t->to_call_history (t, size, flags);
4244 return;
4245 }
4246
4247 tcomplain ();
4248}
4249
4250/* See target.h. */
4251
4252void
4253target_call_history_from (ULONGEST begin, int size, int flags)
4254{
4255 struct target_ops *t;
4256
4257 for (t = current_target.beneath; t != NULL; t = t->beneath)
4258 if (t->to_call_history_from != NULL)
4259 {
4260 t->to_call_history_from (t, begin, size, flags);
4261 return;
4262 }
4263
4264 tcomplain ();
4265}
4266
4267/* See target.h. */
4268
4269void
4270target_call_history_range (ULONGEST begin, ULONGEST end, int flags)
4271{
4272 struct target_ops *t;
4273
4274 for (t = current_target.beneath; t != NULL; t = t->beneath)
4275 if (t->to_call_history_range != NULL)
4276 {
4277 t->to_call_history_range (t, begin, end, flags);
4278 return;
4279 }
4280
4281 tcomplain ();
4282}
4283
4284static void
4285debug_to_prepare_to_store (struct target_ops *self, struct regcache *regcache)
4286{
4287 debug_target.to_prepare_to_store (&debug_target, regcache);
4288
4289 fprintf_unfiltered (gdb_stdlog, "target_prepare_to_store ()\n");
4290}
4291
4292/* See target.h. */
4293
4294const struct frame_unwind *
4295target_get_unwinder (void)
4296{
4297 struct target_ops *t;
4298
4299 for (t = current_target.beneath; t != NULL; t = t->beneath)
4300 if (t->to_get_unwinder != NULL)
4301 return t->to_get_unwinder;
4302
4303 return NULL;
4304}
4305
4306/* See target.h. */
4307
4308const struct frame_unwind *
4309target_get_tailcall_unwinder (void)
4310{
4311 struct target_ops *t;
4312
4313 for (t = current_target.beneath; t != NULL; t = t->beneath)
4314 if (t->to_get_tailcall_unwinder != NULL)
4315 return t->to_get_tailcall_unwinder;
4316
4317 return NULL;
4318}
4319
4320/* See target.h. */
4321
4322CORE_ADDR
4323forward_target_decr_pc_after_break (struct target_ops *ops,
4324 struct gdbarch *gdbarch)
4325{
4326 for (; ops != NULL; ops = ops->beneath)
4327 if (ops->to_decr_pc_after_break != NULL)
4328 return ops->to_decr_pc_after_break (ops, gdbarch);
4329
4330 return gdbarch_decr_pc_after_break (gdbarch);
4331}
4332
4333/* See target.h. */
4334
4335CORE_ADDR
4336target_decr_pc_after_break (struct gdbarch *gdbarch)
4337{
4338 return forward_target_decr_pc_after_break (current_target.beneath, gdbarch);
4339}
4340
4341static int
4342deprecated_debug_xfer_memory (CORE_ADDR memaddr, bfd_byte *myaddr, int len,
4343 int write, struct mem_attrib *attrib,
4344 struct target_ops *target)
4345{
4346 int retval;
4347
4348 retval = debug_target.deprecated_xfer_memory (memaddr, myaddr, len, write,
4349 attrib, target);
4350
4351 fprintf_unfiltered (gdb_stdlog,
4352 "target_xfer_memory (%s, xxx, %d, %s, xxx) = %d",
4353 paddress (target_gdbarch (), memaddr), len,
4354 write ? "write" : "read", retval);
4355
4356 if (retval > 0)
4357 {
4358 int i;
4359
4360 fputs_unfiltered (", bytes =", gdb_stdlog);
4361 for (i = 0; i < retval; i++)
4362 {
4363 if ((((intptr_t) &(myaddr[i])) & 0xf) == 0)
4364 {
4365 if (targetdebug < 2 && i > 0)
4366 {
4367 fprintf_unfiltered (gdb_stdlog, " ...");
4368 break;
4369 }
4370 fprintf_unfiltered (gdb_stdlog, "\n");
4371 }
4372
4373 fprintf_unfiltered (gdb_stdlog, " %02x", myaddr[i] & 0xff);
4374 }
4375 }
4376
4377 fputc_unfiltered ('\n', gdb_stdlog);
4378
4379 return retval;
4380}
4381
4382static void
4383debug_to_files_info (struct target_ops *target)
4384{
4385 debug_target.to_files_info (target);
4386
4387 fprintf_unfiltered (gdb_stdlog, "target_files_info (xxx)\n");
4388}
4389
4390static int
4391debug_to_insert_breakpoint (struct target_ops *ops, struct gdbarch *gdbarch,
4392 struct bp_target_info *bp_tgt)
4393{
4394 int retval;
4395
4396 retval = debug_target.to_insert_breakpoint (&debug_target, gdbarch, bp_tgt);
4397
4398 fprintf_unfiltered (gdb_stdlog,
4399 "target_insert_breakpoint (%s, xxx) = %ld\n",
4400 core_addr_to_string (bp_tgt->placed_address),
4401 (unsigned long) retval);
4402 return retval;
4403}
4404
4405static int
4406debug_to_remove_breakpoint (struct target_ops *ops, struct gdbarch *gdbarch,
4407 struct bp_target_info *bp_tgt)
4408{
4409 int retval;
4410
4411 retval = debug_target.to_remove_breakpoint (&debug_target, gdbarch, bp_tgt);
4412
4413 fprintf_unfiltered (gdb_stdlog,
4414 "target_remove_breakpoint (%s, xxx) = %ld\n",
4415 core_addr_to_string (bp_tgt->placed_address),
4416 (unsigned long) retval);
4417 return retval;
4418}
4419
4420static int
4421debug_to_can_use_hw_breakpoint (struct target_ops *self,
4422 int type, int cnt, int from_tty)
4423{
4424 int retval;
4425
4426 retval = debug_target.to_can_use_hw_breakpoint (&debug_target,
4427 type, cnt, from_tty);
4428
4429 fprintf_unfiltered (gdb_stdlog,
4430 "target_can_use_hw_breakpoint (%ld, %ld, %ld) = %ld\n",
4431 (unsigned long) type,
4432 (unsigned long) cnt,
4433 (unsigned long) from_tty,
4434 (unsigned long) retval);
4435 return retval;
4436}
4437
4438static int
4439debug_to_region_ok_for_hw_watchpoint (struct target_ops *self,
4440 CORE_ADDR addr, int len)
4441{
4442 CORE_ADDR retval;
4443
4444 retval = debug_target.to_region_ok_for_hw_watchpoint (&debug_target,
4445 addr, len);
4446
4447 fprintf_unfiltered (gdb_stdlog,
4448 "target_region_ok_for_hw_watchpoint (%s, %ld) = %s\n",
4449 core_addr_to_string (addr), (unsigned long) len,
4450 core_addr_to_string (retval));
4451 return retval;
4452}
4453
4454static int
4455debug_to_can_accel_watchpoint_condition (struct target_ops *self,
4456 CORE_ADDR addr, int len, int rw,
4457 struct expression *cond)
4458{
4459 int retval;
4460
4461 retval = debug_target.to_can_accel_watchpoint_condition (&debug_target,
4462 addr, len,
4463 rw, cond);
4464
4465 fprintf_unfiltered (gdb_stdlog,
4466 "target_can_accel_watchpoint_condition "
4467 "(%s, %d, %d, %s) = %ld\n",
4468 core_addr_to_string (addr), len, rw,
4469 host_address_to_string (cond), (unsigned long) retval);
4470 return retval;
4471}
4472
4473static int
4474debug_to_stopped_by_watchpoint (struct target_ops *ops)
4475{
4476 int retval;
4477
4478 retval = debug_target.to_stopped_by_watchpoint (&debug_target);
4479
4480 fprintf_unfiltered (gdb_stdlog,
4481 "target_stopped_by_watchpoint () = %ld\n",
4482 (unsigned long) retval);
4483 return retval;
4484}
4485
4486static int
4487debug_to_stopped_data_address (struct target_ops *target, CORE_ADDR *addr)
4488{
4489 int retval;
4490
4491 retval = debug_target.to_stopped_data_address (target, addr);
4492
4493 fprintf_unfiltered (gdb_stdlog,
4494 "target_stopped_data_address ([%s]) = %ld\n",
4495 core_addr_to_string (*addr),
4496 (unsigned long)retval);
4497 return retval;
4498}
4499
4500static int
4501debug_to_watchpoint_addr_within_range (struct target_ops *target,
4502 CORE_ADDR addr,
4503 CORE_ADDR start, int length)
4504{
4505 int retval;
4506
4507 retval = debug_target.to_watchpoint_addr_within_range (target, addr,
4508 start, length);
4509
4510 fprintf_filtered (gdb_stdlog,
4511 "target_watchpoint_addr_within_range (%s, %s, %d) = %d\n",
4512 core_addr_to_string (addr), core_addr_to_string (start),
4513 length, retval);
4514 return retval;
4515}
4516
4517static int
4518debug_to_insert_hw_breakpoint (struct target_ops *self,
4519 struct gdbarch *gdbarch,
4520 struct bp_target_info *bp_tgt)
4521{
4522 int retval;
4523
4524 retval = debug_target.to_insert_hw_breakpoint (&debug_target,
4525 gdbarch, bp_tgt);
4526
4527 fprintf_unfiltered (gdb_stdlog,
4528 "target_insert_hw_breakpoint (%s, xxx) = %ld\n",
4529 core_addr_to_string (bp_tgt->placed_address),
4530 (unsigned long) retval);
4531 return retval;
4532}
4533
4534static int
4535debug_to_remove_hw_breakpoint (struct target_ops *self,
4536 struct gdbarch *gdbarch,
4537 struct bp_target_info *bp_tgt)
4538{
4539 int retval;
4540
4541 retval = debug_target.to_remove_hw_breakpoint (&debug_target,
4542 gdbarch, bp_tgt);
4543
4544 fprintf_unfiltered (gdb_stdlog,
4545 "target_remove_hw_breakpoint (%s, xxx) = %ld\n",
4546 core_addr_to_string (bp_tgt->placed_address),
4547 (unsigned long) retval);
4548 return retval;
4549}
4550
4551static int
4552debug_to_insert_watchpoint (struct target_ops *self,
4553 CORE_ADDR addr, int len, int type,
4554 struct expression *cond)
4555{
4556 int retval;
4557
4558 retval = debug_target.to_insert_watchpoint (&debug_target,
4559 addr, len, type, cond);
4560
4561 fprintf_unfiltered (gdb_stdlog,
4562 "target_insert_watchpoint (%s, %d, %d, %s) = %ld\n",
4563 core_addr_to_string (addr), len, type,
4564 host_address_to_string (cond), (unsigned long) retval);
4565 return retval;
4566}
4567
4568static int
4569debug_to_remove_watchpoint (struct target_ops *self,
4570 CORE_ADDR addr, int len, int type,
4571 struct expression *cond)
4572{
4573 int retval;
4574
4575 retval = debug_target.to_remove_watchpoint (&debug_target,
4576 addr, len, type, cond);
4577
4578 fprintf_unfiltered (gdb_stdlog,
4579 "target_remove_watchpoint (%s, %d, %d, %s) = %ld\n",
4580 core_addr_to_string (addr), len, type,
4581 host_address_to_string (cond), (unsigned long) retval);
4582 return retval;
4583}
4584
4585static void
4586debug_to_terminal_init (struct target_ops *self)
4587{
4588 debug_target.to_terminal_init (&debug_target);
4589
4590 fprintf_unfiltered (gdb_stdlog, "target_terminal_init ()\n");
4591}
4592
4593static void
4594debug_to_terminal_inferior (struct target_ops *self)
4595{
4596 debug_target.to_terminal_inferior (&debug_target);
4597
4598 fprintf_unfiltered (gdb_stdlog, "target_terminal_inferior ()\n");
4599}
4600
4601static void
4602debug_to_terminal_ours_for_output (struct target_ops *self)
4603{
4604 debug_target.to_terminal_ours_for_output (&debug_target);
4605
4606 fprintf_unfiltered (gdb_stdlog, "target_terminal_ours_for_output ()\n");
4607}
4608
4609static void
4610debug_to_terminal_ours (struct target_ops *self)
4611{
4612 debug_target.to_terminal_ours (&debug_target);
4613
4614 fprintf_unfiltered (gdb_stdlog, "target_terminal_ours ()\n");
4615}
4616
4617static void
4618debug_to_terminal_save_ours (struct target_ops *self)
4619{
4620 debug_target.to_terminal_save_ours (&debug_target);
4621
4622 fprintf_unfiltered (gdb_stdlog, "target_terminal_save_ours ()\n");
4623}
4624
4625static void
4626debug_to_terminal_info (struct target_ops *self,
4627 const char *arg, int from_tty)
4628{
4629 debug_target.to_terminal_info (&debug_target, arg, from_tty);
4630
4631 fprintf_unfiltered (gdb_stdlog, "target_terminal_info (%s, %d)\n", arg,
4632 from_tty);
4633}
4634
4635static void
4636debug_to_load (struct target_ops *self, char *args, int from_tty)
4637{
4638 debug_target.to_load (&debug_target, args, from_tty);
4639
4640 fprintf_unfiltered (gdb_stdlog, "target_load (%s, %d)\n", args, from_tty);
4641}
4642
4643static void
4644debug_to_post_startup_inferior (struct target_ops *self, ptid_t ptid)
4645{
4646 debug_target.to_post_startup_inferior (&debug_target, ptid);
4647
4648 fprintf_unfiltered (gdb_stdlog, "target_post_startup_inferior (%d)\n",
4649 ptid_get_pid (ptid));
4650}
4651
4652static int
4653debug_to_insert_fork_catchpoint (struct target_ops *self, int pid)
4654{
4655 int retval;
4656
4657 retval = debug_target.to_insert_fork_catchpoint (&debug_target, pid);
4658
4659 fprintf_unfiltered (gdb_stdlog, "target_insert_fork_catchpoint (%d) = %d\n",
4660 pid, retval);
4661
4662 return retval;
4663}
4664
4665static int
4666debug_to_remove_fork_catchpoint (struct target_ops *self, int pid)
4667{
4668 int retval;
4669
4670 retval = debug_target.to_remove_fork_catchpoint (&debug_target, pid);
4671
4672 fprintf_unfiltered (gdb_stdlog, "target_remove_fork_catchpoint (%d) = %d\n",
4673 pid, retval);
4674
4675 return retval;
4676}
4677
4678static int
4679debug_to_insert_vfork_catchpoint (struct target_ops *self, int pid)
4680{
4681 int retval;
4682
4683 retval = debug_target.to_insert_vfork_catchpoint (&debug_target, pid);
4684
4685 fprintf_unfiltered (gdb_stdlog, "target_insert_vfork_catchpoint (%d) = %d\n",
4686 pid, retval);
4687
4688 return retval;
4689}
4690
4691static int
4692debug_to_remove_vfork_catchpoint (struct target_ops *self, int pid)
4693{
4694 int retval;
4695
4696 retval = debug_target.to_remove_vfork_catchpoint (&debug_target, pid);
4697
4698 fprintf_unfiltered (gdb_stdlog, "target_remove_vfork_catchpoint (%d) = %d\n",
4699 pid, retval);
4700
4701 return retval;
4702}
4703
4704static int
4705debug_to_insert_exec_catchpoint (struct target_ops *self, int pid)
4706{
4707 int retval;
4708
4709 retval = debug_target.to_insert_exec_catchpoint (&debug_target, pid);
4710
4711 fprintf_unfiltered (gdb_stdlog, "target_insert_exec_catchpoint (%d) = %d\n",
4712 pid, retval);
4713
4714 return retval;
4715}
4716
4717static int
4718debug_to_remove_exec_catchpoint (struct target_ops *self, int pid)
4719{
4720 int retval;
4721
4722 retval = debug_target.to_remove_exec_catchpoint (&debug_target, pid);
4723
4724 fprintf_unfiltered (gdb_stdlog, "target_remove_exec_catchpoint (%d) = %d\n",
4725 pid, retval);
4726
4727 return retval;
4728}
4729
4730static int
4731debug_to_has_exited (struct target_ops *self,
4732 int pid, int wait_status, int *exit_status)
4733{
4734 int has_exited;
4735
4736 has_exited = debug_target.to_has_exited (&debug_target,
4737 pid, wait_status, exit_status);
4738
4739 fprintf_unfiltered (gdb_stdlog, "target_has_exited (%d, %d, %d) = %d\n",
4740 pid, wait_status, *exit_status, has_exited);
4741
4742 return has_exited;
4743}
4744
4745static int
4746debug_to_can_run (struct target_ops *self)
4747{
4748 int retval;
4749
4750 retval = debug_target.to_can_run (&debug_target);
4751
4752 fprintf_unfiltered (gdb_stdlog, "target_can_run () = %d\n", retval);
4753
4754 return retval;
4755}
4756
4757static struct gdbarch *
4758debug_to_thread_architecture (struct target_ops *ops, ptid_t ptid)
4759{
4760 struct gdbarch *retval;
4761
4762 retval = debug_target.to_thread_architecture (ops, ptid);
4763
4764 fprintf_unfiltered (gdb_stdlog,
4765 "target_thread_architecture (%s) = %s [%s]\n",
4766 target_pid_to_str (ptid),
4767 host_address_to_string (retval),
4768 gdbarch_bfd_arch_info (retval)->printable_name);
4769 return retval;
4770}
4771
4772static void
4773debug_to_stop (struct target_ops *self, ptid_t ptid)
4774{
4775 debug_target.to_stop (&debug_target, ptid);
4776
4777 fprintf_unfiltered (gdb_stdlog, "target_stop (%s)\n",
4778 target_pid_to_str (ptid));
4779}
4780
4781static void
4782debug_to_rcmd (struct target_ops *self, char *command,
4783 struct ui_file *outbuf)
4784{
4785 debug_target.to_rcmd (&debug_target, command, outbuf);
4786 fprintf_unfiltered (gdb_stdlog, "target_rcmd (%s, ...)\n", command);
4787}
4788
4789static char *
4790debug_to_pid_to_exec_file (struct target_ops *self, int pid)
4791{
4792 char *exec_file;
4793
4794 exec_file = debug_target.to_pid_to_exec_file (&debug_target, pid);
4795
4796 fprintf_unfiltered (gdb_stdlog, "target_pid_to_exec_file (%d) = %s\n",
4797 pid, exec_file);
4798
4799 return exec_file;
4800}
4801
4802static void
4803setup_target_debug (void)
4804{
4805 memcpy (&debug_target, &current_target, sizeof debug_target);
4806
4807 current_target.to_open = debug_to_open;
4808 current_target.to_post_attach = debug_to_post_attach;
4809 current_target.to_prepare_to_store = debug_to_prepare_to_store;
4810 current_target.deprecated_xfer_memory = deprecated_debug_xfer_memory;
4811 current_target.to_files_info = debug_to_files_info;
4812 current_target.to_insert_breakpoint = debug_to_insert_breakpoint;
4813 current_target.to_remove_breakpoint = debug_to_remove_breakpoint;
4814 current_target.to_can_use_hw_breakpoint = debug_to_can_use_hw_breakpoint;
4815 current_target.to_insert_hw_breakpoint = debug_to_insert_hw_breakpoint;
4816 current_target.to_remove_hw_breakpoint = debug_to_remove_hw_breakpoint;
4817 current_target.to_insert_watchpoint = debug_to_insert_watchpoint;
4818 current_target.to_remove_watchpoint = debug_to_remove_watchpoint;
4819 current_target.to_stopped_by_watchpoint = debug_to_stopped_by_watchpoint;
4820 current_target.to_stopped_data_address = debug_to_stopped_data_address;
4821 current_target.to_watchpoint_addr_within_range
4822 = debug_to_watchpoint_addr_within_range;
4823 current_target.to_region_ok_for_hw_watchpoint
4824 = debug_to_region_ok_for_hw_watchpoint;
4825 current_target.to_can_accel_watchpoint_condition
4826 = debug_to_can_accel_watchpoint_condition;
4827 current_target.to_terminal_init = debug_to_terminal_init;
4828 current_target.to_terminal_inferior = debug_to_terminal_inferior;
4829 current_target.to_terminal_ours_for_output
4830 = debug_to_terminal_ours_for_output;
4831 current_target.to_terminal_ours = debug_to_terminal_ours;
4832 current_target.to_terminal_save_ours = debug_to_terminal_save_ours;
4833 current_target.to_terminal_info = debug_to_terminal_info;
4834 current_target.to_load = debug_to_load;
4835 current_target.to_post_startup_inferior = debug_to_post_startup_inferior;
4836 current_target.to_insert_fork_catchpoint = debug_to_insert_fork_catchpoint;
4837 current_target.to_remove_fork_catchpoint = debug_to_remove_fork_catchpoint;
4838 current_target.to_insert_vfork_catchpoint = debug_to_insert_vfork_catchpoint;
4839 current_target.to_remove_vfork_catchpoint = debug_to_remove_vfork_catchpoint;
4840 current_target.to_insert_exec_catchpoint = debug_to_insert_exec_catchpoint;
4841 current_target.to_remove_exec_catchpoint = debug_to_remove_exec_catchpoint;
4842 current_target.to_has_exited = debug_to_has_exited;
4843 current_target.to_can_run = debug_to_can_run;
4844 current_target.to_stop = debug_to_stop;
4845 current_target.to_rcmd = debug_to_rcmd;
4846 current_target.to_pid_to_exec_file = debug_to_pid_to_exec_file;
4847 current_target.to_thread_architecture = debug_to_thread_architecture;
4848}
4849\f
4850
4851static char targ_desc[] =
4852"Names of targets and files being debugged.\nShows the entire \
4853stack of targets currently in use (including the exec-file,\n\
4854core-file, and process, if any), as well as the symbol file name.";
4855
4856static void
4857default_rcmd (struct target_ops *self, char *command, struct ui_file *output)
4858{
4859 error (_("\"monitor\" command not supported by this target."));
4860}
4861
4862static void
4863do_monitor_command (char *cmd,
4864 int from_tty)
4865{
4866 target_rcmd (cmd, gdb_stdtarg);
4867}
4868
4869/* Print the name of each layers of our target stack. */
4870
4871static void
4872maintenance_print_target_stack (char *cmd, int from_tty)
4873{
4874 struct target_ops *t;
4875
4876 printf_filtered (_("The current target stack is:\n"));
4877
4878 for (t = target_stack; t != NULL; t = t->beneath)
4879 {
4880 printf_filtered (" - %s (%s)\n", t->to_shortname, t->to_longname);
4881 }
4882}
4883
4884/* Controls if async mode is permitted. */
4885int target_async_permitted = 0;
4886
4887/* The set command writes to this variable. If the inferior is
4888 executing, target_async_permitted is *not* updated. */
4889static int target_async_permitted_1 = 0;
4890
4891static void
4892set_target_async_command (char *args, int from_tty,
4893 struct cmd_list_element *c)
4894{
4895 if (have_live_inferiors ())
4896 {
4897 target_async_permitted_1 = target_async_permitted;
4898 error (_("Cannot change this setting while the inferior is running."));
4899 }
4900
4901 target_async_permitted = target_async_permitted_1;
4902}
4903
4904static void
4905show_target_async_command (struct ui_file *file, int from_tty,
4906 struct cmd_list_element *c,
4907 const char *value)
4908{
4909 fprintf_filtered (file,
4910 _("Controlling the inferior in "
4911 "asynchronous mode is %s.\n"), value);
4912}
4913
4914/* Temporary copies of permission settings. */
4915
4916static int may_write_registers_1 = 1;
4917static int may_write_memory_1 = 1;
4918static int may_insert_breakpoints_1 = 1;
4919static int may_insert_tracepoints_1 = 1;
4920static int may_insert_fast_tracepoints_1 = 1;
4921static int may_stop_1 = 1;
4922
4923/* Make the user-set values match the real values again. */
4924
4925void
4926update_target_permissions (void)
4927{
4928 may_write_registers_1 = may_write_registers;
4929 may_write_memory_1 = may_write_memory;
4930 may_insert_breakpoints_1 = may_insert_breakpoints;
4931 may_insert_tracepoints_1 = may_insert_tracepoints;
4932 may_insert_fast_tracepoints_1 = may_insert_fast_tracepoints;
4933 may_stop_1 = may_stop;
4934}
4935
4936/* The one function handles (most of) the permission flags in the same
4937 way. */
4938
4939static void
4940set_target_permissions (char *args, int from_tty,
4941 struct cmd_list_element *c)
4942{
4943 if (target_has_execution)
4944 {
4945 update_target_permissions ();
4946 error (_("Cannot change this setting while the inferior is running."));
4947 }
4948
4949 /* Make the real values match the user-changed values. */
4950 may_write_registers = may_write_registers_1;
4951 may_insert_breakpoints = may_insert_breakpoints_1;
4952 may_insert_tracepoints = may_insert_tracepoints_1;
4953 may_insert_fast_tracepoints = may_insert_fast_tracepoints_1;
4954 may_stop = may_stop_1;
4955 update_observer_mode ();
4956}
4957
4958/* Set memory write permission independently of observer mode. */
4959
4960static void
4961set_write_memory_permission (char *args, int from_tty,
4962 struct cmd_list_element *c)
4963{
4964 /* Make the real values match the user-changed values. */
4965 may_write_memory = may_write_memory_1;
4966 update_observer_mode ();
4967}
4968
4969
4970void
4971initialize_targets (void)
4972{
4973 init_dummy_target ();
4974 push_target (&dummy_target);
4975
4976 add_info ("target", target_info, targ_desc);
4977 add_info ("files", target_info, targ_desc);
4978
4979 add_setshow_zuinteger_cmd ("target", class_maintenance, &targetdebug, _("\
4980Set target debugging."), _("\
4981Show target debugging."), _("\
4982When non-zero, target debugging is enabled. Higher numbers are more\n\
4983verbose. Changes do not take effect until the next \"run\" or \"target\"\n\
4984command."),
4985 NULL,
4986 show_targetdebug,
4987 &setdebuglist, &showdebuglist);
4988
4989 add_setshow_boolean_cmd ("trust-readonly-sections", class_support,
4990 &trust_readonly, _("\
4991Set mode for reading from readonly sections."), _("\
4992Show mode for reading from readonly sections."), _("\
4993When this mode is on, memory reads from readonly sections (such as .text)\n\
4994will be read from the object file instead of from the target. This will\n\
4995result in significant performance improvement for remote targets."),
4996 NULL,
4997 show_trust_readonly,
4998 &setlist, &showlist);
4999
5000 add_com ("monitor", class_obscure, do_monitor_command,
5001 _("Send a command to the remote monitor (remote targets only)."));
5002
5003 add_cmd ("target-stack", class_maintenance, maintenance_print_target_stack,
5004 _("Print the name of each layer of the internal target stack."),
5005 &maintenanceprintlist);
5006
5007 add_setshow_boolean_cmd ("target-async", no_class,
5008 &target_async_permitted_1, _("\
5009Set whether gdb controls the inferior in asynchronous mode."), _("\
5010Show whether gdb controls the inferior in asynchronous mode."), _("\
5011Tells gdb whether to control the inferior in asynchronous mode."),
5012 set_target_async_command,
5013 show_target_async_command,
5014 &setlist,
5015 &showlist);
5016
5017 add_setshow_boolean_cmd ("may-write-registers", class_support,
5018 &may_write_registers_1, _("\
5019Set permission to write into registers."), _("\
5020Show permission to write into registers."), _("\
5021When this permission is on, GDB may write into the target's registers.\n\
5022Otherwise, any sort of write attempt will result in an error."),
5023 set_target_permissions, NULL,
5024 &setlist, &showlist);
5025
5026 add_setshow_boolean_cmd ("may-write-memory", class_support,
5027 &may_write_memory_1, _("\
5028Set permission to write into target memory."), _("\
5029Show permission to write into target memory."), _("\
5030When this permission is on, GDB may write into the target's memory.\n\
5031Otherwise, any sort of write attempt will result in an error."),
5032 set_write_memory_permission, NULL,
5033 &setlist, &showlist);
5034
5035 add_setshow_boolean_cmd ("may-insert-breakpoints", class_support,
5036 &may_insert_breakpoints_1, _("\
5037Set permission to insert breakpoints in the target."), _("\
5038Show permission to insert breakpoints in the target."), _("\
5039When this permission is on, GDB may insert breakpoints in the program.\n\
5040Otherwise, any sort of insertion attempt will result in an error."),
5041 set_target_permissions, NULL,
5042 &setlist, &showlist);
5043
5044 add_setshow_boolean_cmd ("may-insert-tracepoints", class_support,
5045 &may_insert_tracepoints_1, _("\
5046Set permission to insert tracepoints in the target."), _("\
5047Show permission to insert tracepoints in the target."), _("\
5048When this permission is on, GDB may insert tracepoints in the program.\n\
5049Otherwise, any sort of insertion attempt will result in an error."),
5050 set_target_permissions, NULL,
5051 &setlist, &showlist);
5052
5053 add_setshow_boolean_cmd ("may-insert-fast-tracepoints", class_support,
5054 &may_insert_fast_tracepoints_1, _("\
5055Set permission to insert fast tracepoints in the target."), _("\
5056Show permission to insert fast tracepoints in the target."), _("\
5057When this permission is on, GDB may insert fast tracepoints.\n\
5058Otherwise, any sort of insertion attempt will result in an error."),
5059 set_target_permissions, NULL,
5060 &setlist, &showlist);
5061
5062 add_setshow_boolean_cmd ("may-interrupt", class_support,
5063 &may_stop_1, _("\
5064Set permission to interrupt or signal the target."), _("\
5065Show permission to interrupt or signal the target."), _("\
5066When this permission is on, GDB may interrupt/stop the target's execution.\n\
5067Otherwise, any attempt to interrupt or stop will be ignored."),
5068 set_target_permissions, NULL,
5069 &setlist, &showlist);
5070}
This page took 0.039163 seconds and 4 git commands to generate.