convert to_get_min_fast_tracepoint_insn_len
[binutils-gdb.git] / gdb / target.c
1 /* Select target systems and architectures at runtime for GDB.
2
3 Copyright (C) 1990-2014 Free Software Foundation, Inc.
4
5 Contributed by Cygnus Support.
6
7 This file is part of GDB.
8
9 This program is free software; you can redistribute it and/or modify
10 it under the terms of the GNU General Public License as published by
11 the Free Software Foundation; either version 3 of the License, or
12 (at your option) any later version.
13
14 This program is distributed in the hope that it will be useful,
15 but WITHOUT ANY WARRANTY; without even the implied warranty of
16 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
17 GNU General Public License for more details.
18
19 You should have received a copy of the GNU General Public License
20 along with this program. If not, see <http://www.gnu.org/licenses/>. */
21
22 #include "defs.h"
23 #include <errno.h>
24 #include <string.h>
25 #include "target.h"
26 #include "target-dcache.h"
27 #include "gdbcmd.h"
28 #include "symtab.h"
29 #include "inferior.h"
30 #include "bfd.h"
31 #include "symfile.h"
32 #include "objfiles.h"
33 #include "dcache.h"
34 #include <signal.h>
35 #include "regcache.h"
36 #include "gdb_assert.h"
37 #include "gdbcore.h"
38 #include "exceptions.h"
39 #include "target-descriptions.h"
40 #include "gdbthread.h"
41 #include "solib.h"
42 #include "exec.h"
43 #include "inline-frame.h"
44 #include "tracepoint.h"
45 #include "gdb/fileio.h"
46 #include "agent.h"
47
48 static void target_info (char *, int);
49
50 static void default_terminal_info (struct target_ops *, const char *, int);
51
52 static int default_watchpoint_addr_within_range (struct target_ops *,
53 CORE_ADDR, CORE_ADDR, int);
54
55 static int default_region_ok_for_hw_watchpoint (struct target_ops *,
56 CORE_ADDR, int);
57
58 static void default_rcmd (struct target_ops *, char *, struct ui_file *);
59
60 static ptid_t default_get_ada_task_ptid (struct target_ops *self,
61 long lwp, long tid);
62
63 static void tcomplain (void) ATTRIBUTE_NORETURN;
64
65 static int nomemory (CORE_ADDR, char *, int, int, struct target_ops *);
66
67 static int return_zero (void);
68
69 static void *return_null (void);
70
71 void target_ignore (void);
72
73 static void target_command (char *, int);
74
75 static struct target_ops *find_default_run_target (char *);
76
77 static target_xfer_partial_ftype default_xfer_partial;
78
79 static struct gdbarch *default_thread_architecture (struct target_ops *ops,
80 ptid_t ptid);
81
82 static int dummy_find_memory_regions (struct target_ops *self,
83 find_memory_region_ftype ignore1,
84 void *ignore2);
85
86 static char *dummy_make_corefile_notes (struct target_ops *self,
87 bfd *ignore1, int *ignore2);
88
89 static int find_default_can_async_p (struct target_ops *ignore);
90
91 static int find_default_is_async_p (struct target_ops *ignore);
92
93 static enum exec_direction_kind default_execution_direction
94 (struct target_ops *self);
95
96 #include "target-delegates.c"
97
98 static void init_dummy_target (void);
99
100 static struct target_ops debug_target;
101
102 static void debug_to_open (char *, int);
103
104 static void debug_to_prepare_to_store (struct target_ops *self,
105 struct regcache *);
106
107 static void debug_to_files_info (struct target_ops *);
108
109 static int debug_to_insert_breakpoint (struct target_ops *, struct gdbarch *,
110 struct bp_target_info *);
111
112 static int debug_to_remove_breakpoint (struct target_ops *, struct gdbarch *,
113 struct bp_target_info *);
114
115 static int debug_to_can_use_hw_breakpoint (struct target_ops *self,
116 int, int, int);
117
118 static int debug_to_insert_hw_breakpoint (struct target_ops *self,
119 struct gdbarch *,
120 struct bp_target_info *);
121
122 static int debug_to_remove_hw_breakpoint (struct target_ops *self,
123 struct gdbarch *,
124 struct bp_target_info *);
125
126 static int debug_to_insert_watchpoint (struct target_ops *self,
127 CORE_ADDR, int, int,
128 struct expression *);
129
130 static int debug_to_remove_watchpoint (struct target_ops *self,
131 CORE_ADDR, int, int,
132 struct expression *);
133
134 static int debug_to_stopped_data_address (struct target_ops *, CORE_ADDR *);
135
136 static int debug_to_watchpoint_addr_within_range (struct target_ops *,
137 CORE_ADDR, CORE_ADDR, int);
138
139 static int debug_to_region_ok_for_hw_watchpoint (struct target_ops *self,
140 CORE_ADDR, int);
141
142 static int debug_to_can_accel_watchpoint_condition (struct target_ops *self,
143 CORE_ADDR, int, int,
144 struct expression *);
145
146 static void debug_to_terminal_init (struct target_ops *self);
147
148 static void debug_to_terminal_inferior (struct target_ops *self);
149
150 static void debug_to_terminal_ours_for_output (struct target_ops *self);
151
152 static void debug_to_terminal_save_ours (struct target_ops *self);
153
154 static void debug_to_terminal_ours (struct target_ops *self);
155
156 static void debug_to_load (struct target_ops *self, char *, int);
157
158 static int debug_to_can_run (struct target_ops *self);
159
160 static void debug_to_stop (struct target_ops *self, ptid_t);
161
162 /* Pointer to array of target architecture structures; the size of the
163 array; the current index into the array; the allocated size of the
164 array. */
165 struct target_ops **target_structs;
166 unsigned target_struct_size;
167 unsigned target_struct_allocsize;
168 #define DEFAULT_ALLOCSIZE 10
169
170 /* The initial current target, so that there is always a semi-valid
171 current target. */
172
173 static struct target_ops dummy_target;
174
175 /* Top of target stack. */
176
177 static struct target_ops *target_stack;
178
179 /* The target structure we are currently using to talk to a process
180 or file or whatever "inferior" we have. */
181
182 struct target_ops current_target;
183
184 /* Command list for target. */
185
186 static struct cmd_list_element *targetlist = NULL;
187
188 /* Nonzero if we should trust readonly sections from the
189 executable when reading memory. */
190
191 static int trust_readonly = 0;
192
193 /* Nonzero if we should show true memory content including
194 memory breakpoint inserted by gdb. */
195
196 static int show_memory_breakpoints = 0;
197
198 /* These globals control whether GDB attempts to perform these
199 operations; they are useful for targets that need to prevent
200 inadvertant disruption, such as in non-stop mode. */
201
202 int may_write_registers = 1;
203
204 int may_write_memory = 1;
205
206 int may_insert_breakpoints = 1;
207
208 int may_insert_tracepoints = 1;
209
210 int may_insert_fast_tracepoints = 1;
211
212 int may_stop = 1;
213
214 /* Non-zero if we want to see trace of target level stuff. */
215
216 static unsigned int targetdebug = 0;
217 static void
218 show_targetdebug (struct ui_file *file, int from_tty,
219 struct cmd_list_element *c, const char *value)
220 {
221 fprintf_filtered (file, _("Target debugging is %s.\n"), value);
222 }
223
224 static void setup_target_debug (void);
225
226 /* The user just typed 'target' without the name of a target. */
227
228 static void
229 target_command (char *arg, int from_tty)
230 {
231 fputs_filtered ("Argument required (target name). Try `help target'\n",
232 gdb_stdout);
233 }
234
235 /* Default target_has_* methods for process_stratum targets. */
236
237 int
238 default_child_has_all_memory (struct target_ops *ops)
239 {
240 /* If no inferior selected, then we can't read memory here. */
241 if (ptid_equal (inferior_ptid, null_ptid))
242 return 0;
243
244 return 1;
245 }
246
247 int
248 default_child_has_memory (struct target_ops *ops)
249 {
250 /* If no inferior selected, then we can't read memory here. */
251 if (ptid_equal (inferior_ptid, null_ptid))
252 return 0;
253
254 return 1;
255 }
256
257 int
258 default_child_has_stack (struct target_ops *ops)
259 {
260 /* If no inferior selected, there's no stack. */
261 if (ptid_equal (inferior_ptid, null_ptid))
262 return 0;
263
264 return 1;
265 }
266
267 int
268 default_child_has_registers (struct target_ops *ops)
269 {
270 /* Can't read registers from no inferior. */
271 if (ptid_equal (inferior_ptid, null_ptid))
272 return 0;
273
274 return 1;
275 }
276
277 int
278 default_child_has_execution (struct target_ops *ops, ptid_t the_ptid)
279 {
280 /* If there's no thread selected, then we can't make it run through
281 hoops. */
282 if (ptid_equal (the_ptid, null_ptid))
283 return 0;
284
285 return 1;
286 }
287
288
289 int
290 target_has_all_memory_1 (void)
291 {
292 struct target_ops *t;
293
294 for (t = current_target.beneath; t != NULL; t = t->beneath)
295 if (t->to_has_all_memory (t))
296 return 1;
297
298 return 0;
299 }
300
301 int
302 target_has_memory_1 (void)
303 {
304 struct target_ops *t;
305
306 for (t = current_target.beneath; t != NULL; t = t->beneath)
307 if (t->to_has_memory (t))
308 return 1;
309
310 return 0;
311 }
312
313 int
314 target_has_stack_1 (void)
315 {
316 struct target_ops *t;
317
318 for (t = current_target.beneath; t != NULL; t = t->beneath)
319 if (t->to_has_stack (t))
320 return 1;
321
322 return 0;
323 }
324
325 int
326 target_has_registers_1 (void)
327 {
328 struct target_ops *t;
329
330 for (t = current_target.beneath; t != NULL; t = t->beneath)
331 if (t->to_has_registers (t))
332 return 1;
333
334 return 0;
335 }
336
337 int
338 target_has_execution_1 (ptid_t the_ptid)
339 {
340 struct target_ops *t;
341
342 for (t = current_target.beneath; t != NULL; t = t->beneath)
343 if (t->to_has_execution (t, the_ptid))
344 return 1;
345
346 return 0;
347 }
348
349 int
350 target_has_execution_current (void)
351 {
352 return target_has_execution_1 (inferior_ptid);
353 }
354
355 /* Complete initialization of T. This ensures that various fields in
356 T are set, if needed by the target implementation. */
357
358 void
359 complete_target_initialization (struct target_ops *t)
360 {
361 /* Provide default values for all "must have" methods. */
362 if (t->to_xfer_partial == NULL)
363 t->to_xfer_partial = default_xfer_partial;
364
365 if (t->to_has_all_memory == NULL)
366 t->to_has_all_memory = (int (*) (struct target_ops *)) return_zero;
367
368 if (t->to_has_memory == NULL)
369 t->to_has_memory = (int (*) (struct target_ops *)) return_zero;
370
371 if (t->to_has_stack == NULL)
372 t->to_has_stack = (int (*) (struct target_ops *)) return_zero;
373
374 if (t->to_has_registers == NULL)
375 t->to_has_registers = (int (*) (struct target_ops *)) return_zero;
376
377 if (t->to_has_execution == NULL)
378 t->to_has_execution = (int (*) (struct target_ops *, ptid_t)) return_zero;
379
380 install_delegators (t);
381 }
382
383 /* Add possible target architecture T to the list and add a new
384 command 'target T->to_shortname'. Set COMPLETER as the command's
385 completer if not NULL. */
386
387 void
388 add_target_with_completer (struct target_ops *t,
389 completer_ftype *completer)
390 {
391 struct cmd_list_element *c;
392
393 complete_target_initialization (t);
394
395 if (!target_structs)
396 {
397 target_struct_allocsize = DEFAULT_ALLOCSIZE;
398 target_structs = (struct target_ops **) xmalloc
399 (target_struct_allocsize * sizeof (*target_structs));
400 }
401 if (target_struct_size >= target_struct_allocsize)
402 {
403 target_struct_allocsize *= 2;
404 target_structs = (struct target_ops **)
405 xrealloc ((char *) target_structs,
406 target_struct_allocsize * sizeof (*target_structs));
407 }
408 target_structs[target_struct_size++] = t;
409
410 if (targetlist == NULL)
411 add_prefix_cmd ("target", class_run, target_command, _("\
412 Connect to a target machine or process.\n\
413 The first argument is the type or protocol of the target machine.\n\
414 Remaining arguments are interpreted by the target protocol. For more\n\
415 information on the arguments for a particular protocol, type\n\
416 `help target ' followed by the protocol name."),
417 &targetlist, "target ", 0, &cmdlist);
418 c = add_cmd (t->to_shortname, no_class, t->to_open, t->to_doc,
419 &targetlist);
420 if (completer != NULL)
421 set_cmd_completer (c, completer);
422 }
423
424 /* Add a possible target architecture to the list. */
425
426 void
427 add_target (struct target_ops *t)
428 {
429 add_target_with_completer (t, NULL);
430 }
431
432 /* See target.h. */
433
434 void
435 add_deprecated_target_alias (struct target_ops *t, char *alias)
436 {
437 struct cmd_list_element *c;
438 char *alt;
439
440 /* If we use add_alias_cmd, here, we do not get the deprecated warning,
441 see PR cli/15104. */
442 c = add_cmd (alias, no_class, t->to_open, t->to_doc, &targetlist);
443 alt = xstrprintf ("target %s", t->to_shortname);
444 deprecate_cmd (c, alt);
445 }
446
447 /* Stub functions */
448
449 void
450 target_ignore (void)
451 {
452 }
453
454 void
455 target_kill (void)
456 {
457 struct target_ops *t;
458
459 for (t = current_target.beneath; t != NULL; t = t->beneath)
460 if (t->to_kill != NULL)
461 {
462 if (targetdebug)
463 fprintf_unfiltered (gdb_stdlog, "target_kill ()\n");
464
465 t->to_kill (t);
466 return;
467 }
468
469 noprocess ();
470 }
471
472 void
473 target_load (char *arg, int from_tty)
474 {
475 target_dcache_invalidate ();
476 (*current_target.to_load) (&current_target, arg, from_tty);
477 }
478
479 void
480 target_create_inferior (char *exec_file, char *args,
481 char **env, int from_tty)
482 {
483 struct target_ops *t;
484
485 for (t = current_target.beneath; t != NULL; t = t->beneath)
486 {
487 if (t->to_create_inferior != NULL)
488 {
489 t->to_create_inferior (t, exec_file, args, env, from_tty);
490 if (targetdebug)
491 fprintf_unfiltered (gdb_stdlog,
492 "target_create_inferior (%s, %s, xxx, %d)\n",
493 exec_file, args, from_tty);
494 return;
495 }
496 }
497
498 internal_error (__FILE__, __LINE__,
499 _("could not find a target to create inferior"));
500 }
501
502 void
503 target_terminal_inferior (void)
504 {
505 /* A background resume (``run&'') should leave GDB in control of the
506 terminal. Use target_can_async_p, not target_is_async_p, since at
507 this point the target is not async yet. However, if sync_execution
508 is not set, we know it will become async prior to resume. */
509 if (target_can_async_p () && !sync_execution)
510 return;
511
512 /* If GDB is resuming the inferior in the foreground, install
513 inferior's terminal modes. */
514 (*current_target.to_terminal_inferior) (&current_target);
515 }
516
517 static int
518 nomemory (CORE_ADDR memaddr, char *myaddr, int len, int write,
519 struct target_ops *t)
520 {
521 errno = EIO; /* Can't read/write this location. */
522 return 0; /* No bytes handled. */
523 }
524
525 static void
526 tcomplain (void)
527 {
528 error (_("You can't do that when your target is `%s'"),
529 current_target.to_shortname);
530 }
531
532 void
533 noprocess (void)
534 {
535 error (_("You can't do that without a process to debug."));
536 }
537
538 static void
539 default_terminal_info (struct target_ops *self, const char *args, int from_tty)
540 {
541 printf_unfiltered (_("No saved terminal information.\n"));
542 }
543
544 /* A default implementation for the to_get_ada_task_ptid target method.
545
546 This function builds the PTID by using both LWP and TID as part of
547 the PTID lwp and tid elements. The pid used is the pid of the
548 inferior_ptid. */
549
550 static ptid_t
551 default_get_ada_task_ptid (struct target_ops *self, long lwp, long tid)
552 {
553 return ptid_build (ptid_get_pid (inferior_ptid), lwp, tid);
554 }
555
556 static enum exec_direction_kind
557 default_execution_direction (struct target_ops *self)
558 {
559 if (!target_can_execute_reverse)
560 return EXEC_FORWARD;
561 else if (!target_can_async_p ())
562 return EXEC_FORWARD;
563 else
564 gdb_assert_not_reached ("\
565 to_execution_direction must be implemented for reverse async");
566 }
567
568 /* Go through the target stack from top to bottom, copying over zero
569 entries in current_target, then filling in still empty entries. In
570 effect, we are doing class inheritance through the pushed target
571 vectors.
572
573 NOTE: cagney/2003-10-17: The problem with this inheritance, as it
574 is currently implemented, is that it discards any knowledge of
575 which target an inherited method originally belonged to.
576 Consequently, new new target methods should instead explicitly and
577 locally search the target stack for the target that can handle the
578 request. */
579
580 static void
581 update_current_target (void)
582 {
583 struct target_ops *t;
584
585 /* First, reset current's contents. */
586 memset (&current_target, 0, sizeof (current_target));
587
588 /* Install the delegators. */
589 install_delegators (&current_target);
590
591 #define INHERIT(FIELD, TARGET) \
592 if (!current_target.FIELD) \
593 current_target.FIELD = (TARGET)->FIELD
594
595 for (t = target_stack; t; t = t->beneath)
596 {
597 INHERIT (to_shortname, t);
598 INHERIT (to_longname, t);
599 INHERIT (to_doc, t);
600 /* Do not inherit to_open. */
601 /* Do not inherit to_close. */
602 /* Do not inherit to_attach. */
603 /* Do not inherit to_post_attach. */
604 INHERIT (to_attach_no_wait, t);
605 /* Do not inherit to_detach. */
606 /* Do not inherit to_disconnect. */
607 /* Do not inherit to_resume. */
608 /* Do not inherit to_wait. */
609 /* Do not inherit to_fetch_registers. */
610 /* Do not inherit to_store_registers. */
611 /* Do not inherit to_prepare_to_store. */
612 INHERIT (deprecated_xfer_memory, t);
613 /* Do not inherit to_files_info. */
614 /* Do not inherit to_insert_breakpoint. */
615 /* Do not inherit to_remove_breakpoint. */
616 /* Do not inherit to_can_use_hw_breakpoint. */
617 /* Do not inherit to_insert_hw_breakpoint. */
618 /* Do not inherit to_remove_hw_breakpoint. */
619 /* Do not inherit to_ranged_break_num_registers. */
620 /* Do not inherit to_insert_watchpoint. */
621 /* Do not inherit to_remove_watchpoint. */
622 /* Do not inherit to_insert_mask_watchpoint. */
623 /* Do not inherit to_remove_mask_watchpoint. */
624 /* Do not inherit to_stopped_data_address. */
625 INHERIT (to_have_steppable_watchpoint, t);
626 INHERIT (to_have_continuable_watchpoint, t);
627 /* Do not inherit to_stopped_by_watchpoint. */
628 /* Do not inherit to_watchpoint_addr_within_range. */
629 /* Do not inherit to_region_ok_for_hw_watchpoint. */
630 /* Do not inherit to_can_accel_watchpoint_condition. */
631 /* Do not inherit to_masked_watch_num_registers. */
632 /* Do not inherit to_terminal_init. */
633 /* Do not inherit to_terminal_inferior. */
634 /* Do not inherit to_terminal_ours_for_output. */
635 /* Do not inherit to_terminal_ours. */
636 /* Do not inherit to_terminal_save_ours. */
637 /* Do not inherit to_terminal_info. */
638 /* Do not inherit to_kill. */
639 /* Do not inherit to_load. */
640 /* Do no inherit to_create_inferior. */
641 /* Do not inherit to_post_startup_inferior. */
642 /* Do not inherit to_insert_fork_catchpoint. */
643 /* Do not inherit to_remove_fork_catchpoint. */
644 /* Do not inherit to_insert_vfork_catchpoint. */
645 /* Do not inherit to_remove_vfork_catchpoint. */
646 /* Do not inherit to_follow_fork. */
647 /* Do not inherit to_insert_exec_catchpoint. */
648 /* Do not inherit to_remove_exec_catchpoint. */
649 /* Do not inherit to_set_syscall_catchpoint. */
650 /* Do not inherit to_has_exited. */
651 /* Do not inherit to_mourn_inferior. */
652 INHERIT (to_can_run, t);
653 /* Do not inherit to_pass_signals. */
654 /* Do not inherit to_program_signals. */
655 /* Do not inherit to_thread_alive. */
656 /* Do not inherit to_find_new_threads. */
657 /* Do not inherit to_pid_to_str. */
658 /* Do not inherit to_extra_thread_info. */
659 /* Do not inherit to_thread_name. */
660 INHERIT (to_stop, t);
661 /* Do not inherit to_xfer_partial. */
662 /* Do not inherit to_rcmd. */
663 /* Do not inherit to_pid_to_exec_file. */
664 /* Do not inherit to_log_command. */
665 INHERIT (to_stratum, t);
666 /* Do not inherit to_has_all_memory. */
667 /* Do not inherit to_has_memory. */
668 /* Do not inherit to_has_stack. */
669 /* Do not inherit to_has_registers. */
670 /* Do not inherit to_has_execution. */
671 INHERIT (to_has_thread_control, t);
672 /* Do not inherit to_can_async_p. */
673 /* Do not inherit to_is_async_p. */
674 /* Do not inherit to_async. */
675 /* Do not inherit to_find_memory_regions. */
676 /* Do not inherit to_make_corefile_notes. */
677 /* Do not inherit to_get_bookmark. */
678 /* Do not inherit to_goto_bookmark. */
679 /* Do not inherit to_get_thread_local_address. */
680 /* Do not inherit to_can_execute_reverse. */
681 /* Do not inherit to_execution_direction. */
682 /* Do not inherit to_thread_architecture. */
683 /* Do not inherit to_read_description. */
684 /* Do not inherit to_get_ada_task_ptid. */
685 /* Do not inherit to_search_memory. */
686 /* Do not inherit to_supports_multi_process. */
687 /* Do not inherit to_supports_enable_disable_tracepoint. */
688 /* Do not inherit to_supports_string_tracing. */
689 /* Do not inherit to_trace_init. */
690 /* Do not inherit to_download_tracepoint. */
691 /* Do not inherit to_can_download_tracepoint. */
692 /* Do not inherit to_download_trace_state_variable. */
693 /* Do not inherit to_enable_tracepoint. */
694 /* Do not inherit to_disable_tracepoint. */
695 /* Do not inherit to_trace_set_readonly_regions. */
696 /* Do not inherit to_trace_start. */
697 /* Do not inherit to_get_trace_status. */
698 /* Do not inherit to_get_tracepoint_status. */
699 /* Do not inherit to_trace_stop. */
700 /* Do not inherit to_trace_find. */
701 /* Do not inherit to_get_trace_state_variable_value. */
702 /* Do not inherit to_save_trace_data. */
703 /* Do not inherit to_upload_tracepoints. */
704 /* Do not inherit to_upload_trace_state_variables. */
705 /* Do not inherit to_get_raw_trace_data. */
706 /* Do not inherit to_get_min_fast_tracepoint_insn_len. */
707 INHERIT (to_set_disconnected_tracing, t);
708 INHERIT (to_set_circular_trace_buffer, t);
709 INHERIT (to_set_trace_buffer_size, t);
710 INHERIT (to_set_trace_notes, t);
711 INHERIT (to_get_tib_address, t);
712 INHERIT (to_set_permissions, t);
713 INHERIT (to_static_tracepoint_marker_at, t);
714 INHERIT (to_static_tracepoint_markers_by_strid, t);
715 INHERIT (to_traceframe_info, t);
716 INHERIT (to_use_agent, t);
717 INHERIT (to_can_use_agent, t);
718 INHERIT (to_augmented_libraries_svr4_read, t);
719 INHERIT (to_magic, t);
720 INHERIT (to_supports_evaluation_of_breakpoint_conditions, t);
721 INHERIT (to_can_run_breakpoint_commands, t);
722 /* Do not inherit to_memory_map. */
723 /* Do not inherit to_flash_erase. */
724 /* Do not inherit to_flash_done. */
725 }
726 #undef INHERIT
727
728 /* Clean up a target struct so it no longer has any zero pointers in
729 it. Some entries are defaulted to a method that print an error,
730 others are hard-wired to a standard recursive default. */
731
732 #define de_fault(field, value) \
733 if (!current_target.field) \
734 current_target.field = value
735
736 de_fault (to_open,
737 (void (*) (char *, int))
738 tcomplain);
739 de_fault (to_close,
740 (void (*) (struct target_ops *))
741 target_ignore);
742 de_fault (deprecated_xfer_memory,
743 (int (*) (CORE_ADDR, gdb_byte *, int, int,
744 struct mem_attrib *, struct target_ops *))
745 nomemory);
746 de_fault (to_can_run,
747 (int (*) (struct target_ops *))
748 return_zero);
749 de_fault (to_stop,
750 (void (*) (struct target_ops *, ptid_t))
751 target_ignore);
752 current_target.to_read_description = NULL;
753 de_fault (to_set_disconnected_tracing,
754 (void (*) (struct target_ops *, int))
755 target_ignore);
756 de_fault (to_set_circular_trace_buffer,
757 (void (*) (struct target_ops *, int))
758 target_ignore);
759 de_fault (to_set_trace_buffer_size,
760 (void (*) (struct target_ops *, LONGEST))
761 target_ignore);
762 de_fault (to_set_trace_notes,
763 (int (*) (struct target_ops *,
764 const char *, const char *, const char *))
765 return_zero);
766 de_fault (to_get_tib_address,
767 (int (*) (struct target_ops *, ptid_t, CORE_ADDR *))
768 tcomplain);
769 de_fault (to_set_permissions,
770 (void (*) (struct target_ops *))
771 target_ignore);
772 de_fault (to_static_tracepoint_marker_at,
773 (int (*) (struct target_ops *,
774 CORE_ADDR, struct static_tracepoint_marker *))
775 return_zero);
776 de_fault (to_static_tracepoint_markers_by_strid,
777 (VEC(static_tracepoint_marker_p) * (*) (struct target_ops *,
778 const char *))
779 tcomplain);
780 de_fault (to_traceframe_info,
781 (struct traceframe_info * (*) (struct target_ops *))
782 return_null);
783 de_fault (to_supports_evaluation_of_breakpoint_conditions,
784 (int (*) (struct target_ops *))
785 return_zero);
786 de_fault (to_can_run_breakpoint_commands,
787 (int (*) (struct target_ops *))
788 return_zero);
789 de_fault (to_use_agent,
790 (int (*) (struct target_ops *, int))
791 tcomplain);
792 de_fault (to_can_use_agent,
793 (int (*) (struct target_ops *))
794 return_zero);
795 de_fault (to_augmented_libraries_svr4_read,
796 (int (*) (struct target_ops *))
797 return_zero);
798
799 #undef de_fault
800
801 /* Finally, position the target-stack beneath the squashed
802 "current_target". That way code looking for a non-inherited
803 target method can quickly and simply find it. */
804 current_target.beneath = target_stack;
805
806 if (targetdebug)
807 setup_target_debug ();
808 }
809
810 /* Push a new target type into the stack of the existing target accessors,
811 possibly superseding some of the existing accessors.
812
813 Rather than allow an empty stack, we always have the dummy target at
814 the bottom stratum, so we can call the function vectors without
815 checking them. */
816
817 void
818 push_target (struct target_ops *t)
819 {
820 struct target_ops **cur;
821
822 /* Check magic number. If wrong, it probably means someone changed
823 the struct definition, but not all the places that initialize one. */
824 if (t->to_magic != OPS_MAGIC)
825 {
826 fprintf_unfiltered (gdb_stderr,
827 "Magic number of %s target struct wrong\n",
828 t->to_shortname);
829 internal_error (__FILE__, __LINE__,
830 _("failed internal consistency check"));
831 }
832
833 /* Find the proper stratum to install this target in. */
834 for (cur = &target_stack; (*cur) != NULL; cur = &(*cur)->beneath)
835 {
836 if ((int) (t->to_stratum) >= (int) (*cur)->to_stratum)
837 break;
838 }
839
840 /* If there's already targets at this stratum, remove them. */
841 /* FIXME: cagney/2003-10-15: I think this should be popping all
842 targets to CUR, and not just those at this stratum level. */
843 while ((*cur) != NULL && t->to_stratum == (*cur)->to_stratum)
844 {
845 /* There's already something at this stratum level. Close it,
846 and un-hook it from the stack. */
847 struct target_ops *tmp = (*cur);
848
849 (*cur) = (*cur)->beneath;
850 tmp->beneath = NULL;
851 target_close (tmp);
852 }
853
854 /* We have removed all targets in our stratum, now add the new one. */
855 t->beneath = (*cur);
856 (*cur) = t;
857
858 update_current_target ();
859 }
860
861 /* Remove a target_ops vector from the stack, wherever it may be.
862 Return how many times it was removed (0 or 1). */
863
864 int
865 unpush_target (struct target_ops *t)
866 {
867 struct target_ops **cur;
868 struct target_ops *tmp;
869
870 if (t->to_stratum == dummy_stratum)
871 internal_error (__FILE__, __LINE__,
872 _("Attempt to unpush the dummy target"));
873
874 /* Look for the specified target. Note that we assume that a target
875 can only occur once in the target stack. */
876
877 for (cur = &target_stack; (*cur) != NULL; cur = &(*cur)->beneath)
878 {
879 if ((*cur) == t)
880 break;
881 }
882
883 /* If we don't find target_ops, quit. Only open targets should be
884 closed. */
885 if ((*cur) == NULL)
886 return 0;
887
888 /* Unchain the target. */
889 tmp = (*cur);
890 (*cur) = (*cur)->beneath;
891 tmp->beneath = NULL;
892
893 update_current_target ();
894
895 /* Finally close the target. Note we do this after unchaining, so
896 any target method calls from within the target_close
897 implementation don't end up in T anymore. */
898 target_close (t);
899
900 return 1;
901 }
902
903 void
904 pop_all_targets_above (enum strata above_stratum)
905 {
906 while ((int) (current_target.to_stratum) > (int) above_stratum)
907 {
908 if (!unpush_target (target_stack))
909 {
910 fprintf_unfiltered (gdb_stderr,
911 "pop_all_targets couldn't find target %s\n",
912 target_stack->to_shortname);
913 internal_error (__FILE__, __LINE__,
914 _("failed internal consistency check"));
915 break;
916 }
917 }
918 }
919
920 void
921 pop_all_targets (void)
922 {
923 pop_all_targets_above (dummy_stratum);
924 }
925
926 /* Return 1 if T is now pushed in the target stack. Return 0 otherwise. */
927
928 int
929 target_is_pushed (struct target_ops *t)
930 {
931 struct target_ops **cur;
932
933 /* Check magic number. If wrong, it probably means someone changed
934 the struct definition, but not all the places that initialize one. */
935 if (t->to_magic != OPS_MAGIC)
936 {
937 fprintf_unfiltered (gdb_stderr,
938 "Magic number of %s target struct wrong\n",
939 t->to_shortname);
940 internal_error (__FILE__, __LINE__,
941 _("failed internal consistency check"));
942 }
943
944 for (cur = &target_stack; (*cur) != NULL; cur = &(*cur)->beneath)
945 if (*cur == t)
946 return 1;
947
948 return 0;
949 }
950
951 /* Using the objfile specified in OBJFILE, find the address for the
952 current thread's thread-local storage with offset OFFSET. */
953 CORE_ADDR
954 target_translate_tls_address (struct objfile *objfile, CORE_ADDR offset)
955 {
956 volatile CORE_ADDR addr = 0;
957 struct target_ops *target;
958
959 for (target = current_target.beneath;
960 target != NULL;
961 target = target->beneath)
962 {
963 if (target->to_get_thread_local_address != NULL)
964 break;
965 }
966
967 if (target != NULL
968 && gdbarch_fetch_tls_load_module_address_p (target_gdbarch ()))
969 {
970 ptid_t ptid = inferior_ptid;
971 volatile struct gdb_exception ex;
972
973 TRY_CATCH (ex, RETURN_MASK_ALL)
974 {
975 CORE_ADDR lm_addr;
976
977 /* Fetch the load module address for this objfile. */
978 lm_addr = gdbarch_fetch_tls_load_module_address (target_gdbarch (),
979 objfile);
980 /* If it's 0, throw the appropriate exception. */
981 if (lm_addr == 0)
982 throw_error (TLS_LOAD_MODULE_NOT_FOUND_ERROR,
983 _("TLS load module not found"));
984
985 addr = target->to_get_thread_local_address (target, ptid,
986 lm_addr, offset);
987 }
988 /* If an error occurred, print TLS related messages here. Otherwise,
989 throw the error to some higher catcher. */
990 if (ex.reason < 0)
991 {
992 int objfile_is_library = (objfile->flags & OBJF_SHARED);
993
994 switch (ex.error)
995 {
996 case TLS_NO_LIBRARY_SUPPORT_ERROR:
997 error (_("Cannot find thread-local variables "
998 "in this thread library."));
999 break;
1000 case TLS_LOAD_MODULE_NOT_FOUND_ERROR:
1001 if (objfile_is_library)
1002 error (_("Cannot find shared library `%s' in dynamic"
1003 " linker's load module list"), objfile_name (objfile));
1004 else
1005 error (_("Cannot find executable file `%s' in dynamic"
1006 " linker's load module list"), objfile_name (objfile));
1007 break;
1008 case TLS_NOT_ALLOCATED_YET_ERROR:
1009 if (objfile_is_library)
1010 error (_("The inferior has not yet allocated storage for"
1011 " thread-local variables in\n"
1012 "the shared library `%s'\n"
1013 "for %s"),
1014 objfile_name (objfile), target_pid_to_str (ptid));
1015 else
1016 error (_("The inferior has not yet allocated storage for"
1017 " thread-local variables in\n"
1018 "the executable `%s'\n"
1019 "for %s"),
1020 objfile_name (objfile), target_pid_to_str (ptid));
1021 break;
1022 case TLS_GENERIC_ERROR:
1023 if (objfile_is_library)
1024 error (_("Cannot find thread-local storage for %s, "
1025 "shared library %s:\n%s"),
1026 target_pid_to_str (ptid),
1027 objfile_name (objfile), ex.message);
1028 else
1029 error (_("Cannot find thread-local storage for %s, "
1030 "executable file %s:\n%s"),
1031 target_pid_to_str (ptid),
1032 objfile_name (objfile), ex.message);
1033 break;
1034 default:
1035 throw_exception (ex);
1036 break;
1037 }
1038 }
1039 }
1040 /* It wouldn't be wrong here to try a gdbarch method, too; finding
1041 TLS is an ABI-specific thing. But we don't do that yet. */
1042 else
1043 error (_("Cannot find thread-local variables on this target"));
1044
1045 return addr;
1046 }
1047
1048 const char *
1049 target_xfer_status_to_string (enum target_xfer_status err)
1050 {
1051 #define CASE(X) case X: return #X
1052 switch (err)
1053 {
1054 CASE(TARGET_XFER_E_IO);
1055 CASE(TARGET_XFER_E_UNAVAILABLE);
1056 default:
1057 return "<unknown>";
1058 }
1059 #undef CASE
1060 };
1061
1062
1063 #undef MIN
1064 #define MIN(A, B) (((A) <= (B)) ? (A) : (B))
1065
1066 /* target_read_string -- read a null terminated string, up to LEN bytes,
1067 from MEMADDR in target. Set *ERRNOP to the errno code, or 0 if successful.
1068 Set *STRING to a pointer to malloc'd memory containing the data; the caller
1069 is responsible for freeing it. Return the number of bytes successfully
1070 read. */
1071
1072 int
1073 target_read_string (CORE_ADDR memaddr, char **string, int len, int *errnop)
1074 {
1075 int tlen, offset, i;
1076 gdb_byte buf[4];
1077 int errcode = 0;
1078 char *buffer;
1079 int buffer_allocated;
1080 char *bufptr;
1081 unsigned int nbytes_read = 0;
1082
1083 gdb_assert (string);
1084
1085 /* Small for testing. */
1086 buffer_allocated = 4;
1087 buffer = xmalloc (buffer_allocated);
1088 bufptr = buffer;
1089
1090 while (len > 0)
1091 {
1092 tlen = MIN (len, 4 - (memaddr & 3));
1093 offset = memaddr & 3;
1094
1095 errcode = target_read_memory (memaddr & ~3, buf, sizeof buf);
1096 if (errcode != 0)
1097 {
1098 /* The transfer request might have crossed the boundary to an
1099 unallocated region of memory. Retry the transfer, requesting
1100 a single byte. */
1101 tlen = 1;
1102 offset = 0;
1103 errcode = target_read_memory (memaddr, buf, 1);
1104 if (errcode != 0)
1105 goto done;
1106 }
1107
1108 if (bufptr - buffer + tlen > buffer_allocated)
1109 {
1110 unsigned int bytes;
1111
1112 bytes = bufptr - buffer;
1113 buffer_allocated *= 2;
1114 buffer = xrealloc (buffer, buffer_allocated);
1115 bufptr = buffer + bytes;
1116 }
1117
1118 for (i = 0; i < tlen; i++)
1119 {
1120 *bufptr++ = buf[i + offset];
1121 if (buf[i + offset] == '\000')
1122 {
1123 nbytes_read += i + 1;
1124 goto done;
1125 }
1126 }
1127
1128 memaddr += tlen;
1129 len -= tlen;
1130 nbytes_read += tlen;
1131 }
1132 done:
1133 *string = buffer;
1134 if (errnop != NULL)
1135 *errnop = errcode;
1136 return nbytes_read;
1137 }
1138
1139 struct target_section_table *
1140 target_get_section_table (struct target_ops *target)
1141 {
1142 struct target_ops *t;
1143
1144 if (targetdebug)
1145 fprintf_unfiltered (gdb_stdlog, "target_get_section_table ()\n");
1146
1147 for (t = target; t != NULL; t = t->beneath)
1148 if (t->to_get_section_table != NULL)
1149 return (*t->to_get_section_table) (t);
1150
1151 return NULL;
1152 }
1153
1154 /* Find a section containing ADDR. */
1155
1156 struct target_section *
1157 target_section_by_addr (struct target_ops *target, CORE_ADDR addr)
1158 {
1159 struct target_section_table *table = target_get_section_table (target);
1160 struct target_section *secp;
1161
1162 if (table == NULL)
1163 return NULL;
1164
1165 for (secp = table->sections; secp < table->sections_end; secp++)
1166 {
1167 if (addr >= secp->addr && addr < secp->endaddr)
1168 return secp;
1169 }
1170 return NULL;
1171 }
1172
1173 /* Read memory from the live target, even if currently inspecting a
1174 traceframe. The return is the same as that of target_read. */
1175
1176 static enum target_xfer_status
1177 target_read_live_memory (enum target_object object,
1178 ULONGEST memaddr, gdb_byte *myaddr, ULONGEST len,
1179 ULONGEST *xfered_len)
1180 {
1181 enum target_xfer_status ret;
1182 struct cleanup *cleanup;
1183
1184 /* Switch momentarily out of tfind mode so to access live memory.
1185 Note that this must not clear global state, such as the frame
1186 cache, which must still remain valid for the previous traceframe.
1187 We may be _building_ the frame cache at this point. */
1188 cleanup = make_cleanup_restore_traceframe_number ();
1189 set_traceframe_number (-1);
1190
1191 ret = target_xfer_partial (current_target.beneath, object, NULL,
1192 myaddr, NULL, memaddr, len, xfered_len);
1193
1194 do_cleanups (cleanup);
1195 return ret;
1196 }
1197
1198 /* Using the set of read-only target sections of OPS, read live
1199 read-only memory. Note that the actual reads start from the
1200 top-most target again.
1201
1202 For interface/parameters/return description see target.h,
1203 to_xfer_partial. */
1204
1205 static enum target_xfer_status
1206 memory_xfer_live_readonly_partial (struct target_ops *ops,
1207 enum target_object object,
1208 gdb_byte *readbuf, ULONGEST memaddr,
1209 ULONGEST len, ULONGEST *xfered_len)
1210 {
1211 struct target_section *secp;
1212 struct target_section_table *table;
1213
1214 secp = target_section_by_addr (ops, memaddr);
1215 if (secp != NULL
1216 && (bfd_get_section_flags (secp->the_bfd_section->owner,
1217 secp->the_bfd_section)
1218 & SEC_READONLY))
1219 {
1220 struct target_section *p;
1221 ULONGEST memend = memaddr + len;
1222
1223 table = target_get_section_table (ops);
1224
1225 for (p = table->sections; p < table->sections_end; p++)
1226 {
1227 if (memaddr >= p->addr)
1228 {
1229 if (memend <= p->endaddr)
1230 {
1231 /* Entire transfer is within this section. */
1232 return target_read_live_memory (object, memaddr,
1233 readbuf, len, xfered_len);
1234 }
1235 else if (memaddr >= p->endaddr)
1236 {
1237 /* This section ends before the transfer starts. */
1238 continue;
1239 }
1240 else
1241 {
1242 /* This section overlaps the transfer. Just do half. */
1243 len = p->endaddr - memaddr;
1244 return target_read_live_memory (object, memaddr,
1245 readbuf, len, xfered_len);
1246 }
1247 }
1248 }
1249 }
1250
1251 return TARGET_XFER_EOF;
1252 }
1253
1254 /* Read memory from more than one valid target. A core file, for
1255 instance, could have some of memory but delegate other bits to
1256 the target below it. So, we must manually try all targets. */
1257
1258 static enum target_xfer_status
1259 raw_memory_xfer_partial (struct target_ops *ops, gdb_byte *readbuf,
1260 const gdb_byte *writebuf, ULONGEST memaddr, LONGEST len,
1261 ULONGEST *xfered_len)
1262 {
1263 enum target_xfer_status res;
1264
1265 do
1266 {
1267 res = ops->to_xfer_partial (ops, TARGET_OBJECT_MEMORY, NULL,
1268 readbuf, writebuf, memaddr, len,
1269 xfered_len);
1270 if (res == TARGET_XFER_OK)
1271 break;
1272
1273 /* Stop if the target reports that the memory is not available. */
1274 if (res == TARGET_XFER_E_UNAVAILABLE)
1275 break;
1276
1277 /* We want to continue past core files to executables, but not
1278 past a running target's memory. */
1279 if (ops->to_has_all_memory (ops))
1280 break;
1281
1282 ops = ops->beneath;
1283 }
1284 while (ops != NULL);
1285
1286 return res;
1287 }
1288
1289 /* Perform a partial memory transfer.
1290 For docs see target.h, to_xfer_partial. */
1291
1292 static enum target_xfer_status
1293 memory_xfer_partial_1 (struct target_ops *ops, enum target_object object,
1294 gdb_byte *readbuf, const gdb_byte *writebuf, ULONGEST memaddr,
1295 ULONGEST len, ULONGEST *xfered_len)
1296 {
1297 enum target_xfer_status res;
1298 int reg_len;
1299 struct mem_region *region;
1300 struct inferior *inf;
1301
1302 /* For accesses to unmapped overlay sections, read directly from
1303 files. Must do this first, as MEMADDR may need adjustment. */
1304 if (readbuf != NULL && overlay_debugging)
1305 {
1306 struct obj_section *section = find_pc_overlay (memaddr);
1307
1308 if (pc_in_unmapped_range (memaddr, section))
1309 {
1310 struct target_section_table *table
1311 = target_get_section_table (ops);
1312 const char *section_name = section->the_bfd_section->name;
1313
1314 memaddr = overlay_mapped_address (memaddr, section);
1315 return section_table_xfer_memory_partial (readbuf, writebuf,
1316 memaddr, len, xfered_len,
1317 table->sections,
1318 table->sections_end,
1319 section_name);
1320 }
1321 }
1322
1323 /* Try the executable files, if "trust-readonly-sections" is set. */
1324 if (readbuf != NULL && trust_readonly)
1325 {
1326 struct target_section *secp;
1327 struct target_section_table *table;
1328
1329 secp = target_section_by_addr (ops, memaddr);
1330 if (secp != NULL
1331 && (bfd_get_section_flags (secp->the_bfd_section->owner,
1332 secp->the_bfd_section)
1333 & SEC_READONLY))
1334 {
1335 table = target_get_section_table (ops);
1336 return section_table_xfer_memory_partial (readbuf, writebuf,
1337 memaddr, len, xfered_len,
1338 table->sections,
1339 table->sections_end,
1340 NULL);
1341 }
1342 }
1343
1344 /* If reading unavailable memory in the context of traceframes, and
1345 this address falls within a read-only section, fallback to
1346 reading from live memory. */
1347 if (readbuf != NULL && get_traceframe_number () != -1)
1348 {
1349 VEC(mem_range_s) *available;
1350
1351 /* If we fail to get the set of available memory, then the
1352 target does not support querying traceframe info, and so we
1353 attempt reading from the traceframe anyway (assuming the
1354 target implements the old QTro packet then). */
1355 if (traceframe_available_memory (&available, memaddr, len))
1356 {
1357 struct cleanup *old_chain;
1358
1359 old_chain = make_cleanup (VEC_cleanup(mem_range_s), &available);
1360
1361 if (VEC_empty (mem_range_s, available)
1362 || VEC_index (mem_range_s, available, 0)->start != memaddr)
1363 {
1364 /* Don't read into the traceframe's available
1365 memory. */
1366 if (!VEC_empty (mem_range_s, available))
1367 {
1368 LONGEST oldlen = len;
1369
1370 len = VEC_index (mem_range_s, available, 0)->start - memaddr;
1371 gdb_assert (len <= oldlen);
1372 }
1373
1374 do_cleanups (old_chain);
1375
1376 /* This goes through the topmost target again. */
1377 res = memory_xfer_live_readonly_partial (ops, object,
1378 readbuf, memaddr,
1379 len, xfered_len);
1380 if (res == TARGET_XFER_OK)
1381 return TARGET_XFER_OK;
1382 else
1383 {
1384 /* No use trying further, we know some memory starting
1385 at MEMADDR isn't available. */
1386 *xfered_len = len;
1387 return TARGET_XFER_E_UNAVAILABLE;
1388 }
1389 }
1390
1391 /* Don't try to read more than how much is available, in
1392 case the target implements the deprecated QTro packet to
1393 cater for older GDBs (the target's knowledge of read-only
1394 sections may be outdated by now). */
1395 len = VEC_index (mem_range_s, available, 0)->length;
1396
1397 do_cleanups (old_chain);
1398 }
1399 }
1400
1401 /* Try GDB's internal data cache. */
1402 region = lookup_mem_region (memaddr);
1403 /* region->hi == 0 means there's no upper bound. */
1404 if (memaddr + len < region->hi || region->hi == 0)
1405 reg_len = len;
1406 else
1407 reg_len = region->hi - memaddr;
1408
1409 switch (region->attrib.mode)
1410 {
1411 case MEM_RO:
1412 if (writebuf != NULL)
1413 return TARGET_XFER_E_IO;
1414 break;
1415
1416 case MEM_WO:
1417 if (readbuf != NULL)
1418 return TARGET_XFER_E_IO;
1419 break;
1420
1421 case MEM_FLASH:
1422 /* We only support writing to flash during "load" for now. */
1423 if (writebuf != NULL)
1424 error (_("Writing to flash memory forbidden in this context"));
1425 break;
1426
1427 case MEM_NONE:
1428 return TARGET_XFER_E_IO;
1429 }
1430
1431 if (!ptid_equal (inferior_ptid, null_ptid))
1432 inf = find_inferior_pid (ptid_get_pid (inferior_ptid));
1433 else
1434 inf = NULL;
1435
1436 if (inf != NULL
1437 /* The dcache reads whole cache lines; that doesn't play well
1438 with reading from a trace buffer, because reading outside of
1439 the collected memory range fails. */
1440 && get_traceframe_number () == -1
1441 && (region->attrib.cache
1442 || (stack_cache_enabled_p () && object == TARGET_OBJECT_STACK_MEMORY)
1443 || (code_cache_enabled_p () && object == TARGET_OBJECT_CODE_MEMORY)))
1444 {
1445 DCACHE *dcache = target_dcache_get_or_init ();
1446 int l;
1447
1448 if (readbuf != NULL)
1449 l = dcache_xfer_memory (ops, dcache, memaddr, readbuf, reg_len, 0);
1450 else
1451 /* FIXME drow/2006-08-09: If we're going to preserve const
1452 correctness dcache_xfer_memory should take readbuf and
1453 writebuf. */
1454 l = dcache_xfer_memory (ops, dcache, memaddr, (void *) writebuf,
1455 reg_len, 1);
1456 if (l <= 0)
1457 return TARGET_XFER_E_IO;
1458 else
1459 {
1460 *xfered_len = (ULONGEST) l;
1461 return TARGET_XFER_OK;
1462 }
1463 }
1464
1465 /* If none of those methods found the memory we wanted, fall back
1466 to a target partial transfer. Normally a single call to
1467 to_xfer_partial is enough; if it doesn't recognize an object
1468 it will call the to_xfer_partial of the next target down.
1469 But for memory this won't do. Memory is the only target
1470 object which can be read from more than one valid target.
1471 A core file, for instance, could have some of memory but
1472 delegate other bits to the target below it. So, we must
1473 manually try all targets. */
1474
1475 res = raw_memory_xfer_partial (ops, readbuf, writebuf, memaddr, reg_len,
1476 xfered_len);
1477
1478 /* Make sure the cache gets updated no matter what - if we are writing
1479 to the stack. Even if this write is not tagged as such, we still need
1480 to update the cache. */
1481
1482 if (res == TARGET_XFER_OK
1483 && inf != NULL
1484 && writebuf != NULL
1485 && target_dcache_init_p ()
1486 && !region->attrib.cache
1487 && ((stack_cache_enabled_p () && object != TARGET_OBJECT_STACK_MEMORY)
1488 || (code_cache_enabled_p () && object != TARGET_OBJECT_CODE_MEMORY)))
1489 {
1490 DCACHE *dcache = target_dcache_get ();
1491
1492 dcache_update (dcache, memaddr, (void *) writebuf, reg_len);
1493 }
1494
1495 /* If we still haven't got anything, return the last error. We
1496 give up. */
1497 return res;
1498 }
1499
1500 /* Perform a partial memory transfer. For docs see target.h,
1501 to_xfer_partial. */
1502
1503 static enum target_xfer_status
1504 memory_xfer_partial (struct target_ops *ops, enum target_object object,
1505 gdb_byte *readbuf, const gdb_byte *writebuf,
1506 ULONGEST memaddr, ULONGEST len, ULONGEST *xfered_len)
1507 {
1508 enum target_xfer_status res;
1509
1510 /* Zero length requests are ok and require no work. */
1511 if (len == 0)
1512 return TARGET_XFER_EOF;
1513
1514 /* Fill in READBUF with breakpoint shadows, or WRITEBUF with
1515 breakpoint insns, thus hiding out from higher layers whether
1516 there are software breakpoints inserted in the code stream. */
1517 if (readbuf != NULL)
1518 {
1519 res = memory_xfer_partial_1 (ops, object, readbuf, NULL, memaddr, len,
1520 xfered_len);
1521
1522 if (res == TARGET_XFER_OK && !show_memory_breakpoints)
1523 breakpoint_xfer_memory (readbuf, NULL, NULL, memaddr, res);
1524 }
1525 else
1526 {
1527 void *buf;
1528 struct cleanup *old_chain;
1529
1530 /* A large write request is likely to be partially satisfied
1531 by memory_xfer_partial_1. We will continually malloc
1532 and free a copy of the entire write request for breakpoint
1533 shadow handling even though we only end up writing a small
1534 subset of it. Cap writes to 4KB to mitigate this. */
1535 len = min (4096, len);
1536
1537 buf = xmalloc (len);
1538 old_chain = make_cleanup (xfree, buf);
1539 memcpy (buf, writebuf, len);
1540
1541 breakpoint_xfer_memory (NULL, buf, writebuf, memaddr, len);
1542 res = memory_xfer_partial_1 (ops, object, NULL, buf, memaddr, len,
1543 xfered_len);
1544
1545 do_cleanups (old_chain);
1546 }
1547
1548 return res;
1549 }
1550
1551 static void
1552 restore_show_memory_breakpoints (void *arg)
1553 {
1554 show_memory_breakpoints = (uintptr_t) arg;
1555 }
1556
1557 struct cleanup *
1558 make_show_memory_breakpoints_cleanup (int show)
1559 {
1560 int current = show_memory_breakpoints;
1561
1562 show_memory_breakpoints = show;
1563 return make_cleanup (restore_show_memory_breakpoints,
1564 (void *) (uintptr_t) current);
1565 }
1566
1567 /* For docs see target.h, to_xfer_partial. */
1568
1569 enum target_xfer_status
1570 target_xfer_partial (struct target_ops *ops,
1571 enum target_object object, const char *annex,
1572 gdb_byte *readbuf, const gdb_byte *writebuf,
1573 ULONGEST offset, ULONGEST len,
1574 ULONGEST *xfered_len)
1575 {
1576 enum target_xfer_status retval;
1577
1578 gdb_assert (ops->to_xfer_partial != NULL);
1579
1580 /* Transfer is done when LEN is zero. */
1581 if (len == 0)
1582 return TARGET_XFER_EOF;
1583
1584 if (writebuf && !may_write_memory)
1585 error (_("Writing to memory is not allowed (addr %s, len %s)"),
1586 core_addr_to_string_nz (offset), plongest (len));
1587
1588 *xfered_len = 0;
1589
1590 /* If this is a memory transfer, let the memory-specific code
1591 have a look at it instead. Memory transfers are more
1592 complicated. */
1593 if (object == TARGET_OBJECT_MEMORY || object == TARGET_OBJECT_STACK_MEMORY
1594 || object == TARGET_OBJECT_CODE_MEMORY)
1595 retval = memory_xfer_partial (ops, object, readbuf,
1596 writebuf, offset, len, xfered_len);
1597 else if (object == TARGET_OBJECT_RAW_MEMORY)
1598 {
1599 /* Request the normal memory object from other layers. */
1600 retval = raw_memory_xfer_partial (ops, readbuf, writebuf, offset, len,
1601 xfered_len);
1602 }
1603 else
1604 retval = ops->to_xfer_partial (ops, object, annex, readbuf,
1605 writebuf, offset, len, xfered_len);
1606
1607 if (targetdebug)
1608 {
1609 const unsigned char *myaddr = NULL;
1610
1611 fprintf_unfiltered (gdb_stdlog,
1612 "%s:target_xfer_partial "
1613 "(%d, %s, %s, %s, %s, %s) = %d, %s",
1614 ops->to_shortname,
1615 (int) object,
1616 (annex ? annex : "(null)"),
1617 host_address_to_string (readbuf),
1618 host_address_to_string (writebuf),
1619 core_addr_to_string_nz (offset),
1620 pulongest (len), retval,
1621 pulongest (*xfered_len));
1622
1623 if (readbuf)
1624 myaddr = readbuf;
1625 if (writebuf)
1626 myaddr = writebuf;
1627 if (retval == TARGET_XFER_OK && myaddr != NULL)
1628 {
1629 int i;
1630
1631 fputs_unfiltered (", bytes =", gdb_stdlog);
1632 for (i = 0; i < *xfered_len; i++)
1633 {
1634 if ((((intptr_t) &(myaddr[i])) & 0xf) == 0)
1635 {
1636 if (targetdebug < 2 && i > 0)
1637 {
1638 fprintf_unfiltered (gdb_stdlog, " ...");
1639 break;
1640 }
1641 fprintf_unfiltered (gdb_stdlog, "\n");
1642 }
1643
1644 fprintf_unfiltered (gdb_stdlog, " %02x", myaddr[i] & 0xff);
1645 }
1646 }
1647
1648 fputc_unfiltered ('\n', gdb_stdlog);
1649 }
1650
1651 /* Check implementations of to_xfer_partial update *XFERED_LEN
1652 properly. Do assertion after printing debug messages, so that we
1653 can find more clues on assertion failure from debugging messages. */
1654 if (retval == TARGET_XFER_OK || retval == TARGET_XFER_E_UNAVAILABLE)
1655 gdb_assert (*xfered_len > 0);
1656
1657 return retval;
1658 }
1659
1660 /* Read LEN bytes of target memory at address MEMADDR, placing the
1661 results in GDB's memory at MYADDR. Returns either 0 for success or
1662 TARGET_XFER_E_IO if any error occurs.
1663
1664 If an error occurs, no guarantee is made about the contents of the data at
1665 MYADDR. In particular, the caller should not depend upon partial reads
1666 filling the buffer with good data. There is no way for the caller to know
1667 how much good data might have been transfered anyway. Callers that can
1668 deal with partial reads should call target_read (which will retry until
1669 it makes no progress, and then return how much was transferred). */
1670
1671 int
1672 target_read_memory (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1673 {
1674 /* Dispatch to the topmost target, not the flattened current_target.
1675 Memory accesses check target->to_has_(all_)memory, and the
1676 flattened target doesn't inherit those. */
1677 if (target_read (current_target.beneath, TARGET_OBJECT_MEMORY, NULL,
1678 myaddr, memaddr, len) == len)
1679 return 0;
1680 else
1681 return TARGET_XFER_E_IO;
1682 }
1683
1684 /* Like target_read_memory, but specify explicitly that this is a read
1685 from the target's raw memory. That is, this read bypasses the
1686 dcache, breakpoint shadowing, etc. */
1687
1688 int
1689 target_read_raw_memory (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1690 {
1691 /* See comment in target_read_memory about why the request starts at
1692 current_target.beneath. */
1693 if (target_read (current_target.beneath, TARGET_OBJECT_RAW_MEMORY, NULL,
1694 myaddr, memaddr, len) == len)
1695 return 0;
1696 else
1697 return TARGET_XFER_E_IO;
1698 }
1699
1700 /* Like target_read_memory, but specify explicitly that this is a read from
1701 the target's stack. This may trigger different cache behavior. */
1702
1703 int
1704 target_read_stack (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1705 {
1706 /* See comment in target_read_memory about why the request starts at
1707 current_target.beneath. */
1708 if (target_read (current_target.beneath, TARGET_OBJECT_STACK_MEMORY, NULL,
1709 myaddr, memaddr, len) == len)
1710 return 0;
1711 else
1712 return TARGET_XFER_E_IO;
1713 }
1714
1715 /* Like target_read_memory, but specify explicitly that this is a read from
1716 the target's code. This may trigger different cache behavior. */
1717
1718 int
1719 target_read_code (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1720 {
1721 /* See comment in target_read_memory about why the request starts at
1722 current_target.beneath. */
1723 if (target_read (current_target.beneath, TARGET_OBJECT_CODE_MEMORY, NULL,
1724 myaddr, memaddr, len) == len)
1725 return 0;
1726 else
1727 return TARGET_XFER_E_IO;
1728 }
1729
1730 /* Write LEN bytes from MYADDR to target memory at address MEMADDR.
1731 Returns either 0 for success or TARGET_XFER_E_IO if any
1732 error occurs. If an error occurs, no guarantee is made about how
1733 much data got written. Callers that can deal with partial writes
1734 should call target_write. */
1735
1736 int
1737 target_write_memory (CORE_ADDR memaddr, const gdb_byte *myaddr, ssize_t len)
1738 {
1739 /* See comment in target_read_memory about why the request starts at
1740 current_target.beneath. */
1741 if (target_write (current_target.beneath, TARGET_OBJECT_MEMORY, NULL,
1742 myaddr, memaddr, len) == len)
1743 return 0;
1744 else
1745 return TARGET_XFER_E_IO;
1746 }
1747
1748 /* Write LEN bytes from MYADDR to target raw memory at address
1749 MEMADDR. Returns either 0 for success or TARGET_XFER_E_IO
1750 if any error occurs. If an error occurs, no guarantee is made
1751 about how much data got written. Callers that can deal with
1752 partial writes should call target_write. */
1753
1754 int
1755 target_write_raw_memory (CORE_ADDR memaddr, const gdb_byte *myaddr, ssize_t len)
1756 {
1757 /* See comment in target_read_memory about why the request starts at
1758 current_target.beneath. */
1759 if (target_write (current_target.beneath, TARGET_OBJECT_RAW_MEMORY, NULL,
1760 myaddr, memaddr, len) == len)
1761 return 0;
1762 else
1763 return TARGET_XFER_E_IO;
1764 }
1765
1766 /* Fetch the target's memory map. */
1767
1768 VEC(mem_region_s) *
1769 target_memory_map (void)
1770 {
1771 VEC(mem_region_s) *result;
1772 struct mem_region *last_one, *this_one;
1773 int ix;
1774 struct target_ops *t;
1775
1776 if (targetdebug)
1777 fprintf_unfiltered (gdb_stdlog, "target_memory_map ()\n");
1778
1779 for (t = current_target.beneath; t != NULL; t = t->beneath)
1780 if (t->to_memory_map != NULL)
1781 break;
1782
1783 if (t == NULL)
1784 return NULL;
1785
1786 result = t->to_memory_map (t);
1787 if (result == NULL)
1788 return NULL;
1789
1790 qsort (VEC_address (mem_region_s, result),
1791 VEC_length (mem_region_s, result),
1792 sizeof (struct mem_region), mem_region_cmp);
1793
1794 /* Check that regions do not overlap. Simultaneously assign
1795 a numbering for the "mem" commands to use to refer to
1796 each region. */
1797 last_one = NULL;
1798 for (ix = 0; VEC_iterate (mem_region_s, result, ix, this_one); ix++)
1799 {
1800 this_one->number = ix;
1801
1802 if (last_one && last_one->hi > this_one->lo)
1803 {
1804 warning (_("Overlapping regions in memory map: ignoring"));
1805 VEC_free (mem_region_s, result);
1806 return NULL;
1807 }
1808 last_one = this_one;
1809 }
1810
1811 return result;
1812 }
1813
1814 void
1815 target_flash_erase (ULONGEST address, LONGEST length)
1816 {
1817 struct target_ops *t;
1818
1819 for (t = current_target.beneath; t != NULL; t = t->beneath)
1820 if (t->to_flash_erase != NULL)
1821 {
1822 if (targetdebug)
1823 fprintf_unfiltered (gdb_stdlog, "target_flash_erase (%s, %s)\n",
1824 hex_string (address), phex (length, 0));
1825 t->to_flash_erase (t, address, length);
1826 return;
1827 }
1828
1829 tcomplain ();
1830 }
1831
1832 void
1833 target_flash_done (void)
1834 {
1835 struct target_ops *t;
1836
1837 for (t = current_target.beneath; t != NULL; t = t->beneath)
1838 if (t->to_flash_done != NULL)
1839 {
1840 if (targetdebug)
1841 fprintf_unfiltered (gdb_stdlog, "target_flash_done\n");
1842 t->to_flash_done (t);
1843 return;
1844 }
1845
1846 tcomplain ();
1847 }
1848
1849 static void
1850 show_trust_readonly (struct ui_file *file, int from_tty,
1851 struct cmd_list_element *c, const char *value)
1852 {
1853 fprintf_filtered (file,
1854 _("Mode for reading from readonly sections is %s.\n"),
1855 value);
1856 }
1857
1858 /* More generic transfers. */
1859
1860 static enum target_xfer_status
1861 default_xfer_partial (struct target_ops *ops, enum target_object object,
1862 const char *annex, gdb_byte *readbuf,
1863 const gdb_byte *writebuf, ULONGEST offset, ULONGEST len,
1864 ULONGEST *xfered_len)
1865 {
1866 if (object == TARGET_OBJECT_MEMORY
1867 && ops->deprecated_xfer_memory != NULL)
1868 /* If available, fall back to the target's
1869 "deprecated_xfer_memory" method. */
1870 {
1871 int xfered = -1;
1872
1873 errno = 0;
1874 if (writebuf != NULL)
1875 {
1876 void *buffer = xmalloc (len);
1877 struct cleanup *cleanup = make_cleanup (xfree, buffer);
1878
1879 memcpy (buffer, writebuf, len);
1880 xfered = ops->deprecated_xfer_memory (offset, buffer, len,
1881 1/*write*/, NULL, ops);
1882 do_cleanups (cleanup);
1883 }
1884 if (readbuf != NULL)
1885 xfered = ops->deprecated_xfer_memory (offset, readbuf, len,
1886 0/*read*/, NULL, ops);
1887 if (xfered > 0)
1888 {
1889 *xfered_len = (ULONGEST) xfered;
1890 return TARGET_XFER_E_IO;
1891 }
1892 else if (xfered == 0 && errno == 0)
1893 /* "deprecated_xfer_memory" uses 0, cross checked against
1894 ERRNO as one indication of an error. */
1895 return TARGET_XFER_EOF;
1896 else
1897 return TARGET_XFER_E_IO;
1898 }
1899 else
1900 {
1901 gdb_assert (ops->beneath != NULL);
1902 return ops->beneath->to_xfer_partial (ops->beneath, object, annex,
1903 readbuf, writebuf, offset, len,
1904 xfered_len);
1905 }
1906 }
1907
1908 /* Target vector read/write partial wrapper functions. */
1909
1910 static enum target_xfer_status
1911 target_read_partial (struct target_ops *ops,
1912 enum target_object object,
1913 const char *annex, gdb_byte *buf,
1914 ULONGEST offset, ULONGEST len,
1915 ULONGEST *xfered_len)
1916 {
1917 return target_xfer_partial (ops, object, annex, buf, NULL, offset, len,
1918 xfered_len);
1919 }
1920
1921 static enum target_xfer_status
1922 target_write_partial (struct target_ops *ops,
1923 enum target_object object,
1924 const char *annex, const gdb_byte *buf,
1925 ULONGEST offset, LONGEST len, ULONGEST *xfered_len)
1926 {
1927 return target_xfer_partial (ops, object, annex, NULL, buf, offset, len,
1928 xfered_len);
1929 }
1930
1931 /* Wrappers to perform the full transfer. */
1932
1933 /* For docs on target_read see target.h. */
1934
1935 LONGEST
1936 target_read (struct target_ops *ops,
1937 enum target_object object,
1938 const char *annex, gdb_byte *buf,
1939 ULONGEST offset, LONGEST len)
1940 {
1941 LONGEST xfered = 0;
1942
1943 while (xfered < len)
1944 {
1945 ULONGEST xfered_len;
1946 enum target_xfer_status status;
1947
1948 status = target_read_partial (ops, object, annex,
1949 (gdb_byte *) buf + xfered,
1950 offset + xfered, len - xfered,
1951 &xfered_len);
1952
1953 /* Call an observer, notifying them of the xfer progress? */
1954 if (status == TARGET_XFER_EOF)
1955 return xfered;
1956 else if (status == TARGET_XFER_OK)
1957 {
1958 xfered += xfered_len;
1959 QUIT;
1960 }
1961 else
1962 return -1;
1963
1964 }
1965 return len;
1966 }
1967
1968 /* Assuming that the entire [begin, end) range of memory cannot be
1969 read, try to read whatever subrange is possible to read.
1970
1971 The function returns, in RESULT, either zero or one memory block.
1972 If there's a readable subrange at the beginning, it is completely
1973 read and returned. Any further readable subrange will not be read.
1974 Otherwise, if there's a readable subrange at the end, it will be
1975 completely read and returned. Any readable subranges before it
1976 (obviously, not starting at the beginning), will be ignored. In
1977 other cases -- either no readable subrange, or readable subrange(s)
1978 that is neither at the beginning, or end, nothing is returned.
1979
1980 The purpose of this function is to handle a read across a boundary
1981 of accessible memory in a case when memory map is not available.
1982 The above restrictions are fine for this case, but will give
1983 incorrect results if the memory is 'patchy'. However, supporting
1984 'patchy' memory would require trying to read every single byte,
1985 and it seems unacceptable solution. Explicit memory map is
1986 recommended for this case -- and target_read_memory_robust will
1987 take care of reading multiple ranges then. */
1988
1989 static void
1990 read_whatever_is_readable (struct target_ops *ops,
1991 ULONGEST begin, ULONGEST end,
1992 VEC(memory_read_result_s) **result)
1993 {
1994 gdb_byte *buf = xmalloc (end - begin);
1995 ULONGEST current_begin = begin;
1996 ULONGEST current_end = end;
1997 int forward;
1998 memory_read_result_s r;
1999 ULONGEST xfered_len;
2000
2001 /* If we previously failed to read 1 byte, nothing can be done here. */
2002 if (end - begin <= 1)
2003 {
2004 xfree (buf);
2005 return;
2006 }
2007
2008 /* Check that either first or the last byte is readable, and give up
2009 if not. This heuristic is meant to permit reading accessible memory
2010 at the boundary of accessible region. */
2011 if (target_read_partial (ops, TARGET_OBJECT_MEMORY, NULL,
2012 buf, begin, 1, &xfered_len) == TARGET_XFER_OK)
2013 {
2014 forward = 1;
2015 ++current_begin;
2016 }
2017 else if (target_read_partial (ops, TARGET_OBJECT_MEMORY, NULL,
2018 buf + (end-begin) - 1, end - 1, 1,
2019 &xfered_len) == TARGET_XFER_OK)
2020 {
2021 forward = 0;
2022 --current_end;
2023 }
2024 else
2025 {
2026 xfree (buf);
2027 return;
2028 }
2029
2030 /* Loop invariant is that the [current_begin, current_end) was previously
2031 found to be not readable as a whole.
2032
2033 Note loop condition -- if the range has 1 byte, we can't divide the range
2034 so there's no point trying further. */
2035 while (current_end - current_begin > 1)
2036 {
2037 ULONGEST first_half_begin, first_half_end;
2038 ULONGEST second_half_begin, second_half_end;
2039 LONGEST xfer;
2040 ULONGEST middle = current_begin + (current_end - current_begin)/2;
2041
2042 if (forward)
2043 {
2044 first_half_begin = current_begin;
2045 first_half_end = middle;
2046 second_half_begin = middle;
2047 second_half_end = current_end;
2048 }
2049 else
2050 {
2051 first_half_begin = middle;
2052 first_half_end = current_end;
2053 second_half_begin = current_begin;
2054 second_half_end = middle;
2055 }
2056
2057 xfer = target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2058 buf + (first_half_begin - begin),
2059 first_half_begin,
2060 first_half_end - first_half_begin);
2061
2062 if (xfer == first_half_end - first_half_begin)
2063 {
2064 /* This half reads up fine. So, the error must be in the
2065 other half. */
2066 current_begin = second_half_begin;
2067 current_end = second_half_end;
2068 }
2069 else
2070 {
2071 /* This half is not readable. Because we've tried one byte, we
2072 know some part of this half if actually redable. Go to the next
2073 iteration to divide again and try to read.
2074
2075 We don't handle the other half, because this function only tries
2076 to read a single readable subrange. */
2077 current_begin = first_half_begin;
2078 current_end = first_half_end;
2079 }
2080 }
2081
2082 if (forward)
2083 {
2084 /* The [begin, current_begin) range has been read. */
2085 r.begin = begin;
2086 r.end = current_begin;
2087 r.data = buf;
2088 }
2089 else
2090 {
2091 /* The [current_end, end) range has been read. */
2092 LONGEST rlen = end - current_end;
2093
2094 r.data = xmalloc (rlen);
2095 memcpy (r.data, buf + current_end - begin, rlen);
2096 r.begin = current_end;
2097 r.end = end;
2098 xfree (buf);
2099 }
2100 VEC_safe_push(memory_read_result_s, (*result), &r);
2101 }
2102
2103 void
2104 free_memory_read_result_vector (void *x)
2105 {
2106 VEC(memory_read_result_s) *v = x;
2107 memory_read_result_s *current;
2108 int ix;
2109
2110 for (ix = 0; VEC_iterate (memory_read_result_s, v, ix, current); ++ix)
2111 {
2112 xfree (current->data);
2113 }
2114 VEC_free (memory_read_result_s, v);
2115 }
2116
2117 VEC(memory_read_result_s) *
2118 read_memory_robust (struct target_ops *ops, ULONGEST offset, LONGEST len)
2119 {
2120 VEC(memory_read_result_s) *result = 0;
2121
2122 LONGEST xfered = 0;
2123 while (xfered < len)
2124 {
2125 struct mem_region *region = lookup_mem_region (offset + xfered);
2126 LONGEST rlen;
2127
2128 /* If there is no explicit region, a fake one should be created. */
2129 gdb_assert (region);
2130
2131 if (region->hi == 0)
2132 rlen = len - xfered;
2133 else
2134 rlen = region->hi - offset;
2135
2136 if (region->attrib.mode == MEM_NONE || region->attrib.mode == MEM_WO)
2137 {
2138 /* Cannot read this region. Note that we can end up here only
2139 if the region is explicitly marked inaccessible, or
2140 'inaccessible-by-default' is in effect. */
2141 xfered += rlen;
2142 }
2143 else
2144 {
2145 LONGEST to_read = min (len - xfered, rlen);
2146 gdb_byte *buffer = (gdb_byte *)xmalloc (to_read);
2147
2148 LONGEST xfer = target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2149 (gdb_byte *) buffer,
2150 offset + xfered, to_read);
2151 /* Call an observer, notifying them of the xfer progress? */
2152 if (xfer <= 0)
2153 {
2154 /* Got an error reading full chunk. See if maybe we can read
2155 some subrange. */
2156 xfree (buffer);
2157 read_whatever_is_readable (ops, offset + xfered,
2158 offset + xfered + to_read, &result);
2159 xfered += to_read;
2160 }
2161 else
2162 {
2163 struct memory_read_result r;
2164 r.data = buffer;
2165 r.begin = offset + xfered;
2166 r.end = r.begin + xfer;
2167 VEC_safe_push (memory_read_result_s, result, &r);
2168 xfered += xfer;
2169 }
2170 QUIT;
2171 }
2172 }
2173 return result;
2174 }
2175
2176
2177 /* An alternative to target_write with progress callbacks. */
2178
2179 LONGEST
2180 target_write_with_progress (struct target_ops *ops,
2181 enum target_object object,
2182 const char *annex, const gdb_byte *buf,
2183 ULONGEST offset, LONGEST len,
2184 void (*progress) (ULONGEST, void *), void *baton)
2185 {
2186 LONGEST xfered = 0;
2187
2188 /* Give the progress callback a chance to set up. */
2189 if (progress)
2190 (*progress) (0, baton);
2191
2192 while (xfered < len)
2193 {
2194 ULONGEST xfered_len;
2195 enum target_xfer_status status;
2196
2197 status = target_write_partial (ops, object, annex,
2198 (gdb_byte *) buf + xfered,
2199 offset + xfered, len - xfered,
2200 &xfered_len);
2201
2202 if (status == TARGET_XFER_EOF)
2203 return xfered;
2204 if (TARGET_XFER_STATUS_ERROR_P (status))
2205 return -1;
2206
2207 gdb_assert (status == TARGET_XFER_OK);
2208 if (progress)
2209 (*progress) (xfered_len, baton);
2210
2211 xfered += xfered_len;
2212 QUIT;
2213 }
2214 return len;
2215 }
2216
2217 /* For docs on target_write see target.h. */
2218
2219 LONGEST
2220 target_write (struct target_ops *ops,
2221 enum target_object object,
2222 const char *annex, const gdb_byte *buf,
2223 ULONGEST offset, LONGEST len)
2224 {
2225 return target_write_with_progress (ops, object, annex, buf, offset, len,
2226 NULL, NULL);
2227 }
2228
2229 /* Read OBJECT/ANNEX using OPS. Store the result in *BUF_P and return
2230 the size of the transferred data. PADDING additional bytes are
2231 available in *BUF_P. This is a helper function for
2232 target_read_alloc; see the declaration of that function for more
2233 information. */
2234
2235 static LONGEST
2236 target_read_alloc_1 (struct target_ops *ops, enum target_object object,
2237 const char *annex, gdb_byte **buf_p, int padding)
2238 {
2239 size_t buf_alloc, buf_pos;
2240 gdb_byte *buf;
2241
2242 /* This function does not have a length parameter; it reads the
2243 entire OBJECT). Also, it doesn't support objects fetched partly
2244 from one target and partly from another (in a different stratum,
2245 e.g. a core file and an executable). Both reasons make it
2246 unsuitable for reading memory. */
2247 gdb_assert (object != TARGET_OBJECT_MEMORY);
2248
2249 /* Start by reading up to 4K at a time. The target will throttle
2250 this number down if necessary. */
2251 buf_alloc = 4096;
2252 buf = xmalloc (buf_alloc);
2253 buf_pos = 0;
2254 while (1)
2255 {
2256 ULONGEST xfered_len;
2257 enum target_xfer_status status;
2258
2259 status = target_read_partial (ops, object, annex, &buf[buf_pos],
2260 buf_pos, buf_alloc - buf_pos - padding,
2261 &xfered_len);
2262
2263 if (status == TARGET_XFER_EOF)
2264 {
2265 /* Read all there was. */
2266 if (buf_pos == 0)
2267 xfree (buf);
2268 else
2269 *buf_p = buf;
2270 return buf_pos;
2271 }
2272 else if (status != TARGET_XFER_OK)
2273 {
2274 /* An error occurred. */
2275 xfree (buf);
2276 return TARGET_XFER_E_IO;
2277 }
2278
2279 buf_pos += xfered_len;
2280
2281 /* If the buffer is filling up, expand it. */
2282 if (buf_alloc < buf_pos * 2)
2283 {
2284 buf_alloc *= 2;
2285 buf = xrealloc (buf, buf_alloc);
2286 }
2287
2288 QUIT;
2289 }
2290 }
2291
2292 /* Read OBJECT/ANNEX using OPS. Store the result in *BUF_P and return
2293 the size of the transferred data. See the declaration in "target.h"
2294 function for more information about the return value. */
2295
2296 LONGEST
2297 target_read_alloc (struct target_ops *ops, enum target_object object,
2298 const char *annex, gdb_byte **buf_p)
2299 {
2300 return target_read_alloc_1 (ops, object, annex, buf_p, 0);
2301 }
2302
2303 /* Read OBJECT/ANNEX using OPS. The result is NUL-terminated and
2304 returned as a string, allocated using xmalloc. If an error occurs
2305 or the transfer is unsupported, NULL is returned. Empty objects
2306 are returned as allocated but empty strings. A warning is issued
2307 if the result contains any embedded NUL bytes. */
2308
2309 char *
2310 target_read_stralloc (struct target_ops *ops, enum target_object object,
2311 const char *annex)
2312 {
2313 gdb_byte *buffer;
2314 char *bufstr;
2315 LONGEST i, transferred;
2316
2317 transferred = target_read_alloc_1 (ops, object, annex, &buffer, 1);
2318 bufstr = (char *) buffer;
2319
2320 if (transferred < 0)
2321 return NULL;
2322
2323 if (transferred == 0)
2324 return xstrdup ("");
2325
2326 bufstr[transferred] = 0;
2327
2328 /* Check for embedded NUL bytes; but allow trailing NULs. */
2329 for (i = strlen (bufstr); i < transferred; i++)
2330 if (bufstr[i] != 0)
2331 {
2332 warning (_("target object %d, annex %s, "
2333 "contained unexpected null characters"),
2334 (int) object, annex ? annex : "(none)");
2335 break;
2336 }
2337
2338 return bufstr;
2339 }
2340
2341 /* Memory transfer methods. */
2342
2343 void
2344 get_target_memory (struct target_ops *ops, CORE_ADDR addr, gdb_byte *buf,
2345 LONGEST len)
2346 {
2347 /* This method is used to read from an alternate, non-current
2348 target. This read must bypass the overlay support (as symbols
2349 don't match this target), and GDB's internal cache (wrong cache
2350 for this target). */
2351 if (target_read (ops, TARGET_OBJECT_RAW_MEMORY, NULL, buf, addr, len)
2352 != len)
2353 memory_error (TARGET_XFER_E_IO, addr);
2354 }
2355
2356 ULONGEST
2357 get_target_memory_unsigned (struct target_ops *ops, CORE_ADDR addr,
2358 int len, enum bfd_endian byte_order)
2359 {
2360 gdb_byte buf[sizeof (ULONGEST)];
2361
2362 gdb_assert (len <= sizeof (buf));
2363 get_target_memory (ops, addr, buf, len);
2364 return extract_unsigned_integer (buf, len, byte_order);
2365 }
2366
2367 /* See target.h. */
2368
2369 int
2370 target_insert_breakpoint (struct gdbarch *gdbarch,
2371 struct bp_target_info *bp_tgt)
2372 {
2373 if (!may_insert_breakpoints)
2374 {
2375 warning (_("May not insert breakpoints"));
2376 return 1;
2377 }
2378
2379 return current_target.to_insert_breakpoint (&current_target,
2380 gdbarch, bp_tgt);
2381 }
2382
2383 /* See target.h. */
2384
2385 int
2386 target_remove_breakpoint (struct gdbarch *gdbarch,
2387 struct bp_target_info *bp_tgt)
2388 {
2389 /* This is kind of a weird case to handle, but the permission might
2390 have been changed after breakpoints were inserted - in which case
2391 we should just take the user literally and assume that any
2392 breakpoints should be left in place. */
2393 if (!may_insert_breakpoints)
2394 {
2395 warning (_("May not remove breakpoints"));
2396 return 1;
2397 }
2398
2399 return current_target.to_remove_breakpoint (&current_target,
2400 gdbarch, bp_tgt);
2401 }
2402
2403 static void
2404 target_info (char *args, int from_tty)
2405 {
2406 struct target_ops *t;
2407 int has_all_mem = 0;
2408
2409 if (symfile_objfile != NULL)
2410 printf_unfiltered (_("Symbols from \"%s\".\n"),
2411 objfile_name (symfile_objfile));
2412
2413 for (t = target_stack; t != NULL; t = t->beneath)
2414 {
2415 if (!(*t->to_has_memory) (t))
2416 continue;
2417
2418 if ((int) (t->to_stratum) <= (int) dummy_stratum)
2419 continue;
2420 if (has_all_mem)
2421 printf_unfiltered (_("\tWhile running this, "
2422 "GDB does not access memory from...\n"));
2423 printf_unfiltered ("%s:\n", t->to_longname);
2424 (t->to_files_info) (t);
2425 has_all_mem = (*t->to_has_all_memory) (t);
2426 }
2427 }
2428
2429 /* This function is called before any new inferior is created, e.g.
2430 by running a program, attaching, or connecting to a target.
2431 It cleans up any state from previous invocations which might
2432 change between runs. This is a subset of what target_preopen
2433 resets (things which might change between targets). */
2434
2435 void
2436 target_pre_inferior (int from_tty)
2437 {
2438 /* Clear out solib state. Otherwise the solib state of the previous
2439 inferior might have survived and is entirely wrong for the new
2440 target. This has been observed on GNU/Linux using glibc 2.3. How
2441 to reproduce:
2442
2443 bash$ ./foo&
2444 [1] 4711
2445 bash$ ./foo&
2446 [1] 4712
2447 bash$ gdb ./foo
2448 [...]
2449 (gdb) attach 4711
2450 (gdb) detach
2451 (gdb) attach 4712
2452 Cannot access memory at address 0xdeadbeef
2453 */
2454
2455 /* In some OSs, the shared library list is the same/global/shared
2456 across inferiors. If code is shared between processes, so are
2457 memory regions and features. */
2458 if (!gdbarch_has_global_solist (target_gdbarch ()))
2459 {
2460 no_shared_libraries (NULL, from_tty);
2461
2462 invalidate_target_mem_regions ();
2463
2464 target_clear_description ();
2465 }
2466
2467 agent_capability_invalidate ();
2468 }
2469
2470 /* Callback for iterate_over_inferiors. Gets rid of the given
2471 inferior. */
2472
2473 static int
2474 dispose_inferior (struct inferior *inf, void *args)
2475 {
2476 struct thread_info *thread;
2477
2478 thread = any_thread_of_process (inf->pid);
2479 if (thread)
2480 {
2481 switch_to_thread (thread->ptid);
2482
2483 /* Core inferiors actually should be detached, not killed. */
2484 if (target_has_execution)
2485 target_kill ();
2486 else
2487 target_detach (NULL, 0);
2488 }
2489
2490 return 0;
2491 }
2492
2493 /* This is to be called by the open routine before it does
2494 anything. */
2495
2496 void
2497 target_preopen (int from_tty)
2498 {
2499 dont_repeat ();
2500
2501 if (have_inferiors ())
2502 {
2503 if (!from_tty
2504 || !have_live_inferiors ()
2505 || query (_("A program is being debugged already. Kill it? ")))
2506 iterate_over_inferiors (dispose_inferior, NULL);
2507 else
2508 error (_("Program not killed."));
2509 }
2510
2511 /* Calling target_kill may remove the target from the stack. But if
2512 it doesn't (which seems like a win for UDI), remove it now. */
2513 /* Leave the exec target, though. The user may be switching from a
2514 live process to a core of the same program. */
2515 pop_all_targets_above (file_stratum);
2516
2517 target_pre_inferior (from_tty);
2518 }
2519
2520 /* Detach a target after doing deferred register stores. */
2521
2522 void
2523 target_detach (const char *args, int from_tty)
2524 {
2525 struct target_ops* t;
2526
2527 if (gdbarch_has_global_breakpoints (target_gdbarch ()))
2528 /* Don't remove global breakpoints here. They're removed on
2529 disconnection from the target. */
2530 ;
2531 else
2532 /* If we're in breakpoints-always-inserted mode, have to remove
2533 them before detaching. */
2534 remove_breakpoints_pid (ptid_get_pid (inferior_ptid));
2535
2536 prepare_for_detach ();
2537
2538 current_target.to_detach (&current_target, args, from_tty);
2539 if (targetdebug)
2540 fprintf_unfiltered (gdb_stdlog, "target_detach (%s, %d)\n",
2541 args, from_tty);
2542 }
2543
2544 void
2545 target_disconnect (char *args, int from_tty)
2546 {
2547 struct target_ops *t;
2548
2549 /* If we're in breakpoints-always-inserted mode or if breakpoints
2550 are global across processes, we have to remove them before
2551 disconnecting. */
2552 remove_breakpoints ();
2553
2554 for (t = current_target.beneath; t != NULL; t = t->beneath)
2555 if (t->to_disconnect != NULL)
2556 {
2557 if (targetdebug)
2558 fprintf_unfiltered (gdb_stdlog, "target_disconnect (%s, %d)\n",
2559 args, from_tty);
2560 t->to_disconnect (t, args, from_tty);
2561 return;
2562 }
2563
2564 tcomplain ();
2565 }
2566
2567 ptid_t
2568 target_wait (ptid_t ptid, struct target_waitstatus *status, int options)
2569 {
2570 struct target_ops *t;
2571 ptid_t retval = (current_target.to_wait) (&current_target, ptid,
2572 status, options);
2573
2574 if (targetdebug)
2575 {
2576 char *status_string;
2577 char *options_string;
2578
2579 status_string = target_waitstatus_to_string (status);
2580 options_string = target_options_to_string (options);
2581 fprintf_unfiltered (gdb_stdlog,
2582 "target_wait (%d, status, options={%s})"
2583 " = %d, %s\n",
2584 ptid_get_pid (ptid), options_string,
2585 ptid_get_pid (retval), status_string);
2586 xfree (status_string);
2587 xfree (options_string);
2588 }
2589
2590 return retval;
2591 }
2592
2593 char *
2594 target_pid_to_str (ptid_t ptid)
2595 {
2596 struct target_ops *t;
2597
2598 for (t = current_target.beneath; t != NULL; t = t->beneath)
2599 {
2600 if (t->to_pid_to_str != NULL)
2601 return (*t->to_pid_to_str) (t, ptid);
2602 }
2603
2604 return normal_pid_to_str (ptid);
2605 }
2606
2607 char *
2608 target_thread_name (struct thread_info *info)
2609 {
2610 return current_target.to_thread_name (&current_target, info);
2611 }
2612
2613 void
2614 target_resume (ptid_t ptid, int step, enum gdb_signal signal)
2615 {
2616 struct target_ops *t;
2617
2618 target_dcache_invalidate ();
2619
2620 current_target.to_resume (&current_target, ptid, step, signal);
2621 if (targetdebug)
2622 fprintf_unfiltered (gdb_stdlog, "target_resume (%d, %s, %s)\n",
2623 ptid_get_pid (ptid),
2624 step ? "step" : "continue",
2625 gdb_signal_to_name (signal));
2626
2627 registers_changed_ptid (ptid);
2628 set_executing (ptid, 1);
2629 set_running (ptid, 1);
2630 clear_inline_frame_state (ptid);
2631 }
2632
2633 void
2634 target_pass_signals (int numsigs, unsigned char *pass_signals)
2635 {
2636 struct target_ops *t;
2637
2638 for (t = current_target.beneath; t != NULL; t = t->beneath)
2639 {
2640 if (t->to_pass_signals != NULL)
2641 {
2642 if (targetdebug)
2643 {
2644 int i;
2645
2646 fprintf_unfiltered (gdb_stdlog, "target_pass_signals (%d, {",
2647 numsigs);
2648
2649 for (i = 0; i < numsigs; i++)
2650 if (pass_signals[i])
2651 fprintf_unfiltered (gdb_stdlog, " %s",
2652 gdb_signal_to_name (i));
2653
2654 fprintf_unfiltered (gdb_stdlog, " })\n");
2655 }
2656
2657 (*t->to_pass_signals) (t, numsigs, pass_signals);
2658 return;
2659 }
2660 }
2661 }
2662
2663 void
2664 target_program_signals (int numsigs, unsigned char *program_signals)
2665 {
2666 struct target_ops *t;
2667
2668 for (t = current_target.beneath; t != NULL; t = t->beneath)
2669 {
2670 if (t->to_program_signals != NULL)
2671 {
2672 if (targetdebug)
2673 {
2674 int i;
2675
2676 fprintf_unfiltered (gdb_stdlog, "target_program_signals (%d, {",
2677 numsigs);
2678
2679 for (i = 0; i < numsigs; i++)
2680 if (program_signals[i])
2681 fprintf_unfiltered (gdb_stdlog, " %s",
2682 gdb_signal_to_name (i));
2683
2684 fprintf_unfiltered (gdb_stdlog, " })\n");
2685 }
2686
2687 (*t->to_program_signals) (t, numsigs, program_signals);
2688 return;
2689 }
2690 }
2691 }
2692
2693 /* Look through the list of possible targets for a target that can
2694 follow forks. */
2695
2696 int
2697 target_follow_fork (int follow_child, int detach_fork)
2698 {
2699 struct target_ops *t;
2700
2701 for (t = current_target.beneath; t != NULL; t = t->beneath)
2702 {
2703 if (t->to_follow_fork != NULL)
2704 {
2705 int retval = t->to_follow_fork (t, follow_child, detach_fork);
2706
2707 if (targetdebug)
2708 fprintf_unfiltered (gdb_stdlog,
2709 "target_follow_fork (%d, %d) = %d\n",
2710 follow_child, detach_fork, retval);
2711 return retval;
2712 }
2713 }
2714
2715 /* Some target returned a fork event, but did not know how to follow it. */
2716 internal_error (__FILE__, __LINE__,
2717 _("could not find a target to follow fork"));
2718 }
2719
2720 void
2721 target_mourn_inferior (void)
2722 {
2723 struct target_ops *t;
2724
2725 for (t = current_target.beneath; t != NULL; t = t->beneath)
2726 {
2727 if (t->to_mourn_inferior != NULL)
2728 {
2729 t->to_mourn_inferior (t);
2730 if (targetdebug)
2731 fprintf_unfiltered (gdb_stdlog, "target_mourn_inferior ()\n");
2732
2733 /* We no longer need to keep handles on any of the object files.
2734 Make sure to release them to avoid unnecessarily locking any
2735 of them while we're not actually debugging. */
2736 bfd_cache_close_all ();
2737
2738 return;
2739 }
2740 }
2741
2742 internal_error (__FILE__, __LINE__,
2743 _("could not find a target to follow mourn inferior"));
2744 }
2745
2746 /* Look for a target which can describe architectural features, starting
2747 from TARGET. If we find one, return its description. */
2748
2749 const struct target_desc *
2750 target_read_description (struct target_ops *target)
2751 {
2752 struct target_ops *t;
2753
2754 for (t = target; t != NULL; t = t->beneath)
2755 if (t->to_read_description != NULL)
2756 {
2757 const struct target_desc *tdesc;
2758
2759 tdesc = t->to_read_description (t);
2760 if (tdesc)
2761 return tdesc;
2762 }
2763
2764 return NULL;
2765 }
2766
2767 /* The default implementation of to_search_memory.
2768 This implements a basic search of memory, reading target memory and
2769 performing the search here (as opposed to performing the search in on the
2770 target side with, for example, gdbserver). */
2771
2772 int
2773 simple_search_memory (struct target_ops *ops,
2774 CORE_ADDR start_addr, ULONGEST search_space_len,
2775 const gdb_byte *pattern, ULONGEST pattern_len,
2776 CORE_ADDR *found_addrp)
2777 {
2778 /* NOTE: also defined in find.c testcase. */
2779 #define SEARCH_CHUNK_SIZE 16000
2780 const unsigned chunk_size = SEARCH_CHUNK_SIZE;
2781 /* Buffer to hold memory contents for searching. */
2782 gdb_byte *search_buf;
2783 unsigned search_buf_size;
2784 struct cleanup *old_cleanups;
2785
2786 search_buf_size = chunk_size + pattern_len - 1;
2787
2788 /* No point in trying to allocate a buffer larger than the search space. */
2789 if (search_space_len < search_buf_size)
2790 search_buf_size = search_space_len;
2791
2792 search_buf = malloc (search_buf_size);
2793 if (search_buf == NULL)
2794 error (_("Unable to allocate memory to perform the search."));
2795 old_cleanups = make_cleanup (free_current_contents, &search_buf);
2796
2797 /* Prime the search buffer. */
2798
2799 if (target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2800 search_buf, start_addr, search_buf_size) != search_buf_size)
2801 {
2802 warning (_("Unable to access %s bytes of target "
2803 "memory at %s, halting search."),
2804 pulongest (search_buf_size), hex_string (start_addr));
2805 do_cleanups (old_cleanups);
2806 return -1;
2807 }
2808
2809 /* Perform the search.
2810
2811 The loop is kept simple by allocating [N + pattern-length - 1] bytes.
2812 When we've scanned N bytes we copy the trailing bytes to the start and
2813 read in another N bytes. */
2814
2815 while (search_space_len >= pattern_len)
2816 {
2817 gdb_byte *found_ptr;
2818 unsigned nr_search_bytes = min (search_space_len, search_buf_size);
2819
2820 found_ptr = memmem (search_buf, nr_search_bytes,
2821 pattern, pattern_len);
2822
2823 if (found_ptr != NULL)
2824 {
2825 CORE_ADDR found_addr = start_addr + (found_ptr - search_buf);
2826
2827 *found_addrp = found_addr;
2828 do_cleanups (old_cleanups);
2829 return 1;
2830 }
2831
2832 /* Not found in this chunk, skip to next chunk. */
2833
2834 /* Don't let search_space_len wrap here, it's unsigned. */
2835 if (search_space_len >= chunk_size)
2836 search_space_len -= chunk_size;
2837 else
2838 search_space_len = 0;
2839
2840 if (search_space_len >= pattern_len)
2841 {
2842 unsigned keep_len = search_buf_size - chunk_size;
2843 CORE_ADDR read_addr = start_addr + chunk_size + keep_len;
2844 int nr_to_read;
2845
2846 /* Copy the trailing part of the previous iteration to the front
2847 of the buffer for the next iteration. */
2848 gdb_assert (keep_len == pattern_len - 1);
2849 memcpy (search_buf, search_buf + chunk_size, keep_len);
2850
2851 nr_to_read = min (search_space_len - keep_len, chunk_size);
2852
2853 if (target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2854 search_buf + keep_len, read_addr,
2855 nr_to_read) != nr_to_read)
2856 {
2857 warning (_("Unable to access %s bytes of target "
2858 "memory at %s, halting search."),
2859 plongest (nr_to_read),
2860 hex_string (read_addr));
2861 do_cleanups (old_cleanups);
2862 return -1;
2863 }
2864
2865 start_addr += chunk_size;
2866 }
2867 }
2868
2869 /* Not found. */
2870
2871 do_cleanups (old_cleanups);
2872 return 0;
2873 }
2874
2875 /* Search SEARCH_SPACE_LEN bytes beginning at START_ADDR for the
2876 sequence of bytes in PATTERN with length PATTERN_LEN.
2877
2878 The result is 1 if found, 0 if not found, and -1 if there was an error
2879 requiring halting of the search (e.g. memory read error).
2880 If the pattern is found the address is recorded in FOUND_ADDRP. */
2881
2882 int
2883 target_search_memory (CORE_ADDR start_addr, ULONGEST search_space_len,
2884 const gdb_byte *pattern, ULONGEST pattern_len,
2885 CORE_ADDR *found_addrp)
2886 {
2887 struct target_ops *t;
2888 int found;
2889
2890 /* We don't use INHERIT to set current_target.to_search_memory,
2891 so we have to scan the target stack and handle targetdebug
2892 ourselves. */
2893
2894 if (targetdebug)
2895 fprintf_unfiltered (gdb_stdlog, "target_search_memory (%s, ...)\n",
2896 hex_string (start_addr));
2897
2898 for (t = current_target.beneath; t != NULL; t = t->beneath)
2899 if (t->to_search_memory != NULL)
2900 break;
2901
2902 if (t != NULL)
2903 {
2904 found = t->to_search_memory (t, start_addr, search_space_len,
2905 pattern, pattern_len, found_addrp);
2906 }
2907 else
2908 {
2909 /* If a special version of to_search_memory isn't available, use the
2910 simple version. */
2911 found = simple_search_memory (current_target.beneath,
2912 start_addr, search_space_len,
2913 pattern, pattern_len, found_addrp);
2914 }
2915
2916 if (targetdebug)
2917 fprintf_unfiltered (gdb_stdlog, " = %d\n", found);
2918
2919 return found;
2920 }
2921
2922 /* Look through the currently pushed targets. If none of them will
2923 be able to restart the currently running process, issue an error
2924 message. */
2925
2926 void
2927 target_require_runnable (void)
2928 {
2929 struct target_ops *t;
2930
2931 for (t = target_stack; t != NULL; t = t->beneath)
2932 {
2933 /* If this target knows how to create a new program, then
2934 assume we will still be able to after killing the current
2935 one. Either killing and mourning will not pop T, or else
2936 find_default_run_target will find it again. */
2937 if (t->to_create_inferior != NULL)
2938 return;
2939
2940 /* Do not worry about thread_stratum targets that can not
2941 create inferiors. Assume they will be pushed again if
2942 necessary, and continue to the process_stratum. */
2943 if (t->to_stratum == thread_stratum
2944 || t->to_stratum == arch_stratum)
2945 continue;
2946
2947 error (_("The \"%s\" target does not support \"run\". "
2948 "Try \"help target\" or \"continue\"."),
2949 t->to_shortname);
2950 }
2951
2952 /* This function is only called if the target is running. In that
2953 case there should have been a process_stratum target and it
2954 should either know how to create inferiors, or not... */
2955 internal_error (__FILE__, __LINE__, _("No targets found"));
2956 }
2957
2958 /* Look through the list of possible targets for a target that can
2959 execute a run or attach command without any other data. This is
2960 used to locate the default process stratum.
2961
2962 If DO_MESG is not NULL, the result is always valid (error() is
2963 called for errors); else, return NULL on error. */
2964
2965 static struct target_ops *
2966 find_default_run_target (char *do_mesg)
2967 {
2968 struct target_ops **t;
2969 struct target_ops *runable = NULL;
2970 int count;
2971
2972 count = 0;
2973
2974 for (t = target_structs; t < target_structs + target_struct_size;
2975 ++t)
2976 {
2977 if ((*t)->to_can_run && target_can_run (*t))
2978 {
2979 runable = *t;
2980 ++count;
2981 }
2982 }
2983
2984 if (count != 1)
2985 {
2986 if (do_mesg)
2987 error (_("Don't know how to %s. Try \"help target\"."), do_mesg);
2988 else
2989 return NULL;
2990 }
2991
2992 return runable;
2993 }
2994
2995 void
2996 find_default_attach (struct target_ops *ops, char *args, int from_tty)
2997 {
2998 struct target_ops *t;
2999
3000 t = find_default_run_target ("attach");
3001 (t->to_attach) (t, args, from_tty);
3002 return;
3003 }
3004
3005 void
3006 find_default_create_inferior (struct target_ops *ops,
3007 char *exec_file, char *allargs, char **env,
3008 int from_tty)
3009 {
3010 struct target_ops *t;
3011
3012 t = find_default_run_target ("run");
3013 (t->to_create_inferior) (t, exec_file, allargs, env, from_tty);
3014 return;
3015 }
3016
3017 static int
3018 find_default_can_async_p (struct target_ops *ignore)
3019 {
3020 struct target_ops *t;
3021
3022 /* This may be called before the target is pushed on the stack;
3023 look for the default process stratum. If there's none, gdb isn't
3024 configured with a native debugger, and target remote isn't
3025 connected yet. */
3026 t = find_default_run_target (NULL);
3027 if (t && t->to_can_async_p != delegate_can_async_p)
3028 return (t->to_can_async_p) (t);
3029 return 0;
3030 }
3031
3032 static int
3033 find_default_is_async_p (struct target_ops *ignore)
3034 {
3035 struct target_ops *t;
3036
3037 /* This may be called before the target is pushed on the stack;
3038 look for the default process stratum. If there's none, gdb isn't
3039 configured with a native debugger, and target remote isn't
3040 connected yet. */
3041 t = find_default_run_target (NULL);
3042 if (t && t->to_is_async_p != delegate_is_async_p)
3043 return (t->to_is_async_p) (t);
3044 return 0;
3045 }
3046
3047 static int
3048 find_default_supports_non_stop (struct target_ops *self)
3049 {
3050 struct target_ops *t;
3051
3052 t = find_default_run_target (NULL);
3053 if (t && t->to_supports_non_stop)
3054 return (t->to_supports_non_stop) (t);
3055 return 0;
3056 }
3057
3058 int
3059 target_supports_non_stop (void)
3060 {
3061 struct target_ops *t;
3062
3063 for (t = &current_target; t != NULL; t = t->beneath)
3064 if (t->to_supports_non_stop)
3065 return t->to_supports_non_stop (t);
3066
3067 return 0;
3068 }
3069
3070 /* Implement the "info proc" command. */
3071
3072 int
3073 target_info_proc (char *args, enum info_proc_what what)
3074 {
3075 struct target_ops *t;
3076
3077 /* If we're already connected to something that can get us OS
3078 related data, use it. Otherwise, try using the native
3079 target. */
3080 if (current_target.to_stratum >= process_stratum)
3081 t = current_target.beneath;
3082 else
3083 t = find_default_run_target (NULL);
3084
3085 for (; t != NULL; t = t->beneath)
3086 {
3087 if (t->to_info_proc != NULL)
3088 {
3089 t->to_info_proc (t, args, what);
3090
3091 if (targetdebug)
3092 fprintf_unfiltered (gdb_stdlog,
3093 "target_info_proc (\"%s\", %d)\n", args, what);
3094
3095 return 1;
3096 }
3097 }
3098
3099 return 0;
3100 }
3101
3102 static int
3103 find_default_supports_disable_randomization (struct target_ops *self)
3104 {
3105 struct target_ops *t;
3106
3107 t = find_default_run_target (NULL);
3108 if (t && t->to_supports_disable_randomization)
3109 return (t->to_supports_disable_randomization) (t);
3110 return 0;
3111 }
3112
3113 int
3114 target_supports_disable_randomization (void)
3115 {
3116 struct target_ops *t;
3117
3118 for (t = &current_target; t != NULL; t = t->beneath)
3119 if (t->to_supports_disable_randomization)
3120 return t->to_supports_disable_randomization (t);
3121
3122 return 0;
3123 }
3124
3125 char *
3126 target_get_osdata (const char *type)
3127 {
3128 struct target_ops *t;
3129
3130 /* If we're already connected to something that can get us OS
3131 related data, use it. Otherwise, try using the native
3132 target. */
3133 if (current_target.to_stratum >= process_stratum)
3134 t = current_target.beneath;
3135 else
3136 t = find_default_run_target ("get OS data");
3137
3138 if (!t)
3139 return NULL;
3140
3141 return target_read_stralloc (t, TARGET_OBJECT_OSDATA, type);
3142 }
3143
3144 /* Determine the current address space of thread PTID. */
3145
3146 struct address_space *
3147 target_thread_address_space (ptid_t ptid)
3148 {
3149 struct address_space *aspace;
3150 struct inferior *inf;
3151 struct target_ops *t;
3152
3153 for (t = current_target.beneath; t != NULL; t = t->beneath)
3154 {
3155 if (t->to_thread_address_space != NULL)
3156 {
3157 aspace = t->to_thread_address_space (t, ptid);
3158 gdb_assert (aspace);
3159
3160 if (targetdebug)
3161 fprintf_unfiltered (gdb_stdlog,
3162 "target_thread_address_space (%s) = %d\n",
3163 target_pid_to_str (ptid),
3164 address_space_num (aspace));
3165 return aspace;
3166 }
3167 }
3168
3169 /* Fall-back to the "main" address space of the inferior. */
3170 inf = find_inferior_pid (ptid_get_pid (ptid));
3171
3172 if (inf == NULL || inf->aspace == NULL)
3173 internal_error (__FILE__, __LINE__,
3174 _("Can't determine the current "
3175 "address space of thread %s\n"),
3176 target_pid_to_str (ptid));
3177
3178 return inf->aspace;
3179 }
3180
3181
3182 /* Target file operations. */
3183
3184 static struct target_ops *
3185 default_fileio_target (void)
3186 {
3187 /* If we're already connected to something that can perform
3188 file I/O, use it. Otherwise, try using the native target. */
3189 if (current_target.to_stratum >= process_stratum)
3190 return current_target.beneath;
3191 else
3192 return find_default_run_target ("file I/O");
3193 }
3194
3195 /* Open FILENAME on the target, using FLAGS and MODE. Return a
3196 target file descriptor, or -1 if an error occurs (and set
3197 *TARGET_ERRNO). */
3198 int
3199 target_fileio_open (const char *filename, int flags, int mode,
3200 int *target_errno)
3201 {
3202 struct target_ops *t;
3203
3204 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3205 {
3206 if (t->to_fileio_open != NULL)
3207 {
3208 int fd = t->to_fileio_open (t, filename, flags, mode, target_errno);
3209
3210 if (targetdebug)
3211 fprintf_unfiltered (gdb_stdlog,
3212 "target_fileio_open (%s,0x%x,0%o) = %d (%d)\n",
3213 filename, flags, mode,
3214 fd, fd != -1 ? 0 : *target_errno);
3215 return fd;
3216 }
3217 }
3218
3219 *target_errno = FILEIO_ENOSYS;
3220 return -1;
3221 }
3222
3223 /* Write up to LEN bytes from WRITE_BUF to FD on the target.
3224 Return the number of bytes written, or -1 if an error occurs
3225 (and set *TARGET_ERRNO). */
3226 int
3227 target_fileio_pwrite (int fd, const gdb_byte *write_buf, int len,
3228 ULONGEST offset, int *target_errno)
3229 {
3230 struct target_ops *t;
3231
3232 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3233 {
3234 if (t->to_fileio_pwrite != NULL)
3235 {
3236 int ret = t->to_fileio_pwrite (t, fd, write_buf, len, offset,
3237 target_errno);
3238
3239 if (targetdebug)
3240 fprintf_unfiltered (gdb_stdlog,
3241 "target_fileio_pwrite (%d,...,%d,%s) "
3242 "= %d (%d)\n",
3243 fd, len, pulongest (offset),
3244 ret, ret != -1 ? 0 : *target_errno);
3245 return ret;
3246 }
3247 }
3248
3249 *target_errno = FILEIO_ENOSYS;
3250 return -1;
3251 }
3252
3253 /* Read up to LEN bytes FD on the target into READ_BUF.
3254 Return the number of bytes read, or -1 if an error occurs
3255 (and set *TARGET_ERRNO). */
3256 int
3257 target_fileio_pread (int fd, gdb_byte *read_buf, int len,
3258 ULONGEST offset, int *target_errno)
3259 {
3260 struct target_ops *t;
3261
3262 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3263 {
3264 if (t->to_fileio_pread != NULL)
3265 {
3266 int ret = t->to_fileio_pread (t, fd, read_buf, len, offset,
3267 target_errno);
3268
3269 if (targetdebug)
3270 fprintf_unfiltered (gdb_stdlog,
3271 "target_fileio_pread (%d,...,%d,%s) "
3272 "= %d (%d)\n",
3273 fd, len, pulongest (offset),
3274 ret, ret != -1 ? 0 : *target_errno);
3275 return ret;
3276 }
3277 }
3278
3279 *target_errno = FILEIO_ENOSYS;
3280 return -1;
3281 }
3282
3283 /* Close FD on the target. Return 0, or -1 if an error occurs
3284 (and set *TARGET_ERRNO). */
3285 int
3286 target_fileio_close (int fd, int *target_errno)
3287 {
3288 struct target_ops *t;
3289
3290 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3291 {
3292 if (t->to_fileio_close != NULL)
3293 {
3294 int ret = t->to_fileio_close (t, fd, target_errno);
3295
3296 if (targetdebug)
3297 fprintf_unfiltered (gdb_stdlog,
3298 "target_fileio_close (%d) = %d (%d)\n",
3299 fd, ret, ret != -1 ? 0 : *target_errno);
3300 return ret;
3301 }
3302 }
3303
3304 *target_errno = FILEIO_ENOSYS;
3305 return -1;
3306 }
3307
3308 /* Unlink FILENAME on the target. Return 0, or -1 if an error
3309 occurs (and set *TARGET_ERRNO). */
3310 int
3311 target_fileio_unlink (const char *filename, int *target_errno)
3312 {
3313 struct target_ops *t;
3314
3315 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3316 {
3317 if (t->to_fileio_unlink != NULL)
3318 {
3319 int ret = t->to_fileio_unlink (t, filename, target_errno);
3320
3321 if (targetdebug)
3322 fprintf_unfiltered (gdb_stdlog,
3323 "target_fileio_unlink (%s) = %d (%d)\n",
3324 filename, ret, ret != -1 ? 0 : *target_errno);
3325 return ret;
3326 }
3327 }
3328
3329 *target_errno = FILEIO_ENOSYS;
3330 return -1;
3331 }
3332
3333 /* Read value of symbolic link FILENAME on the target. Return a
3334 null-terminated string allocated via xmalloc, or NULL if an error
3335 occurs (and set *TARGET_ERRNO). */
3336 char *
3337 target_fileio_readlink (const char *filename, int *target_errno)
3338 {
3339 struct target_ops *t;
3340
3341 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3342 {
3343 if (t->to_fileio_readlink != NULL)
3344 {
3345 char *ret = t->to_fileio_readlink (t, filename, target_errno);
3346
3347 if (targetdebug)
3348 fprintf_unfiltered (gdb_stdlog,
3349 "target_fileio_readlink (%s) = %s (%d)\n",
3350 filename, ret? ret : "(nil)",
3351 ret? 0 : *target_errno);
3352 return ret;
3353 }
3354 }
3355
3356 *target_errno = FILEIO_ENOSYS;
3357 return NULL;
3358 }
3359
3360 static void
3361 target_fileio_close_cleanup (void *opaque)
3362 {
3363 int fd = *(int *) opaque;
3364 int target_errno;
3365
3366 target_fileio_close (fd, &target_errno);
3367 }
3368
3369 /* Read target file FILENAME. Store the result in *BUF_P and
3370 return the size of the transferred data. PADDING additional bytes are
3371 available in *BUF_P. This is a helper function for
3372 target_fileio_read_alloc; see the declaration of that function for more
3373 information. */
3374
3375 static LONGEST
3376 target_fileio_read_alloc_1 (const char *filename,
3377 gdb_byte **buf_p, int padding)
3378 {
3379 struct cleanup *close_cleanup;
3380 size_t buf_alloc, buf_pos;
3381 gdb_byte *buf;
3382 LONGEST n;
3383 int fd;
3384 int target_errno;
3385
3386 fd = target_fileio_open (filename, FILEIO_O_RDONLY, 0700, &target_errno);
3387 if (fd == -1)
3388 return -1;
3389
3390 close_cleanup = make_cleanup (target_fileio_close_cleanup, &fd);
3391
3392 /* Start by reading up to 4K at a time. The target will throttle
3393 this number down if necessary. */
3394 buf_alloc = 4096;
3395 buf = xmalloc (buf_alloc);
3396 buf_pos = 0;
3397 while (1)
3398 {
3399 n = target_fileio_pread (fd, &buf[buf_pos],
3400 buf_alloc - buf_pos - padding, buf_pos,
3401 &target_errno);
3402 if (n < 0)
3403 {
3404 /* An error occurred. */
3405 do_cleanups (close_cleanup);
3406 xfree (buf);
3407 return -1;
3408 }
3409 else if (n == 0)
3410 {
3411 /* Read all there was. */
3412 do_cleanups (close_cleanup);
3413 if (buf_pos == 0)
3414 xfree (buf);
3415 else
3416 *buf_p = buf;
3417 return buf_pos;
3418 }
3419
3420 buf_pos += n;
3421
3422 /* If the buffer is filling up, expand it. */
3423 if (buf_alloc < buf_pos * 2)
3424 {
3425 buf_alloc *= 2;
3426 buf = xrealloc (buf, buf_alloc);
3427 }
3428
3429 QUIT;
3430 }
3431 }
3432
3433 /* Read target file FILENAME. Store the result in *BUF_P and return
3434 the size of the transferred data. See the declaration in "target.h"
3435 function for more information about the return value. */
3436
3437 LONGEST
3438 target_fileio_read_alloc (const char *filename, gdb_byte **buf_p)
3439 {
3440 return target_fileio_read_alloc_1 (filename, buf_p, 0);
3441 }
3442
3443 /* Read target file FILENAME. The result is NUL-terminated and
3444 returned as a string, allocated using xmalloc. If an error occurs
3445 or the transfer is unsupported, NULL is returned. Empty objects
3446 are returned as allocated but empty strings. A warning is issued
3447 if the result contains any embedded NUL bytes. */
3448
3449 char *
3450 target_fileio_read_stralloc (const char *filename)
3451 {
3452 gdb_byte *buffer;
3453 char *bufstr;
3454 LONGEST i, transferred;
3455
3456 transferred = target_fileio_read_alloc_1 (filename, &buffer, 1);
3457 bufstr = (char *) buffer;
3458
3459 if (transferred < 0)
3460 return NULL;
3461
3462 if (transferred == 0)
3463 return xstrdup ("");
3464
3465 bufstr[transferred] = 0;
3466
3467 /* Check for embedded NUL bytes; but allow trailing NULs. */
3468 for (i = strlen (bufstr); i < transferred; i++)
3469 if (bufstr[i] != 0)
3470 {
3471 warning (_("target file %s "
3472 "contained unexpected null characters"),
3473 filename);
3474 break;
3475 }
3476
3477 return bufstr;
3478 }
3479
3480
3481 static int
3482 default_region_ok_for_hw_watchpoint (struct target_ops *self,
3483 CORE_ADDR addr, int len)
3484 {
3485 return (len <= gdbarch_ptr_bit (target_gdbarch ()) / TARGET_CHAR_BIT);
3486 }
3487
3488 static int
3489 default_watchpoint_addr_within_range (struct target_ops *target,
3490 CORE_ADDR addr,
3491 CORE_ADDR start, int length)
3492 {
3493 return addr >= start && addr < start + length;
3494 }
3495
3496 static struct gdbarch *
3497 default_thread_architecture (struct target_ops *ops, ptid_t ptid)
3498 {
3499 return target_gdbarch ();
3500 }
3501
3502 static int
3503 return_zero (void)
3504 {
3505 return 0;
3506 }
3507
3508 static void *
3509 return_null (void)
3510 {
3511 return 0;
3512 }
3513
3514 /*
3515 * Find the next target down the stack from the specified target.
3516 */
3517
3518 struct target_ops *
3519 find_target_beneath (struct target_ops *t)
3520 {
3521 return t->beneath;
3522 }
3523
3524 /* See target.h. */
3525
3526 struct target_ops *
3527 find_target_at (enum strata stratum)
3528 {
3529 struct target_ops *t;
3530
3531 for (t = current_target.beneath; t != NULL; t = t->beneath)
3532 if (t->to_stratum == stratum)
3533 return t;
3534
3535 return NULL;
3536 }
3537
3538 \f
3539 /* The inferior process has died. Long live the inferior! */
3540
3541 void
3542 generic_mourn_inferior (void)
3543 {
3544 ptid_t ptid;
3545
3546 ptid = inferior_ptid;
3547 inferior_ptid = null_ptid;
3548
3549 /* Mark breakpoints uninserted in case something tries to delete a
3550 breakpoint while we delete the inferior's threads (which would
3551 fail, since the inferior is long gone). */
3552 mark_breakpoints_out ();
3553
3554 if (!ptid_equal (ptid, null_ptid))
3555 {
3556 int pid = ptid_get_pid (ptid);
3557 exit_inferior (pid);
3558 }
3559
3560 /* Note this wipes step-resume breakpoints, so needs to be done
3561 after exit_inferior, which ends up referencing the step-resume
3562 breakpoints through clear_thread_inferior_resources. */
3563 breakpoint_init_inferior (inf_exited);
3564
3565 registers_changed ();
3566
3567 reopen_exec_file ();
3568 reinit_frame_cache ();
3569
3570 if (deprecated_detach_hook)
3571 deprecated_detach_hook ();
3572 }
3573 \f
3574 /* Convert a normal process ID to a string. Returns the string in a
3575 static buffer. */
3576
3577 char *
3578 normal_pid_to_str (ptid_t ptid)
3579 {
3580 static char buf[32];
3581
3582 xsnprintf (buf, sizeof buf, "process %d", ptid_get_pid (ptid));
3583 return buf;
3584 }
3585
3586 static char *
3587 dummy_pid_to_str (struct target_ops *ops, ptid_t ptid)
3588 {
3589 return normal_pid_to_str (ptid);
3590 }
3591
3592 /* Error-catcher for target_find_memory_regions. */
3593 static int
3594 dummy_find_memory_regions (struct target_ops *self,
3595 find_memory_region_ftype ignore1, void *ignore2)
3596 {
3597 error (_("Command not implemented for this target."));
3598 return 0;
3599 }
3600
3601 /* Error-catcher for target_make_corefile_notes. */
3602 static char *
3603 dummy_make_corefile_notes (struct target_ops *self,
3604 bfd *ignore1, int *ignore2)
3605 {
3606 error (_("Command not implemented for this target."));
3607 return NULL;
3608 }
3609
3610 /* Set up the handful of non-empty slots needed by the dummy target
3611 vector. */
3612
3613 static void
3614 init_dummy_target (void)
3615 {
3616 dummy_target.to_shortname = "None";
3617 dummy_target.to_longname = "None";
3618 dummy_target.to_doc = "";
3619 dummy_target.to_create_inferior = find_default_create_inferior;
3620 dummy_target.to_supports_non_stop = find_default_supports_non_stop;
3621 dummy_target.to_supports_disable_randomization
3622 = find_default_supports_disable_randomization;
3623 dummy_target.to_pid_to_str = dummy_pid_to_str;
3624 dummy_target.to_stratum = dummy_stratum;
3625 dummy_target.to_has_all_memory = (int (*) (struct target_ops *)) return_zero;
3626 dummy_target.to_has_memory = (int (*) (struct target_ops *)) return_zero;
3627 dummy_target.to_has_stack = (int (*) (struct target_ops *)) return_zero;
3628 dummy_target.to_has_registers = (int (*) (struct target_ops *)) return_zero;
3629 dummy_target.to_has_execution
3630 = (int (*) (struct target_ops *, ptid_t)) return_zero;
3631 dummy_target.to_magic = OPS_MAGIC;
3632
3633 install_dummy_methods (&dummy_target);
3634 }
3635 \f
3636 static void
3637 debug_to_open (char *args, int from_tty)
3638 {
3639 debug_target.to_open (args, from_tty);
3640
3641 fprintf_unfiltered (gdb_stdlog, "target_open (%s, %d)\n", args, from_tty);
3642 }
3643
3644 void
3645 target_close (struct target_ops *targ)
3646 {
3647 gdb_assert (!target_is_pushed (targ));
3648
3649 if (targ->to_xclose != NULL)
3650 targ->to_xclose (targ);
3651 else if (targ->to_close != NULL)
3652 targ->to_close (targ);
3653
3654 if (targetdebug)
3655 fprintf_unfiltered (gdb_stdlog, "target_close ()\n");
3656 }
3657
3658 void
3659 target_attach (char *args, int from_tty)
3660 {
3661 current_target.to_attach (&current_target, args, from_tty);
3662 if (targetdebug)
3663 fprintf_unfiltered (gdb_stdlog, "target_attach (%s, %d)\n",
3664 args, from_tty);
3665 }
3666
3667 int
3668 target_thread_alive (ptid_t ptid)
3669 {
3670 struct target_ops *t;
3671
3672 for (t = current_target.beneath; t != NULL; t = t->beneath)
3673 {
3674 if (t->to_thread_alive != NULL)
3675 {
3676 int retval;
3677
3678 retval = t->to_thread_alive (t, ptid);
3679 if (targetdebug)
3680 fprintf_unfiltered (gdb_stdlog, "target_thread_alive (%d) = %d\n",
3681 ptid_get_pid (ptid), retval);
3682
3683 return retval;
3684 }
3685 }
3686
3687 return 0;
3688 }
3689
3690 void
3691 target_find_new_threads (void)
3692 {
3693 struct target_ops *t;
3694
3695 for (t = current_target.beneath; t != NULL; t = t->beneath)
3696 {
3697 if (t->to_find_new_threads != NULL)
3698 {
3699 t->to_find_new_threads (t);
3700 if (targetdebug)
3701 fprintf_unfiltered (gdb_stdlog, "target_find_new_threads ()\n");
3702
3703 return;
3704 }
3705 }
3706 }
3707
3708 void
3709 target_stop (ptid_t ptid)
3710 {
3711 if (!may_stop)
3712 {
3713 warning (_("May not interrupt or stop the target, ignoring attempt"));
3714 return;
3715 }
3716
3717 (*current_target.to_stop) (&current_target, ptid);
3718 }
3719
3720 static void
3721 debug_to_post_attach (struct target_ops *self, int pid)
3722 {
3723 debug_target.to_post_attach (&debug_target, pid);
3724
3725 fprintf_unfiltered (gdb_stdlog, "target_post_attach (%d)\n", pid);
3726 }
3727
3728 /* Concatenate ELEM to LIST, a comma separate list, and return the
3729 result. The LIST incoming argument is released. */
3730
3731 static char *
3732 str_comma_list_concat_elem (char *list, const char *elem)
3733 {
3734 if (list == NULL)
3735 return xstrdup (elem);
3736 else
3737 return reconcat (list, list, ", ", elem, (char *) NULL);
3738 }
3739
3740 /* Helper for target_options_to_string. If OPT is present in
3741 TARGET_OPTIONS, append the OPT_STR (string version of OPT) in RET.
3742 Returns the new resulting string. OPT is removed from
3743 TARGET_OPTIONS. */
3744
3745 static char *
3746 do_option (int *target_options, char *ret,
3747 int opt, char *opt_str)
3748 {
3749 if ((*target_options & opt) != 0)
3750 {
3751 ret = str_comma_list_concat_elem (ret, opt_str);
3752 *target_options &= ~opt;
3753 }
3754
3755 return ret;
3756 }
3757
3758 char *
3759 target_options_to_string (int target_options)
3760 {
3761 char *ret = NULL;
3762
3763 #define DO_TARG_OPTION(OPT) \
3764 ret = do_option (&target_options, ret, OPT, #OPT)
3765
3766 DO_TARG_OPTION (TARGET_WNOHANG);
3767
3768 if (target_options != 0)
3769 ret = str_comma_list_concat_elem (ret, "unknown???");
3770
3771 if (ret == NULL)
3772 ret = xstrdup ("");
3773 return ret;
3774 }
3775
3776 static void
3777 debug_print_register (const char * func,
3778 struct regcache *regcache, int regno)
3779 {
3780 struct gdbarch *gdbarch = get_regcache_arch (regcache);
3781
3782 fprintf_unfiltered (gdb_stdlog, "%s ", func);
3783 if (regno >= 0 && regno < gdbarch_num_regs (gdbarch)
3784 && gdbarch_register_name (gdbarch, regno) != NULL
3785 && gdbarch_register_name (gdbarch, regno)[0] != '\0')
3786 fprintf_unfiltered (gdb_stdlog, "(%s)",
3787 gdbarch_register_name (gdbarch, regno));
3788 else
3789 fprintf_unfiltered (gdb_stdlog, "(%d)", regno);
3790 if (regno >= 0 && regno < gdbarch_num_regs (gdbarch))
3791 {
3792 enum bfd_endian byte_order = gdbarch_byte_order (gdbarch);
3793 int i, size = register_size (gdbarch, regno);
3794 gdb_byte buf[MAX_REGISTER_SIZE];
3795
3796 regcache_raw_collect (regcache, regno, buf);
3797 fprintf_unfiltered (gdb_stdlog, " = ");
3798 for (i = 0; i < size; i++)
3799 {
3800 fprintf_unfiltered (gdb_stdlog, "%02x", buf[i]);
3801 }
3802 if (size <= sizeof (LONGEST))
3803 {
3804 ULONGEST val = extract_unsigned_integer (buf, size, byte_order);
3805
3806 fprintf_unfiltered (gdb_stdlog, " %s %s",
3807 core_addr_to_string_nz (val), plongest (val));
3808 }
3809 }
3810 fprintf_unfiltered (gdb_stdlog, "\n");
3811 }
3812
3813 void
3814 target_fetch_registers (struct regcache *regcache, int regno)
3815 {
3816 struct target_ops *t;
3817
3818 for (t = current_target.beneath; t != NULL; t = t->beneath)
3819 {
3820 if (t->to_fetch_registers != NULL)
3821 {
3822 t->to_fetch_registers (t, regcache, regno);
3823 if (targetdebug)
3824 debug_print_register ("target_fetch_registers", regcache, regno);
3825 return;
3826 }
3827 }
3828 }
3829
3830 void
3831 target_store_registers (struct regcache *regcache, int regno)
3832 {
3833 struct target_ops *t;
3834
3835 if (!may_write_registers)
3836 error (_("Writing to registers is not allowed (regno %d)"), regno);
3837
3838 current_target.to_store_registers (&current_target, regcache, regno);
3839 if (targetdebug)
3840 {
3841 debug_print_register ("target_store_registers", regcache, regno);
3842 }
3843 }
3844
3845 int
3846 target_core_of_thread (ptid_t ptid)
3847 {
3848 struct target_ops *t;
3849
3850 for (t = current_target.beneath; t != NULL; t = t->beneath)
3851 {
3852 if (t->to_core_of_thread != NULL)
3853 {
3854 int retval = t->to_core_of_thread (t, ptid);
3855
3856 if (targetdebug)
3857 fprintf_unfiltered (gdb_stdlog,
3858 "target_core_of_thread (%d) = %d\n",
3859 ptid_get_pid (ptid), retval);
3860 return retval;
3861 }
3862 }
3863
3864 return -1;
3865 }
3866
3867 int
3868 target_verify_memory (const gdb_byte *data, CORE_ADDR memaddr, ULONGEST size)
3869 {
3870 struct target_ops *t;
3871
3872 for (t = current_target.beneath; t != NULL; t = t->beneath)
3873 {
3874 if (t->to_verify_memory != NULL)
3875 {
3876 int retval = t->to_verify_memory (t, data, memaddr, size);
3877
3878 if (targetdebug)
3879 fprintf_unfiltered (gdb_stdlog,
3880 "target_verify_memory (%s, %s) = %d\n",
3881 paddress (target_gdbarch (), memaddr),
3882 pulongest (size),
3883 retval);
3884 return retval;
3885 }
3886 }
3887
3888 tcomplain ();
3889 }
3890
3891 /* The documentation for this function is in its prototype declaration in
3892 target.h. */
3893
3894 int
3895 target_insert_mask_watchpoint (CORE_ADDR addr, CORE_ADDR mask, int rw)
3896 {
3897 struct target_ops *t;
3898
3899 for (t = current_target.beneath; t != NULL; t = t->beneath)
3900 if (t->to_insert_mask_watchpoint != NULL)
3901 {
3902 int ret;
3903
3904 ret = t->to_insert_mask_watchpoint (t, addr, mask, rw);
3905
3906 if (targetdebug)
3907 fprintf_unfiltered (gdb_stdlog, "\
3908 target_insert_mask_watchpoint (%s, %s, %d) = %d\n",
3909 core_addr_to_string (addr),
3910 core_addr_to_string (mask), rw, ret);
3911
3912 return ret;
3913 }
3914
3915 return 1;
3916 }
3917
3918 /* The documentation for this function is in its prototype declaration in
3919 target.h. */
3920
3921 int
3922 target_remove_mask_watchpoint (CORE_ADDR addr, CORE_ADDR mask, int rw)
3923 {
3924 struct target_ops *t;
3925
3926 for (t = current_target.beneath; t != NULL; t = t->beneath)
3927 if (t->to_remove_mask_watchpoint != NULL)
3928 {
3929 int ret;
3930
3931 ret = t->to_remove_mask_watchpoint (t, addr, mask, rw);
3932
3933 if (targetdebug)
3934 fprintf_unfiltered (gdb_stdlog, "\
3935 target_remove_mask_watchpoint (%s, %s, %d) = %d\n",
3936 core_addr_to_string (addr),
3937 core_addr_to_string (mask), rw, ret);
3938
3939 return ret;
3940 }
3941
3942 return 1;
3943 }
3944
3945 /* The documentation for this function is in its prototype declaration
3946 in target.h. */
3947
3948 int
3949 target_masked_watch_num_registers (CORE_ADDR addr, CORE_ADDR mask)
3950 {
3951 struct target_ops *t;
3952
3953 for (t = current_target.beneath; t != NULL; t = t->beneath)
3954 if (t->to_masked_watch_num_registers != NULL)
3955 return t->to_masked_watch_num_registers (t, addr, mask);
3956
3957 return -1;
3958 }
3959
3960 /* The documentation for this function is in its prototype declaration
3961 in target.h. */
3962
3963 int
3964 target_ranged_break_num_registers (void)
3965 {
3966 struct target_ops *t;
3967
3968 for (t = current_target.beneath; t != NULL; t = t->beneath)
3969 if (t->to_ranged_break_num_registers != NULL)
3970 return t->to_ranged_break_num_registers (t);
3971
3972 return -1;
3973 }
3974
3975 /* See target.h. */
3976
3977 struct btrace_target_info *
3978 target_enable_btrace (ptid_t ptid)
3979 {
3980 struct target_ops *t;
3981
3982 for (t = current_target.beneath; t != NULL; t = t->beneath)
3983 if (t->to_enable_btrace != NULL)
3984 return t->to_enable_btrace (t, ptid);
3985
3986 tcomplain ();
3987 return NULL;
3988 }
3989
3990 /* See target.h. */
3991
3992 void
3993 target_disable_btrace (struct btrace_target_info *btinfo)
3994 {
3995 struct target_ops *t;
3996
3997 for (t = current_target.beneath; t != NULL; t = t->beneath)
3998 if (t->to_disable_btrace != NULL)
3999 {
4000 t->to_disable_btrace (t, btinfo);
4001 return;
4002 }
4003
4004 tcomplain ();
4005 }
4006
4007 /* See target.h. */
4008
4009 void
4010 target_teardown_btrace (struct btrace_target_info *btinfo)
4011 {
4012 struct target_ops *t;
4013
4014 for (t = current_target.beneath; t != NULL; t = t->beneath)
4015 if (t->to_teardown_btrace != NULL)
4016 {
4017 t->to_teardown_btrace (t, btinfo);
4018 return;
4019 }
4020
4021 tcomplain ();
4022 }
4023
4024 /* See target.h. */
4025
4026 enum btrace_error
4027 target_read_btrace (VEC (btrace_block_s) **btrace,
4028 struct btrace_target_info *btinfo,
4029 enum btrace_read_type type)
4030 {
4031 struct target_ops *t;
4032
4033 for (t = current_target.beneath; t != NULL; t = t->beneath)
4034 if (t->to_read_btrace != NULL)
4035 return t->to_read_btrace (t, btrace, btinfo, type);
4036
4037 tcomplain ();
4038 return BTRACE_ERR_NOT_SUPPORTED;
4039 }
4040
4041 /* See target.h. */
4042
4043 void
4044 target_stop_recording (void)
4045 {
4046 struct target_ops *t;
4047
4048 for (t = current_target.beneath; t != NULL; t = t->beneath)
4049 if (t->to_stop_recording != NULL)
4050 {
4051 t->to_stop_recording (t);
4052 return;
4053 }
4054
4055 /* This is optional. */
4056 }
4057
4058 /* See target.h. */
4059
4060 void
4061 target_info_record (void)
4062 {
4063 struct target_ops *t;
4064
4065 for (t = current_target.beneath; t != NULL; t = t->beneath)
4066 if (t->to_info_record != NULL)
4067 {
4068 t->to_info_record (t);
4069 return;
4070 }
4071
4072 tcomplain ();
4073 }
4074
4075 /* See target.h. */
4076
4077 void
4078 target_save_record (const char *filename)
4079 {
4080 struct target_ops *t;
4081
4082 for (t = current_target.beneath; t != NULL; t = t->beneath)
4083 if (t->to_save_record != NULL)
4084 {
4085 t->to_save_record (t, filename);
4086 return;
4087 }
4088
4089 tcomplain ();
4090 }
4091
4092 /* See target.h. */
4093
4094 int
4095 target_supports_delete_record (void)
4096 {
4097 struct target_ops *t;
4098
4099 for (t = current_target.beneath; t != NULL; t = t->beneath)
4100 if (t->to_delete_record != NULL)
4101 return 1;
4102
4103 return 0;
4104 }
4105
4106 /* See target.h. */
4107
4108 void
4109 target_delete_record (void)
4110 {
4111 struct target_ops *t;
4112
4113 for (t = current_target.beneath; t != NULL; t = t->beneath)
4114 if (t->to_delete_record != NULL)
4115 {
4116 t->to_delete_record (t);
4117 return;
4118 }
4119
4120 tcomplain ();
4121 }
4122
4123 /* See target.h. */
4124
4125 int
4126 target_record_is_replaying (void)
4127 {
4128 struct target_ops *t;
4129
4130 for (t = current_target.beneath; t != NULL; t = t->beneath)
4131 if (t->to_record_is_replaying != NULL)
4132 return t->to_record_is_replaying (t);
4133
4134 return 0;
4135 }
4136
4137 /* See target.h. */
4138
4139 void
4140 target_goto_record_begin (void)
4141 {
4142 struct target_ops *t;
4143
4144 for (t = current_target.beneath; t != NULL; t = t->beneath)
4145 if (t->to_goto_record_begin != NULL)
4146 {
4147 t->to_goto_record_begin (t);
4148 return;
4149 }
4150
4151 tcomplain ();
4152 }
4153
4154 /* See target.h. */
4155
4156 void
4157 target_goto_record_end (void)
4158 {
4159 struct target_ops *t;
4160
4161 for (t = current_target.beneath; t != NULL; t = t->beneath)
4162 if (t->to_goto_record_end != NULL)
4163 {
4164 t->to_goto_record_end (t);
4165 return;
4166 }
4167
4168 tcomplain ();
4169 }
4170
4171 /* See target.h. */
4172
4173 void
4174 target_goto_record (ULONGEST insn)
4175 {
4176 struct target_ops *t;
4177
4178 for (t = current_target.beneath; t != NULL; t = t->beneath)
4179 if (t->to_goto_record != NULL)
4180 {
4181 t->to_goto_record (t, insn);
4182 return;
4183 }
4184
4185 tcomplain ();
4186 }
4187
4188 /* See target.h. */
4189
4190 void
4191 target_insn_history (int size, int flags)
4192 {
4193 struct target_ops *t;
4194
4195 for (t = current_target.beneath; t != NULL; t = t->beneath)
4196 if (t->to_insn_history != NULL)
4197 {
4198 t->to_insn_history (t, size, flags);
4199 return;
4200 }
4201
4202 tcomplain ();
4203 }
4204
4205 /* See target.h. */
4206
4207 void
4208 target_insn_history_from (ULONGEST from, int size, int flags)
4209 {
4210 struct target_ops *t;
4211
4212 for (t = current_target.beneath; t != NULL; t = t->beneath)
4213 if (t->to_insn_history_from != NULL)
4214 {
4215 t->to_insn_history_from (t, from, size, flags);
4216 return;
4217 }
4218
4219 tcomplain ();
4220 }
4221
4222 /* See target.h. */
4223
4224 void
4225 target_insn_history_range (ULONGEST begin, ULONGEST end, int flags)
4226 {
4227 struct target_ops *t;
4228
4229 for (t = current_target.beneath; t != NULL; t = t->beneath)
4230 if (t->to_insn_history_range != NULL)
4231 {
4232 t->to_insn_history_range (t, begin, end, flags);
4233 return;
4234 }
4235
4236 tcomplain ();
4237 }
4238
4239 /* See target.h. */
4240
4241 void
4242 target_call_history (int size, int flags)
4243 {
4244 struct target_ops *t;
4245
4246 for (t = current_target.beneath; t != NULL; t = t->beneath)
4247 if (t->to_call_history != NULL)
4248 {
4249 t->to_call_history (t, size, flags);
4250 return;
4251 }
4252
4253 tcomplain ();
4254 }
4255
4256 /* See target.h. */
4257
4258 void
4259 target_call_history_from (ULONGEST begin, int size, int flags)
4260 {
4261 struct target_ops *t;
4262
4263 for (t = current_target.beneath; t != NULL; t = t->beneath)
4264 if (t->to_call_history_from != NULL)
4265 {
4266 t->to_call_history_from (t, begin, size, flags);
4267 return;
4268 }
4269
4270 tcomplain ();
4271 }
4272
4273 /* See target.h. */
4274
4275 void
4276 target_call_history_range (ULONGEST begin, ULONGEST end, int flags)
4277 {
4278 struct target_ops *t;
4279
4280 for (t = current_target.beneath; t != NULL; t = t->beneath)
4281 if (t->to_call_history_range != NULL)
4282 {
4283 t->to_call_history_range (t, begin, end, flags);
4284 return;
4285 }
4286
4287 tcomplain ();
4288 }
4289
4290 static void
4291 debug_to_prepare_to_store (struct target_ops *self, struct regcache *regcache)
4292 {
4293 debug_target.to_prepare_to_store (&debug_target, regcache);
4294
4295 fprintf_unfiltered (gdb_stdlog, "target_prepare_to_store ()\n");
4296 }
4297
4298 /* See target.h. */
4299
4300 const struct frame_unwind *
4301 target_get_unwinder (void)
4302 {
4303 struct target_ops *t;
4304
4305 for (t = current_target.beneath; t != NULL; t = t->beneath)
4306 if (t->to_get_unwinder != NULL)
4307 return t->to_get_unwinder;
4308
4309 return NULL;
4310 }
4311
4312 /* See target.h. */
4313
4314 const struct frame_unwind *
4315 target_get_tailcall_unwinder (void)
4316 {
4317 struct target_ops *t;
4318
4319 for (t = current_target.beneath; t != NULL; t = t->beneath)
4320 if (t->to_get_tailcall_unwinder != NULL)
4321 return t->to_get_tailcall_unwinder;
4322
4323 return NULL;
4324 }
4325
4326 /* See target.h. */
4327
4328 CORE_ADDR
4329 forward_target_decr_pc_after_break (struct target_ops *ops,
4330 struct gdbarch *gdbarch)
4331 {
4332 for (; ops != NULL; ops = ops->beneath)
4333 if (ops->to_decr_pc_after_break != NULL)
4334 return ops->to_decr_pc_after_break (ops, gdbarch);
4335
4336 return gdbarch_decr_pc_after_break (gdbarch);
4337 }
4338
4339 /* See target.h. */
4340
4341 CORE_ADDR
4342 target_decr_pc_after_break (struct gdbarch *gdbarch)
4343 {
4344 return forward_target_decr_pc_after_break (current_target.beneath, gdbarch);
4345 }
4346
4347 static int
4348 deprecated_debug_xfer_memory (CORE_ADDR memaddr, bfd_byte *myaddr, int len,
4349 int write, struct mem_attrib *attrib,
4350 struct target_ops *target)
4351 {
4352 int retval;
4353
4354 retval = debug_target.deprecated_xfer_memory (memaddr, myaddr, len, write,
4355 attrib, target);
4356
4357 fprintf_unfiltered (gdb_stdlog,
4358 "target_xfer_memory (%s, xxx, %d, %s, xxx) = %d",
4359 paddress (target_gdbarch (), memaddr), len,
4360 write ? "write" : "read", retval);
4361
4362 if (retval > 0)
4363 {
4364 int i;
4365
4366 fputs_unfiltered (", bytes =", gdb_stdlog);
4367 for (i = 0; i < retval; i++)
4368 {
4369 if ((((intptr_t) &(myaddr[i])) & 0xf) == 0)
4370 {
4371 if (targetdebug < 2 && i > 0)
4372 {
4373 fprintf_unfiltered (gdb_stdlog, " ...");
4374 break;
4375 }
4376 fprintf_unfiltered (gdb_stdlog, "\n");
4377 }
4378
4379 fprintf_unfiltered (gdb_stdlog, " %02x", myaddr[i] & 0xff);
4380 }
4381 }
4382
4383 fputc_unfiltered ('\n', gdb_stdlog);
4384
4385 return retval;
4386 }
4387
4388 static void
4389 debug_to_files_info (struct target_ops *target)
4390 {
4391 debug_target.to_files_info (target);
4392
4393 fprintf_unfiltered (gdb_stdlog, "target_files_info (xxx)\n");
4394 }
4395
4396 static int
4397 debug_to_insert_breakpoint (struct target_ops *ops, struct gdbarch *gdbarch,
4398 struct bp_target_info *bp_tgt)
4399 {
4400 int retval;
4401
4402 retval = debug_target.to_insert_breakpoint (&debug_target, gdbarch, bp_tgt);
4403
4404 fprintf_unfiltered (gdb_stdlog,
4405 "target_insert_breakpoint (%s, xxx) = %ld\n",
4406 core_addr_to_string (bp_tgt->placed_address),
4407 (unsigned long) retval);
4408 return retval;
4409 }
4410
4411 static int
4412 debug_to_remove_breakpoint (struct target_ops *ops, struct gdbarch *gdbarch,
4413 struct bp_target_info *bp_tgt)
4414 {
4415 int retval;
4416
4417 retval = debug_target.to_remove_breakpoint (&debug_target, gdbarch, bp_tgt);
4418
4419 fprintf_unfiltered (gdb_stdlog,
4420 "target_remove_breakpoint (%s, xxx) = %ld\n",
4421 core_addr_to_string (bp_tgt->placed_address),
4422 (unsigned long) retval);
4423 return retval;
4424 }
4425
4426 static int
4427 debug_to_can_use_hw_breakpoint (struct target_ops *self,
4428 int type, int cnt, int from_tty)
4429 {
4430 int retval;
4431
4432 retval = debug_target.to_can_use_hw_breakpoint (&debug_target,
4433 type, cnt, from_tty);
4434
4435 fprintf_unfiltered (gdb_stdlog,
4436 "target_can_use_hw_breakpoint (%ld, %ld, %ld) = %ld\n",
4437 (unsigned long) type,
4438 (unsigned long) cnt,
4439 (unsigned long) from_tty,
4440 (unsigned long) retval);
4441 return retval;
4442 }
4443
4444 static int
4445 debug_to_region_ok_for_hw_watchpoint (struct target_ops *self,
4446 CORE_ADDR addr, int len)
4447 {
4448 CORE_ADDR retval;
4449
4450 retval = debug_target.to_region_ok_for_hw_watchpoint (&debug_target,
4451 addr, len);
4452
4453 fprintf_unfiltered (gdb_stdlog,
4454 "target_region_ok_for_hw_watchpoint (%s, %ld) = %s\n",
4455 core_addr_to_string (addr), (unsigned long) len,
4456 core_addr_to_string (retval));
4457 return retval;
4458 }
4459
4460 static int
4461 debug_to_can_accel_watchpoint_condition (struct target_ops *self,
4462 CORE_ADDR addr, int len, int rw,
4463 struct expression *cond)
4464 {
4465 int retval;
4466
4467 retval = debug_target.to_can_accel_watchpoint_condition (&debug_target,
4468 addr, len,
4469 rw, cond);
4470
4471 fprintf_unfiltered (gdb_stdlog,
4472 "target_can_accel_watchpoint_condition "
4473 "(%s, %d, %d, %s) = %ld\n",
4474 core_addr_to_string (addr), len, rw,
4475 host_address_to_string (cond), (unsigned long) retval);
4476 return retval;
4477 }
4478
4479 static int
4480 debug_to_stopped_by_watchpoint (struct target_ops *ops)
4481 {
4482 int retval;
4483
4484 retval = debug_target.to_stopped_by_watchpoint (&debug_target);
4485
4486 fprintf_unfiltered (gdb_stdlog,
4487 "target_stopped_by_watchpoint () = %ld\n",
4488 (unsigned long) retval);
4489 return retval;
4490 }
4491
4492 static int
4493 debug_to_stopped_data_address (struct target_ops *target, CORE_ADDR *addr)
4494 {
4495 int retval;
4496
4497 retval = debug_target.to_stopped_data_address (target, addr);
4498
4499 fprintf_unfiltered (gdb_stdlog,
4500 "target_stopped_data_address ([%s]) = %ld\n",
4501 core_addr_to_string (*addr),
4502 (unsigned long)retval);
4503 return retval;
4504 }
4505
4506 static int
4507 debug_to_watchpoint_addr_within_range (struct target_ops *target,
4508 CORE_ADDR addr,
4509 CORE_ADDR start, int length)
4510 {
4511 int retval;
4512
4513 retval = debug_target.to_watchpoint_addr_within_range (target, addr,
4514 start, length);
4515
4516 fprintf_filtered (gdb_stdlog,
4517 "target_watchpoint_addr_within_range (%s, %s, %d) = %d\n",
4518 core_addr_to_string (addr), core_addr_to_string (start),
4519 length, retval);
4520 return retval;
4521 }
4522
4523 static int
4524 debug_to_insert_hw_breakpoint (struct target_ops *self,
4525 struct gdbarch *gdbarch,
4526 struct bp_target_info *bp_tgt)
4527 {
4528 int retval;
4529
4530 retval = debug_target.to_insert_hw_breakpoint (&debug_target,
4531 gdbarch, bp_tgt);
4532
4533 fprintf_unfiltered (gdb_stdlog,
4534 "target_insert_hw_breakpoint (%s, xxx) = %ld\n",
4535 core_addr_to_string (bp_tgt->placed_address),
4536 (unsigned long) retval);
4537 return retval;
4538 }
4539
4540 static int
4541 debug_to_remove_hw_breakpoint (struct target_ops *self,
4542 struct gdbarch *gdbarch,
4543 struct bp_target_info *bp_tgt)
4544 {
4545 int retval;
4546
4547 retval = debug_target.to_remove_hw_breakpoint (&debug_target,
4548 gdbarch, bp_tgt);
4549
4550 fprintf_unfiltered (gdb_stdlog,
4551 "target_remove_hw_breakpoint (%s, xxx) = %ld\n",
4552 core_addr_to_string (bp_tgt->placed_address),
4553 (unsigned long) retval);
4554 return retval;
4555 }
4556
4557 static int
4558 debug_to_insert_watchpoint (struct target_ops *self,
4559 CORE_ADDR addr, int len, int type,
4560 struct expression *cond)
4561 {
4562 int retval;
4563
4564 retval = debug_target.to_insert_watchpoint (&debug_target,
4565 addr, len, type, cond);
4566
4567 fprintf_unfiltered (gdb_stdlog,
4568 "target_insert_watchpoint (%s, %d, %d, %s) = %ld\n",
4569 core_addr_to_string (addr), len, type,
4570 host_address_to_string (cond), (unsigned long) retval);
4571 return retval;
4572 }
4573
4574 static int
4575 debug_to_remove_watchpoint (struct target_ops *self,
4576 CORE_ADDR addr, int len, int type,
4577 struct expression *cond)
4578 {
4579 int retval;
4580
4581 retval = debug_target.to_remove_watchpoint (&debug_target,
4582 addr, len, type, cond);
4583
4584 fprintf_unfiltered (gdb_stdlog,
4585 "target_remove_watchpoint (%s, %d, %d, %s) = %ld\n",
4586 core_addr_to_string (addr), len, type,
4587 host_address_to_string (cond), (unsigned long) retval);
4588 return retval;
4589 }
4590
4591 static void
4592 debug_to_terminal_init (struct target_ops *self)
4593 {
4594 debug_target.to_terminal_init (&debug_target);
4595
4596 fprintf_unfiltered (gdb_stdlog, "target_terminal_init ()\n");
4597 }
4598
4599 static void
4600 debug_to_terminal_inferior (struct target_ops *self)
4601 {
4602 debug_target.to_terminal_inferior (&debug_target);
4603
4604 fprintf_unfiltered (gdb_stdlog, "target_terminal_inferior ()\n");
4605 }
4606
4607 static void
4608 debug_to_terminal_ours_for_output (struct target_ops *self)
4609 {
4610 debug_target.to_terminal_ours_for_output (&debug_target);
4611
4612 fprintf_unfiltered (gdb_stdlog, "target_terminal_ours_for_output ()\n");
4613 }
4614
4615 static void
4616 debug_to_terminal_ours (struct target_ops *self)
4617 {
4618 debug_target.to_terminal_ours (&debug_target);
4619
4620 fprintf_unfiltered (gdb_stdlog, "target_terminal_ours ()\n");
4621 }
4622
4623 static void
4624 debug_to_terminal_save_ours (struct target_ops *self)
4625 {
4626 debug_target.to_terminal_save_ours (&debug_target);
4627
4628 fprintf_unfiltered (gdb_stdlog, "target_terminal_save_ours ()\n");
4629 }
4630
4631 static void
4632 debug_to_terminal_info (struct target_ops *self,
4633 const char *arg, int from_tty)
4634 {
4635 debug_target.to_terminal_info (&debug_target, arg, from_tty);
4636
4637 fprintf_unfiltered (gdb_stdlog, "target_terminal_info (%s, %d)\n", arg,
4638 from_tty);
4639 }
4640
4641 static void
4642 debug_to_load (struct target_ops *self, char *args, int from_tty)
4643 {
4644 debug_target.to_load (&debug_target, args, from_tty);
4645
4646 fprintf_unfiltered (gdb_stdlog, "target_load (%s, %d)\n", args, from_tty);
4647 }
4648
4649 static void
4650 debug_to_post_startup_inferior (struct target_ops *self, ptid_t ptid)
4651 {
4652 debug_target.to_post_startup_inferior (&debug_target, ptid);
4653
4654 fprintf_unfiltered (gdb_stdlog, "target_post_startup_inferior (%d)\n",
4655 ptid_get_pid (ptid));
4656 }
4657
4658 static int
4659 debug_to_insert_fork_catchpoint (struct target_ops *self, int pid)
4660 {
4661 int retval;
4662
4663 retval = debug_target.to_insert_fork_catchpoint (&debug_target, pid);
4664
4665 fprintf_unfiltered (gdb_stdlog, "target_insert_fork_catchpoint (%d) = %d\n",
4666 pid, retval);
4667
4668 return retval;
4669 }
4670
4671 static int
4672 debug_to_remove_fork_catchpoint (struct target_ops *self, int pid)
4673 {
4674 int retval;
4675
4676 retval = debug_target.to_remove_fork_catchpoint (&debug_target, pid);
4677
4678 fprintf_unfiltered (gdb_stdlog, "target_remove_fork_catchpoint (%d) = %d\n",
4679 pid, retval);
4680
4681 return retval;
4682 }
4683
4684 static int
4685 debug_to_insert_vfork_catchpoint (struct target_ops *self, int pid)
4686 {
4687 int retval;
4688
4689 retval = debug_target.to_insert_vfork_catchpoint (&debug_target, pid);
4690
4691 fprintf_unfiltered (gdb_stdlog, "target_insert_vfork_catchpoint (%d) = %d\n",
4692 pid, retval);
4693
4694 return retval;
4695 }
4696
4697 static int
4698 debug_to_remove_vfork_catchpoint (struct target_ops *self, int pid)
4699 {
4700 int retval;
4701
4702 retval = debug_target.to_remove_vfork_catchpoint (&debug_target, pid);
4703
4704 fprintf_unfiltered (gdb_stdlog, "target_remove_vfork_catchpoint (%d) = %d\n",
4705 pid, retval);
4706
4707 return retval;
4708 }
4709
4710 static int
4711 debug_to_insert_exec_catchpoint (struct target_ops *self, int pid)
4712 {
4713 int retval;
4714
4715 retval = debug_target.to_insert_exec_catchpoint (&debug_target, pid);
4716
4717 fprintf_unfiltered (gdb_stdlog, "target_insert_exec_catchpoint (%d) = %d\n",
4718 pid, retval);
4719
4720 return retval;
4721 }
4722
4723 static int
4724 debug_to_remove_exec_catchpoint (struct target_ops *self, int pid)
4725 {
4726 int retval;
4727
4728 retval = debug_target.to_remove_exec_catchpoint (&debug_target, pid);
4729
4730 fprintf_unfiltered (gdb_stdlog, "target_remove_exec_catchpoint (%d) = %d\n",
4731 pid, retval);
4732
4733 return retval;
4734 }
4735
4736 static int
4737 debug_to_has_exited (struct target_ops *self,
4738 int pid, int wait_status, int *exit_status)
4739 {
4740 int has_exited;
4741
4742 has_exited = debug_target.to_has_exited (&debug_target,
4743 pid, wait_status, exit_status);
4744
4745 fprintf_unfiltered (gdb_stdlog, "target_has_exited (%d, %d, %d) = %d\n",
4746 pid, wait_status, *exit_status, has_exited);
4747
4748 return has_exited;
4749 }
4750
4751 static int
4752 debug_to_can_run (struct target_ops *self)
4753 {
4754 int retval;
4755
4756 retval = debug_target.to_can_run (&debug_target);
4757
4758 fprintf_unfiltered (gdb_stdlog, "target_can_run () = %d\n", retval);
4759
4760 return retval;
4761 }
4762
4763 static struct gdbarch *
4764 debug_to_thread_architecture (struct target_ops *ops, ptid_t ptid)
4765 {
4766 struct gdbarch *retval;
4767
4768 retval = debug_target.to_thread_architecture (ops, ptid);
4769
4770 fprintf_unfiltered (gdb_stdlog,
4771 "target_thread_architecture (%s) = %s [%s]\n",
4772 target_pid_to_str (ptid),
4773 host_address_to_string (retval),
4774 gdbarch_bfd_arch_info (retval)->printable_name);
4775 return retval;
4776 }
4777
4778 static void
4779 debug_to_stop (struct target_ops *self, ptid_t ptid)
4780 {
4781 debug_target.to_stop (&debug_target, ptid);
4782
4783 fprintf_unfiltered (gdb_stdlog, "target_stop (%s)\n",
4784 target_pid_to_str (ptid));
4785 }
4786
4787 static void
4788 debug_to_rcmd (struct target_ops *self, char *command,
4789 struct ui_file *outbuf)
4790 {
4791 debug_target.to_rcmd (&debug_target, command, outbuf);
4792 fprintf_unfiltered (gdb_stdlog, "target_rcmd (%s, ...)\n", command);
4793 }
4794
4795 static char *
4796 debug_to_pid_to_exec_file (struct target_ops *self, int pid)
4797 {
4798 char *exec_file;
4799
4800 exec_file = debug_target.to_pid_to_exec_file (&debug_target, pid);
4801
4802 fprintf_unfiltered (gdb_stdlog, "target_pid_to_exec_file (%d) = %s\n",
4803 pid, exec_file);
4804
4805 return exec_file;
4806 }
4807
4808 static void
4809 setup_target_debug (void)
4810 {
4811 memcpy (&debug_target, &current_target, sizeof debug_target);
4812
4813 current_target.to_open = debug_to_open;
4814 current_target.to_post_attach = debug_to_post_attach;
4815 current_target.to_prepare_to_store = debug_to_prepare_to_store;
4816 current_target.deprecated_xfer_memory = deprecated_debug_xfer_memory;
4817 current_target.to_files_info = debug_to_files_info;
4818 current_target.to_insert_breakpoint = debug_to_insert_breakpoint;
4819 current_target.to_remove_breakpoint = debug_to_remove_breakpoint;
4820 current_target.to_can_use_hw_breakpoint = debug_to_can_use_hw_breakpoint;
4821 current_target.to_insert_hw_breakpoint = debug_to_insert_hw_breakpoint;
4822 current_target.to_remove_hw_breakpoint = debug_to_remove_hw_breakpoint;
4823 current_target.to_insert_watchpoint = debug_to_insert_watchpoint;
4824 current_target.to_remove_watchpoint = debug_to_remove_watchpoint;
4825 current_target.to_stopped_by_watchpoint = debug_to_stopped_by_watchpoint;
4826 current_target.to_stopped_data_address = debug_to_stopped_data_address;
4827 current_target.to_watchpoint_addr_within_range
4828 = debug_to_watchpoint_addr_within_range;
4829 current_target.to_region_ok_for_hw_watchpoint
4830 = debug_to_region_ok_for_hw_watchpoint;
4831 current_target.to_can_accel_watchpoint_condition
4832 = debug_to_can_accel_watchpoint_condition;
4833 current_target.to_terminal_init = debug_to_terminal_init;
4834 current_target.to_terminal_inferior = debug_to_terminal_inferior;
4835 current_target.to_terminal_ours_for_output
4836 = debug_to_terminal_ours_for_output;
4837 current_target.to_terminal_ours = debug_to_terminal_ours;
4838 current_target.to_terminal_save_ours = debug_to_terminal_save_ours;
4839 current_target.to_terminal_info = debug_to_terminal_info;
4840 current_target.to_load = debug_to_load;
4841 current_target.to_post_startup_inferior = debug_to_post_startup_inferior;
4842 current_target.to_insert_fork_catchpoint = debug_to_insert_fork_catchpoint;
4843 current_target.to_remove_fork_catchpoint = debug_to_remove_fork_catchpoint;
4844 current_target.to_insert_vfork_catchpoint = debug_to_insert_vfork_catchpoint;
4845 current_target.to_remove_vfork_catchpoint = debug_to_remove_vfork_catchpoint;
4846 current_target.to_insert_exec_catchpoint = debug_to_insert_exec_catchpoint;
4847 current_target.to_remove_exec_catchpoint = debug_to_remove_exec_catchpoint;
4848 current_target.to_has_exited = debug_to_has_exited;
4849 current_target.to_can_run = debug_to_can_run;
4850 current_target.to_stop = debug_to_stop;
4851 current_target.to_rcmd = debug_to_rcmd;
4852 current_target.to_pid_to_exec_file = debug_to_pid_to_exec_file;
4853 current_target.to_thread_architecture = debug_to_thread_architecture;
4854 }
4855 \f
4856
4857 static char targ_desc[] =
4858 "Names of targets and files being debugged.\nShows the entire \
4859 stack of targets currently in use (including the exec-file,\n\
4860 core-file, and process, if any), as well as the symbol file name.";
4861
4862 static void
4863 default_rcmd (struct target_ops *self, char *command, struct ui_file *output)
4864 {
4865 error (_("\"monitor\" command not supported by this target."));
4866 }
4867
4868 static void
4869 do_monitor_command (char *cmd,
4870 int from_tty)
4871 {
4872 target_rcmd (cmd, gdb_stdtarg);
4873 }
4874
4875 /* Print the name of each layers of our target stack. */
4876
4877 static void
4878 maintenance_print_target_stack (char *cmd, int from_tty)
4879 {
4880 struct target_ops *t;
4881
4882 printf_filtered (_("The current target stack is:\n"));
4883
4884 for (t = target_stack; t != NULL; t = t->beneath)
4885 {
4886 printf_filtered (" - %s (%s)\n", t->to_shortname, t->to_longname);
4887 }
4888 }
4889
4890 /* Controls if async mode is permitted. */
4891 int target_async_permitted = 0;
4892
4893 /* The set command writes to this variable. If the inferior is
4894 executing, target_async_permitted is *not* updated. */
4895 static int target_async_permitted_1 = 0;
4896
4897 static void
4898 set_target_async_command (char *args, int from_tty,
4899 struct cmd_list_element *c)
4900 {
4901 if (have_live_inferiors ())
4902 {
4903 target_async_permitted_1 = target_async_permitted;
4904 error (_("Cannot change this setting while the inferior is running."));
4905 }
4906
4907 target_async_permitted = target_async_permitted_1;
4908 }
4909
4910 static void
4911 show_target_async_command (struct ui_file *file, int from_tty,
4912 struct cmd_list_element *c,
4913 const char *value)
4914 {
4915 fprintf_filtered (file,
4916 _("Controlling the inferior in "
4917 "asynchronous mode is %s.\n"), value);
4918 }
4919
4920 /* Temporary copies of permission settings. */
4921
4922 static int may_write_registers_1 = 1;
4923 static int may_write_memory_1 = 1;
4924 static int may_insert_breakpoints_1 = 1;
4925 static int may_insert_tracepoints_1 = 1;
4926 static int may_insert_fast_tracepoints_1 = 1;
4927 static int may_stop_1 = 1;
4928
4929 /* Make the user-set values match the real values again. */
4930
4931 void
4932 update_target_permissions (void)
4933 {
4934 may_write_registers_1 = may_write_registers;
4935 may_write_memory_1 = may_write_memory;
4936 may_insert_breakpoints_1 = may_insert_breakpoints;
4937 may_insert_tracepoints_1 = may_insert_tracepoints;
4938 may_insert_fast_tracepoints_1 = may_insert_fast_tracepoints;
4939 may_stop_1 = may_stop;
4940 }
4941
4942 /* The one function handles (most of) the permission flags in the same
4943 way. */
4944
4945 static void
4946 set_target_permissions (char *args, int from_tty,
4947 struct cmd_list_element *c)
4948 {
4949 if (target_has_execution)
4950 {
4951 update_target_permissions ();
4952 error (_("Cannot change this setting while the inferior is running."));
4953 }
4954
4955 /* Make the real values match the user-changed values. */
4956 may_write_registers = may_write_registers_1;
4957 may_insert_breakpoints = may_insert_breakpoints_1;
4958 may_insert_tracepoints = may_insert_tracepoints_1;
4959 may_insert_fast_tracepoints = may_insert_fast_tracepoints_1;
4960 may_stop = may_stop_1;
4961 update_observer_mode ();
4962 }
4963
4964 /* Set memory write permission independently of observer mode. */
4965
4966 static void
4967 set_write_memory_permission (char *args, int from_tty,
4968 struct cmd_list_element *c)
4969 {
4970 /* Make the real values match the user-changed values. */
4971 may_write_memory = may_write_memory_1;
4972 update_observer_mode ();
4973 }
4974
4975
4976 void
4977 initialize_targets (void)
4978 {
4979 init_dummy_target ();
4980 push_target (&dummy_target);
4981
4982 add_info ("target", target_info, targ_desc);
4983 add_info ("files", target_info, targ_desc);
4984
4985 add_setshow_zuinteger_cmd ("target", class_maintenance, &targetdebug, _("\
4986 Set target debugging."), _("\
4987 Show target debugging."), _("\
4988 When non-zero, target debugging is enabled. Higher numbers are more\n\
4989 verbose. Changes do not take effect until the next \"run\" or \"target\"\n\
4990 command."),
4991 NULL,
4992 show_targetdebug,
4993 &setdebuglist, &showdebuglist);
4994
4995 add_setshow_boolean_cmd ("trust-readonly-sections", class_support,
4996 &trust_readonly, _("\
4997 Set mode for reading from readonly sections."), _("\
4998 Show mode for reading from readonly sections."), _("\
4999 When this mode is on, memory reads from readonly sections (such as .text)\n\
5000 will be read from the object file instead of from the target. This will\n\
5001 result in significant performance improvement for remote targets."),
5002 NULL,
5003 show_trust_readonly,
5004 &setlist, &showlist);
5005
5006 add_com ("monitor", class_obscure, do_monitor_command,
5007 _("Send a command to the remote monitor (remote targets only)."));
5008
5009 add_cmd ("target-stack", class_maintenance, maintenance_print_target_stack,
5010 _("Print the name of each layer of the internal target stack."),
5011 &maintenanceprintlist);
5012
5013 add_setshow_boolean_cmd ("target-async", no_class,
5014 &target_async_permitted_1, _("\
5015 Set whether gdb controls the inferior in asynchronous mode."), _("\
5016 Show whether gdb controls the inferior in asynchronous mode."), _("\
5017 Tells gdb whether to control the inferior in asynchronous mode."),
5018 set_target_async_command,
5019 show_target_async_command,
5020 &setlist,
5021 &showlist);
5022
5023 add_setshow_boolean_cmd ("may-write-registers", class_support,
5024 &may_write_registers_1, _("\
5025 Set permission to write into registers."), _("\
5026 Show permission to write into registers."), _("\
5027 When this permission is on, GDB may write into the target's registers.\n\
5028 Otherwise, any sort of write attempt will result in an error."),
5029 set_target_permissions, NULL,
5030 &setlist, &showlist);
5031
5032 add_setshow_boolean_cmd ("may-write-memory", class_support,
5033 &may_write_memory_1, _("\
5034 Set permission to write into target memory."), _("\
5035 Show permission to write into target memory."), _("\
5036 When this permission is on, GDB may write into the target's memory.\n\
5037 Otherwise, any sort of write attempt will result in an error."),
5038 set_write_memory_permission, NULL,
5039 &setlist, &showlist);
5040
5041 add_setshow_boolean_cmd ("may-insert-breakpoints", class_support,
5042 &may_insert_breakpoints_1, _("\
5043 Set permission to insert breakpoints in the target."), _("\
5044 Show permission to insert breakpoints in the target."), _("\
5045 When this permission is on, GDB may insert breakpoints in the program.\n\
5046 Otherwise, any sort of insertion attempt will result in an error."),
5047 set_target_permissions, NULL,
5048 &setlist, &showlist);
5049
5050 add_setshow_boolean_cmd ("may-insert-tracepoints", class_support,
5051 &may_insert_tracepoints_1, _("\
5052 Set permission to insert tracepoints in the target."), _("\
5053 Show permission to insert tracepoints in the target."), _("\
5054 When this permission is on, GDB may insert tracepoints in the program.\n\
5055 Otherwise, any sort of insertion attempt will result in an error."),
5056 set_target_permissions, NULL,
5057 &setlist, &showlist);
5058
5059 add_setshow_boolean_cmd ("may-insert-fast-tracepoints", class_support,
5060 &may_insert_fast_tracepoints_1, _("\
5061 Set permission to insert fast tracepoints in the target."), _("\
5062 Show permission to insert fast tracepoints in the target."), _("\
5063 When this permission is on, GDB may insert fast tracepoints.\n\
5064 Otherwise, any sort of insertion attempt will result in an error."),
5065 set_target_permissions, NULL,
5066 &setlist, &showlist);
5067
5068 add_setshow_boolean_cmd ("may-interrupt", class_support,
5069 &may_stop_1, _("\
5070 Set permission to interrupt or signal the target."), _("\
5071 Show permission to interrupt or signal the target."), _("\
5072 When this permission is on, GDB may interrupt/stop the target's execution.\n\
5073 Otherwise, any attempt to interrupt or stop will be ignored."),
5074 set_target_permissions, NULL,
5075 &setlist, &showlist);
5076 }