convert to_trace_init
[binutils-gdb.git] / gdb / target.c
1 /* Select target systems and architectures at runtime for GDB.
2
3 Copyright (C) 1990-2014 Free Software Foundation, Inc.
4
5 Contributed by Cygnus Support.
6
7 This file is part of GDB.
8
9 This program is free software; you can redistribute it and/or modify
10 it under the terms of the GNU General Public License as published by
11 the Free Software Foundation; either version 3 of the License, or
12 (at your option) any later version.
13
14 This program is distributed in the hope that it will be useful,
15 but WITHOUT ANY WARRANTY; without even the implied warranty of
16 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
17 GNU General Public License for more details.
18
19 You should have received a copy of the GNU General Public License
20 along with this program. If not, see <http://www.gnu.org/licenses/>. */
21
22 #include "defs.h"
23 #include <errno.h>
24 #include <string.h>
25 #include "target.h"
26 #include "target-dcache.h"
27 #include "gdbcmd.h"
28 #include "symtab.h"
29 #include "inferior.h"
30 #include "bfd.h"
31 #include "symfile.h"
32 #include "objfiles.h"
33 #include "dcache.h"
34 #include <signal.h>
35 #include "regcache.h"
36 #include "gdb_assert.h"
37 #include "gdbcore.h"
38 #include "exceptions.h"
39 #include "target-descriptions.h"
40 #include "gdbthread.h"
41 #include "solib.h"
42 #include "exec.h"
43 #include "inline-frame.h"
44 #include "tracepoint.h"
45 #include "gdb/fileio.h"
46 #include "agent.h"
47
48 static void target_info (char *, int);
49
50 static void default_terminal_info (struct target_ops *, const char *, int);
51
52 static int default_watchpoint_addr_within_range (struct target_ops *,
53 CORE_ADDR, CORE_ADDR, int);
54
55 static int default_region_ok_for_hw_watchpoint (struct target_ops *,
56 CORE_ADDR, int);
57
58 static void default_rcmd (struct target_ops *, char *, struct ui_file *);
59
60 static ptid_t default_get_ada_task_ptid (struct target_ops *self,
61 long lwp, long tid);
62
63 static void tcomplain (void) ATTRIBUTE_NORETURN;
64
65 static int nomemory (CORE_ADDR, char *, int, int, struct target_ops *);
66
67 static int return_zero (void);
68
69 static int return_minus_one (void);
70
71 static void *return_null (void);
72
73 void target_ignore (void);
74
75 static void target_command (char *, int);
76
77 static struct target_ops *find_default_run_target (char *);
78
79 static target_xfer_partial_ftype default_xfer_partial;
80
81 static struct gdbarch *default_thread_architecture (struct target_ops *ops,
82 ptid_t ptid);
83
84 static int dummy_find_memory_regions (struct target_ops *self,
85 find_memory_region_ftype ignore1,
86 void *ignore2);
87
88 static char *dummy_make_corefile_notes (struct target_ops *self,
89 bfd *ignore1, int *ignore2);
90
91 static int find_default_can_async_p (struct target_ops *ignore);
92
93 static int find_default_is_async_p (struct target_ops *ignore);
94
95 static enum exec_direction_kind default_execution_direction
96 (struct target_ops *self);
97
98 #include "target-delegates.c"
99
100 static void init_dummy_target (void);
101
102 static struct target_ops debug_target;
103
104 static void debug_to_open (char *, int);
105
106 static void debug_to_prepare_to_store (struct target_ops *self,
107 struct regcache *);
108
109 static void debug_to_files_info (struct target_ops *);
110
111 static int debug_to_insert_breakpoint (struct target_ops *, struct gdbarch *,
112 struct bp_target_info *);
113
114 static int debug_to_remove_breakpoint (struct target_ops *, struct gdbarch *,
115 struct bp_target_info *);
116
117 static int debug_to_can_use_hw_breakpoint (struct target_ops *self,
118 int, int, int);
119
120 static int debug_to_insert_hw_breakpoint (struct target_ops *self,
121 struct gdbarch *,
122 struct bp_target_info *);
123
124 static int debug_to_remove_hw_breakpoint (struct target_ops *self,
125 struct gdbarch *,
126 struct bp_target_info *);
127
128 static int debug_to_insert_watchpoint (struct target_ops *self,
129 CORE_ADDR, int, int,
130 struct expression *);
131
132 static int debug_to_remove_watchpoint (struct target_ops *self,
133 CORE_ADDR, int, int,
134 struct expression *);
135
136 static int debug_to_stopped_data_address (struct target_ops *, CORE_ADDR *);
137
138 static int debug_to_watchpoint_addr_within_range (struct target_ops *,
139 CORE_ADDR, CORE_ADDR, int);
140
141 static int debug_to_region_ok_for_hw_watchpoint (struct target_ops *self,
142 CORE_ADDR, int);
143
144 static int debug_to_can_accel_watchpoint_condition (struct target_ops *self,
145 CORE_ADDR, int, int,
146 struct expression *);
147
148 static void debug_to_terminal_init (struct target_ops *self);
149
150 static void debug_to_terminal_inferior (struct target_ops *self);
151
152 static void debug_to_terminal_ours_for_output (struct target_ops *self);
153
154 static void debug_to_terminal_save_ours (struct target_ops *self);
155
156 static void debug_to_terminal_ours (struct target_ops *self);
157
158 static void debug_to_load (struct target_ops *self, char *, int);
159
160 static int debug_to_can_run (struct target_ops *self);
161
162 static void debug_to_stop (struct target_ops *self, ptid_t);
163
164 /* Pointer to array of target architecture structures; the size of the
165 array; the current index into the array; the allocated size of the
166 array. */
167 struct target_ops **target_structs;
168 unsigned target_struct_size;
169 unsigned target_struct_allocsize;
170 #define DEFAULT_ALLOCSIZE 10
171
172 /* The initial current target, so that there is always a semi-valid
173 current target. */
174
175 static struct target_ops dummy_target;
176
177 /* Top of target stack. */
178
179 static struct target_ops *target_stack;
180
181 /* The target structure we are currently using to talk to a process
182 or file or whatever "inferior" we have. */
183
184 struct target_ops current_target;
185
186 /* Command list for target. */
187
188 static struct cmd_list_element *targetlist = NULL;
189
190 /* Nonzero if we should trust readonly sections from the
191 executable when reading memory. */
192
193 static int trust_readonly = 0;
194
195 /* Nonzero if we should show true memory content including
196 memory breakpoint inserted by gdb. */
197
198 static int show_memory_breakpoints = 0;
199
200 /* These globals control whether GDB attempts to perform these
201 operations; they are useful for targets that need to prevent
202 inadvertant disruption, such as in non-stop mode. */
203
204 int may_write_registers = 1;
205
206 int may_write_memory = 1;
207
208 int may_insert_breakpoints = 1;
209
210 int may_insert_tracepoints = 1;
211
212 int may_insert_fast_tracepoints = 1;
213
214 int may_stop = 1;
215
216 /* Non-zero if we want to see trace of target level stuff. */
217
218 static unsigned int targetdebug = 0;
219 static void
220 show_targetdebug (struct ui_file *file, int from_tty,
221 struct cmd_list_element *c, const char *value)
222 {
223 fprintf_filtered (file, _("Target debugging is %s.\n"), value);
224 }
225
226 static void setup_target_debug (void);
227
228 /* The user just typed 'target' without the name of a target. */
229
230 static void
231 target_command (char *arg, int from_tty)
232 {
233 fputs_filtered ("Argument required (target name). Try `help target'\n",
234 gdb_stdout);
235 }
236
237 /* Default target_has_* methods for process_stratum targets. */
238
239 int
240 default_child_has_all_memory (struct target_ops *ops)
241 {
242 /* If no inferior selected, then we can't read memory here. */
243 if (ptid_equal (inferior_ptid, null_ptid))
244 return 0;
245
246 return 1;
247 }
248
249 int
250 default_child_has_memory (struct target_ops *ops)
251 {
252 /* If no inferior selected, then we can't read memory here. */
253 if (ptid_equal (inferior_ptid, null_ptid))
254 return 0;
255
256 return 1;
257 }
258
259 int
260 default_child_has_stack (struct target_ops *ops)
261 {
262 /* If no inferior selected, there's no stack. */
263 if (ptid_equal (inferior_ptid, null_ptid))
264 return 0;
265
266 return 1;
267 }
268
269 int
270 default_child_has_registers (struct target_ops *ops)
271 {
272 /* Can't read registers from no inferior. */
273 if (ptid_equal (inferior_ptid, null_ptid))
274 return 0;
275
276 return 1;
277 }
278
279 int
280 default_child_has_execution (struct target_ops *ops, ptid_t the_ptid)
281 {
282 /* If there's no thread selected, then we can't make it run through
283 hoops. */
284 if (ptid_equal (the_ptid, null_ptid))
285 return 0;
286
287 return 1;
288 }
289
290
291 int
292 target_has_all_memory_1 (void)
293 {
294 struct target_ops *t;
295
296 for (t = current_target.beneath; t != NULL; t = t->beneath)
297 if (t->to_has_all_memory (t))
298 return 1;
299
300 return 0;
301 }
302
303 int
304 target_has_memory_1 (void)
305 {
306 struct target_ops *t;
307
308 for (t = current_target.beneath; t != NULL; t = t->beneath)
309 if (t->to_has_memory (t))
310 return 1;
311
312 return 0;
313 }
314
315 int
316 target_has_stack_1 (void)
317 {
318 struct target_ops *t;
319
320 for (t = current_target.beneath; t != NULL; t = t->beneath)
321 if (t->to_has_stack (t))
322 return 1;
323
324 return 0;
325 }
326
327 int
328 target_has_registers_1 (void)
329 {
330 struct target_ops *t;
331
332 for (t = current_target.beneath; t != NULL; t = t->beneath)
333 if (t->to_has_registers (t))
334 return 1;
335
336 return 0;
337 }
338
339 int
340 target_has_execution_1 (ptid_t the_ptid)
341 {
342 struct target_ops *t;
343
344 for (t = current_target.beneath; t != NULL; t = t->beneath)
345 if (t->to_has_execution (t, the_ptid))
346 return 1;
347
348 return 0;
349 }
350
351 int
352 target_has_execution_current (void)
353 {
354 return target_has_execution_1 (inferior_ptid);
355 }
356
357 /* Complete initialization of T. This ensures that various fields in
358 T are set, if needed by the target implementation. */
359
360 void
361 complete_target_initialization (struct target_ops *t)
362 {
363 /* Provide default values for all "must have" methods. */
364 if (t->to_xfer_partial == NULL)
365 t->to_xfer_partial = default_xfer_partial;
366
367 if (t->to_has_all_memory == NULL)
368 t->to_has_all_memory = (int (*) (struct target_ops *)) return_zero;
369
370 if (t->to_has_memory == NULL)
371 t->to_has_memory = (int (*) (struct target_ops *)) return_zero;
372
373 if (t->to_has_stack == NULL)
374 t->to_has_stack = (int (*) (struct target_ops *)) return_zero;
375
376 if (t->to_has_registers == NULL)
377 t->to_has_registers = (int (*) (struct target_ops *)) return_zero;
378
379 if (t->to_has_execution == NULL)
380 t->to_has_execution = (int (*) (struct target_ops *, ptid_t)) return_zero;
381
382 install_delegators (t);
383 }
384
385 /* Add possible target architecture T to the list and add a new
386 command 'target T->to_shortname'. Set COMPLETER as the command's
387 completer if not NULL. */
388
389 void
390 add_target_with_completer (struct target_ops *t,
391 completer_ftype *completer)
392 {
393 struct cmd_list_element *c;
394
395 complete_target_initialization (t);
396
397 if (!target_structs)
398 {
399 target_struct_allocsize = DEFAULT_ALLOCSIZE;
400 target_structs = (struct target_ops **) xmalloc
401 (target_struct_allocsize * sizeof (*target_structs));
402 }
403 if (target_struct_size >= target_struct_allocsize)
404 {
405 target_struct_allocsize *= 2;
406 target_structs = (struct target_ops **)
407 xrealloc ((char *) target_structs,
408 target_struct_allocsize * sizeof (*target_structs));
409 }
410 target_structs[target_struct_size++] = t;
411
412 if (targetlist == NULL)
413 add_prefix_cmd ("target", class_run, target_command, _("\
414 Connect to a target machine or process.\n\
415 The first argument is the type or protocol of the target machine.\n\
416 Remaining arguments are interpreted by the target protocol. For more\n\
417 information on the arguments for a particular protocol, type\n\
418 `help target ' followed by the protocol name."),
419 &targetlist, "target ", 0, &cmdlist);
420 c = add_cmd (t->to_shortname, no_class, t->to_open, t->to_doc,
421 &targetlist);
422 if (completer != NULL)
423 set_cmd_completer (c, completer);
424 }
425
426 /* Add a possible target architecture to the list. */
427
428 void
429 add_target (struct target_ops *t)
430 {
431 add_target_with_completer (t, NULL);
432 }
433
434 /* See target.h. */
435
436 void
437 add_deprecated_target_alias (struct target_ops *t, char *alias)
438 {
439 struct cmd_list_element *c;
440 char *alt;
441
442 /* If we use add_alias_cmd, here, we do not get the deprecated warning,
443 see PR cli/15104. */
444 c = add_cmd (alias, no_class, t->to_open, t->to_doc, &targetlist);
445 alt = xstrprintf ("target %s", t->to_shortname);
446 deprecate_cmd (c, alt);
447 }
448
449 /* Stub functions */
450
451 void
452 target_ignore (void)
453 {
454 }
455
456 void
457 target_kill (void)
458 {
459 struct target_ops *t;
460
461 for (t = current_target.beneath; t != NULL; t = t->beneath)
462 if (t->to_kill != NULL)
463 {
464 if (targetdebug)
465 fprintf_unfiltered (gdb_stdlog, "target_kill ()\n");
466
467 t->to_kill (t);
468 return;
469 }
470
471 noprocess ();
472 }
473
474 void
475 target_load (char *arg, int from_tty)
476 {
477 target_dcache_invalidate ();
478 (*current_target.to_load) (&current_target, arg, from_tty);
479 }
480
481 void
482 target_create_inferior (char *exec_file, char *args,
483 char **env, int from_tty)
484 {
485 struct target_ops *t;
486
487 for (t = current_target.beneath; t != NULL; t = t->beneath)
488 {
489 if (t->to_create_inferior != NULL)
490 {
491 t->to_create_inferior (t, exec_file, args, env, from_tty);
492 if (targetdebug)
493 fprintf_unfiltered (gdb_stdlog,
494 "target_create_inferior (%s, %s, xxx, %d)\n",
495 exec_file, args, from_tty);
496 return;
497 }
498 }
499
500 internal_error (__FILE__, __LINE__,
501 _("could not find a target to create inferior"));
502 }
503
504 void
505 target_terminal_inferior (void)
506 {
507 /* A background resume (``run&'') should leave GDB in control of the
508 terminal. Use target_can_async_p, not target_is_async_p, since at
509 this point the target is not async yet. However, if sync_execution
510 is not set, we know it will become async prior to resume. */
511 if (target_can_async_p () && !sync_execution)
512 return;
513
514 /* If GDB is resuming the inferior in the foreground, install
515 inferior's terminal modes. */
516 (*current_target.to_terminal_inferior) (&current_target);
517 }
518
519 static int
520 nomemory (CORE_ADDR memaddr, char *myaddr, int len, int write,
521 struct target_ops *t)
522 {
523 errno = EIO; /* Can't read/write this location. */
524 return 0; /* No bytes handled. */
525 }
526
527 static void
528 tcomplain (void)
529 {
530 error (_("You can't do that when your target is `%s'"),
531 current_target.to_shortname);
532 }
533
534 void
535 noprocess (void)
536 {
537 error (_("You can't do that without a process to debug."));
538 }
539
540 static void
541 default_terminal_info (struct target_ops *self, const char *args, int from_tty)
542 {
543 printf_unfiltered (_("No saved terminal information.\n"));
544 }
545
546 /* A default implementation for the to_get_ada_task_ptid target method.
547
548 This function builds the PTID by using both LWP and TID as part of
549 the PTID lwp and tid elements. The pid used is the pid of the
550 inferior_ptid. */
551
552 static ptid_t
553 default_get_ada_task_ptid (struct target_ops *self, long lwp, long tid)
554 {
555 return ptid_build (ptid_get_pid (inferior_ptid), lwp, tid);
556 }
557
558 static enum exec_direction_kind
559 default_execution_direction (struct target_ops *self)
560 {
561 if (!target_can_execute_reverse)
562 return EXEC_FORWARD;
563 else if (!target_can_async_p ())
564 return EXEC_FORWARD;
565 else
566 gdb_assert_not_reached ("\
567 to_execution_direction must be implemented for reverse async");
568 }
569
570 /* Go through the target stack from top to bottom, copying over zero
571 entries in current_target, then filling in still empty entries. In
572 effect, we are doing class inheritance through the pushed target
573 vectors.
574
575 NOTE: cagney/2003-10-17: The problem with this inheritance, as it
576 is currently implemented, is that it discards any knowledge of
577 which target an inherited method originally belonged to.
578 Consequently, new new target methods should instead explicitly and
579 locally search the target stack for the target that can handle the
580 request. */
581
582 static void
583 update_current_target (void)
584 {
585 struct target_ops *t;
586
587 /* First, reset current's contents. */
588 memset (&current_target, 0, sizeof (current_target));
589
590 /* Install the delegators. */
591 install_delegators (&current_target);
592
593 #define INHERIT(FIELD, TARGET) \
594 if (!current_target.FIELD) \
595 current_target.FIELD = (TARGET)->FIELD
596
597 for (t = target_stack; t; t = t->beneath)
598 {
599 INHERIT (to_shortname, t);
600 INHERIT (to_longname, t);
601 INHERIT (to_doc, t);
602 /* Do not inherit to_open. */
603 /* Do not inherit to_close. */
604 /* Do not inherit to_attach. */
605 /* Do not inherit to_post_attach. */
606 INHERIT (to_attach_no_wait, t);
607 /* Do not inherit to_detach. */
608 /* Do not inherit to_disconnect. */
609 /* Do not inherit to_resume. */
610 /* Do not inherit to_wait. */
611 /* Do not inherit to_fetch_registers. */
612 /* Do not inherit to_store_registers. */
613 /* Do not inherit to_prepare_to_store. */
614 INHERIT (deprecated_xfer_memory, t);
615 /* Do not inherit to_files_info. */
616 /* Do not inherit to_insert_breakpoint. */
617 /* Do not inherit to_remove_breakpoint. */
618 /* Do not inherit to_can_use_hw_breakpoint. */
619 /* Do not inherit to_insert_hw_breakpoint. */
620 /* Do not inherit to_remove_hw_breakpoint. */
621 /* Do not inherit to_ranged_break_num_registers. */
622 /* Do not inherit to_insert_watchpoint. */
623 /* Do not inherit to_remove_watchpoint. */
624 /* Do not inherit to_insert_mask_watchpoint. */
625 /* Do not inherit to_remove_mask_watchpoint. */
626 /* Do not inherit to_stopped_data_address. */
627 INHERIT (to_have_steppable_watchpoint, t);
628 INHERIT (to_have_continuable_watchpoint, t);
629 /* Do not inherit to_stopped_by_watchpoint. */
630 /* Do not inherit to_watchpoint_addr_within_range. */
631 /* Do not inherit to_region_ok_for_hw_watchpoint. */
632 /* Do not inherit to_can_accel_watchpoint_condition. */
633 /* Do not inherit to_masked_watch_num_registers. */
634 /* Do not inherit to_terminal_init. */
635 /* Do not inherit to_terminal_inferior. */
636 /* Do not inherit to_terminal_ours_for_output. */
637 /* Do not inherit to_terminal_ours. */
638 /* Do not inherit to_terminal_save_ours. */
639 /* Do not inherit to_terminal_info. */
640 /* Do not inherit to_kill. */
641 /* Do not inherit to_load. */
642 /* Do no inherit to_create_inferior. */
643 /* Do not inherit to_post_startup_inferior. */
644 /* Do not inherit to_insert_fork_catchpoint. */
645 /* Do not inherit to_remove_fork_catchpoint. */
646 /* Do not inherit to_insert_vfork_catchpoint. */
647 /* Do not inherit to_remove_vfork_catchpoint. */
648 /* Do not inherit to_follow_fork. */
649 /* Do not inherit to_insert_exec_catchpoint. */
650 /* Do not inherit to_remove_exec_catchpoint. */
651 /* Do not inherit to_set_syscall_catchpoint. */
652 /* Do not inherit to_has_exited. */
653 /* Do not inherit to_mourn_inferior. */
654 INHERIT (to_can_run, t);
655 /* Do not inherit to_pass_signals. */
656 /* Do not inherit to_program_signals. */
657 /* Do not inherit to_thread_alive. */
658 /* Do not inherit to_find_new_threads. */
659 /* Do not inherit to_pid_to_str. */
660 /* Do not inherit to_extra_thread_info. */
661 /* Do not inherit to_thread_name. */
662 INHERIT (to_stop, t);
663 /* Do not inherit to_xfer_partial. */
664 /* Do not inherit to_rcmd. */
665 /* Do not inherit to_pid_to_exec_file. */
666 /* Do not inherit to_log_command. */
667 INHERIT (to_stratum, t);
668 /* Do not inherit to_has_all_memory. */
669 /* Do not inherit to_has_memory. */
670 /* Do not inherit to_has_stack. */
671 /* Do not inherit to_has_registers. */
672 /* Do not inherit to_has_execution. */
673 INHERIT (to_has_thread_control, t);
674 /* Do not inherit to_can_async_p. */
675 /* Do not inherit to_is_async_p. */
676 /* Do not inherit to_async. */
677 /* Do not inherit to_find_memory_regions. */
678 /* Do not inherit to_make_corefile_notes. */
679 /* Do not inherit to_get_bookmark. */
680 /* Do not inherit to_goto_bookmark. */
681 /* Do not inherit to_get_thread_local_address. */
682 /* Do not inherit to_can_execute_reverse. */
683 /* Do not inherit to_execution_direction. */
684 /* Do not inherit to_thread_architecture. */
685 /* Do not inherit to_read_description. */
686 /* Do not inherit to_get_ada_task_ptid. */
687 /* Do not inherit to_search_memory. */
688 /* Do not inherit to_supports_multi_process. */
689 /* Do not inherit to_supports_enable_disable_tracepoint. */
690 /* Do not inherit to_supports_string_tracing. */
691 /* Do not inherit to_trace_init. */
692 INHERIT (to_download_tracepoint, t);
693 INHERIT (to_can_download_tracepoint, t);
694 INHERIT (to_download_trace_state_variable, t);
695 INHERIT (to_enable_tracepoint, t);
696 INHERIT (to_disable_tracepoint, t);
697 INHERIT (to_trace_set_readonly_regions, t);
698 INHERIT (to_trace_start, t);
699 INHERIT (to_get_trace_status, t);
700 INHERIT (to_get_tracepoint_status, t);
701 INHERIT (to_trace_stop, t);
702 INHERIT (to_trace_find, t);
703 INHERIT (to_get_trace_state_variable_value, t);
704 INHERIT (to_save_trace_data, t);
705 INHERIT (to_upload_tracepoints, t);
706 INHERIT (to_upload_trace_state_variables, t);
707 INHERIT (to_get_raw_trace_data, t);
708 INHERIT (to_get_min_fast_tracepoint_insn_len, t);
709 INHERIT (to_set_disconnected_tracing, t);
710 INHERIT (to_set_circular_trace_buffer, t);
711 INHERIT (to_set_trace_buffer_size, t);
712 INHERIT (to_set_trace_notes, t);
713 INHERIT (to_get_tib_address, t);
714 INHERIT (to_set_permissions, t);
715 INHERIT (to_static_tracepoint_marker_at, t);
716 INHERIT (to_static_tracepoint_markers_by_strid, t);
717 INHERIT (to_traceframe_info, t);
718 INHERIT (to_use_agent, t);
719 INHERIT (to_can_use_agent, t);
720 INHERIT (to_augmented_libraries_svr4_read, t);
721 INHERIT (to_magic, t);
722 INHERIT (to_supports_evaluation_of_breakpoint_conditions, t);
723 INHERIT (to_can_run_breakpoint_commands, t);
724 /* Do not inherit to_memory_map. */
725 /* Do not inherit to_flash_erase. */
726 /* Do not inherit to_flash_done. */
727 }
728 #undef INHERIT
729
730 /* Clean up a target struct so it no longer has any zero pointers in
731 it. Some entries are defaulted to a method that print an error,
732 others are hard-wired to a standard recursive default. */
733
734 #define de_fault(field, value) \
735 if (!current_target.field) \
736 current_target.field = value
737
738 de_fault (to_open,
739 (void (*) (char *, int))
740 tcomplain);
741 de_fault (to_close,
742 (void (*) (struct target_ops *))
743 target_ignore);
744 de_fault (deprecated_xfer_memory,
745 (int (*) (CORE_ADDR, gdb_byte *, int, int,
746 struct mem_attrib *, struct target_ops *))
747 nomemory);
748 de_fault (to_can_run,
749 (int (*) (struct target_ops *))
750 return_zero);
751 de_fault (to_stop,
752 (void (*) (struct target_ops *, ptid_t))
753 target_ignore);
754 current_target.to_read_description = NULL;
755 de_fault (to_download_tracepoint,
756 (void (*) (struct target_ops *, struct bp_location *))
757 tcomplain);
758 de_fault (to_can_download_tracepoint,
759 (int (*) (struct target_ops *))
760 return_zero);
761 de_fault (to_download_trace_state_variable,
762 (void (*) (struct target_ops *, struct trace_state_variable *))
763 tcomplain);
764 de_fault (to_enable_tracepoint,
765 (void (*) (struct target_ops *, struct bp_location *))
766 tcomplain);
767 de_fault (to_disable_tracepoint,
768 (void (*) (struct target_ops *, struct bp_location *))
769 tcomplain);
770 de_fault (to_trace_set_readonly_regions,
771 (void (*) (struct target_ops *))
772 tcomplain);
773 de_fault (to_trace_start,
774 (void (*) (struct target_ops *))
775 tcomplain);
776 de_fault (to_get_trace_status,
777 (int (*) (struct target_ops *, struct trace_status *))
778 return_minus_one);
779 de_fault (to_get_tracepoint_status,
780 (void (*) (struct target_ops *, struct breakpoint *,
781 struct uploaded_tp *))
782 tcomplain);
783 de_fault (to_trace_stop,
784 (void (*) (struct target_ops *))
785 tcomplain);
786 de_fault (to_trace_find,
787 (int (*) (struct target_ops *,
788 enum trace_find_type, int, CORE_ADDR, CORE_ADDR, int *))
789 return_minus_one);
790 de_fault (to_get_trace_state_variable_value,
791 (int (*) (struct target_ops *, int, LONGEST *))
792 return_zero);
793 de_fault (to_save_trace_data,
794 (int (*) (struct target_ops *, const char *))
795 tcomplain);
796 de_fault (to_upload_tracepoints,
797 (int (*) (struct target_ops *, struct uploaded_tp **))
798 return_zero);
799 de_fault (to_upload_trace_state_variables,
800 (int (*) (struct target_ops *, struct uploaded_tsv **))
801 return_zero);
802 de_fault (to_get_raw_trace_data,
803 (LONGEST (*) (struct target_ops *, gdb_byte *, ULONGEST, LONGEST))
804 tcomplain);
805 de_fault (to_get_min_fast_tracepoint_insn_len,
806 (int (*) (struct target_ops *))
807 return_minus_one);
808 de_fault (to_set_disconnected_tracing,
809 (void (*) (struct target_ops *, int))
810 target_ignore);
811 de_fault (to_set_circular_trace_buffer,
812 (void (*) (struct target_ops *, int))
813 target_ignore);
814 de_fault (to_set_trace_buffer_size,
815 (void (*) (struct target_ops *, LONGEST))
816 target_ignore);
817 de_fault (to_set_trace_notes,
818 (int (*) (struct target_ops *,
819 const char *, const char *, const char *))
820 return_zero);
821 de_fault (to_get_tib_address,
822 (int (*) (struct target_ops *, ptid_t, CORE_ADDR *))
823 tcomplain);
824 de_fault (to_set_permissions,
825 (void (*) (struct target_ops *))
826 target_ignore);
827 de_fault (to_static_tracepoint_marker_at,
828 (int (*) (struct target_ops *,
829 CORE_ADDR, struct static_tracepoint_marker *))
830 return_zero);
831 de_fault (to_static_tracepoint_markers_by_strid,
832 (VEC(static_tracepoint_marker_p) * (*) (struct target_ops *,
833 const char *))
834 tcomplain);
835 de_fault (to_traceframe_info,
836 (struct traceframe_info * (*) (struct target_ops *))
837 return_null);
838 de_fault (to_supports_evaluation_of_breakpoint_conditions,
839 (int (*) (struct target_ops *))
840 return_zero);
841 de_fault (to_can_run_breakpoint_commands,
842 (int (*) (struct target_ops *))
843 return_zero);
844 de_fault (to_use_agent,
845 (int (*) (struct target_ops *, int))
846 tcomplain);
847 de_fault (to_can_use_agent,
848 (int (*) (struct target_ops *))
849 return_zero);
850 de_fault (to_augmented_libraries_svr4_read,
851 (int (*) (struct target_ops *))
852 return_zero);
853
854 #undef de_fault
855
856 /* Finally, position the target-stack beneath the squashed
857 "current_target". That way code looking for a non-inherited
858 target method can quickly and simply find it. */
859 current_target.beneath = target_stack;
860
861 if (targetdebug)
862 setup_target_debug ();
863 }
864
865 /* Push a new target type into the stack of the existing target accessors,
866 possibly superseding some of the existing accessors.
867
868 Rather than allow an empty stack, we always have the dummy target at
869 the bottom stratum, so we can call the function vectors without
870 checking them. */
871
872 void
873 push_target (struct target_ops *t)
874 {
875 struct target_ops **cur;
876
877 /* Check magic number. If wrong, it probably means someone changed
878 the struct definition, but not all the places that initialize one. */
879 if (t->to_magic != OPS_MAGIC)
880 {
881 fprintf_unfiltered (gdb_stderr,
882 "Magic number of %s target struct wrong\n",
883 t->to_shortname);
884 internal_error (__FILE__, __LINE__,
885 _("failed internal consistency check"));
886 }
887
888 /* Find the proper stratum to install this target in. */
889 for (cur = &target_stack; (*cur) != NULL; cur = &(*cur)->beneath)
890 {
891 if ((int) (t->to_stratum) >= (int) (*cur)->to_stratum)
892 break;
893 }
894
895 /* If there's already targets at this stratum, remove them. */
896 /* FIXME: cagney/2003-10-15: I think this should be popping all
897 targets to CUR, and not just those at this stratum level. */
898 while ((*cur) != NULL && t->to_stratum == (*cur)->to_stratum)
899 {
900 /* There's already something at this stratum level. Close it,
901 and un-hook it from the stack. */
902 struct target_ops *tmp = (*cur);
903
904 (*cur) = (*cur)->beneath;
905 tmp->beneath = NULL;
906 target_close (tmp);
907 }
908
909 /* We have removed all targets in our stratum, now add the new one. */
910 t->beneath = (*cur);
911 (*cur) = t;
912
913 update_current_target ();
914 }
915
916 /* Remove a target_ops vector from the stack, wherever it may be.
917 Return how many times it was removed (0 or 1). */
918
919 int
920 unpush_target (struct target_ops *t)
921 {
922 struct target_ops **cur;
923 struct target_ops *tmp;
924
925 if (t->to_stratum == dummy_stratum)
926 internal_error (__FILE__, __LINE__,
927 _("Attempt to unpush the dummy target"));
928
929 /* Look for the specified target. Note that we assume that a target
930 can only occur once in the target stack. */
931
932 for (cur = &target_stack; (*cur) != NULL; cur = &(*cur)->beneath)
933 {
934 if ((*cur) == t)
935 break;
936 }
937
938 /* If we don't find target_ops, quit. Only open targets should be
939 closed. */
940 if ((*cur) == NULL)
941 return 0;
942
943 /* Unchain the target. */
944 tmp = (*cur);
945 (*cur) = (*cur)->beneath;
946 tmp->beneath = NULL;
947
948 update_current_target ();
949
950 /* Finally close the target. Note we do this after unchaining, so
951 any target method calls from within the target_close
952 implementation don't end up in T anymore. */
953 target_close (t);
954
955 return 1;
956 }
957
958 void
959 pop_all_targets_above (enum strata above_stratum)
960 {
961 while ((int) (current_target.to_stratum) > (int) above_stratum)
962 {
963 if (!unpush_target (target_stack))
964 {
965 fprintf_unfiltered (gdb_stderr,
966 "pop_all_targets couldn't find target %s\n",
967 target_stack->to_shortname);
968 internal_error (__FILE__, __LINE__,
969 _("failed internal consistency check"));
970 break;
971 }
972 }
973 }
974
975 void
976 pop_all_targets (void)
977 {
978 pop_all_targets_above (dummy_stratum);
979 }
980
981 /* Return 1 if T is now pushed in the target stack. Return 0 otherwise. */
982
983 int
984 target_is_pushed (struct target_ops *t)
985 {
986 struct target_ops **cur;
987
988 /* Check magic number. If wrong, it probably means someone changed
989 the struct definition, but not all the places that initialize one. */
990 if (t->to_magic != OPS_MAGIC)
991 {
992 fprintf_unfiltered (gdb_stderr,
993 "Magic number of %s target struct wrong\n",
994 t->to_shortname);
995 internal_error (__FILE__, __LINE__,
996 _("failed internal consistency check"));
997 }
998
999 for (cur = &target_stack; (*cur) != NULL; cur = &(*cur)->beneath)
1000 if (*cur == t)
1001 return 1;
1002
1003 return 0;
1004 }
1005
1006 /* Using the objfile specified in OBJFILE, find the address for the
1007 current thread's thread-local storage with offset OFFSET. */
1008 CORE_ADDR
1009 target_translate_tls_address (struct objfile *objfile, CORE_ADDR offset)
1010 {
1011 volatile CORE_ADDR addr = 0;
1012 struct target_ops *target;
1013
1014 for (target = current_target.beneath;
1015 target != NULL;
1016 target = target->beneath)
1017 {
1018 if (target->to_get_thread_local_address != NULL)
1019 break;
1020 }
1021
1022 if (target != NULL
1023 && gdbarch_fetch_tls_load_module_address_p (target_gdbarch ()))
1024 {
1025 ptid_t ptid = inferior_ptid;
1026 volatile struct gdb_exception ex;
1027
1028 TRY_CATCH (ex, RETURN_MASK_ALL)
1029 {
1030 CORE_ADDR lm_addr;
1031
1032 /* Fetch the load module address for this objfile. */
1033 lm_addr = gdbarch_fetch_tls_load_module_address (target_gdbarch (),
1034 objfile);
1035 /* If it's 0, throw the appropriate exception. */
1036 if (lm_addr == 0)
1037 throw_error (TLS_LOAD_MODULE_NOT_FOUND_ERROR,
1038 _("TLS load module not found"));
1039
1040 addr = target->to_get_thread_local_address (target, ptid,
1041 lm_addr, offset);
1042 }
1043 /* If an error occurred, print TLS related messages here. Otherwise,
1044 throw the error to some higher catcher. */
1045 if (ex.reason < 0)
1046 {
1047 int objfile_is_library = (objfile->flags & OBJF_SHARED);
1048
1049 switch (ex.error)
1050 {
1051 case TLS_NO_LIBRARY_SUPPORT_ERROR:
1052 error (_("Cannot find thread-local variables "
1053 "in this thread library."));
1054 break;
1055 case TLS_LOAD_MODULE_NOT_FOUND_ERROR:
1056 if (objfile_is_library)
1057 error (_("Cannot find shared library `%s' in dynamic"
1058 " linker's load module list"), objfile_name (objfile));
1059 else
1060 error (_("Cannot find executable file `%s' in dynamic"
1061 " linker's load module list"), objfile_name (objfile));
1062 break;
1063 case TLS_NOT_ALLOCATED_YET_ERROR:
1064 if (objfile_is_library)
1065 error (_("The inferior has not yet allocated storage for"
1066 " thread-local variables in\n"
1067 "the shared library `%s'\n"
1068 "for %s"),
1069 objfile_name (objfile), target_pid_to_str (ptid));
1070 else
1071 error (_("The inferior has not yet allocated storage for"
1072 " thread-local variables in\n"
1073 "the executable `%s'\n"
1074 "for %s"),
1075 objfile_name (objfile), target_pid_to_str (ptid));
1076 break;
1077 case TLS_GENERIC_ERROR:
1078 if (objfile_is_library)
1079 error (_("Cannot find thread-local storage for %s, "
1080 "shared library %s:\n%s"),
1081 target_pid_to_str (ptid),
1082 objfile_name (objfile), ex.message);
1083 else
1084 error (_("Cannot find thread-local storage for %s, "
1085 "executable file %s:\n%s"),
1086 target_pid_to_str (ptid),
1087 objfile_name (objfile), ex.message);
1088 break;
1089 default:
1090 throw_exception (ex);
1091 break;
1092 }
1093 }
1094 }
1095 /* It wouldn't be wrong here to try a gdbarch method, too; finding
1096 TLS is an ABI-specific thing. But we don't do that yet. */
1097 else
1098 error (_("Cannot find thread-local variables on this target"));
1099
1100 return addr;
1101 }
1102
1103 const char *
1104 target_xfer_status_to_string (enum target_xfer_status err)
1105 {
1106 #define CASE(X) case X: return #X
1107 switch (err)
1108 {
1109 CASE(TARGET_XFER_E_IO);
1110 CASE(TARGET_XFER_E_UNAVAILABLE);
1111 default:
1112 return "<unknown>";
1113 }
1114 #undef CASE
1115 };
1116
1117
1118 #undef MIN
1119 #define MIN(A, B) (((A) <= (B)) ? (A) : (B))
1120
1121 /* target_read_string -- read a null terminated string, up to LEN bytes,
1122 from MEMADDR in target. Set *ERRNOP to the errno code, or 0 if successful.
1123 Set *STRING to a pointer to malloc'd memory containing the data; the caller
1124 is responsible for freeing it. Return the number of bytes successfully
1125 read. */
1126
1127 int
1128 target_read_string (CORE_ADDR memaddr, char **string, int len, int *errnop)
1129 {
1130 int tlen, offset, i;
1131 gdb_byte buf[4];
1132 int errcode = 0;
1133 char *buffer;
1134 int buffer_allocated;
1135 char *bufptr;
1136 unsigned int nbytes_read = 0;
1137
1138 gdb_assert (string);
1139
1140 /* Small for testing. */
1141 buffer_allocated = 4;
1142 buffer = xmalloc (buffer_allocated);
1143 bufptr = buffer;
1144
1145 while (len > 0)
1146 {
1147 tlen = MIN (len, 4 - (memaddr & 3));
1148 offset = memaddr & 3;
1149
1150 errcode = target_read_memory (memaddr & ~3, buf, sizeof buf);
1151 if (errcode != 0)
1152 {
1153 /* The transfer request might have crossed the boundary to an
1154 unallocated region of memory. Retry the transfer, requesting
1155 a single byte. */
1156 tlen = 1;
1157 offset = 0;
1158 errcode = target_read_memory (memaddr, buf, 1);
1159 if (errcode != 0)
1160 goto done;
1161 }
1162
1163 if (bufptr - buffer + tlen > buffer_allocated)
1164 {
1165 unsigned int bytes;
1166
1167 bytes = bufptr - buffer;
1168 buffer_allocated *= 2;
1169 buffer = xrealloc (buffer, buffer_allocated);
1170 bufptr = buffer + bytes;
1171 }
1172
1173 for (i = 0; i < tlen; i++)
1174 {
1175 *bufptr++ = buf[i + offset];
1176 if (buf[i + offset] == '\000')
1177 {
1178 nbytes_read += i + 1;
1179 goto done;
1180 }
1181 }
1182
1183 memaddr += tlen;
1184 len -= tlen;
1185 nbytes_read += tlen;
1186 }
1187 done:
1188 *string = buffer;
1189 if (errnop != NULL)
1190 *errnop = errcode;
1191 return nbytes_read;
1192 }
1193
1194 struct target_section_table *
1195 target_get_section_table (struct target_ops *target)
1196 {
1197 struct target_ops *t;
1198
1199 if (targetdebug)
1200 fprintf_unfiltered (gdb_stdlog, "target_get_section_table ()\n");
1201
1202 for (t = target; t != NULL; t = t->beneath)
1203 if (t->to_get_section_table != NULL)
1204 return (*t->to_get_section_table) (t);
1205
1206 return NULL;
1207 }
1208
1209 /* Find a section containing ADDR. */
1210
1211 struct target_section *
1212 target_section_by_addr (struct target_ops *target, CORE_ADDR addr)
1213 {
1214 struct target_section_table *table = target_get_section_table (target);
1215 struct target_section *secp;
1216
1217 if (table == NULL)
1218 return NULL;
1219
1220 for (secp = table->sections; secp < table->sections_end; secp++)
1221 {
1222 if (addr >= secp->addr && addr < secp->endaddr)
1223 return secp;
1224 }
1225 return NULL;
1226 }
1227
1228 /* Read memory from the live target, even if currently inspecting a
1229 traceframe. The return is the same as that of target_read. */
1230
1231 static enum target_xfer_status
1232 target_read_live_memory (enum target_object object,
1233 ULONGEST memaddr, gdb_byte *myaddr, ULONGEST len,
1234 ULONGEST *xfered_len)
1235 {
1236 enum target_xfer_status ret;
1237 struct cleanup *cleanup;
1238
1239 /* Switch momentarily out of tfind mode so to access live memory.
1240 Note that this must not clear global state, such as the frame
1241 cache, which must still remain valid for the previous traceframe.
1242 We may be _building_ the frame cache at this point. */
1243 cleanup = make_cleanup_restore_traceframe_number ();
1244 set_traceframe_number (-1);
1245
1246 ret = target_xfer_partial (current_target.beneath, object, NULL,
1247 myaddr, NULL, memaddr, len, xfered_len);
1248
1249 do_cleanups (cleanup);
1250 return ret;
1251 }
1252
1253 /* Using the set of read-only target sections of OPS, read live
1254 read-only memory. Note that the actual reads start from the
1255 top-most target again.
1256
1257 For interface/parameters/return description see target.h,
1258 to_xfer_partial. */
1259
1260 static enum target_xfer_status
1261 memory_xfer_live_readonly_partial (struct target_ops *ops,
1262 enum target_object object,
1263 gdb_byte *readbuf, ULONGEST memaddr,
1264 ULONGEST len, ULONGEST *xfered_len)
1265 {
1266 struct target_section *secp;
1267 struct target_section_table *table;
1268
1269 secp = target_section_by_addr (ops, memaddr);
1270 if (secp != NULL
1271 && (bfd_get_section_flags (secp->the_bfd_section->owner,
1272 secp->the_bfd_section)
1273 & SEC_READONLY))
1274 {
1275 struct target_section *p;
1276 ULONGEST memend = memaddr + len;
1277
1278 table = target_get_section_table (ops);
1279
1280 for (p = table->sections; p < table->sections_end; p++)
1281 {
1282 if (memaddr >= p->addr)
1283 {
1284 if (memend <= p->endaddr)
1285 {
1286 /* Entire transfer is within this section. */
1287 return target_read_live_memory (object, memaddr,
1288 readbuf, len, xfered_len);
1289 }
1290 else if (memaddr >= p->endaddr)
1291 {
1292 /* This section ends before the transfer starts. */
1293 continue;
1294 }
1295 else
1296 {
1297 /* This section overlaps the transfer. Just do half. */
1298 len = p->endaddr - memaddr;
1299 return target_read_live_memory (object, memaddr,
1300 readbuf, len, xfered_len);
1301 }
1302 }
1303 }
1304 }
1305
1306 return TARGET_XFER_EOF;
1307 }
1308
1309 /* Read memory from more than one valid target. A core file, for
1310 instance, could have some of memory but delegate other bits to
1311 the target below it. So, we must manually try all targets. */
1312
1313 static enum target_xfer_status
1314 raw_memory_xfer_partial (struct target_ops *ops, gdb_byte *readbuf,
1315 const gdb_byte *writebuf, ULONGEST memaddr, LONGEST len,
1316 ULONGEST *xfered_len)
1317 {
1318 enum target_xfer_status res;
1319
1320 do
1321 {
1322 res = ops->to_xfer_partial (ops, TARGET_OBJECT_MEMORY, NULL,
1323 readbuf, writebuf, memaddr, len,
1324 xfered_len);
1325 if (res == TARGET_XFER_OK)
1326 break;
1327
1328 /* Stop if the target reports that the memory is not available. */
1329 if (res == TARGET_XFER_E_UNAVAILABLE)
1330 break;
1331
1332 /* We want to continue past core files to executables, but not
1333 past a running target's memory. */
1334 if (ops->to_has_all_memory (ops))
1335 break;
1336
1337 ops = ops->beneath;
1338 }
1339 while (ops != NULL);
1340
1341 return res;
1342 }
1343
1344 /* Perform a partial memory transfer.
1345 For docs see target.h, to_xfer_partial. */
1346
1347 static enum target_xfer_status
1348 memory_xfer_partial_1 (struct target_ops *ops, enum target_object object,
1349 gdb_byte *readbuf, const gdb_byte *writebuf, ULONGEST memaddr,
1350 ULONGEST len, ULONGEST *xfered_len)
1351 {
1352 enum target_xfer_status res;
1353 int reg_len;
1354 struct mem_region *region;
1355 struct inferior *inf;
1356
1357 /* For accesses to unmapped overlay sections, read directly from
1358 files. Must do this first, as MEMADDR may need adjustment. */
1359 if (readbuf != NULL && overlay_debugging)
1360 {
1361 struct obj_section *section = find_pc_overlay (memaddr);
1362
1363 if (pc_in_unmapped_range (memaddr, section))
1364 {
1365 struct target_section_table *table
1366 = target_get_section_table (ops);
1367 const char *section_name = section->the_bfd_section->name;
1368
1369 memaddr = overlay_mapped_address (memaddr, section);
1370 return section_table_xfer_memory_partial (readbuf, writebuf,
1371 memaddr, len, xfered_len,
1372 table->sections,
1373 table->sections_end,
1374 section_name);
1375 }
1376 }
1377
1378 /* Try the executable files, if "trust-readonly-sections" is set. */
1379 if (readbuf != NULL && trust_readonly)
1380 {
1381 struct target_section *secp;
1382 struct target_section_table *table;
1383
1384 secp = target_section_by_addr (ops, memaddr);
1385 if (secp != NULL
1386 && (bfd_get_section_flags (secp->the_bfd_section->owner,
1387 secp->the_bfd_section)
1388 & SEC_READONLY))
1389 {
1390 table = target_get_section_table (ops);
1391 return section_table_xfer_memory_partial (readbuf, writebuf,
1392 memaddr, len, xfered_len,
1393 table->sections,
1394 table->sections_end,
1395 NULL);
1396 }
1397 }
1398
1399 /* If reading unavailable memory in the context of traceframes, and
1400 this address falls within a read-only section, fallback to
1401 reading from live memory. */
1402 if (readbuf != NULL && get_traceframe_number () != -1)
1403 {
1404 VEC(mem_range_s) *available;
1405
1406 /* If we fail to get the set of available memory, then the
1407 target does not support querying traceframe info, and so we
1408 attempt reading from the traceframe anyway (assuming the
1409 target implements the old QTro packet then). */
1410 if (traceframe_available_memory (&available, memaddr, len))
1411 {
1412 struct cleanup *old_chain;
1413
1414 old_chain = make_cleanup (VEC_cleanup(mem_range_s), &available);
1415
1416 if (VEC_empty (mem_range_s, available)
1417 || VEC_index (mem_range_s, available, 0)->start != memaddr)
1418 {
1419 /* Don't read into the traceframe's available
1420 memory. */
1421 if (!VEC_empty (mem_range_s, available))
1422 {
1423 LONGEST oldlen = len;
1424
1425 len = VEC_index (mem_range_s, available, 0)->start - memaddr;
1426 gdb_assert (len <= oldlen);
1427 }
1428
1429 do_cleanups (old_chain);
1430
1431 /* This goes through the topmost target again. */
1432 res = memory_xfer_live_readonly_partial (ops, object,
1433 readbuf, memaddr,
1434 len, xfered_len);
1435 if (res == TARGET_XFER_OK)
1436 return TARGET_XFER_OK;
1437 else
1438 {
1439 /* No use trying further, we know some memory starting
1440 at MEMADDR isn't available. */
1441 *xfered_len = len;
1442 return TARGET_XFER_E_UNAVAILABLE;
1443 }
1444 }
1445
1446 /* Don't try to read more than how much is available, in
1447 case the target implements the deprecated QTro packet to
1448 cater for older GDBs (the target's knowledge of read-only
1449 sections may be outdated by now). */
1450 len = VEC_index (mem_range_s, available, 0)->length;
1451
1452 do_cleanups (old_chain);
1453 }
1454 }
1455
1456 /* Try GDB's internal data cache. */
1457 region = lookup_mem_region (memaddr);
1458 /* region->hi == 0 means there's no upper bound. */
1459 if (memaddr + len < region->hi || region->hi == 0)
1460 reg_len = len;
1461 else
1462 reg_len = region->hi - memaddr;
1463
1464 switch (region->attrib.mode)
1465 {
1466 case MEM_RO:
1467 if (writebuf != NULL)
1468 return TARGET_XFER_E_IO;
1469 break;
1470
1471 case MEM_WO:
1472 if (readbuf != NULL)
1473 return TARGET_XFER_E_IO;
1474 break;
1475
1476 case MEM_FLASH:
1477 /* We only support writing to flash during "load" for now. */
1478 if (writebuf != NULL)
1479 error (_("Writing to flash memory forbidden in this context"));
1480 break;
1481
1482 case MEM_NONE:
1483 return TARGET_XFER_E_IO;
1484 }
1485
1486 if (!ptid_equal (inferior_ptid, null_ptid))
1487 inf = find_inferior_pid (ptid_get_pid (inferior_ptid));
1488 else
1489 inf = NULL;
1490
1491 if (inf != NULL
1492 /* The dcache reads whole cache lines; that doesn't play well
1493 with reading from a trace buffer, because reading outside of
1494 the collected memory range fails. */
1495 && get_traceframe_number () == -1
1496 && (region->attrib.cache
1497 || (stack_cache_enabled_p () && object == TARGET_OBJECT_STACK_MEMORY)
1498 || (code_cache_enabled_p () && object == TARGET_OBJECT_CODE_MEMORY)))
1499 {
1500 DCACHE *dcache = target_dcache_get_or_init ();
1501 int l;
1502
1503 if (readbuf != NULL)
1504 l = dcache_xfer_memory (ops, dcache, memaddr, readbuf, reg_len, 0);
1505 else
1506 /* FIXME drow/2006-08-09: If we're going to preserve const
1507 correctness dcache_xfer_memory should take readbuf and
1508 writebuf. */
1509 l = dcache_xfer_memory (ops, dcache, memaddr, (void *) writebuf,
1510 reg_len, 1);
1511 if (l <= 0)
1512 return TARGET_XFER_E_IO;
1513 else
1514 {
1515 *xfered_len = (ULONGEST) l;
1516 return TARGET_XFER_OK;
1517 }
1518 }
1519
1520 /* If none of those methods found the memory we wanted, fall back
1521 to a target partial transfer. Normally a single call to
1522 to_xfer_partial is enough; if it doesn't recognize an object
1523 it will call the to_xfer_partial of the next target down.
1524 But for memory this won't do. Memory is the only target
1525 object which can be read from more than one valid target.
1526 A core file, for instance, could have some of memory but
1527 delegate other bits to the target below it. So, we must
1528 manually try all targets. */
1529
1530 res = raw_memory_xfer_partial (ops, readbuf, writebuf, memaddr, reg_len,
1531 xfered_len);
1532
1533 /* Make sure the cache gets updated no matter what - if we are writing
1534 to the stack. Even if this write is not tagged as such, we still need
1535 to update the cache. */
1536
1537 if (res == TARGET_XFER_OK
1538 && inf != NULL
1539 && writebuf != NULL
1540 && target_dcache_init_p ()
1541 && !region->attrib.cache
1542 && ((stack_cache_enabled_p () && object != TARGET_OBJECT_STACK_MEMORY)
1543 || (code_cache_enabled_p () && object != TARGET_OBJECT_CODE_MEMORY)))
1544 {
1545 DCACHE *dcache = target_dcache_get ();
1546
1547 dcache_update (dcache, memaddr, (void *) writebuf, reg_len);
1548 }
1549
1550 /* If we still haven't got anything, return the last error. We
1551 give up. */
1552 return res;
1553 }
1554
1555 /* Perform a partial memory transfer. For docs see target.h,
1556 to_xfer_partial. */
1557
1558 static enum target_xfer_status
1559 memory_xfer_partial (struct target_ops *ops, enum target_object object,
1560 gdb_byte *readbuf, const gdb_byte *writebuf,
1561 ULONGEST memaddr, ULONGEST len, ULONGEST *xfered_len)
1562 {
1563 enum target_xfer_status res;
1564
1565 /* Zero length requests are ok and require no work. */
1566 if (len == 0)
1567 return TARGET_XFER_EOF;
1568
1569 /* Fill in READBUF with breakpoint shadows, or WRITEBUF with
1570 breakpoint insns, thus hiding out from higher layers whether
1571 there are software breakpoints inserted in the code stream. */
1572 if (readbuf != NULL)
1573 {
1574 res = memory_xfer_partial_1 (ops, object, readbuf, NULL, memaddr, len,
1575 xfered_len);
1576
1577 if (res == TARGET_XFER_OK && !show_memory_breakpoints)
1578 breakpoint_xfer_memory (readbuf, NULL, NULL, memaddr, res);
1579 }
1580 else
1581 {
1582 void *buf;
1583 struct cleanup *old_chain;
1584
1585 /* A large write request is likely to be partially satisfied
1586 by memory_xfer_partial_1. We will continually malloc
1587 and free a copy of the entire write request for breakpoint
1588 shadow handling even though we only end up writing a small
1589 subset of it. Cap writes to 4KB to mitigate this. */
1590 len = min (4096, len);
1591
1592 buf = xmalloc (len);
1593 old_chain = make_cleanup (xfree, buf);
1594 memcpy (buf, writebuf, len);
1595
1596 breakpoint_xfer_memory (NULL, buf, writebuf, memaddr, len);
1597 res = memory_xfer_partial_1 (ops, object, NULL, buf, memaddr, len,
1598 xfered_len);
1599
1600 do_cleanups (old_chain);
1601 }
1602
1603 return res;
1604 }
1605
1606 static void
1607 restore_show_memory_breakpoints (void *arg)
1608 {
1609 show_memory_breakpoints = (uintptr_t) arg;
1610 }
1611
1612 struct cleanup *
1613 make_show_memory_breakpoints_cleanup (int show)
1614 {
1615 int current = show_memory_breakpoints;
1616
1617 show_memory_breakpoints = show;
1618 return make_cleanup (restore_show_memory_breakpoints,
1619 (void *) (uintptr_t) current);
1620 }
1621
1622 /* For docs see target.h, to_xfer_partial. */
1623
1624 enum target_xfer_status
1625 target_xfer_partial (struct target_ops *ops,
1626 enum target_object object, const char *annex,
1627 gdb_byte *readbuf, const gdb_byte *writebuf,
1628 ULONGEST offset, ULONGEST len,
1629 ULONGEST *xfered_len)
1630 {
1631 enum target_xfer_status retval;
1632
1633 gdb_assert (ops->to_xfer_partial != NULL);
1634
1635 /* Transfer is done when LEN is zero. */
1636 if (len == 0)
1637 return TARGET_XFER_EOF;
1638
1639 if (writebuf && !may_write_memory)
1640 error (_("Writing to memory is not allowed (addr %s, len %s)"),
1641 core_addr_to_string_nz (offset), plongest (len));
1642
1643 *xfered_len = 0;
1644
1645 /* If this is a memory transfer, let the memory-specific code
1646 have a look at it instead. Memory transfers are more
1647 complicated. */
1648 if (object == TARGET_OBJECT_MEMORY || object == TARGET_OBJECT_STACK_MEMORY
1649 || object == TARGET_OBJECT_CODE_MEMORY)
1650 retval = memory_xfer_partial (ops, object, readbuf,
1651 writebuf, offset, len, xfered_len);
1652 else if (object == TARGET_OBJECT_RAW_MEMORY)
1653 {
1654 /* Request the normal memory object from other layers. */
1655 retval = raw_memory_xfer_partial (ops, readbuf, writebuf, offset, len,
1656 xfered_len);
1657 }
1658 else
1659 retval = ops->to_xfer_partial (ops, object, annex, readbuf,
1660 writebuf, offset, len, xfered_len);
1661
1662 if (targetdebug)
1663 {
1664 const unsigned char *myaddr = NULL;
1665
1666 fprintf_unfiltered (gdb_stdlog,
1667 "%s:target_xfer_partial "
1668 "(%d, %s, %s, %s, %s, %s) = %d, %s",
1669 ops->to_shortname,
1670 (int) object,
1671 (annex ? annex : "(null)"),
1672 host_address_to_string (readbuf),
1673 host_address_to_string (writebuf),
1674 core_addr_to_string_nz (offset),
1675 pulongest (len), retval,
1676 pulongest (*xfered_len));
1677
1678 if (readbuf)
1679 myaddr = readbuf;
1680 if (writebuf)
1681 myaddr = writebuf;
1682 if (retval == TARGET_XFER_OK && myaddr != NULL)
1683 {
1684 int i;
1685
1686 fputs_unfiltered (", bytes =", gdb_stdlog);
1687 for (i = 0; i < *xfered_len; i++)
1688 {
1689 if ((((intptr_t) &(myaddr[i])) & 0xf) == 0)
1690 {
1691 if (targetdebug < 2 && i > 0)
1692 {
1693 fprintf_unfiltered (gdb_stdlog, " ...");
1694 break;
1695 }
1696 fprintf_unfiltered (gdb_stdlog, "\n");
1697 }
1698
1699 fprintf_unfiltered (gdb_stdlog, " %02x", myaddr[i] & 0xff);
1700 }
1701 }
1702
1703 fputc_unfiltered ('\n', gdb_stdlog);
1704 }
1705
1706 /* Check implementations of to_xfer_partial update *XFERED_LEN
1707 properly. Do assertion after printing debug messages, so that we
1708 can find more clues on assertion failure from debugging messages. */
1709 if (retval == TARGET_XFER_OK || retval == TARGET_XFER_E_UNAVAILABLE)
1710 gdb_assert (*xfered_len > 0);
1711
1712 return retval;
1713 }
1714
1715 /* Read LEN bytes of target memory at address MEMADDR, placing the
1716 results in GDB's memory at MYADDR. Returns either 0 for success or
1717 TARGET_XFER_E_IO if any error occurs.
1718
1719 If an error occurs, no guarantee is made about the contents of the data at
1720 MYADDR. In particular, the caller should not depend upon partial reads
1721 filling the buffer with good data. There is no way for the caller to know
1722 how much good data might have been transfered anyway. Callers that can
1723 deal with partial reads should call target_read (which will retry until
1724 it makes no progress, and then return how much was transferred). */
1725
1726 int
1727 target_read_memory (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1728 {
1729 /* Dispatch to the topmost target, not the flattened current_target.
1730 Memory accesses check target->to_has_(all_)memory, and the
1731 flattened target doesn't inherit those. */
1732 if (target_read (current_target.beneath, TARGET_OBJECT_MEMORY, NULL,
1733 myaddr, memaddr, len) == len)
1734 return 0;
1735 else
1736 return TARGET_XFER_E_IO;
1737 }
1738
1739 /* Like target_read_memory, but specify explicitly that this is a read
1740 from the target's raw memory. That is, this read bypasses the
1741 dcache, breakpoint shadowing, etc. */
1742
1743 int
1744 target_read_raw_memory (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1745 {
1746 /* See comment in target_read_memory about why the request starts at
1747 current_target.beneath. */
1748 if (target_read (current_target.beneath, TARGET_OBJECT_RAW_MEMORY, NULL,
1749 myaddr, memaddr, len) == len)
1750 return 0;
1751 else
1752 return TARGET_XFER_E_IO;
1753 }
1754
1755 /* Like target_read_memory, but specify explicitly that this is a read from
1756 the target's stack. This may trigger different cache behavior. */
1757
1758 int
1759 target_read_stack (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1760 {
1761 /* See comment in target_read_memory about why the request starts at
1762 current_target.beneath. */
1763 if (target_read (current_target.beneath, TARGET_OBJECT_STACK_MEMORY, NULL,
1764 myaddr, memaddr, len) == len)
1765 return 0;
1766 else
1767 return TARGET_XFER_E_IO;
1768 }
1769
1770 /* Like target_read_memory, but specify explicitly that this is a read from
1771 the target's code. This may trigger different cache behavior. */
1772
1773 int
1774 target_read_code (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1775 {
1776 /* See comment in target_read_memory about why the request starts at
1777 current_target.beneath. */
1778 if (target_read (current_target.beneath, TARGET_OBJECT_CODE_MEMORY, NULL,
1779 myaddr, memaddr, len) == len)
1780 return 0;
1781 else
1782 return TARGET_XFER_E_IO;
1783 }
1784
1785 /* Write LEN bytes from MYADDR to target memory at address MEMADDR.
1786 Returns either 0 for success or TARGET_XFER_E_IO if any
1787 error occurs. If an error occurs, no guarantee is made about how
1788 much data got written. Callers that can deal with partial writes
1789 should call target_write. */
1790
1791 int
1792 target_write_memory (CORE_ADDR memaddr, const gdb_byte *myaddr, ssize_t len)
1793 {
1794 /* See comment in target_read_memory about why the request starts at
1795 current_target.beneath. */
1796 if (target_write (current_target.beneath, TARGET_OBJECT_MEMORY, NULL,
1797 myaddr, memaddr, len) == len)
1798 return 0;
1799 else
1800 return TARGET_XFER_E_IO;
1801 }
1802
1803 /* Write LEN bytes from MYADDR to target raw memory at address
1804 MEMADDR. Returns either 0 for success or TARGET_XFER_E_IO
1805 if any error occurs. If an error occurs, no guarantee is made
1806 about how much data got written. Callers that can deal with
1807 partial writes should call target_write. */
1808
1809 int
1810 target_write_raw_memory (CORE_ADDR memaddr, const gdb_byte *myaddr, ssize_t len)
1811 {
1812 /* See comment in target_read_memory about why the request starts at
1813 current_target.beneath. */
1814 if (target_write (current_target.beneath, TARGET_OBJECT_RAW_MEMORY, NULL,
1815 myaddr, memaddr, len) == len)
1816 return 0;
1817 else
1818 return TARGET_XFER_E_IO;
1819 }
1820
1821 /* Fetch the target's memory map. */
1822
1823 VEC(mem_region_s) *
1824 target_memory_map (void)
1825 {
1826 VEC(mem_region_s) *result;
1827 struct mem_region *last_one, *this_one;
1828 int ix;
1829 struct target_ops *t;
1830
1831 if (targetdebug)
1832 fprintf_unfiltered (gdb_stdlog, "target_memory_map ()\n");
1833
1834 for (t = current_target.beneath; t != NULL; t = t->beneath)
1835 if (t->to_memory_map != NULL)
1836 break;
1837
1838 if (t == NULL)
1839 return NULL;
1840
1841 result = t->to_memory_map (t);
1842 if (result == NULL)
1843 return NULL;
1844
1845 qsort (VEC_address (mem_region_s, result),
1846 VEC_length (mem_region_s, result),
1847 sizeof (struct mem_region), mem_region_cmp);
1848
1849 /* Check that regions do not overlap. Simultaneously assign
1850 a numbering for the "mem" commands to use to refer to
1851 each region. */
1852 last_one = NULL;
1853 for (ix = 0; VEC_iterate (mem_region_s, result, ix, this_one); ix++)
1854 {
1855 this_one->number = ix;
1856
1857 if (last_one && last_one->hi > this_one->lo)
1858 {
1859 warning (_("Overlapping regions in memory map: ignoring"));
1860 VEC_free (mem_region_s, result);
1861 return NULL;
1862 }
1863 last_one = this_one;
1864 }
1865
1866 return result;
1867 }
1868
1869 void
1870 target_flash_erase (ULONGEST address, LONGEST length)
1871 {
1872 struct target_ops *t;
1873
1874 for (t = current_target.beneath; t != NULL; t = t->beneath)
1875 if (t->to_flash_erase != NULL)
1876 {
1877 if (targetdebug)
1878 fprintf_unfiltered (gdb_stdlog, "target_flash_erase (%s, %s)\n",
1879 hex_string (address), phex (length, 0));
1880 t->to_flash_erase (t, address, length);
1881 return;
1882 }
1883
1884 tcomplain ();
1885 }
1886
1887 void
1888 target_flash_done (void)
1889 {
1890 struct target_ops *t;
1891
1892 for (t = current_target.beneath; t != NULL; t = t->beneath)
1893 if (t->to_flash_done != NULL)
1894 {
1895 if (targetdebug)
1896 fprintf_unfiltered (gdb_stdlog, "target_flash_done\n");
1897 t->to_flash_done (t);
1898 return;
1899 }
1900
1901 tcomplain ();
1902 }
1903
1904 static void
1905 show_trust_readonly (struct ui_file *file, int from_tty,
1906 struct cmd_list_element *c, const char *value)
1907 {
1908 fprintf_filtered (file,
1909 _("Mode for reading from readonly sections is %s.\n"),
1910 value);
1911 }
1912
1913 /* More generic transfers. */
1914
1915 static enum target_xfer_status
1916 default_xfer_partial (struct target_ops *ops, enum target_object object,
1917 const char *annex, gdb_byte *readbuf,
1918 const gdb_byte *writebuf, ULONGEST offset, ULONGEST len,
1919 ULONGEST *xfered_len)
1920 {
1921 if (object == TARGET_OBJECT_MEMORY
1922 && ops->deprecated_xfer_memory != NULL)
1923 /* If available, fall back to the target's
1924 "deprecated_xfer_memory" method. */
1925 {
1926 int xfered = -1;
1927
1928 errno = 0;
1929 if (writebuf != NULL)
1930 {
1931 void *buffer = xmalloc (len);
1932 struct cleanup *cleanup = make_cleanup (xfree, buffer);
1933
1934 memcpy (buffer, writebuf, len);
1935 xfered = ops->deprecated_xfer_memory (offset, buffer, len,
1936 1/*write*/, NULL, ops);
1937 do_cleanups (cleanup);
1938 }
1939 if (readbuf != NULL)
1940 xfered = ops->deprecated_xfer_memory (offset, readbuf, len,
1941 0/*read*/, NULL, ops);
1942 if (xfered > 0)
1943 {
1944 *xfered_len = (ULONGEST) xfered;
1945 return TARGET_XFER_E_IO;
1946 }
1947 else if (xfered == 0 && errno == 0)
1948 /* "deprecated_xfer_memory" uses 0, cross checked against
1949 ERRNO as one indication of an error. */
1950 return TARGET_XFER_EOF;
1951 else
1952 return TARGET_XFER_E_IO;
1953 }
1954 else
1955 {
1956 gdb_assert (ops->beneath != NULL);
1957 return ops->beneath->to_xfer_partial (ops->beneath, object, annex,
1958 readbuf, writebuf, offset, len,
1959 xfered_len);
1960 }
1961 }
1962
1963 /* Target vector read/write partial wrapper functions. */
1964
1965 static enum target_xfer_status
1966 target_read_partial (struct target_ops *ops,
1967 enum target_object object,
1968 const char *annex, gdb_byte *buf,
1969 ULONGEST offset, ULONGEST len,
1970 ULONGEST *xfered_len)
1971 {
1972 return target_xfer_partial (ops, object, annex, buf, NULL, offset, len,
1973 xfered_len);
1974 }
1975
1976 static enum target_xfer_status
1977 target_write_partial (struct target_ops *ops,
1978 enum target_object object,
1979 const char *annex, const gdb_byte *buf,
1980 ULONGEST offset, LONGEST len, ULONGEST *xfered_len)
1981 {
1982 return target_xfer_partial (ops, object, annex, NULL, buf, offset, len,
1983 xfered_len);
1984 }
1985
1986 /* Wrappers to perform the full transfer. */
1987
1988 /* For docs on target_read see target.h. */
1989
1990 LONGEST
1991 target_read (struct target_ops *ops,
1992 enum target_object object,
1993 const char *annex, gdb_byte *buf,
1994 ULONGEST offset, LONGEST len)
1995 {
1996 LONGEST xfered = 0;
1997
1998 while (xfered < len)
1999 {
2000 ULONGEST xfered_len;
2001 enum target_xfer_status status;
2002
2003 status = target_read_partial (ops, object, annex,
2004 (gdb_byte *) buf + xfered,
2005 offset + xfered, len - xfered,
2006 &xfered_len);
2007
2008 /* Call an observer, notifying them of the xfer progress? */
2009 if (status == TARGET_XFER_EOF)
2010 return xfered;
2011 else if (status == TARGET_XFER_OK)
2012 {
2013 xfered += xfered_len;
2014 QUIT;
2015 }
2016 else
2017 return -1;
2018
2019 }
2020 return len;
2021 }
2022
2023 /* Assuming that the entire [begin, end) range of memory cannot be
2024 read, try to read whatever subrange is possible to read.
2025
2026 The function returns, in RESULT, either zero or one memory block.
2027 If there's a readable subrange at the beginning, it is completely
2028 read and returned. Any further readable subrange will not be read.
2029 Otherwise, if there's a readable subrange at the end, it will be
2030 completely read and returned. Any readable subranges before it
2031 (obviously, not starting at the beginning), will be ignored. In
2032 other cases -- either no readable subrange, or readable subrange(s)
2033 that is neither at the beginning, or end, nothing is returned.
2034
2035 The purpose of this function is to handle a read across a boundary
2036 of accessible memory in a case when memory map is not available.
2037 The above restrictions are fine for this case, but will give
2038 incorrect results if the memory is 'patchy'. However, supporting
2039 'patchy' memory would require trying to read every single byte,
2040 and it seems unacceptable solution. Explicit memory map is
2041 recommended for this case -- and target_read_memory_robust will
2042 take care of reading multiple ranges then. */
2043
2044 static void
2045 read_whatever_is_readable (struct target_ops *ops,
2046 ULONGEST begin, ULONGEST end,
2047 VEC(memory_read_result_s) **result)
2048 {
2049 gdb_byte *buf = xmalloc (end - begin);
2050 ULONGEST current_begin = begin;
2051 ULONGEST current_end = end;
2052 int forward;
2053 memory_read_result_s r;
2054 ULONGEST xfered_len;
2055
2056 /* If we previously failed to read 1 byte, nothing can be done here. */
2057 if (end - begin <= 1)
2058 {
2059 xfree (buf);
2060 return;
2061 }
2062
2063 /* Check that either first or the last byte is readable, and give up
2064 if not. This heuristic is meant to permit reading accessible memory
2065 at the boundary of accessible region. */
2066 if (target_read_partial (ops, TARGET_OBJECT_MEMORY, NULL,
2067 buf, begin, 1, &xfered_len) == TARGET_XFER_OK)
2068 {
2069 forward = 1;
2070 ++current_begin;
2071 }
2072 else if (target_read_partial (ops, TARGET_OBJECT_MEMORY, NULL,
2073 buf + (end-begin) - 1, end - 1, 1,
2074 &xfered_len) == TARGET_XFER_OK)
2075 {
2076 forward = 0;
2077 --current_end;
2078 }
2079 else
2080 {
2081 xfree (buf);
2082 return;
2083 }
2084
2085 /* Loop invariant is that the [current_begin, current_end) was previously
2086 found to be not readable as a whole.
2087
2088 Note loop condition -- if the range has 1 byte, we can't divide the range
2089 so there's no point trying further. */
2090 while (current_end - current_begin > 1)
2091 {
2092 ULONGEST first_half_begin, first_half_end;
2093 ULONGEST second_half_begin, second_half_end;
2094 LONGEST xfer;
2095 ULONGEST middle = current_begin + (current_end - current_begin)/2;
2096
2097 if (forward)
2098 {
2099 first_half_begin = current_begin;
2100 first_half_end = middle;
2101 second_half_begin = middle;
2102 second_half_end = current_end;
2103 }
2104 else
2105 {
2106 first_half_begin = middle;
2107 first_half_end = current_end;
2108 second_half_begin = current_begin;
2109 second_half_end = middle;
2110 }
2111
2112 xfer = target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2113 buf + (first_half_begin - begin),
2114 first_half_begin,
2115 first_half_end - first_half_begin);
2116
2117 if (xfer == first_half_end - first_half_begin)
2118 {
2119 /* This half reads up fine. So, the error must be in the
2120 other half. */
2121 current_begin = second_half_begin;
2122 current_end = second_half_end;
2123 }
2124 else
2125 {
2126 /* This half is not readable. Because we've tried one byte, we
2127 know some part of this half if actually redable. Go to the next
2128 iteration to divide again and try to read.
2129
2130 We don't handle the other half, because this function only tries
2131 to read a single readable subrange. */
2132 current_begin = first_half_begin;
2133 current_end = first_half_end;
2134 }
2135 }
2136
2137 if (forward)
2138 {
2139 /* The [begin, current_begin) range has been read. */
2140 r.begin = begin;
2141 r.end = current_begin;
2142 r.data = buf;
2143 }
2144 else
2145 {
2146 /* The [current_end, end) range has been read. */
2147 LONGEST rlen = end - current_end;
2148
2149 r.data = xmalloc (rlen);
2150 memcpy (r.data, buf + current_end - begin, rlen);
2151 r.begin = current_end;
2152 r.end = end;
2153 xfree (buf);
2154 }
2155 VEC_safe_push(memory_read_result_s, (*result), &r);
2156 }
2157
2158 void
2159 free_memory_read_result_vector (void *x)
2160 {
2161 VEC(memory_read_result_s) *v = x;
2162 memory_read_result_s *current;
2163 int ix;
2164
2165 for (ix = 0; VEC_iterate (memory_read_result_s, v, ix, current); ++ix)
2166 {
2167 xfree (current->data);
2168 }
2169 VEC_free (memory_read_result_s, v);
2170 }
2171
2172 VEC(memory_read_result_s) *
2173 read_memory_robust (struct target_ops *ops, ULONGEST offset, LONGEST len)
2174 {
2175 VEC(memory_read_result_s) *result = 0;
2176
2177 LONGEST xfered = 0;
2178 while (xfered < len)
2179 {
2180 struct mem_region *region = lookup_mem_region (offset + xfered);
2181 LONGEST rlen;
2182
2183 /* If there is no explicit region, a fake one should be created. */
2184 gdb_assert (region);
2185
2186 if (region->hi == 0)
2187 rlen = len - xfered;
2188 else
2189 rlen = region->hi - offset;
2190
2191 if (region->attrib.mode == MEM_NONE || region->attrib.mode == MEM_WO)
2192 {
2193 /* Cannot read this region. Note that we can end up here only
2194 if the region is explicitly marked inaccessible, or
2195 'inaccessible-by-default' is in effect. */
2196 xfered += rlen;
2197 }
2198 else
2199 {
2200 LONGEST to_read = min (len - xfered, rlen);
2201 gdb_byte *buffer = (gdb_byte *)xmalloc (to_read);
2202
2203 LONGEST xfer = target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2204 (gdb_byte *) buffer,
2205 offset + xfered, to_read);
2206 /* Call an observer, notifying them of the xfer progress? */
2207 if (xfer <= 0)
2208 {
2209 /* Got an error reading full chunk. See if maybe we can read
2210 some subrange. */
2211 xfree (buffer);
2212 read_whatever_is_readable (ops, offset + xfered,
2213 offset + xfered + to_read, &result);
2214 xfered += to_read;
2215 }
2216 else
2217 {
2218 struct memory_read_result r;
2219 r.data = buffer;
2220 r.begin = offset + xfered;
2221 r.end = r.begin + xfer;
2222 VEC_safe_push (memory_read_result_s, result, &r);
2223 xfered += xfer;
2224 }
2225 QUIT;
2226 }
2227 }
2228 return result;
2229 }
2230
2231
2232 /* An alternative to target_write with progress callbacks. */
2233
2234 LONGEST
2235 target_write_with_progress (struct target_ops *ops,
2236 enum target_object object,
2237 const char *annex, const gdb_byte *buf,
2238 ULONGEST offset, LONGEST len,
2239 void (*progress) (ULONGEST, void *), void *baton)
2240 {
2241 LONGEST xfered = 0;
2242
2243 /* Give the progress callback a chance to set up. */
2244 if (progress)
2245 (*progress) (0, baton);
2246
2247 while (xfered < len)
2248 {
2249 ULONGEST xfered_len;
2250 enum target_xfer_status status;
2251
2252 status = target_write_partial (ops, object, annex,
2253 (gdb_byte *) buf + xfered,
2254 offset + xfered, len - xfered,
2255 &xfered_len);
2256
2257 if (status == TARGET_XFER_EOF)
2258 return xfered;
2259 if (TARGET_XFER_STATUS_ERROR_P (status))
2260 return -1;
2261
2262 gdb_assert (status == TARGET_XFER_OK);
2263 if (progress)
2264 (*progress) (xfered_len, baton);
2265
2266 xfered += xfered_len;
2267 QUIT;
2268 }
2269 return len;
2270 }
2271
2272 /* For docs on target_write see target.h. */
2273
2274 LONGEST
2275 target_write (struct target_ops *ops,
2276 enum target_object object,
2277 const char *annex, const gdb_byte *buf,
2278 ULONGEST offset, LONGEST len)
2279 {
2280 return target_write_with_progress (ops, object, annex, buf, offset, len,
2281 NULL, NULL);
2282 }
2283
2284 /* Read OBJECT/ANNEX using OPS. Store the result in *BUF_P and return
2285 the size of the transferred data. PADDING additional bytes are
2286 available in *BUF_P. This is a helper function for
2287 target_read_alloc; see the declaration of that function for more
2288 information. */
2289
2290 static LONGEST
2291 target_read_alloc_1 (struct target_ops *ops, enum target_object object,
2292 const char *annex, gdb_byte **buf_p, int padding)
2293 {
2294 size_t buf_alloc, buf_pos;
2295 gdb_byte *buf;
2296
2297 /* This function does not have a length parameter; it reads the
2298 entire OBJECT). Also, it doesn't support objects fetched partly
2299 from one target and partly from another (in a different stratum,
2300 e.g. a core file and an executable). Both reasons make it
2301 unsuitable for reading memory. */
2302 gdb_assert (object != TARGET_OBJECT_MEMORY);
2303
2304 /* Start by reading up to 4K at a time. The target will throttle
2305 this number down if necessary. */
2306 buf_alloc = 4096;
2307 buf = xmalloc (buf_alloc);
2308 buf_pos = 0;
2309 while (1)
2310 {
2311 ULONGEST xfered_len;
2312 enum target_xfer_status status;
2313
2314 status = target_read_partial (ops, object, annex, &buf[buf_pos],
2315 buf_pos, buf_alloc - buf_pos - padding,
2316 &xfered_len);
2317
2318 if (status == TARGET_XFER_EOF)
2319 {
2320 /* Read all there was. */
2321 if (buf_pos == 0)
2322 xfree (buf);
2323 else
2324 *buf_p = buf;
2325 return buf_pos;
2326 }
2327 else if (status != TARGET_XFER_OK)
2328 {
2329 /* An error occurred. */
2330 xfree (buf);
2331 return TARGET_XFER_E_IO;
2332 }
2333
2334 buf_pos += xfered_len;
2335
2336 /* If the buffer is filling up, expand it. */
2337 if (buf_alloc < buf_pos * 2)
2338 {
2339 buf_alloc *= 2;
2340 buf = xrealloc (buf, buf_alloc);
2341 }
2342
2343 QUIT;
2344 }
2345 }
2346
2347 /* Read OBJECT/ANNEX using OPS. Store the result in *BUF_P and return
2348 the size of the transferred data. See the declaration in "target.h"
2349 function for more information about the return value. */
2350
2351 LONGEST
2352 target_read_alloc (struct target_ops *ops, enum target_object object,
2353 const char *annex, gdb_byte **buf_p)
2354 {
2355 return target_read_alloc_1 (ops, object, annex, buf_p, 0);
2356 }
2357
2358 /* Read OBJECT/ANNEX using OPS. The result is NUL-terminated and
2359 returned as a string, allocated using xmalloc. If an error occurs
2360 or the transfer is unsupported, NULL is returned. Empty objects
2361 are returned as allocated but empty strings. A warning is issued
2362 if the result contains any embedded NUL bytes. */
2363
2364 char *
2365 target_read_stralloc (struct target_ops *ops, enum target_object object,
2366 const char *annex)
2367 {
2368 gdb_byte *buffer;
2369 char *bufstr;
2370 LONGEST i, transferred;
2371
2372 transferred = target_read_alloc_1 (ops, object, annex, &buffer, 1);
2373 bufstr = (char *) buffer;
2374
2375 if (transferred < 0)
2376 return NULL;
2377
2378 if (transferred == 0)
2379 return xstrdup ("");
2380
2381 bufstr[transferred] = 0;
2382
2383 /* Check for embedded NUL bytes; but allow trailing NULs. */
2384 for (i = strlen (bufstr); i < transferred; i++)
2385 if (bufstr[i] != 0)
2386 {
2387 warning (_("target object %d, annex %s, "
2388 "contained unexpected null characters"),
2389 (int) object, annex ? annex : "(none)");
2390 break;
2391 }
2392
2393 return bufstr;
2394 }
2395
2396 /* Memory transfer methods. */
2397
2398 void
2399 get_target_memory (struct target_ops *ops, CORE_ADDR addr, gdb_byte *buf,
2400 LONGEST len)
2401 {
2402 /* This method is used to read from an alternate, non-current
2403 target. This read must bypass the overlay support (as symbols
2404 don't match this target), and GDB's internal cache (wrong cache
2405 for this target). */
2406 if (target_read (ops, TARGET_OBJECT_RAW_MEMORY, NULL, buf, addr, len)
2407 != len)
2408 memory_error (TARGET_XFER_E_IO, addr);
2409 }
2410
2411 ULONGEST
2412 get_target_memory_unsigned (struct target_ops *ops, CORE_ADDR addr,
2413 int len, enum bfd_endian byte_order)
2414 {
2415 gdb_byte buf[sizeof (ULONGEST)];
2416
2417 gdb_assert (len <= sizeof (buf));
2418 get_target_memory (ops, addr, buf, len);
2419 return extract_unsigned_integer (buf, len, byte_order);
2420 }
2421
2422 /* See target.h. */
2423
2424 int
2425 target_insert_breakpoint (struct gdbarch *gdbarch,
2426 struct bp_target_info *bp_tgt)
2427 {
2428 if (!may_insert_breakpoints)
2429 {
2430 warning (_("May not insert breakpoints"));
2431 return 1;
2432 }
2433
2434 return current_target.to_insert_breakpoint (&current_target,
2435 gdbarch, bp_tgt);
2436 }
2437
2438 /* See target.h. */
2439
2440 int
2441 target_remove_breakpoint (struct gdbarch *gdbarch,
2442 struct bp_target_info *bp_tgt)
2443 {
2444 /* This is kind of a weird case to handle, but the permission might
2445 have been changed after breakpoints were inserted - in which case
2446 we should just take the user literally and assume that any
2447 breakpoints should be left in place. */
2448 if (!may_insert_breakpoints)
2449 {
2450 warning (_("May not remove breakpoints"));
2451 return 1;
2452 }
2453
2454 return current_target.to_remove_breakpoint (&current_target,
2455 gdbarch, bp_tgt);
2456 }
2457
2458 static void
2459 target_info (char *args, int from_tty)
2460 {
2461 struct target_ops *t;
2462 int has_all_mem = 0;
2463
2464 if (symfile_objfile != NULL)
2465 printf_unfiltered (_("Symbols from \"%s\".\n"),
2466 objfile_name (symfile_objfile));
2467
2468 for (t = target_stack; t != NULL; t = t->beneath)
2469 {
2470 if (!(*t->to_has_memory) (t))
2471 continue;
2472
2473 if ((int) (t->to_stratum) <= (int) dummy_stratum)
2474 continue;
2475 if (has_all_mem)
2476 printf_unfiltered (_("\tWhile running this, "
2477 "GDB does not access memory from...\n"));
2478 printf_unfiltered ("%s:\n", t->to_longname);
2479 (t->to_files_info) (t);
2480 has_all_mem = (*t->to_has_all_memory) (t);
2481 }
2482 }
2483
2484 /* This function is called before any new inferior is created, e.g.
2485 by running a program, attaching, or connecting to a target.
2486 It cleans up any state from previous invocations which might
2487 change between runs. This is a subset of what target_preopen
2488 resets (things which might change between targets). */
2489
2490 void
2491 target_pre_inferior (int from_tty)
2492 {
2493 /* Clear out solib state. Otherwise the solib state of the previous
2494 inferior might have survived and is entirely wrong for the new
2495 target. This has been observed on GNU/Linux using glibc 2.3. How
2496 to reproduce:
2497
2498 bash$ ./foo&
2499 [1] 4711
2500 bash$ ./foo&
2501 [1] 4712
2502 bash$ gdb ./foo
2503 [...]
2504 (gdb) attach 4711
2505 (gdb) detach
2506 (gdb) attach 4712
2507 Cannot access memory at address 0xdeadbeef
2508 */
2509
2510 /* In some OSs, the shared library list is the same/global/shared
2511 across inferiors. If code is shared between processes, so are
2512 memory regions and features. */
2513 if (!gdbarch_has_global_solist (target_gdbarch ()))
2514 {
2515 no_shared_libraries (NULL, from_tty);
2516
2517 invalidate_target_mem_regions ();
2518
2519 target_clear_description ();
2520 }
2521
2522 agent_capability_invalidate ();
2523 }
2524
2525 /* Callback for iterate_over_inferiors. Gets rid of the given
2526 inferior. */
2527
2528 static int
2529 dispose_inferior (struct inferior *inf, void *args)
2530 {
2531 struct thread_info *thread;
2532
2533 thread = any_thread_of_process (inf->pid);
2534 if (thread)
2535 {
2536 switch_to_thread (thread->ptid);
2537
2538 /* Core inferiors actually should be detached, not killed. */
2539 if (target_has_execution)
2540 target_kill ();
2541 else
2542 target_detach (NULL, 0);
2543 }
2544
2545 return 0;
2546 }
2547
2548 /* This is to be called by the open routine before it does
2549 anything. */
2550
2551 void
2552 target_preopen (int from_tty)
2553 {
2554 dont_repeat ();
2555
2556 if (have_inferiors ())
2557 {
2558 if (!from_tty
2559 || !have_live_inferiors ()
2560 || query (_("A program is being debugged already. Kill it? ")))
2561 iterate_over_inferiors (dispose_inferior, NULL);
2562 else
2563 error (_("Program not killed."));
2564 }
2565
2566 /* Calling target_kill may remove the target from the stack. But if
2567 it doesn't (which seems like a win for UDI), remove it now. */
2568 /* Leave the exec target, though. The user may be switching from a
2569 live process to a core of the same program. */
2570 pop_all_targets_above (file_stratum);
2571
2572 target_pre_inferior (from_tty);
2573 }
2574
2575 /* Detach a target after doing deferred register stores. */
2576
2577 void
2578 target_detach (const char *args, int from_tty)
2579 {
2580 struct target_ops* t;
2581
2582 if (gdbarch_has_global_breakpoints (target_gdbarch ()))
2583 /* Don't remove global breakpoints here. They're removed on
2584 disconnection from the target. */
2585 ;
2586 else
2587 /* If we're in breakpoints-always-inserted mode, have to remove
2588 them before detaching. */
2589 remove_breakpoints_pid (ptid_get_pid (inferior_ptid));
2590
2591 prepare_for_detach ();
2592
2593 current_target.to_detach (&current_target, args, from_tty);
2594 if (targetdebug)
2595 fprintf_unfiltered (gdb_stdlog, "target_detach (%s, %d)\n",
2596 args, from_tty);
2597 }
2598
2599 void
2600 target_disconnect (char *args, int from_tty)
2601 {
2602 struct target_ops *t;
2603
2604 /* If we're in breakpoints-always-inserted mode or if breakpoints
2605 are global across processes, we have to remove them before
2606 disconnecting. */
2607 remove_breakpoints ();
2608
2609 for (t = current_target.beneath; t != NULL; t = t->beneath)
2610 if (t->to_disconnect != NULL)
2611 {
2612 if (targetdebug)
2613 fprintf_unfiltered (gdb_stdlog, "target_disconnect (%s, %d)\n",
2614 args, from_tty);
2615 t->to_disconnect (t, args, from_tty);
2616 return;
2617 }
2618
2619 tcomplain ();
2620 }
2621
2622 ptid_t
2623 target_wait (ptid_t ptid, struct target_waitstatus *status, int options)
2624 {
2625 struct target_ops *t;
2626 ptid_t retval = (current_target.to_wait) (&current_target, ptid,
2627 status, options);
2628
2629 if (targetdebug)
2630 {
2631 char *status_string;
2632 char *options_string;
2633
2634 status_string = target_waitstatus_to_string (status);
2635 options_string = target_options_to_string (options);
2636 fprintf_unfiltered (gdb_stdlog,
2637 "target_wait (%d, status, options={%s})"
2638 " = %d, %s\n",
2639 ptid_get_pid (ptid), options_string,
2640 ptid_get_pid (retval), status_string);
2641 xfree (status_string);
2642 xfree (options_string);
2643 }
2644
2645 return retval;
2646 }
2647
2648 char *
2649 target_pid_to_str (ptid_t ptid)
2650 {
2651 struct target_ops *t;
2652
2653 for (t = current_target.beneath; t != NULL; t = t->beneath)
2654 {
2655 if (t->to_pid_to_str != NULL)
2656 return (*t->to_pid_to_str) (t, ptid);
2657 }
2658
2659 return normal_pid_to_str (ptid);
2660 }
2661
2662 char *
2663 target_thread_name (struct thread_info *info)
2664 {
2665 return current_target.to_thread_name (&current_target, info);
2666 }
2667
2668 void
2669 target_resume (ptid_t ptid, int step, enum gdb_signal signal)
2670 {
2671 struct target_ops *t;
2672
2673 target_dcache_invalidate ();
2674
2675 current_target.to_resume (&current_target, ptid, step, signal);
2676 if (targetdebug)
2677 fprintf_unfiltered (gdb_stdlog, "target_resume (%d, %s, %s)\n",
2678 ptid_get_pid (ptid),
2679 step ? "step" : "continue",
2680 gdb_signal_to_name (signal));
2681
2682 registers_changed_ptid (ptid);
2683 set_executing (ptid, 1);
2684 set_running (ptid, 1);
2685 clear_inline_frame_state (ptid);
2686 }
2687
2688 void
2689 target_pass_signals (int numsigs, unsigned char *pass_signals)
2690 {
2691 struct target_ops *t;
2692
2693 for (t = current_target.beneath; t != NULL; t = t->beneath)
2694 {
2695 if (t->to_pass_signals != NULL)
2696 {
2697 if (targetdebug)
2698 {
2699 int i;
2700
2701 fprintf_unfiltered (gdb_stdlog, "target_pass_signals (%d, {",
2702 numsigs);
2703
2704 for (i = 0; i < numsigs; i++)
2705 if (pass_signals[i])
2706 fprintf_unfiltered (gdb_stdlog, " %s",
2707 gdb_signal_to_name (i));
2708
2709 fprintf_unfiltered (gdb_stdlog, " })\n");
2710 }
2711
2712 (*t->to_pass_signals) (t, numsigs, pass_signals);
2713 return;
2714 }
2715 }
2716 }
2717
2718 void
2719 target_program_signals (int numsigs, unsigned char *program_signals)
2720 {
2721 struct target_ops *t;
2722
2723 for (t = current_target.beneath; t != NULL; t = t->beneath)
2724 {
2725 if (t->to_program_signals != NULL)
2726 {
2727 if (targetdebug)
2728 {
2729 int i;
2730
2731 fprintf_unfiltered (gdb_stdlog, "target_program_signals (%d, {",
2732 numsigs);
2733
2734 for (i = 0; i < numsigs; i++)
2735 if (program_signals[i])
2736 fprintf_unfiltered (gdb_stdlog, " %s",
2737 gdb_signal_to_name (i));
2738
2739 fprintf_unfiltered (gdb_stdlog, " })\n");
2740 }
2741
2742 (*t->to_program_signals) (t, numsigs, program_signals);
2743 return;
2744 }
2745 }
2746 }
2747
2748 /* Look through the list of possible targets for a target that can
2749 follow forks. */
2750
2751 int
2752 target_follow_fork (int follow_child, int detach_fork)
2753 {
2754 struct target_ops *t;
2755
2756 for (t = current_target.beneath; t != NULL; t = t->beneath)
2757 {
2758 if (t->to_follow_fork != NULL)
2759 {
2760 int retval = t->to_follow_fork (t, follow_child, detach_fork);
2761
2762 if (targetdebug)
2763 fprintf_unfiltered (gdb_stdlog,
2764 "target_follow_fork (%d, %d) = %d\n",
2765 follow_child, detach_fork, retval);
2766 return retval;
2767 }
2768 }
2769
2770 /* Some target returned a fork event, but did not know how to follow it. */
2771 internal_error (__FILE__, __LINE__,
2772 _("could not find a target to follow fork"));
2773 }
2774
2775 void
2776 target_mourn_inferior (void)
2777 {
2778 struct target_ops *t;
2779
2780 for (t = current_target.beneath; t != NULL; t = t->beneath)
2781 {
2782 if (t->to_mourn_inferior != NULL)
2783 {
2784 t->to_mourn_inferior (t);
2785 if (targetdebug)
2786 fprintf_unfiltered (gdb_stdlog, "target_mourn_inferior ()\n");
2787
2788 /* We no longer need to keep handles on any of the object files.
2789 Make sure to release them to avoid unnecessarily locking any
2790 of them while we're not actually debugging. */
2791 bfd_cache_close_all ();
2792
2793 return;
2794 }
2795 }
2796
2797 internal_error (__FILE__, __LINE__,
2798 _("could not find a target to follow mourn inferior"));
2799 }
2800
2801 /* Look for a target which can describe architectural features, starting
2802 from TARGET. If we find one, return its description. */
2803
2804 const struct target_desc *
2805 target_read_description (struct target_ops *target)
2806 {
2807 struct target_ops *t;
2808
2809 for (t = target; t != NULL; t = t->beneath)
2810 if (t->to_read_description != NULL)
2811 {
2812 const struct target_desc *tdesc;
2813
2814 tdesc = t->to_read_description (t);
2815 if (tdesc)
2816 return tdesc;
2817 }
2818
2819 return NULL;
2820 }
2821
2822 /* The default implementation of to_search_memory.
2823 This implements a basic search of memory, reading target memory and
2824 performing the search here (as opposed to performing the search in on the
2825 target side with, for example, gdbserver). */
2826
2827 int
2828 simple_search_memory (struct target_ops *ops,
2829 CORE_ADDR start_addr, ULONGEST search_space_len,
2830 const gdb_byte *pattern, ULONGEST pattern_len,
2831 CORE_ADDR *found_addrp)
2832 {
2833 /* NOTE: also defined in find.c testcase. */
2834 #define SEARCH_CHUNK_SIZE 16000
2835 const unsigned chunk_size = SEARCH_CHUNK_SIZE;
2836 /* Buffer to hold memory contents for searching. */
2837 gdb_byte *search_buf;
2838 unsigned search_buf_size;
2839 struct cleanup *old_cleanups;
2840
2841 search_buf_size = chunk_size + pattern_len - 1;
2842
2843 /* No point in trying to allocate a buffer larger than the search space. */
2844 if (search_space_len < search_buf_size)
2845 search_buf_size = search_space_len;
2846
2847 search_buf = malloc (search_buf_size);
2848 if (search_buf == NULL)
2849 error (_("Unable to allocate memory to perform the search."));
2850 old_cleanups = make_cleanup (free_current_contents, &search_buf);
2851
2852 /* Prime the search buffer. */
2853
2854 if (target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2855 search_buf, start_addr, search_buf_size) != search_buf_size)
2856 {
2857 warning (_("Unable to access %s bytes of target "
2858 "memory at %s, halting search."),
2859 pulongest (search_buf_size), hex_string (start_addr));
2860 do_cleanups (old_cleanups);
2861 return -1;
2862 }
2863
2864 /* Perform the search.
2865
2866 The loop is kept simple by allocating [N + pattern-length - 1] bytes.
2867 When we've scanned N bytes we copy the trailing bytes to the start and
2868 read in another N bytes. */
2869
2870 while (search_space_len >= pattern_len)
2871 {
2872 gdb_byte *found_ptr;
2873 unsigned nr_search_bytes = min (search_space_len, search_buf_size);
2874
2875 found_ptr = memmem (search_buf, nr_search_bytes,
2876 pattern, pattern_len);
2877
2878 if (found_ptr != NULL)
2879 {
2880 CORE_ADDR found_addr = start_addr + (found_ptr - search_buf);
2881
2882 *found_addrp = found_addr;
2883 do_cleanups (old_cleanups);
2884 return 1;
2885 }
2886
2887 /* Not found in this chunk, skip to next chunk. */
2888
2889 /* Don't let search_space_len wrap here, it's unsigned. */
2890 if (search_space_len >= chunk_size)
2891 search_space_len -= chunk_size;
2892 else
2893 search_space_len = 0;
2894
2895 if (search_space_len >= pattern_len)
2896 {
2897 unsigned keep_len = search_buf_size - chunk_size;
2898 CORE_ADDR read_addr = start_addr + chunk_size + keep_len;
2899 int nr_to_read;
2900
2901 /* Copy the trailing part of the previous iteration to the front
2902 of the buffer for the next iteration. */
2903 gdb_assert (keep_len == pattern_len - 1);
2904 memcpy (search_buf, search_buf + chunk_size, keep_len);
2905
2906 nr_to_read = min (search_space_len - keep_len, chunk_size);
2907
2908 if (target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2909 search_buf + keep_len, read_addr,
2910 nr_to_read) != nr_to_read)
2911 {
2912 warning (_("Unable to access %s bytes of target "
2913 "memory at %s, halting search."),
2914 plongest (nr_to_read),
2915 hex_string (read_addr));
2916 do_cleanups (old_cleanups);
2917 return -1;
2918 }
2919
2920 start_addr += chunk_size;
2921 }
2922 }
2923
2924 /* Not found. */
2925
2926 do_cleanups (old_cleanups);
2927 return 0;
2928 }
2929
2930 /* Search SEARCH_SPACE_LEN bytes beginning at START_ADDR for the
2931 sequence of bytes in PATTERN with length PATTERN_LEN.
2932
2933 The result is 1 if found, 0 if not found, and -1 if there was an error
2934 requiring halting of the search (e.g. memory read error).
2935 If the pattern is found the address is recorded in FOUND_ADDRP. */
2936
2937 int
2938 target_search_memory (CORE_ADDR start_addr, ULONGEST search_space_len,
2939 const gdb_byte *pattern, ULONGEST pattern_len,
2940 CORE_ADDR *found_addrp)
2941 {
2942 struct target_ops *t;
2943 int found;
2944
2945 /* We don't use INHERIT to set current_target.to_search_memory,
2946 so we have to scan the target stack and handle targetdebug
2947 ourselves. */
2948
2949 if (targetdebug)
2950 fprintf_unfiltered (gdb_stdlog, "target_search_memory (%s, ...)\n",
2951 hex_string (start_addr));
2952
2953 for (t = current_target.beneath; t != NULL; t = t->beneath)
2954 if (t->to_search_memory != NULL)
2955 break;
2956
2957 if (t != NULL)
2958 {
2959 found = t->to_search_memory (t, start_addr, search_space_len,
2960 pattern, pattern_len, found_addrp);
2961 }
2962 else
2963 {
2964 /* If a special version of to_search_memory isn't available, use the
2965 simple version. */
2966 found = simple_search_memory (current_target.beneath,
2967 start_addr, search_space_len,
2968 pattern, pattern_len, found_addrp);
2969 }
2970
2971 if (targetdebug)
2972 fprintf_unfiltered (gdb_stdlog, " = %d\n", found);
2973
2974 return found;
2975 }
2976
2977 /* Look through the currently pushed targets. If none of them will
2978 be able to restart the currently running process, issue an error
2979 message. */
2980
2981 void
2982 target_require_runnable (void)
2983 {
2984 struct target_ops *t;
2985
2986 for (t = target_stack; t != NULL; t = t->beneath)
2987 {
2988 /* If this target knows how to create a new program, then
2989 assume we will still be able to after killing the current
2990 one. Either killing and mourning will not pop T, or else
2991 find_default_run_target will find it again. */
2992 if (t->to_create_inferior != NULL)
2993 return;
2994
2995 /* Do not worry about thread_stratum targets that can not
2996 create inferiors. Assume they will be pushed again if
2997 necessary, and continue to the process_stratum. */
2998 if (t->to_stratum == thread_stratum
2999 || t->to_stratum == arch_stratum)
3000 continue;
3001
3002 error (_("The \"%s\" target does not support \"run\". "
3003 "Try \"help target\" or \"continue\"."),
3004 t->to_shortname);
3005 }
3006
3007 /* This function is only called if the target is running. In that
3008 case there should have been a process_stratum target and it
3009 should either know how to create inferiors, or not... */
3010 internal_error (__FILE__, __LINE__, _("No targets found"));
3011 }
3012
3013 /* Look through the list of possible targets for a target that can
3014 execute a run or attach command without any other data. This is
3015 used to locate the default process stratum.
3016
3017 If DO_MESG is not NULL, the result is always valid (error() is
3018 called for errors); else, return NULL on error. */
3019
3020 static struct target_ops *
3021 find_default_run_target (char *do_mesg)
3022 {
3023 struct target_ops **t;
3024 struct target_ops *runable = NULL;
3025 int count;
3026
3027 count = 0;
3028
3029 for (t = target_structs; t < target_structs + target_struct_size;
3030 ++t)
3031 {
3032 if ((*t)->to_can_run && target_can_run (*t))
3033 {
3034 runable = *t;
3035 ++count;
3036 }
3037 }
3038
3039 if (count != 1)
3040 {
3041 if (do_mesg)
3042 error (_("Don't know how to %s. Try \"help target\"."), do_mesg);
3043 else
3044 return NULL;
3045 }
3046
3047 return runable;
3048 }
3049
3050 void
3051 find_default_attach (struct target_ops *ops, char *args, int from_tty)
3052 {
3053 struct target_ops *t;
3054
3055 t = find_default_run_target ("attach");
3056 (t->to_attach) (t, args, from_tty);
3057 return;
3058 }
3059
3060 void
3061 find_default_create_inferior (struct target_ops *ops,
3062 char *exec_file, char *allargs, char **env,
3063 int from_tty)
3064 {
3065 struct target_ops *t;
3066
3067 t = find_default_run_target ("run");
3068 (t->to_create_inferior) (t, exec_file, allargs, env, from_tty);
3069 return;
3070 }
3071
3072 static int
3073 find_default_can_async_p (struct target_ops *ignore)
3074 {
3075 struct target_ops *t;
3076
3077 /* This may be called before the target is pushed on the stack;
3078 look for the default process stratum. If there's none, gdb isn't
3079 configured with a native debugger, and target remote isn't
3080 connected yet. */
3081 t = find_default_run_target (NULL);
3082 if (t && t->to_can_async_p != delegate_can_async_p)
3083 return (t->to_can_async_p) (t);
3084 return 0;
3085 }
3086
3087 static int
3088 find_default_is_async_p (struct target_ops *ignore)
3089 {
3090 struct target_ops *t;
3091
3092 /* This may be called before the target is pushed on the stack;
3093 look for the default process stratum. If there's none, gdb isn't
3094 configured with a native debugger, and target remote isn't
3095 connected yet. */
3096 t = find_default_run_target (NULL);
3097 if (t && t->to_is_async_p != delegate_is_async_p)
3098 return (t->to_is_async_p) (t);
3099 return 0;
3100 }
3101
3102 static int
3103 find_default_supports_non_stop (struct target_ops *self)
3104 {
3105 struct target_ops *t;
3106
3107 t = find_default_run_target (NULL);
3108 if (t && t->to_supports_non_stop)
3109 return (t->to_supports_non_stop) (t);
3110 return 0;
3111 }
3112
3113 int
3114 target_supports_non_stop (void)
3115 {
3116 struct target_ops *t;
3117
3118 for (t = &current_target; t != NULL; t = t->beneath)
3119 if (t->to_supports_non_stop)
3120 return t->to_supports_non_stop (t);
3121
3122 return 0;
3123 }
3124
3125 /* Implement the "info proc" command. */
3126
3127 int
3128 target_info_proc (char *args, enum info_proc_what what)
3129 {
3130 struct target_ops *t;
3131
3132 /* If we're already connected to something that can get us OS
3133 related data, use it. Otherwise, try using the native
3134 target. */
3135 if (current_target.to_stratum >= process_stratum)
3136 t = current_target.beneath;
3137 else
3138 t = find_default_run_target (NULL);
3139
3140 for (; t != NULL; t = t->beneath)
3141 {
3142 if (t->to_info_proc != NULL)
3143 {
3144 t->to_info_proc (t, args, what);
3145
3146 if (targetdebug)
3147 fprintf_unfiltered (gdb_stdlog,
3148 "target_info_proc (\"%s\", %d)\n", args, what);
3149
3150 return 1;
3151 }
3152 }
3153
3154 return 0;
3155 }
3156
3157 static int
3158 find_default_supports_disable_randomization (struct target_ops *self)
3159 {
3160 struct target_ops *t;
3161
3162 t = find_default_run_target (NULL);
3163 if (t && t->to_supports_disable_randomization)
3164 return (t->to_supports_disable_randomization) (t);
3165 return 0;
3166 }
3167
3168 int
3169 target_supports_disable_randomization (void)
3170 {
3171 struct target_ops *t;
3172
3173 for (t = &current_target; t != NULL; t = t->beneath)
3174 if (t->to_supports_disable_randomization)
3175 return t->to_supports_disable_randomization (t);
3176
3177 return 0;
3178 }
3179
3180 char *
3181 target_get_osdata (const char *type)
3182 {
3183 struct target_ops *t;
3184
3185 /* If we're already connected to something that can get us OS
3186 related data, use it. Otherwise, try using the native
3187 target. */
3188 if (current_target.to_stratum >= process_stratum)
3189 t = current_target.beneath;
3190 else
3191 t = find_default_run_target ("get OS data");
3192
3193 if (!t)
3194 return NULL;
3195
3196 return target_read_stralloc (t, TARGET_OBJECT_OSDATA, type);
3197 }
3198
3199 /* Determine the current address space of thread PTID. */
3200
3201 struct address_space *
3202 target_thread_address_space (ptid_t ptid)
3203 {
3204 struct address_space *aspace;
3205 struct inferior *inf;
3206 struct target_ops *t;
3207
3208 for (t = current_target.beneath; t != NULL; t = t->beneath)
3209 {
3210 if (t->to_thread_address_space != NULL)
3211 {
3212 aspace = t->to_thread_address_space (t, ptid);
3213 gdb_assert (aspace);
3214
3215 if (targetdebug)
3216 fprintf_unfiltered (gdb_stdlog,
3217 "target_thread_address_space (%s) = %d\n",
3218 target_pid_to_str (ptid),
3219 address_space_num (aspace));
3220 return aspace;
3221 }
3222 }
3223
3224 /* Fall-back to the "main" address space of the inferior. */
3225 inf = find_inferior_pid (ptid_get_pid (ptid));
3226
3227 if (inf == NULL || inf->aspace == NULL)
3228 internal_error (__FILE__, __LINE__,
3229 _("Can't determine the current "
3230 "address space of thread %s\n"),
3231 target_pid_to_str (ptid));
3232
3233 return inf->aspace;
3234 }
3235
3236
3237 /* Target file operations. */
3238
3239 static struct target_ops *
3240 default_fileio_target (void)
3241 {
3242 /* If we're already connected to something that can perform
3243 file I/O, use it. Otherwise, try using the native target. */
3244 if (current_target.to_stratum >= process_stratum)
3245 return current_target.beneath;
3246 else
3247 return find_default_run_target ("file I/O");
3248 }
3249
3250 /* Open FILENAME on the target, using FLAGS and MODE. Return a
3251 target file descriptor, or -1 if an error occurs (and set
3252 *TARGET_ERRNO). */
3253 int
3254 target_fileio_open (const char *filename, int flags, int mode,
3255 int *target_errno)
3256 {
3257 struct target_ops *t;
3258
3259 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3260 {
3261 if (t->to_fileio_open != NULL)
3262 {
3263 int fd = t->to_fileio_open (t, filename, flags, mode, target_errno);
3264
3265 if (targetdebug)
3266 fprintf_unfiltered (gdb_stdlog,
3267 "target_fileio_open (%s,0x%x,0%o) = %d (%d)\n",
3268 filename, flags, mode,
3269 fd, fd != -1 ? 0 : *target_errno);
3270 return fd;
3271 }
3272 }
3273
3274 *target_errno = FILEIO_ENOSYS;
3275 return -1;
3276 }
3277
3278 /* Write up to LEN bytes from WRITE_BUF to FD on the target.
3279 Return the number of bytes written, or -1 if an error occurs
3280 (and set *TARGET_ERRNO). */
3281 int
3282 target_fileio_pwrite (int fd, const gdb_byte *write_buf, int len,
3283 ULONGEST offset, int *target_errno)
3284 {
3285 struct target_ops *t;
3286
3287 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3288 {
3289 if (t->to_fileio_pwrite != NULL)
3290 {
3291 int ret = t->to_fileio_pwrite (t, fd, write_buf, len, offset,
3292 target_errno);
3293
3294 if (targetdebug)
3295 fprintf_unfiltered (gdb_stdlog,
3296 "target_fileio_pwrite (%d,...,%d,%s) "
3297 "= %d (%d)\n",
3298 fd, len, pulongest (offset),
3299 ret, ret != -1 ? 0 : *target_errno);
3300 return ret;
3301 }
3302 }
3303
3304 *target_errno = FILEIO_ENOSYS;
3305 return -1;
3306 }
3307
3308 /* Read up to LEN bytes FD on the target into READ_BUF.
3309 Return the number of bytes read, or -1 if an error occurs
3310 (and set *TARGET_ERRNO). */
3311 int
3312 target_fileio_pread (int fd, gdb_byte *read_buf, int len,
3313 ULONGEST offset, int *target_errno)
3314 {
3315 struct target_ops *t;
3316
3317 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3318 {
3319 if (t->to_fileio_pread != NULL)
3320 {
3321 int ret = t->to_fileio_pread (t, fd, read_buf, len, offset,
3322 target_errno);
3323
3324 if (targetdebug)
3325 fprintf_unfiltered (gdb_stdlog,
3326 "target_fileio_pread (%d,...,%d,%s) "
3327 "= %d (%d)\n",
3328 fd, len, pulongest (offset),
3329 ret, ret != -1 ? 0 : *target_errno);
3330 return ret;
3331 }
3332 }
3333
3334 *target_errno = FILEIO_ENOSYS;
3335 return -1;
3336 }
3337
3338 /* Close FD on the target. Return 0, or -1 if an error occurs
3339 (and set *TARGET_ERRNO). */
3340 int
3341 target_fileio_close (int fd, int *target_errno)
3342 {
3343 struct target_ops *t;
3344
3345 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3346 {
3347 if (t->to_fileio_close != NULL)
3348 {
3349 int ret = t->to_fileio_close (t, fd, target_errno);
3350
3351 if (targetdebug)
3352 fprintf_unfiltered (gdb_stdlog,
3353 "target_fileio_close (%d) = %d (%d)\n",
3354 fd, ret, ret != -1 ? 0 : *target_errno);
3355 return ret;
3356 }
3357 }
3358
3359 *target_errno = FILEIO_ENOSYS;
3360 return -1;
3361 }
3362
3363 /* Unlink FILENAME on the target. Return 0, or -1 if an error
3364 occurs (and set *TARGET_ERRNO). */
3365 int
3366 target_fileio_unlink (const char *filename, int *target_errno)
3367 {
3368 struct target_ops *t;
3369
3370 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3371 {
3372 if (t->to_fileio_unlink != NULL)
3373 {
3374 int ret = t->to_fileio_unlink (t, filename, target_errno);
3375
3376 if (targetdebug)
3377 fprintf_unfiltered (gdb_stdlog,
3378 "target_fileio_unlink (%s) = %d (%d)\n",
3379 filename, ret, ret != -1 ? 0 : *target_errno);
3380 return ret;
3381 }
3382 }
3383
3384 *target_errno = FILEIO_ENOSYS;
3385 return -1;
3386 }
3387
3388 /* Read value of symbolic link FILENAME on the target. Return a
3389 null-terminated string allocated via xmalloc, or NULL if an error
3390 occurs (and set *TARGET_ERRNO). */
3391 char *
3392 target_fileio_readlink (const char *filename, int *target_errno)
3393 {
3394 struct target_ops *t;
3395
3396 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3397 {
3398 if (t->to_fileio_readlink != NULL)
3399 {
3400 char *ret = t->to_fileio_readlink (t, filename, target_errno);
3401
3402 if (targetdebug)
3403 fprintf_unfiltered (gdb_stdlog,
3404 "target_fileio_readlink (%s) = %s (%d)\n",
3405 filename, ret? ret : "(nil)",
3406 ret? 0 : *target_errno);
3407 return ret;
3408 }
3409 }
3410
3411 *target_errno = FILEIO_ENOSYS;
3412 return NULL;
3413 }
3414
3415 static void
3416 target_fileio_close_cleanup (void *opaque)
3417 {
3418 int fd = *(int *) opaque;
3419 int target_errno;
3420
3421 target_fileio_close (fd, &target_errno);
3422 }
3423
3424 /* Read target file FILENAME. Store the result in *BUF_P and
3425 return the size of the transferred data. PADDING additional bytes are
3426 available in *BUF_P. This is a helper function for
3427 target_fileio_read_alloc; see the declaration of that function for more
3428 information. */
3429
3430 static LONGEST
3431 target_fileio_read_alloc_1 (const char *filename,
3432 gdb_byte **buf_p, int padding)
3433 {
3434 struct cleanup *close_cleanup;
3435 size_t buf_alloc, buf_pos;
3436 gdb_byte *buf;
3437 LONGEST n;
3438 int fd;
3439 int target_errno;
3440
3441 fd = target_fileio_open (filename, FILEIO_O_RDONLY, 0700, &target_errno);
3442 if (fd == -1)
3443 return -1;
3444
3445 close_cleanup = make_cleanup (target_fileio_close_cleanup, &fd);
3446
3447 /* Start by reading up to 4K at a time. The target will throttle
3448 this number down if necessary. */
3449 buf_alloc = 4096;
3450 buf = xmalloc (buf_alloc);
3451 buf_pos = 0;
3452 while (1)
3453 {
3454 n = target_fileio_pread (fd, &buf[buf_pos],
3455 buf_alloc - buf_pos - padding, buf_pos,
3456 &target_errno);
3457 if (n < 0)
3458 {
3459 /* An error occurred. */
3460 do_cleanups (close_cleanup);
3461 xfree (buf);
3462 return -1;
3463 }
3464 else if (n == 0)
3465 {
3466 /* Read all there was. */
3467 do_cleanups (close_cleanup);
3468 if (buf_pos == 0)
3469 xfree (buf);
3470 else
3471 *buf_p = buf;
3472 return buf_pos;
3473 }
3474
3475 buf_pos += n;
3476
3477 /* If the buffer is filling up, expand it. */
3478 if (buf_alloc < buf_pos * 2)
3479 {
3480 buf_alloc *= 2;
3481 buf = xrealloc (buf, buf_alloc);
3482 }
3483
3484 QUIT;
3485 }
3486 }
3487
3488 /* Read target file FILENAME. Store the result in *BUF_P and return
3489 the size of the transferred data. See the declaration in "target.h"
3490 function for more information about the return value. */
3491
3492 LONGEST
3493 target_fileio_read_alloc (const char *filename, gdb_byte **buf_p)
3494 {
3495 return target_fileio_read_alloc_1 (filename, buf_p, 0);
3496 }
3497
3498 /* Read target file FILENAME. The result is NUL-terminated and
3499 returned as a string, allocated using xmalloc. If an error occurs
3500 or the transfer is unsupported, NULL is returned. Empty objects
3501 are returned as allocated but empty strings. A warning is issued
3502 if the result contains any embedded NUL bytes. */
3503
3504 char *
3505 target_fileio_read_stralloc (const char *filename)
3506 {
3507 gdb_byte *buffer;
3508 char *bufstr;
3509 LONGEST i, transferred;
3510
3511 transferred = target_fileio_read_alloc_1 (filename, &buffer, 1);
3512 bufstr = (char *) buffer;
3513
3514 if (transferred < 0)
3515 return NULL;
3516
3517 if (transferred == 0)
3518 return xstrdup ("");
3519
3520 bufstr[transferred] = 0;
3521
3522 /* Check for embedded NUL bytes; but allow trailing NULs. */
3523 for (i = strlen (bufstr); i < transferred; i++)
3524 if (bufstr[i] != 0)
3525 {
3526 warning (_("target file %s "
3527 "contained unexpected null characters"),
3528 filename);
3529 break;
3530 }
3531
3532 return bufstr;
3533 }
3534
3535
3536 static int
3537 default_region_ok_for_hw_watchpoint (struct target_ops *self,
3538 CORE_ADDR addr, int len)
3539 {
3540 return (len <= gdbarch_ptr_bit (target_gdbarch ()) / TARGET_CHAR_BIT);
3541 }
3542
3543 static int
3544 default_watchpoint_addr_within_range (struct target_ops *target,
3545 CORE_ADDR addr,
3546 CORE_ADDR start, int length)
3547 {
3548 return addr >= start && addr < start + length;
3549 }
3550
3551 static struct gdbarch *
3552 default_thread_architecture (struct target_ops *ops, ptid_t ptid)
3553 {
3554 return target_gdbarch ();
3555 }
3556
3557 static int
3558 return_zero (void)
3559 {
3560 return 0;
3561 }
3562
3563 static int
3564 return_minus_one (void)
3565 {
3566 return -1;
3567 }
3568
3569 static void *
3570 return_null (void)
3571 {
3572 return 0;
3573 }
3574
3575 /*
3576 * Find the next target down the stack from the specified target.
3577 */
3578
3579 struct target_ops *
3580 find_target_beneath (struct target_ops *t)
3581 {
3582 return t->beneath;
3583 }
3584
3585 /* See target.h. */
3586
3587 struct target_ops *
3588 find_target_at (enum strata stratum)
3589 {
3590 struct target_ops *t;
3591
3592 for (t = current_target.beneath; t != NULL; t = t->beneath)
3593 if (t->to_stratum == stratum)
3594 return t;
3595
3596 return NULL;
3597 }
3598
3599 \f
3600 /* The inferior process has died. Long live the inferior! */
3601
3602 void
3603 generic_mourn_inferior (void)
3604 {
3605 ptid_t ptid;
3606
3607 ptid = inferior_ptid;
3608 inferior_ptid = null_ptid;
3609
3610 /* Mark breakpoints uninserted in case something tries to delete a
3611 breakpoint while we delete the inferior's threads (which would
3612 fail, since the inferior is long gone). */
3613 mark_breakpoints_out ();
3614
3615 if (!ptid_equal (ptid, null_ptid))
3616 {
3617 int pid = ptid_get_pid (ptid);
3618 exit_inferior (pid);
3619 }
3620
3621 /* Note this wipes step-resume breakpoints, so needs to be done
3622 after exit_inferior, which ends up referencing the step-resume
3623 breakpoints through clear_thread_inferior_resources. */
3624 breakpoint_init_inferior (inf_exited);
3625
3626 registers_changed ();
3627
3628 reopen_exec_file ();
3629 reinit_frame_cache ();
3630
3631 if (deprecated_detach_hook)
3632 deprecated_detach_hook ();
3633 }
3634 \f
3635 /* Convert a normal process ID to a string. Returns the string in a
3636 static buffer. */
3637
3638 char *
3639 normal_pid_to_str (ptid_t ptid)
3640 {
3641 static char buf[32];
3642
3643 xsnprintf (buf, sizeof buf, "process %d", ptid_get_pid (ptid));
3644 return buf;
3645 }
3646
3647 static char *
3648 dummy_pid_to_str (struct target_ops *ops, ptid_t ptid)
3649 {
3650 return normal_pid_to_str (ptid);
3651 }
3652
3653 /* Error-catcher for target_find_memory_regions. */
3654 static int
3655 dummy_find_memory_regions (struct target_ops *self,
3656 find_memory_region_ftype ignore1, void *ignore2)
3657 {
3658 error (_("Command not implemented for this target."));
3659 return 0;
3660 }
3661
3662 /* Error-catcher for target_make_corefile_notes. */
3663 static char *
3664 dummy_make_corefile_notes (struct target_ops *self,
3665 bfd *ignore1, int *ignore2)
3666 {
3667 error (_("Command not implemented for this target."));
3668 return NULL;
3669 }
3670
3671 /* Set up the handful of non-empty slots needed by the dummy target
3672 vector. */
3673
3674 static void
3675 init_dummy_target (void)
3676 {
3677 dummy_target.to_shortname = "None";
3678 dummy_target.to_longname = "None";
3679 dummy_target.to_doc = "";
3680 dummy_target.to_create_inferior = find_default_create_inferior;
3681 dummy_target.to_supports_non_stop = find_default_supports_non_stop;
3682 dummy_target.to_supports_disable_randomization
3683 = find_default_supports_disable_randomization;
3684 dummy_target.to_pid_to_str = dummy_pid_to_str;
3685 dummy_target.to_stratum = dummy_stratum;
3686 dummy_target.to_has_all_memory = (int (*) (struct target_ops *)) return_zero;
3687 dummy_target.to_has_memory = (int (*) (struct target_ops *)) return_zero;
3688 dummy_target.to_has_stack = (int (*) (struct target_ops *)) return_zero;
3689 dummy_target.to_has_registers = (int (*) (struct target_ops *)) return_zero;
3690 dummy_target.to_has_execution
3691 = (int (*) (struct target_ops *, ptid_t)) return_zero;
3692 dummy_target.to_magic = OPS_MAGIC;
3693
3694 install_dummy_methods (&dummy_target);
3695 }
3696 \f
3697 static void
3698 debug_to_open (char *args, int from_tty)
3699 {
3700 debug_target.to_open (args, from_tty);
3701
3702 fprintf_unfiltered (gdb_stdlog, "target_open (%s, %d)\n", args, from_tty);
3703 }
3704
3705 void
3706 target_close (struct target_ops *targ)
3707 {
3708 gdb_assert (!target_is_pushed (targ));
3709
3710 if (targ->to_xclose != NULL)
3711 targ->to_xclose (targ);
3712 else if (targ->to_close != NULL)
3713 targ->to_close (targ);
3714
3715 if (targetdebug)
3716 fprintf_unfiltered (gdb_stdlog, "target_close ()\n");
3717 }
3718
3719 void
3720 target_attach (char *args, int from_tty)
3721 {
3722 current_target.to_attach (&current_target, args, from_tty);
3723 if (targetdebug)
3724 fprintf_unfiltered (gdb_stdlog, "target_attach (%s, %d)\n",
3725 args, from_tty);
3726 }
3727
3728 int
3729 target_thread_alive (ptid_t ptid)
3730 {
3731 struct target_ops *t;
3732
3733 for (t = current_target.beneath; t != NULL; t = t->beneath)
3734 {
3735 if (t->to_thread_alive != NULL)
3736 {
3737 int retval;
3738
3739 retval = t->to_thread_alive (t, ptid);
3740 if (targetdebug)
3741 fprintf_unfiltered (gdb_stdlog, "target_thread_alive (%d) = %d\n",
3742 ptid_get_pid (ptid), retval);
3743
3744 return retval;
3745 }
3746 }
3747
3748 return 0;
3749 }
3750
3751 void
3752 target_find_new_threads (void)
3753 {
3754 struct target_ops *t;
3755
3756 for (t = current_target.beneath; t != NULL; t = t->beneath)
3757 {
3758 if (t->to_find_new_threads != NULL)
3759 {
3760 t->to_find_new_threads (t);
3761 if (targetdebug)
3762 fprintf_unfiltered (gdb_stdlog, "target_find_new_threads ()\n");
3763
3764 return;
3765 }
3766 }
3767 }
3768
3769 void
3770 target_stop (ptid_t ptid)
3771 {
3772 if (!may_stop)
3773 {
3774 warning (_("May not interrupt or stop the target, ignoring attempt"));
3775 return;
3776 }
3777
3778 (*current_target.to_stop) (&current_target, ptid);
3779 }
3780
3781 static void
3782 debug_to_post_attach (struct target_ops *self, int pid)
3783 {
3784 debug_target.to_post_attach (&debug_target, pid);
3785
3786 fprintf_unfiltered (gdb_stdlog, "target_post_attach (%d)\n", pid);
3787 }
3788
3789 /* Concatenate ELEM to LIST, a comma separate list, and return the
3790 result. The LIST incoming argument is released. */
3791
3792 static char *
3793 str_comma_list_concat_elem (char *list, const char *elem)
3794 {
3795 if (list == NULL)
3796 return xstrdup (elem);
3797 else
3798 return reconcat (list, list, ", ", elem, (char *) NULL);
3799 }
3800
3801 /* Helper for target_options_to_string. If OPT is present in
3802 TARGET_OPTIONS, append the OPT_STR (string version of OPT) in RET.
3803 Returns the new resulting string. OPT is removed from
3804 TARGET_OPTIONS. */
3805
3806 static char *
3807 do_option (int *target_options, char *ret,
3808 int opt, char *opt_str)
3809 {
3810 if ((*target_options & opt) != 0)
3811 {
3812 ret = str_comma_list_concat_elem (ret, opt_str);
3813 *target_options &= ~opt;
3814 }
3815
3816 return ret;
3817 }
3818
3819 char *
3820 target_options_to_string (int target_options)
3821 {
3822 char *ret = NULL;
3823
3824 #define DO_TARG_OPTION(OPT) \
3825 ret = do_option (&target_options, ret, OPT, #OPT)
3826
3827 DO_TARG_OPTION (TARGET_WNOHANG);
3828
3829 if (target_options != 0)
3830 ret = str_comma_list_concat_elem (ret, "unknown???");
3831
3832 if (ret == NULL)
3833 ret = xstrdup ("");
3834 return ret;
3835 }
3836
3837 static void
3838 debug_print_register (const char * func,
3839 struct regcache *regcache, int regno)
3840 {
3841 struct gdbarch *gdbarch = get_regcache_arch (regcache);
3842
3843 fprintf_unfiltered (gdb_stdlog, "%s ", func);
3844 if (regno >= 0 && regno < gdbarch_num_regs (gdbarch)
3845 && gdbarch_register_name (gdbarch, regno) != NULL
3846 && gdbarch_register_name (gdbarch, regno)[0] != '\0')
3847 fprintf_unfiltered (gdb_stdlog, "(%s)",
3848 gdbarch_register_name (gdbarch, regno));
3849 else
3850 fprintf_unfiltered (gdb_stdlog, "(%d)", regno);
3851 if (regno >= 0 && regno < gdbarch_num_regs (gdbarch))
3852 {
3853 enum bfd_endian byte_order = gdbarch_byte_order (gdbarch);
3854 int i, size = register_size (gdbarch, regno);
3855 gdb_byte buf[MAX_REGISTER_SIZE];
3856
3857 regcache_raw_collect (regcache, regno, buf);
3858 fprintf_unfiltered (gdb_stdlog, " = ");
3859 for (i = 0; i < size; i++)
3860 {
3861 fprintf_unfiltered (gdb_stdlog, "%02x", buf[i]);
3862 }
3863 if (size <= sizeof (LONGEST))
3864 {
3865 ULONGEST val = extract_unsigned_integer (buf, size, byte_order);
3866
3867 fprintf_unfiltered (gdb_stdlog, " %s %s",
3868 core_addr_to_string_nz (val), plongest (val));
3869 }
3870 }
3871 fprintf_unfiltered (gdb_stdlog, "\n");
3872 }
3873
3874 void
3875 target_fetch_registers (struct regcache *regcache, int regno)
3876 {
3877 struct target_ops *t;
3878
3879 for (t = current_target.beneath; t != NULL; t = t->beneath)
3880 {
3881 if (t->to_fetch_registers != NULL)
3882 {
3883 t->to_fetch_registers (t, regcache, regno);
3884 if (targetdebug)
3885 debug_print_register ("target_fetch_registers", regcache, regno);
3886 return;
3887 }
3888 }
3889 }
3890
3891 void
3892 target_store_registers (struct regcache *regcache, int regno)
3893 {
3894 struct target_ops *t;
3895
3896 if (!may_write_registers)
3897 error (_("Writing to registers is not allowed (regno %d)"), regno);
3898
3899 current_target.to_store_registers (&current_target, regcache, regno);
3900 if (targetdebug)
3901 {
3902 debug_print_register ("target_store_registers", regcache, regno);
3903 }
3904 }
3905
3906 int
3907 target_core_of_thread (ptid_t ptid)
3908 {
3909 struct target_ops *t;
3910
3911 for (t = current_target.beneath; t != NULL; t = t->beneath)
3912 {
3913 if (t->to_core_of_thread != NULL)
3914 {
3915 int retval = t->to_core_of_thread (t, ptid);
3916
3917 if (targetdebug)
3918 fprintf_unfiltered (gdb_stdlog,
3919 "target_core_of_thread (%d) = %d\n",
3920 ptid_get_pid (ptid), retval);
3921 return retval;
3922 }
3923 }
3924
3925 return -1;
3926 }
3927
3928 int
3929 target_verify_memory (const gdb_byte *data, CORE_ADDR memaddr, ULONGEST size)
3930 {
3931 struct target_ops *t;
3932
3933 for (t = current_target.beneath; t != NULL; t = t->beneath)
3934 {
3935 if (t->to_verify_memory != NULL)
3936 {
3937 int retval = t->to_verify_memory (t, data, memaddr, size);
3938
3939 if (targetdebug)
3940 fprintf_unfiltered (gdb_stdlog,
3941 "target_verify_memory (%s, %s) = %d\n",
3942 paddress (target_gdbarch (), memaddr),
3943 pulongest (size),
3944 retval);
3945 return retval;
3946 }
3947 }
3948
3949 tcomplain ();
3950 }
3951
3952 /* The documentation for this function is in its prototype declaration in
3953 target.h. */
3954
3955 int
3956 target_insert_mask_watchpoint (CORE_ADDR addr, CORE_ADDR mask, int rw)
3957 {
3958 struct target_ops *t;
3959
3960 for (t = current_target.beneath; t != NULL; t = t->beneath)
3961 if (t->to_insert_mask_watchpoint != NULL)
3962 {
3963 int ret;
3964
3965 ret = t->to_insert_mask_watchpoint (t, addr, mask, rw);
3966
3967 if (targetdebug)
3968 fprintf_unfiltered (gdb_stdlog, "\
3969 target_insert_mask_watchpoint (%s, %s, %d) = %d\n",
3970 core_addr_to_string (addr),
3971 core_addr_to_string (mask), rw, ret);
3972
3973 return ret;
3974 }
3975
3976 return 1;
3977 }
3978
3979 /* The documentation for this function is in its prototype declaration in
3980 target.h. */
3981
3982 int
3983 target_remove_mask_watchpoint (CORE_ADDR addr, CORE_ADDR mask, int rw)
3984 {
3985 struct target_ops *t;
3986
3987 for (t = current_target.beneath; t != NULL; t = t->beneath)
3988 if (t->to_remove_mask_watchpoint != NULL)
3989 {
3990 int ret;
3991
3992 ret = t->to_remove_mask_watchpoint (t, addr, mask, rw);
3993
3994 if (targetdebug)
3995 fprintf_unfiltered (gdb_stdlog, "\
3996 target_remove_mask_watchpoint (%s, %s, %d) = %d\n",
3997 core_addr_to_string (addr),
3998 core_addr_to_string (mask), rw, ret);
3999
4000 return ret;
4001 }
4002
4003 return 1;
4004 }
4005
4006 /* The documentation for this function is in its prototype declaration
4007 in target.h. */
4008
4009 int
4010 target_masked_watch_num_registers (CORE_ADDR addr, CORE_ADDR mask)
4011 {
4012 struct target_ops *t;
4013
4014 for (t = current_target.beneath; t != NULL; t = t->beneath)
4015 if (t->to_masked_watch_num_registers != NULL)
4016 return t->to_masked_watch_num_registers (t, addr, mask);
4017
4018 return -1;
4019 }
4020
4021 /* The documentation for this function is in its prototype declaration
4022 in target.h. */
4023
4024 int
4025 target_ranged_break_num_registers (void)
4026 {
4027 struct target_ops *t;
4028
4029 for (t = current_target.beneath; t != NULL; t = t->beneath)
4030 if (t->to_ranged_break_num_registers != NULL)
4031 return t->to_ranged_break_num_registers (t);
4032
4033 return -1;
4034 }
4035
4036 /* See target.h. */
4037
4038 struct btrace_target_info *
4039 target_enable_btrace (ptid_t ptid)
4040 {
4041 struct target_ops *t;
4042
4043 for (t = current_target.beneath; t != NULL; t = t->beneath)
4044 if (t->to_enable_btrace != NULL)
4045 return t->to_enable_btrace (t, ptid);
4046
4047 tcomplain ();
4048 return NULL;
4049 }
4050
4051 /* See target.h. */
4052
4053 void
4054 target_disable_btrace (struct btrace_target_info *btinfo)
4055 {
4056 struct target_ops *t;
4057
4058 for (t = current_target.beneath; t != NULL; t = t->beneath)
4059 if (t->to_disable_btrace != NULL)
4060 {
4061 t->to_disable_btrace (t, btinfo);
4062 return;
4063 }
4064
4065 tcomplain ();
4066 }
4067
4068 /* See target.h. */
4069
4070 void
4071 target_teardown_btrace (struct btrace_target_info *btinfo)
4072 {
4073 struct target_ops *t;
4074
4075 for (t = current_target.beneath; t != NULL; t = t->beneath)
4076 if (t->to_teardown_btrace != NULL)
4077 {
4078 t->to_teardown_btrace (t, btinfo);
4079 return;
4080 }
4081
4082 tcomplain ();
4083 }
4084
4085 /* See target.h. */
4086
4087 enum btrace_error
4088 target_read_btrace (VEC (btrace_block_s) **btrace,
4089 struct btrace_target_info *btinfo,
4090 enum btrace_read_type type)
4091 {
4092 struct target_ops *t;
4093
4094 for (t = current_target.beneath; t != NULL; t = t->beneath)
4095 if (t->to_read_btrace != NULL)
4096 return t->to_read_btrace (t, btrace, btinfo, type);
4097
4098 tcomplain ();
4099 return BTRACE_ERR_NOT_SUPPORTED;
4100 }
4101
4102 /* See target.h. */
4103
4104 void
4105 target_stop_recording (void)
4106 {
4107 struct target_ops *t;
4108
4109 for (t = current_target.beneath; t != NULL; t = t->beneath)
4110 if (t->to_stop_recording != NULL)
4111 {
4112 t->to_stop_recording (t);
4113 return;
4114 }
4115
4116 /* This is optional. */
4117 }
4118
4119 /* See target.h. */
4120
4121 void
4122 target_info_record (void)
4123 {
4124 struct target_ops *t;
4125
4126 for (t = current_target.beneath; t != NULL; t = t->beneath)
4127 if (t->to_info_record != NULL)
4128 {
4129 t->to_info_record (t);
4130 return;
4131 }
4132
4133 tcomplain ();
4134 }
4135
4136 /* See target.h. */
4137
4138 void
4139 target_save_record (const char *filename)
4140 {
4141 struct target_ops *t;
4142
4143 for (t = current_target.beneath; t != NULL; t = t->beneath)
4144 if (t->to_save_record != NULL)
4145 {
4146 t->to_save_record (t, filename);
4147 return;
4148 }
4149
4150 tcomplain ();
4151 }
4152
4153 /* See target.h. */
4154
4155 int
4156 target_supports_delete_record (void)
4157 {
4158 struct target_ops *t;
4159
4160 for (t = current_target.beneath; t != NULL; t = t->beneath)
4161 if (t->to_delete_record != NULL)
4162 return 1;
4163
4164 return 0;
4165 }
4166
4167 /* See target.h. */
4168
4169 void
4170 target_delete_record (void)
4171 {
4172 struct target_ops *t;
4173
4174 for (t = current_target.beneath; t != NULL; t = t->beneath)
4175 if (t->to_delete_record != NULL)
4176 {
4177 t->to_delete_record (t);
4178 return;
4179 }
4180
4181 tcomplain ();
4182 }
4183
4184 /* See target.h. */
4185
4186 int
4187 target_record_is_replaying (void)
4188 {
4189 struct target_ops *t;
4190
4191 for (t = current_target.beneath; t != NULL; t = t->beneath)
4192 if (t->to_record_is_replaying != NULL)
4193 return t->to_record_is_replaying (t);
4194
4195 return 0;
4196 }
4197
4198 /* See target.h. */
4199
4200 void
4201 target_goto_record_begin (void)
4202 {
4203 struct target_ops *t;
4204
4205 for (t = current_target.beneath; t != NULL; t = t->beneath)
4206 if (t->to_goto_record_begin != NULL)
4207 {
4208 t->to_goto_record_begin (t);
4209 return;
4210 }
4211
4212 tcomplain ();
4213 }
4214
4215 /* See target.h. */
4216
4217 void
4218 target_goto_record_end (void)
4219 {
4220 struct target_ops *t;
4221
4222 for (t = current_target.beneath; t != NULL; t = t->beneath)
4223 if (t->to_goto_record_end != NULL)
4224 {
4225 t->to_goto_record_end (t);
4226 return;
4227 }
4228
4229 tcomplain ();
4230 }
4231
4232 /* See target.h. */
4233
4234 void
4235 target_goto_record (ULONGEST insn)
4236 {
4237 struct target_ops *t;
4238
4239 for (t = current_target.beneath; t != NULL; t = t->beneath)
4240 if (t->to_goto_record != NULL)
4241 {
4242 t->to_goto_record (t, insn);
4243 return;
4244 }
4245
4246 tcomplain ();
4247 }
4248
4249 /* See target.h. */
4250
4251 void
4252 target_insn_history (int size, int flags)
4253 {
4254 struct target_ops *t;
4255
4256 for (t = current_target.beneath; t != NULL; t = t->beneath)
4257 if (t->to_insn_history != NULL)
4258 {
4259 t->to_insn_history (t, size, flags);
4260 return;
4261 }
4262
4263 tcomplain ();
4264 }
4265
4266 /* See target.h. */
4267
4268 void
4269 target_insn_history_from (ULONGEST from, int size, int flags)
4270 {
4271 struct target_ops *t;
4272
4273 for (t = current_target.beneath; t != NULL; t = t->beneath)
4274 if (t->to_insn_history_from != NULL)
4275 {
4276 t->to_insn_history_from (t, from, size, flags);
4277 return;
4278 }
4279
4280 tcomplain ();
4281 }
4282
4283 /* See target.h. */
4284
4285 void
4286 target_insn_history_range (ULONGEST begin, ULONGEST end, int flags)
4287 {
4288 struct target_ops *t;
4289
4290 for (t = current_target.beneath; t != NULL; t = t->beneath)
4291 if (t->to_insn_history_range != NULL)
4292 {
4293 t->to_insn_history_range (t, begin, end, flags);
4294 return;
4295 }
4296
4297 tcomplain ();
4298 }
4299
4300 /* See target.h. */
4301
4302 void
4303 target_call_history (int size, int flags)
4304 {
4305 struct target_ops *t;
4306
4307 for (t = current_target.beneath; t != NULL; t = t->beneath)
4308 if (t->to_call_history != NULL)
4309 {
4310 t->to_call_history (t, size, flags);
4311 return;
4312 }
4313
4314 tcomplain ();
4315 }
4316
4317 /* See target.h. */
4318
4319 void
4320 target_call_history_from (ULONGEST begin, int size, int flags)
4321 {
4322 struct target_ops *t;
4323
4324 for (t = current_target.beneath; t != NULL; t = t->beneath)
4325 if (t->to_call_history_from != NULL)
4326 {
4327 t->to_call_history_from (t, begin, size, flags);
4328 return;
4329 }
4330
4331 tcomplain ();
4332 }
4333
4334 /* See target.h. */
4335
4336 void
4337 target_call_history_range (ULONGEST begin, ULONGEST end, int flags)
4338 {
4339 struct target_ops *t;
4340
4341 for (t = current_target.beneath; t != NULL; t = t->beneath)
4342 if (t->to_call_history_range != NULL)
4343 {
4344 t->to_call_history_range (t, begin, end, flags);
4345 return;
4346 }
4347
4348 tcomplain ();
4349 }
4350
4351 static void
4352 debug_to_prepare_to_store (struct target_ops *self, struct regcache *regcache)
4353 {
4354 debug_target.to_prepare_to_store (&debug_target, regcache);
4355
4356 fprintf_unfiltered (gdb_stdlog, "target_prepare_to_store ()\n");
4357 }
4358
4359 /* See target.h. */
4360
4361 const struct frame_unwind *
4362 target_get_unwinder (void)
4363 {
4364 struct target_ops *t;
4365
4366 for (t = current_target.beneath; t != NULL; t = t->beneath)
4367 if (t->to_get_unwinder != NULL)
4368 return t->to_get_unwinder;
4369
4370 return NULL;
4371 }
4372
4373 /* See target.h. */
4374
4375 const struct frame_unwind *
4376 target_get_tailcall_unwinder (void)
4377 {
4378 struct target_ops *t;
4379
4380 for (t = current_target.beneath; t != NULL; t = t->beneath)
4381 if (t->to_get_tailcall_unwinder != NULL)
4382 return t->to_get_tailcall_unwinder;
4383
4384 return NULL;
4385 }
4386
4387 /* See target.h. */
4388
4389 CORE_ADDR
4390 forward_target_decr_pc_after_break (struct target_ops *ops,
4391 struct gdbarch *gdbarch)
4392 {
4393 for (; ops != NULL; ops = ops->beneath)
4394 if (ops->to_decr_pc_after_break != NULL)
4395 return ops->to_decr_pc_after_break (ops, gdbarch);
4396
4397 return gdbarch_decr_pc_after_break (gdbarch);
4398 }
4399
4400 /* See target.h. */
4401
4402 CORE_ADDR
4403 target_decr_pc_after_break (struct gdbarch *gdbarch)
4404 {
4405 return forward_target_decr_pc_after_break (current_target.beneath, gdbarch);
4406 }
4407
4408 static int
4409 deprecated_debug_xfer_memory (CORE_ADDR memaddr, bfd_byte *myaddr, int len,
4410 int write, struct mem_attrib *attrib,
4411 struct target_ops *target)
4412 {
4413 int retval;
4414
4415 retval = debug_target.deprecated_xfer_memory (memaddr, myaddr, len, write,
4416 attrib, target);
4417
4418 fprintf_unfiltered (gdb_stdlog,
4419 "target_xfer_memory (%s, xxx, %d, %s, xxx) = %d",
4420 paddress (target_gdbarch (), memaddr), len,
4421 write ? "write" : "read", retval);
4422
4423 if (retval > 0)
4424 {
4425 int i;
4426
4427 fputs_unfiltered (", bytes =", gdb_stdlog);
4428 for (i = 0; i < retval; i++)
4429 {
4430 if ((((intptr_t) &(myaddr[i])) & 0xf) == 0)
4431 {
4432 if (targetdebug < 2 && i > 0)
4433 {
4434 fprintf_unfiltered (gdb_stdlog, " ...");
4435 break;
4436 }
4437 fprintf_unfiltered (gdb_stdlog, "\n");
4438 }
4439
4440 fprintf_unfiltered (gdb_stdlog, " %02x", myaddr[i] & 0xff);
4441 }
4442 }
4443
4444 fputc_unfiltered ('\n', gdb_stdlog);
4445
4446 return retval;
4447 }
4448
4449 static void
4450 debug_to_files_info (struct target_ops *target)
4451 {
4452 debug_target.to_files_info (target);
4453
4454 fprintf_unfiltered (gdb_stdlog, "target_files_info (xxx)\n");
4455 }
4456
4457 static int
4458 debug_to_insert_breakpoint (struct target_ops *ops, struct gdbarch *gdbarch,
4459 struct bp_target_info *bp_tgt)
4460 {
4461 int retval;
4462
4463 retval = debug_target.to_insert_breakpoint (&debug_target, gdbarch, bp_tgt);
4464
4465 fprintf_unfiltered (gdb_stdlog,
4466 "target_insert_breakpoint (%s, xxx) = %ld\n",
4467 core_addr_to_string (bp_tgt->placed_address),
4468 (unsigned long) retval);
4469 return retval;
4470 }
4471
4472 static int
4473 debug_to_remove_breakpoint (struct target_ops *ops, struct gdbarch *gdbarch,
4474 struct bp_target_info *bp_tgt)
4475 {
4476 int retval;
4477
4478 retval = debug_target.to_remove_breakpoint (&debug_target, gdbarch, bp_tgt);
4479
4480 fprintf_unfiltered (gdb_stdlog,
4481 "target_remove_breakpoint (%s, xxx) = %ld\n",
4482 core_addr_to_string (bp_tgt->placed_address),
4483 (unsigned long) retval);
4484 return retval;
4485 }
4486
4487 static int
4488 debug_to_can_use_hw_breakpoint (struct target_ops *self,
4489 int type, int cnt, int from_tty)
4490 {
4491 int retval;
4492
4493 retval = debug_target.to_can_use_hw_breakpoint (&debug_target,
4494 type, cnt, from_tty);
4495
4496 fprintf_unfiltered (gdb_stdlog,
4497 "target_can_use_hw_breakpoint (%ld, %ld, %ld) = %ld\n",
4498 (unsigned long) type,
4499 (unsigned long) cnt,
4500 (unsigned long) from_tty,
4501 (unsigned long) retval);
4502 return retval;
4503 }
4504
4505 static int
4506 debug_to_region_ok_for_hw_watchpoint (struct target_ops *self,
4507 CORE_ADDR addr, int len)
4508 {
4509 CORE_ADDR retval;
4510
4511 retval = debug_target.to_region_ok_for_hw_watchpoint (&debug_target,
4512 addr, len);
4513
4514 fprintf_unfiltered (gdb_stdlog,
4515 "target_region_ok_for_hw_watchpoint (%s, %ld) = %s\n",
4516 core_addr_to_string (addr), (unsigned long) len,
4517 core_addr_to_string (retval));
4518 return retval;
4519 }
4520
4521 static int
4522 debug_to_can_accel_watchpoint_condition (struct target_ops *self,
4523 CORE_ADDR addr, int len, int rw,
4524 struct expression *cond)
4525 {
4526 int retval;
4527
4528 retval = debug_target.to_can_accel_watchpoint_condition (&debug_target,
4529 addr, len,
4530 rw, cond);
4531
4532 fprintf_unfiltered (gdb_stdlog,
4533 "target_can_accel_watchpoint_condition "
4534 "(%s, %d, %d, %s) = %ld\n",
4535 core_addr_to_string (addr), len, rw,
4536 host_address_to_string (cond), (unsigned long) retval);
4537 return retval;
4538 }
4539
4540 static int
4541 debug_to_stopped_by_watchpoint (struct target_ops *ops)
4542 {
4543 int retval;
4544
4545 retval = debug_target.to_stopped_by_watchpoint (&debug_target);
4546
4547 fprintf_unfiltered (gdb_stdlog,
4548 "target_stopped_by_watchpoint () = %ld\n",
4549 (unsigned long) retval);
4550 return retval;
4551 }
4552
4553 static int
4554 debug_to_stopped_data_address (struct target_ops *target, CORE_ADDR *addr)
4555 {
4556 int retval;
4557
4558 retval = debug_target.to_stopped_data_address (target, addr);
4559
4560 fprintf_unfiltered (gdb_stdlog,
4561 "target_stopped_data_address ([%s]) = %ld\n",
4562 core_addr_to_string (*addr),
4563 (unsigned long)retval);
4564 return retval;
4565 }
4566
4567 static int
4568 debug_to_watchpoint_addr_within_range (struct target_ops *target,
4569 CORE_ADDR addr,
4570 CORE_ADDR start, int length)
4571 {
4572 int retval;
4573
4574 retval = debug_target.to_watchpoint_addr_within_range (target, addr,
4575 start, length);
4576
4577 fprintf_filtered (gdb_stdlog,
4578 "target_watchpoint_addr_within_range (%s, %s, %d) = %d\n",
4579 core_addr_to_string (addr), core_addr_to_string (start),
4580 length, retval);
4581 return retval;
4582 }
4583
4584 static int
4585 debug_to_insert_hw_breakpoint (struct target_ops *self,
4586 struct gdbarch *gdbarch,
4587 struct bp_target_info *bp_tgt)
4588 {
4589 int retval;
4590
4591 retval = debug_target.to_insert_hw_breakpoint (&debug_target,
4592 gdbarch, bp_tgt);
4593
4594 fprintf_unfiltered (gdb_stdlog,
4595 "target_insert_hw_breakpoint (%s, xxx) = %ld\n",
4596 core_addr_to_string (bp_tgt->placed_address),
4597 (unsigned long) retval);
4598 return retval;
4599 }
4600
4601 static int
4602 debug_to_remove_hw_breakpoint (struct target_ops *self,
4603 struct gdbarch *gdbarch,
4604 struct bp_target_info *bp_tgt)
4605 {
4606 int retval;
4607
4608 retval = debug_target.to_remove_hw_breakpoint (&debug_target,
4609 gdbarch, bp_tgt);
4610
4611 fprintf_unfiltered (gdb_stdlog,
4612 "target_remove_hw_breakpoint (%s, xxx) = %ld\n",
4613 core_addr_to_string (bp_tgt->placed_address),
4614 (unsigned long) retval);
4615 return retval;
4616 }
4617
4618 static int
4619 debug_to_insert_watchpoint (struct target_ops *self,
4620 CORE_ADDR addr, int len, int type,
4621 struct expression *cond)
4622 {
4623 int retval;
4624
4625 retval = debug_target.to_insert_watchpoint (&debug_target,
4626 addr, len, type, cond);
4627
4628 fprintf_unfiltered (gdb_stdlog,
4629 "target_insert_watchpoint (%s, %d, %d, %s) = %ld\n",
4630 core_addr_to_string (addr), len, type,
4631 host_address_to_string (cond), (unsigned long) retval);
4632 return retval;
4633 }
4634
4635 static int
4636 debug_to_remove_watchpoint (struct target_ops *self,
4637 CORE_ADDR addr, int len, int type,
4638 struct expression *cond)
4639 {
4640 int retval;
4641
4642 retval = debug_target.to_remove_watchpoint (&debug_target,
4643 addr, len, type, cond);
4644
4645 fprintf_unfiltered (gdb_stdlog,
4646 "target_remove_watchpoint (%s, %d, %d, %s) = %ld\n",
4647 core_addr_to_string (addr), len, type,
4648 host_address_to_string (cond), (unsigned long) retval);
4649 return retval;
4650 }
4651
4652 static void
4653 debug_to_terminal_init (struct target_ops *self)
4654 {
4655 debug_target.to_terminal_init (&debug_target);
4656
4657 fprintf_unfiltered (gdb_stdlog, "target_terminal_init ()\n");
4658 }
4659
4660 static void
4661 debug_to_terminal_inferior (struct target_ops *self)
4662 {
4663 debug_target.to_terminal_inferior (&debug_target);
4664
4665 fprintf_unfiltered (gdb_stdlog, "target_terminal_inferior ()\n");
4666 }
4667
4668 static void
4669 debug_to_terminal_ours_for_output (struct target_ops *self)
4670 {
4671 debug_target.to_terminal_ours_for_output (&debug_target);
4672
4673 fprintf_unfiltered (gdb_stdlog, "target_terminal_ours_for_output ()\n");
4674 }
4675
4676 static void
4677 debug_to_terminal_ours (struct target_ops *self)
4678 {
4679 debug_target.to_terminal_ours (&debug_target);
4680
4681 fprintf_unfiltered (gdb_stdlog, "target_terminal_ours ()\n");
4682 }
4683
4684 static void
4685 debug_to_terminal_save_ours (struct target_ops *self)
4686 {
4687 debug_target.to_terminal_save_ours (&debug_target);
4688
4689 fprintf_unfiltered (gdb_stdlog, "target_terminal_save_ours ()\n");
4690 }
4691
4692 static void
4693 debug_to_terminal_info (struct target_ops *self,
4694 const char *arg, int from_tty)
4695 {
4696 debug_target.to_terminal_info (&debug_target, arg, from_tty);
4697
4698 fprintf_unfiltered (gdb_stdlog, "target_terminal_info (%s, %d)\n", arg,
4699 from_tty);
4700 }
4701
4702 static void
4703 debug_to_load (struct target_ops *self, char *args, int from_tty)
4704 {
4705 debug_target.to_load (&debug_target, args, from_tty);
4706
4707 fprintf_unfiltered (gdb_stdlog, "target_load (%s, %d)\n", args, from_tty);
4708 }
4709
4710 static void
4711 debug_to_post_startup_inferior (struct target_ops *self, ptid_t ptid)
4712 {
4713 debug_target.to_post_startup_inferior (&debug_target, ptid);
4714
4715 fprintf_unfiltered (gdb_stdlog, "target_post_startup_inferior (%d)\n",
4716 ptid_get_pid (ptid));
4717 }
4718
4719 static int
4720 debug_to_insert_fork_catchpoint (struct target_ops *self, int pid)
4721 {
4722 int retval;
4723
4724 retval = debug_target.to_insert_fork_catchpoint (&debug_target, pid);
4725
4726 fprintf_unfiltered (gdb_stdlog, "target_insert_fork_catchpoint (%d) = %d\n",
4727 pid, retval);
4728
4729 return retval;
4730 }
4731
4732 static int
4733 debug_to_remove_fork_catchpoint (struct target_ops *self, int pid)
4734 {
4735 int retval;
4736
4737 retval = debug_target.to_remove_fork_catchpoint (&debug_target, pid);
4738
4739 fprintf_unfiltered (gdb_stdlog, "target_remove_fork_catchpoint (%d) = %d\n",
4740 pid, retval);
4741
4742 return retval;
4743 }
4744
4745 static int
4746 debug_to_insert_vfork_catchpoint (struct target_ops *self, int pid)
4747 {
4748 int retval;
4749
4750 retval = debug_target.to_insert_vfork_catchpoint (&debug_target, pid);
4751
4752 fprintf_unfiltered (gdb_stdlog, "target_insert_vfork_catchpoint (%d) = %d\n",
4753 pid, retval);
4754
4755 return retval;
4756 }
4757
4758 static int
4759 debug_to_remove_vfork_catchpoint (struct target_ops *self, int pid)
4760 {
4761 int retval;
4762
4763 retval = debug_target.to_remove_vfork_catchpoint (&debug_target, pid);
4764
4765 fprintf_unfiltered (gdb_stdlog, "target_remove_vfork_catchpoint (%d) = %d\n",
4766 pid, retval);
4767
4768 return retval;
4769 }
4770
4771 static int
4772 debug_to_insert_exec_catchpoint (struct target_ops *self, int pid)
4773 {
4774 int retval;
4775
4776 retval = debug_target.to_insert_exec_catchpoint (&debug_target, pid);
4777
4778 fprintf_unfiltered (gdb_stdlog, "target_insert_exec_catchpoint (%d) = %d\n",
4779 pid, retval);
4780
4781 return retval;
4782 }
4783
4784 static int
4785 debug_to_remove_exec_catchpoint (struct target_ops *self, int pid)
4786 {
4787 int retval;
4788
4789 retval = debug_target.to_remove_exec_catchpoint (&debug_target, pid);
4790
4791 fprintf_unfiltered (gdb_stdlog, "target_remove_exec_catchpoint (%d) = %d\n",
4792 pid, retval);
4793
4794 return retval;
4795 }
4796
4797 static int
4798 debug_to_has_exited (struct target_ops *self,
4799 int pid, int wait_status, int *exit_status)
4800 {
4801 int has_exited;
4802
4803 has_exited = debug_target.to_has_exited (&debug_target,
4804 pid, wait_status, exit_status);
4805
4806 fprintf_unfiltered (gdb_stdlog, "target_has_exited (%d, %d, %d) = %d\n",
4807 pid, wait_status, *exit_status, has_exited);
4808
4809 return has_exited;
4810 }
4811
4812 static int
4813 debug_to_can_run (struct target_ops *self)
4814 {
4815 int retval;
4816
4817 retval = debug_target.to_can_run (&debug_target);
4818
4819 fprintf_unfiltered (gdb_stdlog, "target_can_run () = %d\n", retval);
4820
4821 return retval;
4822 }
4823
4824 static struct gdbarch *
4825 debug_to_thread_architecture (struct target_ops *ops, ptid_t ptid)
4826 {
4827 struct gdbarch *retval;
4828
4829 retval = debug_target.to_thread_architecture (ops, ptid);
4830
4831 fprintf_unfiltered (gdb_stdlog,
4832 "target_thread_architecture (%s) = %s [%s]\n",
4833 target_pid_to_str (ptid),
4834 host_address_to_string (retval),
4835 gdbarch_bfd_arch_info (retval)->printable_name);
4836 return retval;
4837 }
4838
4839 static void
4840 debug_to_stop (struct target_ops *self, ptid_t ptid)
4841 {
4842 debug_target.to_stop (&debug_target, ptid);
4843
4844 fprintf_unfiltered (gdb_stdlog, "target_stop (%s)\n",
4845 target_pid_to_str (ptid));
4846 }
4847
4848 static void
4849 debug_to_rcmd (struct target_ops *self, char *command,
4850 struct ui_file *outbuf)
4851 {
4852 debug_target.to_rcmd (&debug_target, command, outbuf);
4853 fprintf_unfiltered (gdb_stdlog, "target_rcmd (%s, ...)\n", command);
4854 }
4855
4856 static char *
4857 debug_to_pid_to_exec_file (struct target_ops *self, int pid)
4858 {
4859 char *exec_file;
4860
4861 exec_file = debug_target.to_pid_to_exec_file (&debug_target, pid);
4862
4863 fprintf_unfiltered (gdb_stdlog, "target_pid_to_exec_file (%d) = %s\n",
4864 pid, exec_file);
4865
4866 return exec_file;
4867 }
4868
4869 static void
4870 setup_target_debug (void)
4871 {
4872 memcpy (&debug_target, &current_target, sizeof debug_target);
4873
4874 current_target.to_open = debug_to_open;
4875 current_target.to_post_attach = debug_to_post_attach;
4876 current_target.to_prepare_to_store = debug_to_prepare_to_store;
4877 current_target.deprecated_xfer_memory = deprecated_debug_xfer_memory;
4878 current_target.to_files_info = debug_to_files_info;
4879 current_target.to_insert_breakpoint = debug_to_insert_breakpoint;
4880 current_target.to_remove_breakpoint = debug_to_remove_breakpoint;
4881 current_target.to_can_use_hw_breakpoint = debug_to_can_use_hw_breakpoint;
4882 current_target.to_insert_hw_breakpoint = debug_to_insert_hw_breakpoint;
4883 current_target.to_remove_hw_breakpoint = debug_to_remove_hw_breakpoint;
4884 current_target.to_insert_watchpoint = debug_to_insert_watchpoint;
4885 current_target.to_remove_watchpoint = debug_to_remove_watchpoint;
4886 current_target.to_stopped_by_watchpoint = debug_to_stopped_by_watchpoint;
4887 current_target.to_stopped_data_address = debug_to_stopped_data_address;
4888 current_target.to_watchpoint_addr_within_range
4889 = debug_to_watchpoint_addr_within_range;
4890 current_target.to_region_ok_for_hw_watchpoint
4891 = debug_to_region_ok_for_hw_watchpoint;
4892 current_target.to_can_accel_watchpoint_condition
4893 = debug_to_can_accel_watchpoint_condition;
4894 current_target.to_terminal_init = debug_to_terminal_init;
4895 current_target.to_terminal_inferior = debug_to_terminal_inferior;
4896 current_target.to_terminal_ours_for_output
4897 = debug_to_terminal_ours_for_output;
4898 current_target.to_terminal_ours = debug_to_terminal_ours;
4899 current_target.to_terminal_save_ours = debug_to_terminal_save_ours;
4900 current_target.to_terminal_info = debug_to_terminal_info;
4901 current_target.to_load = debug_to_load;
4902 current_target.to_post_startup_inferior = debug_to_post_startup_inferior;
4903 current_target.to_insert_fork_catchpoint = debug_to_insert_fork_catchpoint;
4904 current_target.to_remove_fork_catchpoint = debug_to_remove_fork_catchpoint;
4905 current_target.to_insert_vfork_catchpoint = debug_to_insert_vfork_catchpoint;
4906 current_target.to_remove_vfork_catchpoint = debug_to_remove_vfork_catchpoint;
4907 current_target.to_insert_exec_catchpoint = debug_to_insert_exec_catchpoint;
4908 current_target.to_remove_exec_catchpoint = debug_to_remove_exec_catchpoint;
4909 current_target.to_has_exited = debug_to_has_exited;
4910 current_target.to_can_run = debug_to_can_run;
4911 current_target.to_stop = debug_to_stop;
4912 current_target.to_rcmd = debug_to_rcmd;
4913 current_target.to_pid_to_exec_file = debug_to_pid_to_exec_file;
4914 current_target.to_thread_architecture = debug_to_thread_architecture;
4915 }
4916 \f
4917
4918 static char targ_desc[] =
4919 "Names of targets and files being debugged.\nShows the entire \
4920 stack of targets currently in use (including the exec-file,\n\
4921 core-file, and process, if any), as well as the symbol file name.";
4922
4923 static void
4924 default_rcmd (struct target_ops *self, char *command, struct ui_file *output)
4925 {
4926 error (_("\"monitor\" command not supported by this target."));
4927 }
4928
4929 static void
4930 do_monitor_command (char *cmd,
4931 int from_tty)
4932 {
4933 target_rcmd (cmd, gdb_stdtarg);
4934 }
4935
4936 /* Print the name of each layers of our target stack. */
4937
4938 static void
4939 maintenance_print_target_stack (char *cmd, int from_tty)
4940 {
4941 struct target_ops *t;
4942
4943 printf_filtered (_("The current target stack is:\n"));
4944
4945 for (t = target_stack; t != NULL; t = t->beneath)
4946 {
4947 printf_filtered (" - %s (%s)\n", t->to_shortname, t->to_longname);
4948 }
4949 }
4950
4951 /* Controls if async mode is permitted. */
4952 int target_async_permitted = 0;
4953
4954 /* The set command writes to this variable. If the inferior is
4955 executing, target_async_permitted is *not* updated. */
4956 static int target_async_permitted_1 = 0;
4957
4958 static void
4959 set_target_async_command (char *args, int from_tty,
4960 struct cmd_list_element *c)
4961 {
4962 if (have_live_inferiors ())
4963 {
4964 target_async_permitted_1 = target_async_permitted;
4965 error (_("Cannot change this setting while the inferior is running."));
4966 }
4967
4968 target_async_permitted = target_async_permitted_1;
4969 }
4970
4971 static void
4972 show_target_async_command (struct ui_file *file, int from_tty,
4973 struct cmd_list_element *c,
4974 const char *value)
4975 {
4976 fprintf_filtered (file,
4977 _("Controlling the inferior in "
4978 "asynchronous mode is %s.\n"), value);
4979 }
4980
4981 /* Temporary copies of permission settings. */
4982
4983 static int may_write_registers_1 = 1;
4984 static int may_write_memory_1 = 1;
4985 static int may_insert_breakpoints_1 = 1;
4986 static int may_insert_tracepoints_1 = 1;
4987 static int may_insert_fast_tracepoints_1 = 1;
4988 static int may_stop_1 = 1;
4989
4990 /* Make the user-set values match the real values again. */
4991
4992 void
4993 update_target_permissions (void)
4994 {
4995 may_write_registers_1 = may_write_registers;
4996 may_write_memory_1 = may_write_memory;
4997 may_insert_breakpoints_1 = may_insert_breakpoints;
4998 may_insert_tracepoints_1 = may_insert_tracepoints;
4999 may_insert_fast_tracepoints_1 = may_insert_fast_tracepoints;
5000 may_stop_1 = may_stop;
5001 }
5002
5003 /* The one function handles (most of) the permission flags in the same
5004 way. */
5005
5006 static void
5007 set_target_permissions (char *args, int from_tty,
5008 struct cmd_list_element *c)
5009 {
5010 if (target_has_execution)
5011 {
5012 update_target_permissions ();
5013 error (_("Cannot change this setting while the inferior is running."));
5014 }
5015
5016 /* Make the real values match the user-changed values. */
5017 may_write_registers = may_write_registers_1;
5018 may_insert_breakpoints = may_insert_breakpoints_1;
5019 may_insert_tracepoints = may_insert_tracepoints_1;
5020 may_insert_fast_tracepoints = may_insert_fast_tracepoints_1;
5021 may_stop = may_stop_1;
5022 update_observer_mode ();
5023 }
5024
5025 /* Set memory write permission independently of observer mode. */
5026
5027 static void
5028 set_write_memory_permission (char *args, int from_tty,
5029 struct cmd_list_element *c)
5030 {
5031 /* Make the real values match the user-changed values. */
5032 may_write_memory = may_write_memory_1;
5033 update_observer_mode ();
5034 }
5035
5036
5037 void
5038 initialize_targets (void)
5039 {
5040 init_dummy_target ();
5041 push_target (&dummy_target);
5042
5043 add_info ("target", target_info, targ_desc);
5044 add_info ("files", target_info, targ_desc);
5045
5046 add_setshow_zuinteger_cmd ("target", class_maintenance, &targetdebug, _("\
5047 Set target debugging."), _("\
5048 Show target debugging."), _("\
5049 When non-zero, target debugging is enabled. Higher numbers are more\n\
5050 verbose. Changes do not take effect until the next \"run\" or \"target\"\n\
5051 command."),
5052 NULL,
5053 show_targetdebug,
5054 &setdebuglist, &showdebuglist);
5055
5056 add_setshow_boolean_cmd ("trust-readonly-sections", class_support,
5057 &trust_readonly, _("\
5058 Set mode for reading from readonly sections."), _("\
5059 Show mode for reading from readonly sections."), _("\
5060 When this mode is on, memory reads from readonly sections (such as .text)\n\
5061 will be read from the object file instead of from the target. This will\n\
5062 result in significant performance improvement for remote targets."),
5063 NULL,
5064 show_trust_readonly,
5065 &setlist, &showlist);
5066
5067 add_com ("monitor", class_obscure, do_monitor_command,
5068 _("Send a command to the remote monitor (remote targets only)."));
5069
5070 add_cmd ("target-stack", class_maintenance, maintenance_print_target_stack,
5071 _("Print the name of each layer of the internal target stack."),
5072 &maintenanceprintlist);
5073
5074 add_setshow_boolean_cmd ("target-async", no_class,
5075 &target_async_permitted_1, _("\
5076 Set whether gdb controls the inferior in asynchronous mode."), _("\
5077 Show whether gdb controls the inferior in asynchronous mode."), _("\
5078 Tells gdb whether to control the inferior in asynchronous mode."),
5079 set_target_async_command,
5080 show_target_async_command,
5081 &setlist,
5082 &showlist);
5083
5084 add_setshow_boolean_cmd ("may-write-registers", class_support,
5085 &may_write_registers_1, _("\
5086 Set permission to write into registers."), _("\
5087 Show permission to write into registers."), _("\
5088 When this permission is on, GDB may write into the target's registers.\n\
5089 Otherwise, any sort of write attempt will result in an error."),
5090 set_target_permissions, NULL,
5091 &setlist, &showlist);
5092
5093 add_setshow_boolean_cmd ("may-write-memory", class_support,
5094 &may_write_memory_1, _("\
5095 Set permission to write into target memory."), _("\
5096 Show permission to write into target memory."), _("\
5097 When this permission is on, GDB may write into the target's memory.\n\
5098 Otherwise, any sort of write attempt will result in an error."),
5099 set_write_memory_permission, NULL,
5100 &setlist, &showlist);
5101
5102 add_setshow_boolean_cmd ("may-insert-breakpoints", class_support,
5103 &may_insert_breakpoints_1, _("\
5104 Set permission to insert breakpoints in the target."), _("\
5105 Show permission to insert breakpoints in the target."), _("\
5106 When this permission is on, GDB may insert breakpoints in the program.\n\
5107 Otherwise, any sort of insertion attempt will result in an error."),
5108 set_target_permissions, NULL,
5109 &setlist, &showlist);
5110
5111 add_setshow_boolean_cmd ("may-insert-tracepoints", class_support,
5112 &may_insert_tracepoints_1, _("\
5113 Set permission to insert tracepoints in the target."), _("\
5114 Show permission to insert tracepoints in the target."), _("\
5115 When this permission is on, GDB may insert tracepoints in the program.\n\
5116 Otherwise, any sort of insertion attempt will result in an error."),
5117 set_target_permissions, NULL,
5118 &setlist, &showlist);
5119
5120 add_setshow_boolean_cmd ("may-insert-fast-tracepoints", class_support,
5121 &may_insert_fast_tracepoints_1, _("\
5122 Set permission to insert fast tracepoints in the target."), _("\
5123 Show permission to insert fast tracepoints in the target."), _("\
5124 When this permission is on, GDB may insert fast tracepoints.\n\
5125 Otherwise, any sort of insertion attempt will result in an error."),
5126 set_target_permissions, NULL,
5127 &setlist, &showlist);
5128
5129 add_setshow_boolean_cmd ("may-interrupt", class_support,
5130 &may_stop_1, _("\
5131 Set permission to interrupt or signal the target."), _("\
5132 Show permission to interrupt or signal the target."), _("\
5133 When this permission is on, GDB may interrupt/stop the target's execution.\n\
5134 Otherwise, any attempt to interrupt or stop will be ignored."),
5135 set_target_permissions, NULL,
5136 &setlist, &showlist);
5137 }