Code cleanup: Add objfile_name accessor
[deliverable/binutils-gdb.git] / gdb / target.c
1 /* Select target systems and architectures at runtime for GDB.
2
3 Copyright (C) 1990-2013 Free Software Foundation, Inc.
4
5 Contributed by Cygnus Support.
6
7 This file is part of GDB.
8
9 This program is free software; you can redistribute it and/or modify
10 it under the terms of the GNU General Public License as published by
11 the Free Software Foundation; either version 3 of the License, or
12 (at your option) any later version.
13
14 This program is distributed in the hope that it will be useful,
15 but WITHOUT ANY WARRANTY; without even the implied warranty of
16 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
17 GNU General Public License for more details.
18
19 You should have received a copy of the GNU General Public License
20 along with this program. If not, see <http://www.gnu.org/licenses/>. */
21
22 #include "defs.h"
23 #include <errno.h>
24 #include "gdb_string.h"
25 #include "target.h"
26 #include "gdbcmd.h"
27 #include "symtab.h"
28 #include "inferior.h"
29 #include "bfd.h"
30 #include "symfile.h"
31 #include "objfiles.h"
32 #include "dcache.h"
33 #include <signal.h>
34 #include "regcache.h"
35 #include "gdb_assert.h"
36 #include "gdbcore.h"
37 #include "exceptions.h"
38 #include "target-descriptions.h"
39 #include "gdbthread.h"
40 #include "solib.h"
41 #include "exec.h"
42 #include "inline-frame.h"
43 #include "tracepoint.h"
44 #include "gdb/fileio.h"
45 #include "agent.h"
46
47 static void target_info (char *, int);
48
49 static void default_terminal_info (const char *, int);
50
51 static int default_watchpoint_addr_within_range (struct target_ops *,
52 CORE_ADDR, CORE_ADDR, int);
53
54 static int default_region_ok_for_hw_watchpoint (CORE_ADDR, int);
55
56 static void tcomplain (void) ATTRIBUTE_NORETURN;
57
58 static int nomemory (CORE_ADDR, char *, int, int, struct target_ops *);
59
60 static int return_zero (void);
61
62 static int return_one (void);
63
64 static int return_minus_one (void);
65
66 void target_ignore (void);
67
68 static void target_command (char *, int);
69
70 static struct target_ops *find_default_run_target (char *);
71
72 static LONGEST default_xfer_partial (struct target_ops *ops,
73 enum target_object object,
74 const char *annex, gdb_byte *readbuf,
75 const gdb_byte *writebuf,
76 ULONGEST offset, LONGEST len);
77
78 static LONGEST current_xfer_partial (struct target_ops *ops,
79 enum target_object object,
80 const char *annex, gdb_byte *readbuf,
81 const gdb_byte *writebuf,
82 ULONGEST offset, LONGEST len);
83
84 static struct gdbarch *default_thread_architecture (struct target_ops *ops,
85 ptid_t ptid);
86
87 static void init_dummy_target (void);
88
89 static struct target_ops debug_target;
90
91 static void debug_to_open (char *, int);
92
93 static void debug_to_prepare_to_store (struct regcache *);
94
95 static void debug_to_files_info (struct target_ops *);
96
97 static int debug_to_insert_breakpoint (struct gdbarch *,
98 struct bp_target_info *);
99
100 static int debug_to_remove_breakpoint (struct gdbarch *,
101 struct bp_target_info *);
102
103 static int debug_to_can_use_hw_breakpoint (int, int, int);
104
105 static int debug_to_insert_hw_breakpoint (struct gdbarch *,
106 struct bp_target_info *);
107
108 static int debug_to_remove_hw_breakpoint (struct gdbarch *,
109 struct bp_target_info *);
110
111 static int debug_to_insert_watchpoint (CORE_ADDR, int, int,
112 struct expression *);
113
114 static int debug_to_remove_watchpoint (CORE_ADDR, int, int,
115 struct expression *);
116
117 static int debug_to_stopped_by_watchpoint (void);
118
119 static int debug_to_stopped_data_address (struct target_ops *, CORE_ADDR *);
120
121 static int debug_to_watchpoint_addr_within_range (struct target_ops *,
122 CORE_ADDR, CORE_ADDR, int);
123
124 static int debug_to_region_ok_for_hw_watchpoint (CORE_ADDR, int);
125
126 static int debug_to_can_accel_watchpoint_condition (CORE_ADDR, int, int,
127 struct expression *);
128
129 static void debug_to_terminal_init (void);
130
131 static void debug_to_terminal_inferior (void);
132
133 static void debug_to_terminal_ours_for_output (void);
134
135 static void debug_to_terminal_save_ours (void);
136
137 static void debug_to_terminal_ours (void);
138
139 static void debug_to_load (char *, int);
140
141 static int debug_to_can_run (void);
142
143 static void debug_to_stop (ptid_t);
144
145 /* Pointer to array of target architecture structures; the size of the
146 array; the current index into the array; the allocated size of the
147 array. */
148 struct target_ops **target_structs;
149 unsigned target_struct_size;
150 unsigned target_struct_allocsize;
151 #define DEFAULT_ALLOCSIZE 10
152
153 /* The initial current target, so that there is always a semi-valid
154 current target. */
155
156 static struct target_ops dummy_target;
157
158 /* Top of target stack. */
159
160 static struct target_ops *target_stack;
161
162 /* The target structure we are currently using to talk to a process
163 or file or whatever "inferior" we have. */
164
165 struct target_ops current_target;
166
167 /* Command list for target. */
168
169 static struct cmd_list_element *targetlist = NULL;
170
171 /* Nonzero if we should trust readonly sections from the
172 executable when reading memory. */
173
174 static int trust_readonly = 0;
175
176 /* Nonzero if we should show true memory content including
177 memory breakpoint inserted by gdb. */
178
179 static int show_memory_breakpoints = 0;
180
181 /* These globals control whether GDB attempts to perform these
182 operations; they are useful for targets that need to prevent
183 inadvertant disruption, such as in non-stop mode. */
184
185 int may_write_registers = 1;
186
187 int may_write_memory = 1;
188
189 int may_insert_breakpoints = 1;
190
191 int may_insert_tracepoints = 1;
192
193 int may_insert_fast_tracepoints = 1;
194
195 int may_stop = 1;
196
197 /* Non-zero if we want to see trace of target level stuff. */
198
199 static unsigned int targetdebug = 0;
200 static void
201 show_targetdebug (struct ui_file *file, int from_tty,
202 struct cmd_list_element *c, const char *value)
203 {
204 fprintf_filtered (file, _("Target debugging is %s.\n"), value);
205 }
206
207 static void setup_target_debug (void);
208
209 /* The option sets this. */
210 static int stack_cache_enabled_p_1 = 1;
211 /* And set_stack_cache_enabled_p updates this.
212 The reason for the separation is so that we don't flush the cache for
213 on->on transitions. */
214 static int stack_cache_enabled_p = 1;
215
216 /* This is called *after* the stack-cache has been set.
217 Flush the cache for off->on and on->off transitions.
218 There's no real need to flush the cache for on->off transitions,
219 except cleanliness. */
220
221 static void
222 set_stack_cache_enabled_p (char *args, int from_tty,
223 struct cmd_list_element *c)
224 {
225 if (stack_cache_enabled_p != stack_cache_enabled_p_1)
226 target_dcache_invalidate ();
227
228 stack_cache_enabled_p = stack_cache_enabled_p_1;
229 }
230
231 static void
232 show_stack_cache_enabled_p (struct ui_file *file, int from_tty,
233 struct cmd_list_element *c, const char *value)
234 {
235 fprintf_filtered (file, _("Cache use for stack accesses is %s.\n"), value);
236 }
237
238 /* Cache of memory operations, to speed up remote access. */
239 static DCACHE *target_dcache;
240
241 /* Invalidate the target dcache. */
242
243 void
244 target_dcache_invalidate (void)
245 {
246 dcache_invalidate (target_dcache);
247 }
248
249 /* The user just typed 'target' without the name of a target. */
250
251 static void
252 target_command (char *arg, int from_tty)
253 {
254 fputs_filtered ("Argument required (target name). Try `help target'\n",
255 gdb_stdout);
256 }
257
258 /* Default target_has_* methods for process_stratum targets. */
259
260 int
261 default_child_has_all_memory (struct target_ops *ops)
262 {
263 /* If no inferior selected, then we can't read memory here. */
264 if (ptid_equal (inferior_ptid, null_ptid))
265 return 0;
266
267 return 1;
268 }
269
270 int
271 default_child_has_memory (struct target_ops *ops)
272 {
273 /* If no inferior selected, then we can't read memory here. */
274 if (ptid_equal (inferior_ptid, null_ptid))
275 return 0;
276
277 return 1;
278 }
279
280 int
281 default_child_has_stack (struct target_ops *ops)
282 {
283 /* If no inferior selected, there's no stack. */
284 if (ptid_equal (inferior_ptid, null_ptid))
285 return 0;
286
287 return 1;
288 }
289
290 int
291 default_child_has_registers (struct target_ops *ops)
292 {
293 /* Can't read registers from no inferior. */
294 if (ptid_equal (inferior_ptid, null_ptid))
295 return 0;
296
297 return 1;
298 }
299
300 int
301 default_child_has_execution (struct target_ops *ops, ptid_t the_ptid)
302 {
303 /* If there's no thread selected, then we can't make it run through
304 hoops. */
305 if (ptid_equal (the_ptid, null_ptid))
306 return 0;
307
308 return 1;
309 }
310
311
312 int
313 target_has_all_memory_1 (void)
314 {
315 struct target_ops *t;
316
317 for (t = current_target.beneath; t != NULL; t = t->beneath)
318 if (t->to_has_all_memory (t))
319 return 1;
320
321 return 0;
322 }
323
324 int
325 target_has_memory_1 (void)
326 {
327 struct target_ops *t;
328
329 for (t = current_target.beneath; t != NULL; t = t->beneath)
330 if (t->to_has_memory (t))
331 return 1;
332
333 return 0;
334 }
335
336 int
337 target_has_stack_1 (void)
338 {
339 struct target_ops *t;
340
341 for (t = current_target.beneath; t != NULL; t = t->beneath)
342 if (t->to_has_stack (t))
343 return 1;
344
345 return 0;
346 }
347
348 int
349 target_has_registers_1 (void)
350 {
351 struct target_ops *t;
352
353 for (t = current_target.beneath; t != NULL; t = t->beneath)
354 if (t->to_has_registers (t))
355 return 1;
356
357 return 0;
358 }
359
360 int
361 target_has_execution_1 (ptid_t the_ptid)
362 {
363 struct target_ops *t;
364
365 for (t = current_target.beneath; t != NULL; t = t->beneath)
366 if (t->to_has_execution (t, the_ptid))
367 return 1;
368
369 return 0;
370 }
371
372 int
373 target_has_execution_current (void)
374 {
375 return target_has_execution_1 (inferior_ptid);
376 }
377
378 /* Complete initialization of T. This ensures that various fields in
379 T are set, if needed by the target implementation. */
380
381 void
382 complete_target_initialization (struct target_ops *t)
383 {
384 /* Provide default values for all "must have" methods. */
385 if (t->to_xfer_partial == NULL)
386 t->to_xfer_partial = default_xfer_partial;
387
388 if (t->to_has_all_memory == NULL)
389 t->to_has_all_memory = (int (*) (struct target_ops *)) return_zero;
390
391 if (t->to_has_memory == NULL)
392 t->to_has_memory = (int (*) (struct target_ops *)) return_zero;
393
394 if (t->to_has_stack == NULL)
395 t->to_has_stack = (int (*) (struct target_ops *)) return_zero;
396
397 if (t->to_has_registers == NULL)
398 t->to_has_registers = (int (*) (struct target_ops *)) return_zero;
399
400 if (t->to_has_execution == NULL)
401 t->to_has_execution = (int (*) (struct target_ops *, ptid_t)) return_zero;
402 }
403
404 /* Add possible target architecture T to the list and add a new
405 command 'target T->to_shortname'. Set COMPLETER as the command's
406 completer if not NULL. */
407
408 void
409 add_target_with_completer (struct target_ops *t,
410 completer_ftype *completer)
411 {
412 struct cmd_list_element *c;
413
414 complete_target_initialization (t);
415
416 if (!target_structs)
417 {
418 target_struct_allocsize = DEFAULT_ALLOCSIZE;
419 target_structs = (struct target_ops **) xmalloc
420 (target_struct_allocsize * sizeof (*target_structs));
421 }
422 if (target_struct_size >= target_struct_allocsize)
423 {
424 target_struct_allocsize *= 2;
425 target_structs = (struct target_ops **)
426 xrealloc ((char *) target_structs,
427 target_struct_allocsize * sizeof (*target_structs));
428 }
429 target_structs[target_struct_size++] = t;
430
431 if (targetlist == NULL)
432 add_prefix_cmd ("target", class_run, target_command, _("\
433 Connect to a target machine or process.\n\
434 The first argument is the type or protocol of the target machine.\n\
435 Remaining arguments are interpreted by the target protocol. For more\n\
436 information on the arguments for a particular protocol, type\n\
437 `help target ' followed by the protocol name."),
438 &targetlist, "target ", 0, &cmdlist);
439 c = add_cmd (t->to_shortname, no_class, t->to_open, t->to_doc,
440 &targetlist);
441 if (completer != NULL)
442 set_cmd_completer (c, completer);
443 }
444
445 /* Add a possible target architecture to the list. */
446
447 void
448 add_target (struct target_ops *t)
449 {
450 add_target_with_completer (t, NULL);
451 }
452
453 /* See target.h. */
454
455 void
456 add_deprecated_target_alias (struct target_ops *t, char *alias)
457 {
458 struct cmd_list_element *c;
459 char *alt;
460
461 /* If we use add_alias_cmd, here, we do not get the deprecated warning,
462 see PR cli/15104. */
463 c = add_cmd (alias, no_class, t->to_open, t->to_doc, &targetlist);
464 alt = xstrprintf ("target %s", t->to_shortname);
465 deprecate_cmd (c, alt);
466 }
467
468 /* Stub functions */
469
470 void
471 target_ignore (void)
472 {
473 }
474
475 void
476 target_kill (void)
477 {
478 struct target_ops *t;
479
480 for (t = current_target.beneath; t != NULL; t = t->beneath)
481 if (t->to_kill != NULL)
482 {
483 if (targetdebug)
484 fprintf_unfiltered (gdb_stdlog, "target_kill ()\n");
485
486 t->to_kill (t);
487 return;
488 }
489
490 noprocess ();
491 }
492
493 void
494 target_load (char *arg, int from_tty)
495 {
496 target_dcache_invalidate ();
497 (*current_target.to_load) (arg, from_tty);
498 }
499
500 void
501 target_create_inferior (char *exec_file, char *args,
502 char **env, int from_tty)
503 {
504 struct target_ops *t;
505
506 for (t = current_target.beneath; t != NULL; t = t->beneath)
507 {
508 if (t->to_create_inferior != NULL)
509 {
510 t->to_create_inferior (t, exec_file, args, env, from_tty);
511 if (targetdebug)
512 fprintf_unfiltered (gdb_stdlog,
513 "target_create_inferior (%s, %s, xxx, %d)\n",
514 exec_file, args, from_tty);
515 return;
516 }
517 }
518
519 internal_error (__FILE__, __LINE__,
520 _("could not find a target to create inferior"));
521 }
522
523 void
524 target_terminal_inferior (void)
525 {
526 /* A background resume (``run&'') should leave GDB in control of the
527 terminal. Use target_can_async_p, not target_is_async_p, since at
528 this point the target is not async yet. However, if sync_execution
529 is not set, we know it will become async prior to resume. */
530 if (target_can_async_p () && !sync_execution)
531 return;
532
533 /* If GDB is resuming the inferior in the foreground, install
534 inferior's terminal modes. */
535 (*current_target.to_terminal_inferior) ();
536 }
537
538 static int
539 nomemory (CORE_ADDR memaddr, char *myaddr, int len, int write,
540 struct target_ops *t)
541 {
542 errno = EIO; /* Can't read/write this location. */
543 return 0; /* No bytes handled. */
544 }
545
546 static void
547 tcomplain (void)
548 {
549 error (_("You can't do that when your target is `%s'"),
550 current_target.to_shortname);
551 }
552
553 void
554 noprocess (void)
555 {
556 error (_("You can't do that without a process to debug."));
557 }
558
559 static void
560 default_terminal_info (const char *args, int from_tty)
561 {
562 printf_unfiltered (_("No saved terminal information.\n"));
563 }
564
565 /* A default implementation for the to_get_ada_task_ptid target method.
566
567 This function builds the PTID by using both LWP and TID as part of
568 the PTID lwp and tid elements. The pid used is the pid of the
569 inferior_ptid. */
570
571 static ptid_t
572 default_get_ada_task_ptid (long lwp, long tid)
573 {
574 return ptid_build (ptid_get_pid (inferior_ptid), lwp, tid);
575 }
576
577 static enum exec_direction_kind
578 default_execution_direction (void)
579 {
580 if (!target_can_execute_reverse)
581 return EXEC_FORWARD;
582 else if (!target_can_async_p ())
583 return EXEC_FORWARD;
584 else
585 gdb_assert_not_reached ("\
586 to_execution_direction must be implemented for reverse async");
587 }
588
589 /* Go through the target stack from top to bottom, copying over zero
590 entries in current_target, then filling in still empty entries. In
591 effect, we are doing class inheritance through the pushed target
592 vectors.
593
594 NOTE: cagney/2003-10-17: The problem with this inheritance, as it
595 is currently implemented, is that it discards any knowledge of
596 which target an inherited method originally belonged to.
597 Consequently, new new target methods should instead explicitly and
598 locally search the target stack for the target that can handle the
599 request. */
600
601 static void
602 update_current_target (void)
603 {
604 struct target_ops *t;
605
606 /* First, reset current's contents. */
607 memset (&current_target, 0, sizeof (current_target));
608
609 #define INHERIT(FIELD, TARGET) \
610 if (!current_target.FIELD) \
611 current_target.FIELD = (TARGET)->FIELD
612
613 for (t = target_stack; t; t = t->beneath)
614 {
615 INHERIT (to_shortname, t);
616 INHERIT (to_longname, t);
617 INHERIT (to_doc, t);
618 /* Do not inherit to_open. */
619 /* Do not inherit to_close. */
620 /* Do not inherit to_attach. */
621 INHERIT (to_post_attach, t);
622 INHERIT (to_attach_no_wait, t);
623 /* Do not inherit to_detach. */
624 /* Do not inherit to_disconnect. */
625 /* Do not inherit to_resume. */
626 /* Do not inherit to_wait. */
627 /* Do not inherit to_fetch_registers. */
628 /* Do not inherit to_store_registers. */
629 INHERIT (to_prepare_to_store, t);
630 INHERIT (deprecated_xfer_memory, t);
631 INHERIT (to_files_info, t);
632 INHERIT (to_insert_breakpoint, t);
633 INHERIT (to_remove_breakpoint, t);
634 INHERIT (to_can_use_hw_breakpoint, t);
635 INHERIT (to_insert_hw_breakpoint, t);
636 INHERIT (to_remove_hw_breakpoint, t);
637 /* Do not inherit to_ranged_break_num_registers. */
638 INHERIT (to_insert_watchpoint, t);
639 INHERIT (to_remove_watchpoint, t);
640 /* Do not inherit to_insert_mask_watchpoint. */
641 /* Do not inherit to_remove_mask_watchpoint. */
642 INHERIT (to_stopped_data_address, t);
643 INHERIT (to_have_steppable_watchpoint, t);
644 INHERIT (to_have_continuable_watchpoint, t);
645 INHERIT (to_stopped_by_watchpoint, t);
646 INHERIT (to_watchpoint_addr_within_range, t);
647 INHERIT (to_region_ok_for_hw_watchpoint, t);
648 INHERIT (to_can_accel_watchpoint_condition, t);
649 /* Do not inherit to_masked_watch_num_registers. */
650 INHERIT (to_terminal_init, t);
651 INHERIT (to_terminal_inferior, t);
652 INHERIT (to_terminal_ours_for_output, t);
653 INHERIT (to_terminal_ours, t);
654 INHERIT (to_terminal_save_ours, t);
655 INHERIT (to_terminal_info, t);
656 /* Do not inherit to_kill. */
657 INHERIT (to_load, t);
658 /* Do no inherit to_create_inferior. */
659 INHERIT (to_post_startup_inferior, t);
660 INHERIT (to_insert_fork_catchpoint, t);
661 INHERIT (to_remove_fork_catchpoint, t);
662 INHERIT (to_insert_vfork_catchpoint, t);
663 INHERIT (to_remove_vfork_catchpoint, t);
664 /* Do not inherit to_follow_fork. */
665 INHERIT (to_insert_exec_catchpoint, t);
666 INHERIT (to_remove_exec_catchpoint, t);
667 INHERIT (to_set_syscall_catchpoint, t);
668 INHERIT (to_has_exited, t);
669 /* Do not inherit to_mourn_inferior. */
670 INHERIT (to_can_run, t);
671 /* Do not inherit to_pass_signals. */
672 /* Do not inherit to_program_signals. */
673 /* Do not inherit to_thread_alive. */
674 /* Do not inherit to_find_new_threads. */
675 /* Do not inherit to_pid_to_str. */
676 INHERIT (to_extra_thread_info, t);
677 INHERIT (to_thread_name, t);
678 INHERIT (to_stop, t);
679 /* Do not inherit to_xfer_partial. */
680 INHERIT (to_rcmd, t);
681 INHERIT (to_pid_to_exec_file, t);
682 INHERIT (to_log_command, t);
683 INHERIT (to_stratum, t);
684 /* Do not inherit to_has_all_memory. */
685 /* Do not inherit to_has_memory. */
686 /* Do not inherit to_has_stack. */
687 /* Do not inherit to_has_registers. */
688 /* Do not inherit to_has_execution. */
689 INHERIT (to_has_thread_control, t);
690 INHERIT (to_can_async_p, t);
691 INHERIT (to_is_async_p, t);
692 INHERIT (to_async, t);
693 INHERIT (to_find_memory_regions, t);
694 INHERIT (to_make_corefile_notes, t);
695 INHERIT (to_get_bookmark, t);
696 INHERIT (to_goto_bookmark, t);
697 /* Do not inherit to_get_thread_local_address. */
698 INHERIT (to_can_execute_reverse, t);
699 INHERIT (to_execution_direction, t);
700 INHERIT (to_thread_architecture, t);
701 /* Do not inherit to_read_description. */
702 INHERIT (to_get_ada_task_ptid, t);
703 /* Do not inherit to_search_memory. */
704 INHERIT (to_supports_multi_process, t);
705 INHERIT (to_supports_enable_disable_tracepoint, t);
706 INHERIT (to_supports_string_tracing, t);
707 INHERIT (to_trace_init, t);
708 INHERIT (to_download_tracepoint, t);
709 INHERIT (to_can_download_tracepoint, t);
710 INHERIT (to_download_trace_state_variable, t);
711 INHERIT (to_enable_tracepoint, t);
712 INHERIT (to_disable_tracepoint, t);
713 INHERIT (to_trace_set_readonly_regions, t);
714 INHERIT (to_trace_start, t);
715 INHERIT (to_get_trace_status, t);
716 INHERIT (to_get_tracepoint_status, t);
717 INHERIT (to_trace_stop, t);
718 INHERIT (to_trace_find, t);
719 INHERIT (to_get_trace_state_variable_value, t);
720 INHERIT (to_save_trace_data, t);
721 INHERIT (to_upload_tracepoints, t);
722 INHERIT (to_upload_trace_state_variables, t);
723 INHERIT (to_get_raw_trace_data, t);
724 INHERIT (to_get_min_fast_tracepoint_insn_len, t);
725 INHERIT (to_set_disconnected_tracing, t);
726 INHERIT (to_set_circular_trace_buffer, t);
727 INHERIT (to_set_trace_buffer_size, t);
728 INHERIT (to_set_trace_notes, t);
729 INHERIT (to_get_tib_address, t);
730 INHERIT (to_set_permissions, t);
731 INHERIT (to_static_tracepoint_marker_at, t);
732 INHERIT (to_static_tracepoint_markers_by_strid, t);
733 INHERIT (to_traceframe_info, t);
734 INHERIT (to_use_agent, t);
735 INHERIT (to_can_use_agent, t);
736 INHERIT (to_augmented_libraries_svr4_read, t);
737 INHERIT (to_magic, t);
738 INHERIT (to_supports_evaluation_of_breakpoint_conditions, t);
739 INHERIT (to_can_run_breakpoint_commands, t);
740 /* Do not inherit to_memory_map. */
741 /* Do not inherit to_flash_erase. */
742 /* Do not inherit to_flash_done. */
743 }
744 #undef INHERIT
745
746 /* Clean up a target struct so it no longer has any zero pointers in
747 it. Some entries are defaulted to a method that print an error,
748 others are hard-wired to a standard recursive default. */
749
750 #define de_fault(field, value) \
751 if (!current_target.field) \
752 current_target.field = value
753
754 de_fault (to_open,
755 (void (*) (char *, int))
756 tcomplain);
757 de_fault (to_close,
758 (void (*) (void))
759 target_ignore);
760 de_fault (to_post_attach,
761 (void (*) (int))
762 target_ignore);
763 de_fault (to_prepare_to_store,
764 (void (*) (struct regcache *))
765 noprocess);
766 de_fault (deprecated_xfer_memory,
767 (int (*) (CORE_ADDR, gdb_byte *, int, int,
768 struct mem_attrib *, struct target_ops *))
769 nomemory);
770 de_fault (to_files_info,
771 (void (*) (struct target_ops *))
772 target_ignore);
773 de_fault (to_insert_breakpoint,
774 memory_insert_breakpoint);
775 de_fault (to_remove_breakpoint,
776 memory_remove_breakpoint);
777 de_fault (to_can_use_hw_breakpoint,
778 (int (*) (int, int, int))
779 return_zero);
780 de_fault (to_insert_hw_breakpoint,
781 (int (*) (struct gdbarch *, struct bp_target_info *))
782 return_minus_one);
783 de_fault (to_remove_hw_breakpoint,
784 (int (*) (struct gdbarch *, struct bp_target_info *))
785 return_minus_one);
786 de_fault (to_insert_watchpoint,
787 (int (*) (CORE_ADDR, int, int, struct expression *))
788 return_minus_one);
789 de_fault (to_remove_watchpoint,
790 (int (*) (CORE_ADDR, int, int, struct expression *))
791 return_minus_one);
792 de_fault (to_stopped_by_watchpoint,
793 (int (*) (void))
794 return_zero);
795 de_fault (to_stopped_data_address,
796 (int (*) (struct target_ops *, CORE_ADDR *))
797 return_zero);
798 de_fault (to_watchpoint_addr_within_range,
799 default_watchpoint_addr_within_range);
800 de_fault (to_region_ok_for_hw_watchpoint,
801 default_region_ok_for_hw_watchpoint);
802 de_fault (to_can_accel_watchpoint_condition,
803 (int (*) (CORE_ADDR, int, int, struct expression *))
804 return_zero);
805 de_fault (to_terminal_init,
806 (void (*) (void))
807 target_ignore);
808 de_fault (to_terminal_inferior,
809 (void (*) (void))
810 target_ignore);
811 de_fault (to_terminal_ours_for_output,
812 (void (*) (void))
813 target_ignore);
814 de_fault (to_terminal_ours,
815 (void (*) (void))
816 target_ignore);
817 de_fault (to_terminal_save_ours,
818 (void (*) (void))
819 target_ignore);
820 de_fault (to_terminal_info,
821 default_terminal_info);
822 de_fault (to_load,
823 (void (*) (char *, int))
824 tcomplain);
825 de_fault (to_post_startup_inferior,
826 (void (*) (ptid_t))
827 target_ignore);
828 de_fault (to_insert_fork_catchpoint,
829 (int (*) (int))
830 return_one);
831 de_fault (to_remove_fork_catchpoint,
832 (int (*) (int))
833 return_one);
834 de_fault (to_insert_vfork_catchpoint,
835 (int (*) (int))
836 return_one);
837 de_fault (to_remove_vfork_catchpoint,
838 (int (*) (int))
839 return_one);
840 de_fault (to_insert_exec_catchpoint,
841 (int (*) (int))
842 return_one);
843 de_fault (to_remove_exec_catchpoint,
844 (int (*) (int))
845 return_one);
846 de_fault (to_set_syscall_catchpoint,
847 (int (*) (int, int, int, int, int *))
848 return_one);
849 de_fault (to_has_exited,
850 (int (*) (int, int, int *))
851 return_zero);
852 de_fault (to_can_run,
853 return_zero);
854 de_fault (to_extra_thread_info,
855 (char *(*) (struct thread_info *))
856 return_zero);
857 de_fault (to_thread_name,
858 (char *(*) (struct thread_info *))
859 return_zero);
860 de_fault (to_stop,
861 (void (*) (ptid_t))
862 target_ignore);
863 current_target.to_xfer_partial = current_xfer_partial;
864 de_fault (to_rcmd,
865 (void (*) (char *, struct ui_file *))
866 tcomplain);
867 de_fault (to_pid_to_exec_file,
868 (char *(*) (int))
869 return_zero);
870 de_fault (to_async,
871 (void (*) (void (*) (enum inferior_event_type, void*), void*))
872 tcomplain);
873 de_fault (to_thread_architecture,
874 default_thread_architecture);
875 current_target.to_read_description = NULL;
876 de_fault (to_get_ada_task_ptid,
877 (ptid_t (*) (long, long))
878 default_get_ada_task_ptid);
879 de_fault (to_supports_multi_process,
880 (int (*) (void))
881 return_zero);
882 de_fault (to_supports_enable_disable_tracepoint,
883 (int (*) (void))
884 return_zero);
885 de_fault (to_supports_string_tracing,
886 (int (*) (void))
887 return_zero);
888 de_fault (to_trace_init,
889 (void (*) (void))
890 tcomplain);
891 de_fault (to_download_tracepoint,
892 (void (*) (struct bp_location *))
893 tcomplain);
894 de_fault (to_can_download_tracepoint,
895 (int (*) (void))
896 return_zero);
897 de_fault (to_download_trace_state_variable,
898 (void (*) (struct trace_state_variable *))
899 tcomplain);
900 de_fault (to_enable_tracepoint,
901 (void (*) (struct bp_location *))
902 tcomplain);
903 de_fault (to_disable_tracepoint,
904 (void (*) (struct bp_location *))
905 tcomplain);
906 de_fault (to_trace_set_readonly_regions,
907 (void (*) (void))
908 tcomplain);
909 de_fault (to_trace_start,
910 (void (*) (void))
911 tcomplain);
912 de_fault (to_get_trace_status,
913 (int (*) (struct trace_status *))
914 return_minus_one);
915 de_fault (to_get_tracepoint_status,
916 (void (*) (struct breakpoint *, struct uploaded_tp *))
917 tcomplain);
918 de_fault (to_trace_stop,
919 (void (*) (void))
920 tcomplain);
921 de_fault (to_trace_find,
922 (int (*) (enum trace_find_type, int, CORE_ADDR, CORE_ADDR, int *))
923 return_minus_one);
924 de_fault (to_get_trace_state_variable_value,
925 (int (*) (int, LONGEST *))
926 return_zero);
927 de_fault (to_save_trace_data,
928 (int (*) (const char *))
929 tcomplain);
930 de_fault (to_upload_tracepoints,
931 (int (*) (struct uploaded_tp **))
932 return_zero);
933 de_fault (to_upload_trace_state_variables,
934 (int (*) (struct uploaded_tsv **))
935 return_zero);
936 de_fault (to_get_raw_trace_data,
937 (LONGEST (*) (gdb_byte *, ULONGEST, LONGEST))
938 tcomplain);
939 de_fault (to_get_min_fast_tracepoint_insn_len,
940 (int (*) (void))
941 return_minus_one);
942 de_fault (to_set_disconnected_tracing,
943 (void (*) (int))
944 target_ignore);
945 de_fault (to_set_circular_trace_buffer,
946 (void (*) (int))
947 target_ignore);
948 de_fault (to_set_trace_buffer_size,
949 (void (*) (LONGEST))
950 target_ignore);
951 de_fault (to_set_trace_notes,
952 (int (*) (const char *, const char *, const char *))
953 return_zero);
954 de_fault (to_get_tib_address,
955 (int (*) (ptid_t, CORE_ADDR *))
956 tcomplain);
957 de_fault (to_set_permissions,
958 (void (*) (void))
959 target_ignore);
960 de_fault (to_static_tracepoint_marker_at,
961 (int (*) (CORE_ADDR, struct static_tracepoint_marker *))
962 return_zero);
963 de_fault (to_static_tracepoint_markers_by_strid,
964 (VEC(static_tracepoint_marker_p) * (*) (const char *))
965 tcomplain);
966 de_fault (to_traceframe_info,
967 (struct traceframe_info * (*) (void))
968 return_zero);
969 de_fault (to_supports_evaluation_of_breakpoint_conditions,
970 (int (*) (void))
971 return_zero);
972 de_fault (to_can_run_breakpoint_commands,
973 (int (*) (void))
974 return_zero);
975 de_fault (to_use_agent,
976 (int (*) (int))
977 tcomplain);
978 de_fault (to_can_use_agent,
979 (int (*) (void))
980 return_zero);
981 de_fault (to_augmented_libraries_svr4_read,
982 (int (*) (void))
983 return_zero);
984 de_fault (to_execution_direction, default_execution_direction);
985
986 #undef de_fault
987
988 /* Finally, position the target-stack beneath the squashed
989 "current_target". That way code looking for a non-inherited
990 target method can quickly and simply find it. */
991 current_target.beneath = target_stack;
992
993 if (targetdebug)
994 setup_target_debug ();
995 }
996
997 /* Push a new target type into the stack of the existing target accessors,
998 possibly superseding some of the existing accessors.
999
1000 Rather than allow an empty stack, we always have the dummy target at
1001 the bottom stratum, so we can call the function vectors without
1002 checking them. */
1003
1004 void
1005 push_target (struct target_ops *t)
1006 {
1007 struct target_ops **cur;
1008
1009 /* Check magic number. If wrong, it probably means someone changed
1010 the struct definition, but not all the places that initialize one. */
1011 if (t->to_magic != OPS_MAGIC)
1012 {
1013 fprintf_unfiltered (gdb_stderr,
1014 "Magic number of %s target struct wrong\n",
1015 t->to_shortname);
1016 internal_error (__FILE__, __LINE__,
1017 _("failed internal consistency check"));
1018 }
1019
1020 /* Find the proper stratum to install this target in. */
1021 for (cur = &target_stack; (*cur) != NULL; cur = &(*cur)->beneath)
1022 {
1023 if ((int) (t->to_stratum) >= (int) (*cur)->to_stratum)
1024 break;
1025 }
1026
1027 /* If there's already targets at this stratum, remove them. */
1028 /* FIXME: cagney/2003-10-15: I think this should be popping all
1029 targets to CUR, and not just those at this stratum level. */
1030 while ((*cur) != NULL && t->to_stratum == (*cur)->to_stratum)
1031 {
1032 /* There's already something at this stratum level. Close it,
1033 and un-hook it from the stack. */
1034 struct target_ops *tmp = (*cur);
1035
1036 (*cur) = (*cur)->beneath;
1037 tmp->beneath = NULL;
1038 target_close (tmp);
1039 }
1040
1041 /* We have removed all targets in our stratum, now add the new one. */
1042 t->beneath = (*cur);
1043 (*cur) = t;
1044
1045 update_current_target ();
1046 }
1047
1048 /* Remove a target_ops vector from the stack, wherever it may be.
1049 Return how many times it was removed (0 or 1). */
1050
1051 int
1052 unpush_target (struct target_ops *t)
1053 {
1054 struct target_ops **cur;
1055 struct target_ops *tmp;
1056
1057 if (t->to_stratum == dummy_stratum)
1058 internal_error (__FILE__, __LINE__,
1059 _("Attempt to unpush the dummy target"));
1060
1061 /* Look for the specified target. Note that we assume that a target
1062 can only occur once in the target stack. */
1063
1064 for (cur = &target_stack; (*cur) != NULL; cur = &(*cur)->beneath)
1065 {
1066 if ((*cur) == t)
1067 break;
1068 }
1069
1070 /* If we don't find target_ops, quit. Only open targets should be
1071 closed. */
1072 if ((*cur) == NULL)
1073 return 0;
1074
1075 /* Unchain the target. */
1076 tmp = (*cur);
1077 (*cur) = (*cur)->beneath;
1078 tmp->beneath = NULL;
1079
1080 update_current_target ();
1081
1082 /* Finally close the target. Note we do this after unchaining, so
1083 any target method calls from within the target_close
1084 implementation don't end up in T anymore. */
1085 target_close (t);
1086
1087 return 1;
1088 }
1089
1090 void
1091 pop_all_targets_above (enum strata above_stratum)
1092 {
1093 while ((int) (current_target.to_stratum) > (int) above_stratum)
1094 {
1095 if (!unpush_target (target_stack))
1096 {
1097 fprintf_unfiltered (gdb_stderr,
1098 "pop_all_targets couldn't find target %s\n",
1099 target_stack->to_shortname);
1100 internal_error (__FILE__, __LINE__,
1101 _("failed internal consistency check"));
1102 break;
1103 }
1104 }
1105 }
1106
1107 void
1108 pop_all_targets (void)
1109 {
1110 pop_all_targets_above (dummy_stratum);
1111 }
1112
1113 /* Return 1 if T is now pushed in the target stack. Return 0 otherwise. */
1114
1115 int
1116 target_is_pushed (struct target_ops *t)
1117 {
1118 struct target_ops **cur;
1119
1120 /* Check magic number. If wrong, it probably means someone changed
1121 the struct definition, but not all the places that initialize one. */
1122 if (t->to_magic != OPS_MAGIC)
1123 {
1124 fprintf_unfiltered (gdb_stderr,
1125 "Magic number of %s target struct wrong\n",
1126 t->to_shortname);
1127 internal_error (__FILE__, __LINE__,
1128 _("failed internal consistency check"));
1129 }
1130
1131 for (cur = &target_stack; (*cur) != NULL; cur = &(*cur)->beneath)
1132 if (*cur == t)
1133 return 1;
1134
1135 return 0;
1136 }
1137
1138 /* Using the objfile specified in OBJFILE, find the address for the
1139 current thread's thread-local storage with offset OFFSET. */
1140 CORE_ADDR
1141 target_translate_tls_address (struct objfile *objfile, CORE_ADDR offset)
1142 {
1143 volatile CORE_ADDR addr = 0;
1144 struct target_ops *target;
1145
1146 for (target = current_target.beneath;
1147 target != NULL;
1148 target = target->beneath)
1149 {
1150 if (target->to_get_thread_local_address != NULL)
1151 break;
1152 }
1153
1154 if (target != NULL
1155 && gdbarch_fetch_tls_load_module_address_p (target_gdbarch ()))
1156 {
1157 ptid_t ptid = inferior_ptid;
1158 volatile struct gdb_exception ex;
1159
1160 TRY_CATCH (ex, RETURN_MASK_ALL)
1161 {
1162 CORE_ADDR lm_addr;
1163
1164 /* Fetch the load module address for this objfile. */
1165 lm_addr = gdbarch_fetch_tls_load_module_address (target_gdbarch (),
1166 objfile);
1167 /* If it's 0, throw the appropriate exception. */
1168 if (lm_addr == 0)
1169 throw_error (TLS_LOAD_MODULE_NOT_FOUND_ERROR,
1170 _("TLS load module not found"));
1171
1172 addr = target->to_get_thread_local_address (target, ptid,
1173 lm_addr, offset);
1174 }
1175 /* If an error occurred, print TLS related messages here. Otherwise,
1176 throw the error to some higher catcher. */
1177 if (ex.reason < 0)
1178 {
1179 int objfile_is_library = (objfile->flags & OBJF_SHARED);
1180
1181 switch (ex.error)
1182 {
1183 case TLS_NO_LIBRARY_SUPPORT_ERROR:
1184 error (_("Cannot find thread-local variables "
1185 "in this thread library."));
1186 break;
1187 case TLS_LOAD_MODULE_NOT_FOUND_ERROR:
1188 if (objfile_is_library)
1189 error (_("Cannot find shared library `%s' in dynamic"
1190 " linker's load module list"), objfile_name (objfile));
1191 else
1192 error (_("Cannot find executable file `%s' in dynamic"
1193 " linker's load module list"), objfile_name (objfile));
1194 break;
1195 case TLS_NOT_ALLOCATED_YET_ERROR:
1196 if (objfile_is_library)
1197 error (_("The inferior has not yet allocated storage for"
1198 " thread-local variables in\n"
1199 "the shared library `%s'\n"
1200 "for %s"),
1201 objfile_name (objfile), target_pid_to_str (ptid));
1202 else
1203 error (_("The inferior has not yet allocated storage for"
1204 " thread-local variables in\n"
1205 "the executable `%s'\n"
1206 "for %s"),
1207 objfile_name (objfile), target_pid_to_str (ptid));
1208 break;
1209 case TLS_GENERIC_ERROR:
1210 if (objfile_is_library)
1211 error (_("Cannot find thread-local storage for %s, "
1212 "shared library %s:\n%s"),
1213 target_pid_to_str (ptid),
1214 objfile_name (objfile), ex.message);
1215 else
1216 error (_("Cannot find thread-local storage for %s, "
1217 "executable file %s:\n%s"),
1218 target_pid_to_str (ptid),
1219 objfile_name (objfile), ex.message);
1220 break;
1221 default:
1222 throw_exception (ex);
1223 break;
1224 }
1225 }
1226 }
1227 /* It wouldn't be wrong here to try a gdbarch method, too; finding
1228 TLS is an ABI-specific thing. But we don't do that yet. */
1229 else
1230 error (_("Cannot find thread-local variables on this target"));
1231
1232 return addr;
1233 }
1234
1235 const char *
1236 target_xfer_error_to_string (enum target_xfer_error err)
1237 {
1238 #define CASE(X) case X: return #X
1239 switch (err)
1240 {
1241 CASE(TARGET_XFER_E_IO);
1242 CASE(TARGET_XFER_E_UNAVAILABLE);
1243 default:
1244 return "<unknown>";
1245 }
1246 #undef CASE
1247 };
1248
1249
1250 #undef MIN
1251 #define MIN(A, B) (((A) <= (B)) ? (A) : (B))
1252
1253 /* target_read_string -- read a null terminated string, up to LEN bytes,
1254 from MEMADDR in target. Set *ERRNOP to the errno code, or 0 if successful.
1255 Set *STRING to a pointer to malloc'd memory containing the data; the caller
1256 is responsible for freeing it. Return the number of bytes successfully
1257 read. */
1258
1259 int
1260 target_read_string (CORE_ADDR memaddr, char **string, int len, int *errnop)
1261 {
1262 int tlen, offset, i;
1263 gdb_byte buf[4];
1264 int errcode = 0;
1265 char *buffer;
1266 int buffer_allocated;
1267 char *bufptr;
1268 unsigned int nbytes_read = 0;
1269
1270 gdb_assert (string);
1271
1272 /* Small for testing. */
1273 buffer_allocated = 4;
1274 buffer = xmalloc (buffer_allocated);
1275 bufptr = buffer;
1276
1277 while (len > 0)
1278 {
1279 tlen = MIN (len, 4 - (memaddr & 3));
1280 offset = memaddr & 3;
1281
1282 errcode = target_read_memory (memaddr & ~3, buf, sizeof buf);
1283 if (errcode != 0)
1284 {
1285 /* The transfer request might have crossed the boundary to an
1286 unallocated region of memory. Retry the transfer, requesting
1287 a single byte. */
1288 tlen = 1;
1289 offset = 0;
1290 errcode = target_read_memory (memaddr, buf, 1);
1291 if (errcode != 0)
1292 goto done;
1293 }
1294
1295 if (bufptr - buffer + tlen > buffer_allocated)
1296 {
1297 unsigned int bytes;
1298
1299 bytes = bufptr - buffer;
1300 buffer_allocated *= 2;
1301 buffer = xrealloc (buffer, buffer_allocated);
1302 bufptr = buffer + bytes;
1303 }
1304
1305 for (i = 0; i < tlen; i++)
1306 {
1307 *bufptr++ = buf[i + offset];
1308 if (buf[i + offset] == '\000')
1309 {
1310 nbytes_read += i + 1;
1311 goto done;
1312 }
1313 }
1314
1315 memaddr += tlen;
1316 len -= tlen;
1317 nbytes_read += tlen;
1318 }
1319 done:
1320 *string = buffer;
1321 if (errnop != NULL)
1322 *errnop = errcode;
1323 return nbytes_read;
1324 }
1325
1326 struct target_section_table *
1327 target_get_section_table (struct target_ops *target)
1328 {
1329 struct target_ops *t;
1330
1331 if (targetdebug)
1332 fprintf_unfiltered (gdb_stdlog, "target_get_section_table ()\n");
1333
1334 for (t = target; t != NULL; t = t->beneath)
1335 if (t->to_get_section_table != NULL)
1336 return (*t->to_get_section_table) (t);
1337
1338 return NULL;
1339 }
1340
1341 /* Find a section containing ADDR. */
1342
1343 struct target_section *
1344 target_section_by_addr (struct target_ops *target, CORE_ADDR addr)
1345 {
1346 struct target_section_table *table = target_get_section_table (target);
1347 struct target_section *secp;
1348
1349 if (table == NULL)
1350 return NULL;
1351
1352 for (secp = table->sections; secp < table->sections_end; secp++)
1353 {
1354 if (addr >= secp->addr && addr < secp->endaddr)
1355 return secp;
1356 }
1357 return NULL;
1358 }
1359
1360 /* Read memory from the live target, even if currently inspecting a
1361 traceframe. The return is the same as that of target_read. */
1362
1363 static LONGEST
1364 target_read_live_memory (enum target_object object,
1365 ULONGEST memaddr, gdb_byte *myaddr, LONGEST len)
1366 {
1367 LONGEST ret;
1368 struct cleanup *cleanup;
1369
1370 /* Switch momentarily out of tfind mode so to access live memory.
1371 Note that this must not clear global state, such as the frame
1372 cache, which must still remain valid for the previous traceframe.
1373 We may be _building_ the frame cache at this point. */
1374 cleanup = make_cleanup_restore_traceframe_number ();
1375 set_traceframe_number (-1);
1376
1377 ret = target_read (current_target.beneath, object, NULL,
1378 myaddr, memaddr, len);
1379
1380 do_cleanups (cleanup);
1381 return ret;
1382 }
1383
1384 /* Using the set of read-only target sections of OPS, read live
1385 read-only memory. Note that the actual reads start from the
1386 top-most target again.
1387
1388 For interface/parameters/return description see target.h,
1389 to_xfer_partial. */
1390
1391 static LONGEST
1392 memory_xfer_live_readonly_partial (struct target_ops *ops,
1393 enum target_object object,
1394 gdb_byte *readbuf, ULONGEST memaddr,
1395 LONGEST len)
1396 {
1397 struct target_section *secp;
1398 struct target_section_table *table;
1399
1400 secp = target_section_by_addr (ops, memaddr);
1401 if (secp != NULL
1402 && (bfd_get_section_flags (secp->the_bfd_section->owner,
1403 secp->the_bfd_section)
1404 & SEC_READONLY))
1405 {
1406 struct target_section *p;
1407 ULONGEST memend = memaddr + len;
1408
1409 table = target_get_section_table (ops);
1410
1411 for (p = table->sections; p < table->sections_end; p++)
1412 {
1413 if (memaddr >= p->addr)
1414 {
1415 if (memend <= p->endaddr)
1416 {
1417 /* Entire transfer is within this section. */
1418 return target_read_live_memory (object, memaddr,
1419 readbuf, len);
1420 }
1421 else if (memaddr >= p->endaddr)
1422 {
1423 /* This section ends before the transfer starts. */
1424 continue;
1425 }
1426 else
1427 {
1428 /* This section overlaps the transfer. Just do half. */
1429 len = p->endaddr - memaddr;
1430 return target_read_live_memory (object, memaddr,
1431 readbuf, len);
1432 }
1433 }
1434 }
1435 }
1436
1437 return 0;
1438 }
1439
1440 /* Perform a partial memory transfer.
1441 For docs see target.h, to_xfer_partial. */
1442
1443 static LONGEST
1444 memory_xfer_partial_1 (struct target_ops *ops, enum target_object object,
1445 void *readbuf, const void *writebuf, ULONGEST memaddr,
1446 LONGEST len)
1447 {
1448 LONGEST res;
1449 int reg_len;
1450 struct mem_region *region;
1451 struct inferior *inf;
1452
1453 /* For accesses to unmapped overlay sections, read directly from
1454 files. Must do this first, as MEMADDR may need adjustment. */
1455 if (readbuf != NULL && overlay_debugging)
1456 {
1457 struct obj_section *section = find_pc_overlay (memaddr);
1458
1459 if (pc_in_unmapped_range (memaddr, section))
1460 {
1461 struct target_section_table *table
1462 = target_get_section_table (ops);
1463 const char *section_name = section->the_bfd_section->name;
1464
1465 memaddr = overlay_mapped_address (memaddr, section);
1466 return section_table_xfer_memory_partial (readbuf, writebuf,
1467 memaddr, len,
1468 table->sections,
1469 table->sections_end,
1470 section_name);
1471 }
1472 }
1473
1474 /* Try the executable files, if "trust-readonly-sections" is set. */
1475 if (readbuf != NULL && trust_readonly)
1476 {
1477 struct target_section *secp;
1478 struct target_section_table *table;
1479
1480 secp = target_section_by_addr (ops, memaddr);
1481 if (secp != NULL
1482 && (bfd_get_section_flags (secp->the_bfd_section->owner,
1483 secp->the_bfd_section)
1484 & SEC_READONLY))
1485 {
1486 table = target_get_section_table (ops);
1487 return section_table_xfer_memory_partial (readbuf, writebuf,
1488 memaddr, len,
1489 table->sections,
1490 table->sections_end,
1491 NULL);
1492 }
1493 }
1494
1495 /* If reading unavailable memory in the context of traceframes, and
1496 this address falls within a read-only section, fallback to
1497 reading from live memory. */
1498 if (readbuf != NULL && get_traceframe_number () != -1)
1499 {
1500 VEC(mem_range_s) *available;
1501
1502 /* If we fail to get the set of available memory, then the
1503 target does not support querying traceframe info, and so we
1504 attempt reading from the traceframe anyway (assuming the
1505 target implements the old QTro packet then). */
1506 if (traceframe_available_memory (&available, memaddr, len))
1507 {
1508 struct cleanup *old_chain;
1509
1510 old_chain = make_cleanup (VEC_cleanup(mem_range_s), &available);
1511
1512 if (VEC_empty (mem_range_s, available)
1513 || VEC_index (mem_range_s, available, 0)->start != memaddr)
1514 {
1515 /* Don't read into the traceframe's available
1516 memory. */
1517 if (!VEC_empty (mem_range_s, available))
1518 {
1519 LONGEST oldlen = len;
1520
1521 len = VEC_index (mem_range_s, available, 0)->start - memaddr;
1522 gdb_assert (len <= oldlen);
1523 }
1524
1525 do_cleanups (old_chain);
1526
1527 /* This goes through the topmost target again. */
1528 res = memory_xfer_live_readonly_partial (ops, object,
1529 readbuf, memaddr, len);
1530 if (res > 0)
1531 return res;
1532
1533 /* No use trying further, we know some memory starting
1534 at MEMADDR isn't available. */
1535 return TARGET_XFER_E_UNAVAILABLE;
1536 }
1537
1538 /* Don't try to read more than how much is available, in
1539 case the target implements the deprecated QTro packet to
1540 cater for older GDBs (the target's knowledge of read-only
1541 sections may be outdated by now). */
1542 len = VEC_index (mem_range_s, available, 0)->length;
1543
1544 do_cleanups (old_chain);
1545 }
1546 }
1547
1548 /* Try GDB's internal data cache. */
1549 region = lookup_mem_region (memaddr);
1550 /* region->hi == 0 means there's no upper bound. */
1551 if (memaddr + len < region->hi || region->hi == 0)
1552 reg_len = len;
1553 else
1554 reg_len = region->hi - memaddr;
1555
1556 switch (region->attrib.mode)
1557 {
1558 case MEM_RO:
1559 if (writebuf != NULL)
1560 return -1;
1561 break;
1562
1563 case MEM_WO:
1564 if (readbuf != NULL)
1565 return -1;
1566 break;
1567
1568 case MEM_FLASH:
1569 /* We only support writing to flash during "load" for now. */
1570 if (writebuf != NULL)
1571 error (_("Writing to flash memory forbidden in this context"));
1572 break;
1573
1574 case MEM_NONE:
1575 return -1;
1576 }
1577
1578 if (!ptid_equal (inferior_ptid, null_ptid))
1579 inf = find_inferior_pid (ptid_get_pid (inferior_ptid));
1580 else
1581 inf = NULL;
1582
1583 if (inf != NULL
1584 /* The dcache reads whole cache lines; that doesn't play well
1585 with reading from a trace buffer, because reading outside of
1586 the collected memory range fails. */
1587 && get_traceframe_number () == -1
1588 && (region->attrib.cache
1589 || (stack_cache_enabled_p && object == TARGET_OBJECT_STACK_MEMORY)))
1590 {
1591 if (readbuf != NULL)
1592 res = dcache_xfer_memory (ops, target_dcache, memaddr, readbuf,
1593 reg_len, 0);
1594 else
1595 /* FIXME drow/2006-08-09: If we're going to preserve const
1596 correctness dcache_xfer_memory should take readbuf and
1597 writebuf. */
1598 res = dcache_xfer_memory (ops, target_dcache, memaddr,
1599 (void *) writebuf,
1600 reg_len, 1);
1601 if (res <= 0)
1602 return -1;
1603 else
1604 return res;
1605 }
1606
1607 /* If none of those methods found the memory we wanted, fall back
1608 to a target partial transfer. Normally a single call to
1609 to_xfer_partial is enough; if it doesn't recognize an object
1610 it will call the to_xfer_partial of the next target down.
1611 But for memory this won't do. Memory is the only target
1612 object which can be read from more than one valid target.
1613 A core file, for instance, could have some of memory but
1614 delegate other bits to the target below it. So, we must
1615 manually try all targets. */
1616
1617 do
1618 {
1619 res = ops->to_xfer_partial (ops, TARGET_OBJECT_MEMORY, NULL,
1620 readbuf, writebuf, memaddr, reg_len);
1621 if (res > 0)
1622 break;
1623
1624 /* We want to continue past core files to executables, but not
1625 past a running target's memory. */
1626 if (ops->to_has_all_memory (ops))
1627 break;
1628
1629 ops = ops->beneath;
1630 }
1631 while (ops != NULL);
1632
1633 /* Make sure the cache gets updated no matter what - if we are writing
1634 to the stack. Even if this write is not tagged as such, we still need
1635 to update the cache. */
1636
1637 if (res > 0
1638 && inf != NULL
1639 && writebuf != NULL
1640 && !region->attrib.cache
1641 && stack_cache_enabled_p
1642 && object != TARGET_OBJECT_STACK_MEMORY)
1643 {
1644 dcache_update (target_dcache, memaddr, (void *) writebuf, res);
1645 }
1646
1647 /* If we still haven't got anything, return the last error. We
1648 give up. */
1649 return res;
1650 }
1651
1652 /* Perform a partial memory transfer. For docs see target.h,
1653 to_xfer_partial. */
1654
1655 static LONGEST
1656 memory_xfer_partial (struct target_ops *ops, enum target_object object,
1657 void *readbuf, const void *writebuf, ULONGEST memaddr,
1658 LONGEST len)
1659 {
1660 int res;
1661
1662 /* Zero length requests are ok and require no work. */
1663 if (len == 0)
1664 return 0;
1665
1666 /* Fill in READBUF with breakpoint shadows, or WRITEBUF with
1667 breakpoint insns, thus hiding out from higher layers whether
1668 there are software breakpoints inserted in the code stream. */
1669 if (readbuf != NULL)
1670 {
1671 res = memory_xfer_partial_1 (ops, object, readbuf, NULL, memaddr, len);
1672
1673 if (res > 0 && !show_memory_breakpoints)
1674 breakpoint_xfer_memory (readbuf, NULL, NULL, memaddr, res);
1675 }
1676 else
1677 {
1678 void *buf;
1679 struct cleanup *old_chain;
1680
1681 buf = xmalloc (len);
1682 old_chain = make_cleanup (xfree, buf);
1683 memcpy (buf, writebuf, len);
1684
1685 breakpoint_xfer_memory (NULL, buf, writebuf, memaddr, len);
1686 res = memory_xfer_partial_1 (ops, object, NULL, buf, memaddr, len);
1687
1688 do_cleanups (old_chain);
1689 }
1690
1691 return res;
1692 }
1693
1694 static void
1695 restore_show_memory_breakpoints (void *arg)
1696 {
1697 show_memory_breakpoints = (uintptr_t) arg;
1698 }
1699
1700 struct cleanup *
1701 make_show_memory_breakpoints_cleanup (int show)
1702 {
1703 int current = show_memory_breakpoints;
1704
1705 show_memory_breakpoints = show;
1706 return make_cleanup (restore_show_memory_breakpoints,
1707 (void *) (uintptr_t) current);
1708 }
1709
1710 /* For docs see target.h, to_xfer_partial. */
1711
1712 LONGEST
1713 target_xfer_partial (struct target_ops *ops,
1714 enum target_object object, const char *annex,
1715 void *readbuf, const void *writebuf,
1716 ULONGEST offset, LONGEST len)
1717 {
1718 LONGEST retval;
1719
1720 gdb_assert (ops->to_xfer_partial != NULL);
1721
1722 if (writebuf && !may_write_memory)
1723 error (_("Writing to memory is not allowed (addr %s, len %s)"),
1724 core_addr_to_string_nz (offset), plongest (len));
1725
1726 /* If this is a memory transfer, let the memory-specific code
1727 have a look at it instead. Memory transfers are more
1728 complicated. */
1729 if (object == TARGET_OBJECT_MEMORY || object == TARGET_OBJECT_STACK_MEMORY)
1730 retval = memory_xfer_partial (ops, object, readbuf,
1731 writebuf, offset, len);
1732 else
1733 {
1734 enum target_object raw_object = object;
1735
1736 /* If this is a raw memory transfer, request the normal
1737 memory object from other layers. */
1738 if (raw_object == TARGET_OBJECT_RAW_MEMORY)
1739 raw_object = TARGET_OBJECT_MEMORY;
1740
1741 retval = ops->to_xfer_partial (ops, raw_object, annex, readbuf,
1742 writebuf, offset, len);
1743 }
1744
1745 if (targetdebug)
1746 {
1747 const unsigned char *myaddr = NULL;
1748
1749 fprintf_unfiltered (gdb_stdlog,
1750 "%s:target_xfer_partial "
1751 "(%d, %s, %s, %s, %s, %s) = %s",
1752 ops->to_shortname,
1753 (int) object,
1754 (annex ? annex : "(null)"),
1755 host_address_to_string (readbuf),
1756 host_address_to_string (writebuf),
1757 core_addr_to_string_nz (offset),
1758 plongest (len), plongest (retval));
1759
1760 if (readbuf)
1761 myaddr = readbuf;
1762 if (writebuf)
1763 myaddr = writebuf;
1764 if (retval > 0 && myaddr != NULL)
1765 {
1766 int i;
1767
1768 fputs_unfiltered (", bytes =", gdb_stdlog);
1769 for (i = 0; i < retval; i++)
1770 {
1771 if ((((intptr_t) &(myaddr[i])) & 0xf) == 0)
1772 {
1773 if (targetdebug < 2 && i > 0)
1774 {
1775 fprintf_unfiltered (gdb_stdlog, " ...");
1776 break;
1777 }
1778 fprintf_unfiltered (gdb_stdlog, "\n");
1779 }
1780
1781 fprintf_unfiltered (gdb_stdlog, " %02x", myaddr[i] & 0xff);
1782 }
1783 }
1784
1785 fputc_unfiltered ('\n', gdb_stdlog);
1786 }
1787 return retval;
1788 }
1789
1790 /* Read LEN bytes of target memory at address MEMADDR, placing the results in
1791 GDB's memory at MYADDR. Returns either 0 for success or an errno value
1792 if any error occurs.
1793
1794 If an error occurs, no guarantee is made about the contents of the data at
1795 MYADDR. In particular, the caller should not depend upon partial reads
1796 filling the buffer with good data. There is no way for the caller to know
1797 how much good data might have been transfered anyway. Callers that can
1798 deal with partial reads should call target_read (which will retry until
1799 it makes no progress, and then return how much was transferred). */
1800
1801 int
1802 target_read_memory (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1803 {
1804 /* Dispatch to the topmost target, not the flattened current_target.
1805 Memory accesses check target->to_has_(all_)memory, and the
1806 flattened target doesn't inherit those. */
1807 if (target_read (current_target.beneath, TARGET_OBJECT_MEMORY, NULL,
1808 myaddr, memaddr, len) == len)
1809 return 0;
1810 else
1811 return EIO;
1812 }
1813
1814 /* Like target_read_memory, but specify explicitly that this is a read from
1815 the target's stack. This may trigger different cache behavior. */
1816
1817 int
1818 target_read_stack (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1819 {
1820 /* Dispatch to the topmost target, not the flattened current_target.
1821 Memory accesses check target->to_has_(all_)memory, and the
1822 flattened target doesn't inherit those. */
1823
1824 if (target_read (current_target.beneath, TARGET_OBJECT_STACK_MEMORY, NULL,
1825 myaddr, memaddr, len) == len)
1826 return 0;
1827 else
1828 return EIO;
1829 }
1830
1831 /* Write LEN bytes from MYADDR to target memory at address MEMADDR.
1832 Returns either 0 for success or an errno value if any error occurs.
1833 If an error occurs, no guarantee is made about how much data got written.
1834 Callers that can deal with partial writes should call target_write. */
1835
1836 int
1837 target_write_memory (CORE_ADDR memaddr, const gdb_byte *myaddr, ssize_t len)
1838 {
1839 /* Dispatch to the topmost target, not the flattened current_target.
1840 Memory accesses check target->to_has_(all_)memory, and the
1841 flattened target doesn't inherit those. */
1842 if (target_write (current_target.beneath, TARGET_OBJECT_MEMORY, NULL,
1843 myaddr, memaddr, len) == len)
1844 return 0;
1845 else
1846 return EIO;
1847 }
1848
1849 /* Write LEN bytes from MYADDR to target raw memory at address
1850 MEMADDR. Returns either 0 for success or an errno value if any
1851 error occurs. If an error occurs, no guarantee is made about how
1852 much data got written. Callers that can deal with partial writes
1853 should call target_write. */
1854
1855 int
1856 target_write_raw_memory (CORE_ADDR memaddr, const gdb_byte *myaddr, ssize_t len)
1857 {
1858 /* Dispatch to the topmost target, not the flattened current_target.
1859 Memory accesses check target->to_has_(all_)memory, and the
1860 flattened target doesn't inherit those. */
1861 if (target_write (current_target.beneath, TARGET_OBJECT_RAW_MEMORY, NULL,
1862 myaddr, memaddr, len) == len)
1863 return 0;
1864 else
1865 return EIO;
1866 }
1867
1868 /* Fetch the target's memory map. */
1869
1870 VEC(mem_region_s) *
1871 target_memory_map (void)
1872 {
1873 VEC(mem_region_s) *result;
1874 struct mem_region *last_one, *this_one;
1875 int ix;
1876 struct target_ops *t;
1877
1878 if (targetdebug)
1879 fprintf_unfiltered (gdb_stdlog, "target_memory_map ()\n");
1880
1881 for (t = current_target.beneath; t != NULL; t = t->beneath)
1882 if (t->to_memory_map != NULL)
1883 break;
1884
1885 if (t == NULL)
1886 return NULL;
1887
1888 result = t->to_memory_map (t);
1889 if (result == NULL)
1890 return NULL;
1891
1892 qsort (VEC_address (mem_region_s, result),
1893 VEC_length (mem_region_s, result),
1894 sizeof (struct mem_region), mem_region_cmp);
1895
1896 /* Check that regions do not overlap. Simultaneously assign
1897 a numbering for the "mem" commands to use to refer to
1898 each region. */
1899 last_one = NULL;
1900 for (ix = 0; VEC_iterate (mem_region_s, result, ix, this_one); ix++)
1901 {
1902 this_one->number = ix;
1903
1904 if (last_one && last_one->hi > this_one->lo)
1905 {
1906 warning (_("Overlapping regions in memory map: ignoring"));
1907 VEC_free (mem_region_s, result);
1908 return NULL;
1909 }
1910 last_one = this_one;
1911 }
1912
1913 return result;
1914 }
1915
1916 void
1917 target_flash_erase (ULONGEST address, LONGEST length)
1918 {
1919 struct target_ops *t;
1920
1921 for (t = current_target.beneath; t != NULL; t = t->beneath)
1922 if (t->to_flash_erase != NULL)
1923 {
1924 if (targetdebug)
1925 fprintf_unfiltered (gdb_stdlog, "target_flash_erase (%s, %s)\n",
1926 hex_string (address), phex (length, 0));
1927 t->to_flash_erase (t, address, length);
1928 return;
1929 }
1930
1931 tcomplain ();
1932 }
1933
1934 void
1935 target_flash_done (void)
1936 {
1937 struct target_ops *t;
1938
1939 for (t = current_target.beneath; t != NULL; t = t->beneath)
1940 if (t->to_flash_done != NULL)
1941 {
1942 if (targetdebug)
1943 fprintf_unfiltered (gdb_stdlog, "target_flash_done\n");
1944 t->to_flash_done (t);
1945 return;
1946 }
1947
1948 tcomplain ();
1949 }
1950
1951 static void
1952 show_trust_readonly (struct ui_file *file, int from_tty,
1953 struct cmd_list_element *c, const char *value)
1954 {
1955 fprintf_filtered (file,
1956 _("Mode for reading from readonly sections is %s.\n"),
1957 value);
1958 }
1959
1960 /* More generic transfers. */
1961
1962 static LONGEST
1963 default_xfer_partial (struct target_ops *ops, enum target_object object,
1964 const char *annex, gdb_byte *readbuf,
1965 const gdb_byte *writebuf, ULONGEST offset, LONGEST len)
1966 {
1967 if (object == TARGET_OBJECT_MEMORY
1968 && ops->deprecated_xfer_memory != NULL)
1969 /* If available, fall back to the target's
1970 "deprecated_xfer_memory" method. */
1971 {
1972 int xfered = -1;
1973
1974 errno = 0;
1975 if (writebuf != NULL)
1976 {
1977 void *buffer = xmalloc (len);
1978 struct cleanup *cleanup = make_cleanup (xfree, buffer);
1979
1980 memcpy (buffer, writebuf, len);
1981 xfered = ops->deprecated_xfer_memory (offset, buffer, len,
1982 1/*write*/, NULL, ops);
1983 do_cleanups (cleanup);
1984 }
1985 if (readbuf != NULL)
1986 xfered = ops->deprecated_xfer_memory (offset, readbuf, len,
1987 0/*read*/, NULL, ops);
1988 if (xfered > 0)
1989 return xfered;
1990 else if (xfered == 0 && errno == 0)
1991 /* "deprecated_xfer_memory" uses 0, cross checked against
1992 ERRNO as one indication of an error. */
1993 return 0;
1994 else
1995 return -1;
1996 }
1997 else if (ops->beneath != NULL)
1998 return ops->beneath->to_xfer_partial (ops->beneath, object, annex,
1999 readbuf, writebuf, offset, len);
2000 else
2001 return -1;
2002 }
2003
2004 /* The xfer_partial handler for the topmost target. Unlike the default,
2005 it does not need to handle memory specially; it just passes all
2006 requests down the stack. */
2007
2008 static LONGEST
2009 current_xfer_partial (struct target_ops *ops, enum target_object object,
2010 const char *annex, gdb_byte *readbuf,
2011 const gdb_byte *writebuf, ULONGEST offset, LONGEST len)
2012 {
2013 if (ops->beneath != NULL)
2014 return ops->beneath->to_xfer_partial (ops->beneath, object, annex,
2015 readbuf, writebuf, offset, len);
2016 else
2017 return -1;
2018 }
2019
2020 /* Target vector read/write partial wrapper functions. */
2021
2022 static LONGEST
2023 target_read_partial (struct target_ops *ops,
2024 enum target_object object,
2025 const char *annex, gdb_byte *buf,
2026 ULONGEST offset, LONGEST len)
2027 {
2028 return target_xfer_partial (ops, object, annex, buf, NULL, offset, len);
2029 }
2030
2031 static LONGEST
2032 target_write_partial (struct target_ops *ops,
2033 enum target_object object,
2034 const char *annex, const gdb_byte *buf,
2035 ULONGEST offset, LONGEST len)
2036 {
2037 return target_xfer_partial (ops, object, annex, NULL, buf, offset, len);
2038 }
2039
2040 /* Wrappers to perform the full transfer. */
2041
2042 /* For docs on target_read see target.h. */
2043
2044 LONGEST
2045 target_read (struct target_ops *ops,
2046 enum target_object object,
2047 const char *annex, gdb_byte *buf,
2048 ULONGEST offset, LONGEST len)
2049 {
2050 LONGEST xfered = 0;
2051
2052 while (xfered < len)
2053 {
2054 LONGEST xfer = target_read_partial (ops, object, annex,
2055 (gdb_byte *) buf + xfered,
2056 offset + xfered, len - xfered);
2057
2058 /* Call an observer, notifying them of the xfer progress? */
2059 if (xfer == 0)
2060 return xfered;
2061 if (xfer < 0)
2062 return -1;
2063 xfered += xfer;
2064 QUIT;
2065 }
2066 return len;
2067 }
2068
2069 /* Assuming that the entire [begin, end) range of memory cannot be
2070 read, try to read whatever subrange is possible to read.
2071
2072 The function returns, in RESULT, either zero or one memory block.
2073 If there's a readable subrange at the beginning, it is completely
2074 read and returned. Any further readable subrange will not be read.
2075 Otherwise, if there's a readable subrange at the end, it will be
2076 completely read and returned. Any readable subranges before it
2077 (obviously, not starting at the beginning), will be ignored. In
2078 other cases -- either no readable subrange, or readable subrange(s)
2079 that is neither at the beginning, or end, nothing is returned.
2080
2081 The purpose of this function is to handle a read across a boundary
2082 of accessible memory in a case when memory map is not available.
2083 The above restrictions are fine for this case, but will give
2084 incorrect results if the memory is 'patchy'. However, supporting
2085 'patchy' memory would require trying to read every single byte,
2086 and it seems unacceptable solution. Explicit memory map is
2087 recommended for this case -- and target_read_memory_robust will
2088 take care of reading multiple ranges then. */
2089
2090 static void
2091 read_whatever_is_readable (struct target_ops *ops,
2092 ULONGEST begin, ULONGEST end,
2093 VEC(memory_read_result_s) **result)
2094 {
2095 gdb_byte *buf = xmalloc (end - begin);
2096 ULONGEST current_begin = begin;
2097 ULONGEST current_end = end;
2098 int forward;
2099 memory_read_result_s r;
2100
2101 /* If we previously failed to read 1 byte, nothing can be done here. */
2102 if (end - begin <= 1)
2103 {
2104 xfree (buf);
2105 return;
2106 }
2107
2108 /* Check that either first or the last byte is readable, and give up
2109 if not. This heuristic is meant to permit reading accessible memory
2110 at the boundary of accessible region. */
2111 if (target_read_partial (ops, TARGET_OBJECT_MEMORY, NULL,
2112 buf, begin, 1) == 1)
2113 {
2114 forward = 1;
2115 ++current_begin;
2116 }
2117 else if (target_read_partial (ops, TARGET_OBJECT_MEMORY, NULL,
2118 buf + (end-begin) - 1, end - 1, 1) == 1)
2119 {
2120 forward = 0;
2121 --current_end;
2122 }
2123 else
2124 {
2125 xfree (buf);
2126 return;
2127 }
2128
2129 /* Loop invariant is that the [current_begin, current_end) was previously
2130 found to be not readable as a whole.
2131
2132 Note loop condition -- if the range has 1 byte, we can't divide the range
2133 so there's no point trying further. */
2134 while (current_end - current_begin > 1)
2135 {
2136 ULONGEST first_half_begin, first_half_end;
2137 ULONGEST second_half_begin, second_half_end;
2138 LONGEST xfer;
2139 ULONGEST middle = current_begin + (current_end - current_begin)/2;
2140
2141 if (forward)
2142 {
2143 first_half_begin = current_begin;
2144 first_half_end = middle;
2145 second_half_begin = middle;
2146 second_half_end = current_end;
2147 }
2148 else
2149 {
2150 first_half_begin = middle;
2151 first_half_end = current_end;
2152 second_half_begin = current_begin;
2153 second_half_end = middle;
2154 }
2155
2156 xfer = target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2157 buf + (first_half_begin - begin),
2158 first_half_begin,
2159 first_half_end - first_half_begin);
2160
2161 if (xfer == first_half_end - first_half_begin)
2162 {
2163 /* This half reads up fine. So, the error must be in the
2164 other half. */
2165 current_begin = second_half_begin;
2166 current_end = second_half_end;
2167 }
2168 else
2169 {
2170 /* This half is not readable. Because we've tried one byte, we
2171 know some part of this half if actually redable. Go to the next
2172 iteration to divide again and try to read.
2173
2174 We don't handle the other half, because this function only tries
2175 to read a single readable subrange. */
2176 current_begin = first_half_begin;
2177 current_end = first_half_end;
2178 }
2179 }
2180
2181 if (forward)
2182 {
2183 /* The [begin, current_begin) range has been read. */
2184 r.begin = begin;
2185 r.end = current_begin;
2186 r.data = buf;
2187 }
2188 else
2189 {
2190 /* The [current_end, end) range has been read. */
2191 LONGEST rlen = end - current_end;
2192
2193 r.data = xmalloc (rlen);
2194 memcpy (r.data, buf + current_end - begin, rlen);
2195 r.begin = current_end;
2196 r.end = end;
2197 xfree (buf);
2198 }
2199 VEC_safe_push(memory_read_result_s, (*result), &r);
2200 }
2201
2202 void
2203 free_memory_read_result_vector (void *x)
2204 {
2205 VEC(memory_read_result_s) *v = x;
2206 memory_read_result_s *current;
2207 int ix;
2208
2209 for (ix = 0; VEC_iterate (memory_read_result_s, v, ix, current); ++ix)
2210 {
2211 xfree (current->data);
2212 }
2213 VEC_free (memory_read_result_s, v);
2214 }
2215
2216 VEC(memory_read_result_s) *
2217 read_memory_robust (struct target_ops *ops, ULONGEST offset, LONGEST len)
2218 {
2219 VEC(memory_read_result_s) *result = 0;
2220
2221 LONGEST xfered = 0;
2222 while (xfered < len)
2223 {
2224 struct mem_region *region = lookup_mem_region (offset + xfered);
2225 LONGEST rlen;
2226
2227 /* If there is no explicit region, a fake one should be created. */
2228 gdb_assert (region);
2229
2230 if (region->hi == 0)
2231 rlen = len - xfered;
2232 else
2233 rlen = region->hi - offset;
2234
2235 if (region->attrib.mode == MEM_NONE || region->attrib.mode == MEM_WO)
2236 {
2237 /* Cannot read this region. Note that we can end up here only
2238 if the region is explicitly marked inaccessible, or
2239 'inaccessible-by-default' is in effect. */
2240 xfered += rlen;
2241 }
2242 else
2243 {
2244 LONGEST to_read = min (len - xfered, rlen);
2245 gdb_byte *buffer = (gdb_byte *)xmalloc (to_read);
2246
2247 LONGEST xfer = target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2248 (gdb_byte *) buffer,
2249 offset + xfered, to_read);
2250 /* Call an observer, notifying them of the xfer progress? */
2251 if (xfer <= 0)
2252 {
2253 /* Got an error reading full chunk. See if maybe we can read
2254 some subrange. */
2255 xfree (buffer);
2256 read_whatever_is_readable (ops, offset + xfered,
2257 offset + xfered + to_read, &result);
2258 xfered += to_read;
2259 }
2260 else
2261 {
2262 struct memory_read_result r;
2263 r.data = buffer;
2264 r.begin = offset + xfered;
2265 r.end = r.begin + xfer;
2266 VEC_safe_push (memory_read_result_s, result, &r);
2267 xfered += xfer;
2268 }
2269 QUIT;
2270 }
2271 }
2272 return result;
2273 }
2274
2275
2276 /* An alternative to target_write with progress callbacks. */
2277
2278 LONGEST
2279 target_write_with_progress (struct target_ops *ops,
2280 enum target_object object,
2281 const char *annex, const gdb_byte *buf,
2282 ULONGEST offset, LONGEST len,
2283 void (*progress) (ULONGEST, void *), void *baton)
2284 {
2285 LONGEST xfered = 0;
2286
2287 /* Give the progress callback a chance to set up. */
2288 if (progress)
2289 (*progress) (0, baton);
2290
2291 while (xfered < len)
2292 {
2293 LONGEST xfer = target_write_partial (ops, object, annex,
2294 (gdb_byte *) buf + xfered,
2295 offset + xfered, len - xfered);
2296
2297 if (xfer == 0)
2298 return xfered;
2299 if (xfer < 0)
2300 return -1;
2301
2302 if (progress)
2303 (*progress) (xfer, baton);
2304
2305 xfered += xfer;
2306 QUIT;
2307 }
2308 return len;
2309 }
2310
2311 /* For docs on target_write see target.h. */
2312
2313 LONGEST
2314 target_write (struct target_ops *ops,
2315 enum target_object object,
2316 const char *annex, const gdb_byte *buf,
2317 ULONGEST offset, LONGEST len)
2318 {
2319 return target_write_with_progress (ops, object, annex, buf, offset, len,
2320 NULL, NULL);
2321 }
2322
2323 /* Read OBJECT/ANNEX using OPS. Store the result in *BUF_P and return
2324 the size of the transferred data. PADDING additional bytes are
2325 available in *BUF_P. This is a helper function for
2326 target_read_alloc; see the declaration of that function for more
2327 information. */
2328
2329 static LONGEST
2330 target_read_alloc_1 (struct target_ops *ops, enum target_object object,
2331 const char *annex, gdb_byte **buf_p, int padding)
2332 {
2333 size_t buf_alloc, buf_pos;
2334 gdb_byte *buf;
2335 LONGEST n;
2336
2337 /* This function does not have a length parameter; it reads the
2338 entire OBJECT). Also, it doesn't support objects fetched partly
2339 from one target and partly from another (in a different stratum,
2340 e.g. a core file and an executable). Both reasons make it
2341 unsuitable for reading memory. */
2342 gdb_assert (object != TARGET_OBJECT_MEMORY);
2343
2344 /* Start by reading up to 4K at a time. The target will throttle
2345 this number down if necessary. */
2346 buf_alloc = 4096;
2347 buf = xmalloc (buf_alloc);
2348 buf_pos = 0;
2349 while (1)
2350 {
2351 n = target_read_partial (ops, object, annex, &buf[buf_pos],
2352 buf_pos, buf_alloc - buf_pos - padding);
2353 if (n < 0)
2354 {
2355 /* An error occurred. */
2356 xfree (buf);
2357 return -1;
2358 }
2359 else if (n == 0)
2360 {
2361 /* Read all there was. */
2362 if (buf_pos == 0)
2363 xfree (buf);
2364 else
2365 *buf_p = buf;
2366 return buf_pos;
2367 }
2368
2369 buf_pos += n;
2370
2371 /* If the buffer is filling up, expand it. */
2372 if (buf_alloc < buf_pos * 2)
2373 {
2374 buf_alloc *= 2;
2375 buf = xrealloc (buf, buf_alloc);
2376 }
2377
2378 QUIT;
2379 }
2380 }
2381
2382 /* Read OBJECT/ANNEX using OPS. Store the result in *BUF_P and return
2383 the size of the transferred data. See the declaration in "target.h"
2384 function for more information about the return value. */
2385
2386 LONGEST
2387 target_read_alloc (struct target_ops *ops, enum target_object object,
2388 const char *annex, gdb_byte **buf_p)
2389 {
2390 return target_read_alloc_1 (ops, object, annex, buf_p, 0);
2391 }
2392
2393 /* Read OBJECT/ANNEX using OPS. The result is NUL-terminated and
2394 returned as a string, allocated using xmalloc. If an error occurs
2395 or the transfer is unsupported, NULL is returned. Empty objects
2396 are returned as allocated but empty strings. A warning is issued
2397 if the result contains any embedded NUL bytes. */
2398
2399 char *
2400 target_read_stralloc (struct target_ops *ops, enum target_object object,
2401 const char *annex)
2402 {
2403 gdb_byte *buffer;
2404 char *bufstr;
2405 LONGEST i, transferred;
2406
2407 transferred = target_read_alloc_1 (ops, object, annex, &buffer, 1);
2408 bufstr = (char *) buffer;
2409
2410 if (transferred < 0)
2411 return NULL;
2412
2413 if (transferred == 0)
2414 return xstrdup ("");
2415
2416 bufstr[transferred] = 0;
2417
2418 /* Check for embedded NUL bytes; but allow trailing NULs. */
2419 for (i = strlen (bufstr); i < transferred; i++)
2420 if (bufstr[i] != 0)
2421 {
2422 warning (_("target object %d, annex %s, "
2423 "contained unexpected null characters"),
2424 (int) object, annex ? annex : "(none)");
2425 break;
2426 }
2427
2428 return bufstr;
2429 }
2430
2431 /* Memory transfer methods. */
2432
2433 void
2434 get_target_memory (struct target_ops *ops, CORE_ADDR addr, gdb_byte *buf,
2435 LONGEST len)
2436 {
2437 /* This method is used to read from an alternate, non-current
2438 target. This read must bypass the overlay support (as symbols
2439 don't match this target), and GDB's internal cache (wrong cache
2440 for this target). */
2441 if (target_read (ops, TARGET_OBJECT_RAW_MEMORY, NULL, buf, addr, len)
2442 != len)
2443 memory_error (EIO, addr);
2444 }
2445
2446 ULONGEST
2447 get_target_memory_unsigned (struct target_ops *ops, CORE_ADDR addr,
2448 int len, enum bfd_endian byte_order)
2449 {
2450 gdb_byte buf[sizeof (ULONGEST)];
2451
2452 gdb_assert (len <= sizeof (buf));
2453 get_target_memory (ops, addr, buf, len);
2454 return extract_unsigned_integer (buf, len, byte_order);
2455 }
2456
2457 int
2458 target_insert_breakpoint (struct gdbarch *gdbarch,
2459 struct bp_target_info *bp_tgt)
2460 {
2461 if (!may_insert_breakpoints)
2462 {
2463 warning (_("May not insert breakpoints"));
2464 return 1;
2465 }
2466
2467 return (*current_target.to_insert_breakpoint) (gdbarch, bp_tgt);
2468 }
2469
2470 int
2471 target_remove_breakpoint (struct gdbarch *gdbarch,
2472 struct bp_target_info *bp_tgt)
2473 {
2474 /* This is kind of a weird case to handle, but the permission might
2475 have been changed after breakpoints were inserted - in which case
2476 we should just take the user literally and assume that any
2477 breakpoints should be left in place. */
2478 if (!may_insert_breakpoints)
2479 {
2480 warning (_("May not remove breakpoints"));
2481 return 1;
2482 }
2483
2484 return (*current_target.to_remove_breakpoint) (gdbarch, bp_tgt);
2485 }
2486
2487 static void
2488 target_info (char *args, int from_tty)
2489 {
2490 struct target_ops *t;
2491 int has_all_mem = 0;
2492
2493 if (symfile_objfile != NULL)
2494 printf_unfiltered (_("Symbols from \"%s\".\n"),
2495 objfile_name (symfile_objfile));
2496
2497 for (t = target_stack; t != NULL; t = t->beneath)
2498 {
2499 if (!(*t->to_has_memory) (t))
2500 continue;
2501
2502 if ((int) (t->to_stratum) <= (int) dummy_stratum)
2503 continue;
2504 if (has_all_mem)
2505 printf_unfiltered (_("\tWhile running this, "
2506 "GDB does not access memory from...\n"));
2507 printf_unfiltered ("%s:\n", t->to_longname);
2508 (t->to_files_info) (t);
2509 has_all_mem = (*t->to_has_all_memory) (t);
2510 }
2511 }
2512
2513 /* This function is called before any new inferior is created, e.g.
2514 by running a program, attaching, or connecting to a target.
2515 It cleans up any state from previous invocations which might
2516 change between runs. This is a subset of what target_preopen
2517 resets (things which might change between targets). */
2518
2519 void
2520 target_pre_inferior (int from_tty)
2521 {
2522 /* Clear out solib state. Otherwise the solib state of the previous
2523 inferior might have survived and is entirely wrong for the new
2524 target. This has been observed on GNU/Linux using glibc 2.3. How
2525 to reproduce:
2526
2527 bash$ ./foo&
2528 [1] 4711
2529 bash$ ./foo&
2530 [1] 4712
2531 bash$ gdb ./foo
2532 [...]
2533 (gdb) attach 4711
2534 (gdb) detach
2535 (gdb) attach 4712
2536 Cannot access memory at address 0xdeadbeef
2537 */
2538
2539 /* In some OSs, the shared library list is the same/global/shared
2540 across inferiors. If code is shared between processes, so are
2541 memory regions and features. */
2542 if (!gdbarch_has_global_solist (target_gdbarch ()))
2543 {
2544 no_shared_libraries (NULL, from_tty);
2545
2546 invalidate_target_mem_regions ();
2547
2548 target_clear_description ();
2549 }
2550
2551 agent_capability_invalidate ();
2552 }
2553
2554 /* Callback for iterate_over_inferiors. Gets rid of the given
2555 inferior. */
2556
2557 static int
2558 dispose_inferior (struct inferior *inf, void *args)
2559 {
2560 struct thread_info *thread;
2561
2562 thread = any_thread_of_process (inf->pid);
2563 if (thread)
2564 {
2565 switch_to_thread (thread->ptid);
2566
2567 /* Core inferiors actually should be detached, not killed. */
2568 if (target_has_execution)
2569 target_kill ();
2570 else
2571 target_detach (NULL, 0);
2572 }
2573
2574 return 0;
2575 }
2576
2577 /* This is to be called by the open routine before it does
2578 anything. */
2579
2580 void
2581 target_preopen (int from_tty)
2582 {
2583 dont_repeat ();
2584
2585 if (have_inferiors ())
2586 {
2587 if (!from_tty
2588 || !have_live_inferiors ()
2589 || query (_("A program is being debugged already. Kill it? ")))
2590 iterate_over_inferiors (dispose_inferior, NULL);
2591 else
2592 error (_("Program not killed."));
2593 }
2594
2595 /* Calling target_kill may remove the target from the stack. But if
2596 it doesn't (which seems like a win for UDI), remove it now. */
2597 /* Leave the exec target, though. The user may be switching from a
2598 live process to a core of the same program. */
2599 pop_all_targets_above (file_stratum);
2600
2601 target_pre_inferior (from_tty);
2602 }
2603
2604 /* Detach a target after doing deferred register stores. */
2605
2606 void
2607 target_detach (char *args, int from_tty)
2608 {
2609 struct target_ops* t;
2610
2611 if (gdbarch_has_global_breakpoints (target_gdbarch ()))
2612 /* Don't remove global breakpoints here. They're removed on
2613 disconnection from the target. */
2614 ;
2615 else
2616 /* If we're in breakpoints-always-inserted mode, have to remove
2617 them before detaching. */
2618 remove_breakpoints_pid (PIDGET (inferior_ptid));
2619
2620 prepare_for_detach ();
2621
2622 for (t = current_target.beneath; t != NULL; t = t->beneath)
2623 {
2624 if (t->to_detach != NULL)
2625 {
2626 t->to_detach (t, args, from_tty);
2627 if (targetdebug)
2628 fprintf_unfiltered (gdb_stdlog, "target_detach (%s, %d)\n",
2629 args, from_tty);
2630 return;
2631 }
2632 }
2633
2634 internal_error (__FILE__, __LINE__, _("could not find a target to detach"));
2635 }
2636
2637 void
2638 target_disconnect (char *args, int from_tty)
2639 {
2640 struct target_ops *t;
2641
2642 /* If we're in breakpoints-always-inserted mode or if breakpoints
2643 are global across processes, we have to remove them before
2644 disconnecting. */
2645 remove_breakpoints ();
2646
2647 for (t = current_target.beneath; t != NULL; t = t->beneath)
2648 if (t->to_disconnect != NULL)
2649 {
2650 if (targetdebug)
2651 fprintf_unfiltered (gdb_stdlog, "target_disconnect (%s, %d)\n",
2652 args, from_tty);
2653 t->to_disconnect (t, args, from_tty);
2654 return;
2655 }
2656
2657 tcomplain ();
2658 }
2659
2660 ptid_t
2661 target_wait (ptid_t ptid, struct target_waitstatus *status, int options)
2662 {
2663 struct target_ops *t;
2664
2665 for (t = current_target.beneath; t != NULL; t = t->beneath)
2666 {
2667 if (t->to_wait != NULL)
2668 {
2669 ptid_t retval = (*t->to_wait) (t, ptid, status, options);
2670
2671 if (targetdebug)
2672 {
2673 char *status_string;
2674 char *options_string;
2675
2676 status_string = target_waitstatus_to_string (status);
2677 options_string = target_options_to_string (options);
2678 fprintf_unfiltered (gdb_stdlog,
2679 "target_wait (%d, status, options={%s})"
2680 " = %d, %s\n",
2681 PIDGET (ptid), options_string,
2682 PIDGET (retval), status_string);
2683 xfree (status_string);
2684 xfree (options_string);
2685 }
2686
2687 return retval;
2688 }
2689 }
2690
2691 noprocess ();
2692 }
2693
2694 char *
2695 target_pid_to_str (ptid_t ptid)
2696 {
2697 struct target_ops *t;
2698
2699 for (t = current_target.beneath; t != NULL; t = t->beneath)
2700 {
2701 if (t->to_pid_to_str != NULL)
2702 return (*t->to_pid_to_str) (t, ptid);
2703 }
2704
2705 return normal_pid_to_str (ptid);
2706 }
2707
2708 char *
2709 target_thread_name (struct thread_info *info)
2710 {
2711 struct target_ops *t;
2712
2713 for (t = current_target.beneath; t != NULL; t = t->beneath)
2714 {
2715 if (t->to_thread_name != NULL)
2716 return (*t->to_thread_name) (info);
2717 }
2718
2719 return NULL;
2720 }
2721
2722 void
2723 target_resume (ptid_t ptid, int step, enum gdb_signal signal)
2724 {
2725 struct target_ops *t;
2726
2727 target_dcache_invalidate ();
2728
2729 for (t = current_target.beneath; t != NULL; t = t->beneath)
2730 {
2731 if (t->to_resume != NULL)
2732 {
2733 t->to_resume (t, ptid, step, signal);
2734 if (targetdebug)
2735 fprintf_unfiltered (gdb_stdlog, "target_resume (%d, %s, %s)\n",
2736 PIDGET (ptid),
2737 step ? "step" : "continue",
2738 gdb_signal_to_name (signal));
2739
2740 registers_changed_ptid (ptid);
2741 set_executing (ptid, 1);
2742 set_running (ptid, 1);
2743 clear_inline_frame_state (ptid);
2744 return;
2745 }
2746 }
2747
2748 noprocess ();
2749 }
2750
2751 void
2752 target_pass_signals (int numsigs, unsigned char *pass_signals)
2753 {
2754 struct target_ops *t;
2755
2756 for (t = current_target.beneath; t != NULL; t = t->beneath)
2757 {
2758 if (t->to_pass_signals != NULL)
2759 {
2760 if (targetdebug)
2761 {
2762 int i;
2763
2764 fprintf_unfiltered (gdb_stdlog, "target_pass_signals (%d, {",
2765 numsigs);
2766
2767 for (i = 0; i < numsigs; i++)
2768 if (pass_signals[i])
2769 fprintf_unfiltered (gdb_stdlog, " %s",
2770 gdb_signal_to_name (i));
2771
2772 fprintf_unfiltered (gdb_stdlog, " })\n");
2773 }
2774
2775 (*t->to_pass_signals) (numsigs, pass_signals);
2776 return;
2777 }
2778 }
2779 }
2780
2781 void
2782 target_program_signals (int numsigs, unsigned char *program_signals)
2783 {
2784 struct target_ops *t;
2785
2786 for (t = current_target.beneath; t != NULL; t = t->beneath)
2787 {
2788 if (t->to_program_signals != NULL)
2789 {
2790 if (targetdebug)
2791 {
2792 int i;
2793
2794 fprintf_unfiltered (gdb_stdlog, "target_program_signals (%d, {",
2795 numsigs);
2796
2797 for (i = 0; i < numsigs; i++)
2798 if (program_signals[i])
2799 fprintf_unfiltered (gdb_stdlog, " %s",
2800 gdb_signal_to_name (i));
2801
2802 fprintf_unfiltered (gdb_stdlog, " })\n");
2803 }
2804
2805 (*t->to_program_signals) (numsigs, program_signals);
2806 return;
2807 }
2808 }
2809 }
2810
2811 /* Look through the list of possible targets for a target that can
2812 follow forks. */
2813
2814 int
2815 target_follow_fork (int follow_child, int detach_fork)
2816 {
2817 struct target_ops *t;
2818
2819 for (t = current_target.beneath; t != NULL; t = t->beneath)
2820 {
2821 if (t->to_follow_fork != NULL)
2822 {
2823 int retval = t->to_follow_fork (t, follow_child, detach_fork);
2824
2825 if (targetdebug)
2826 fprintf_unfiltered (gdb_stdlog,
2827 "target_follow_fork (%d, %d) = %d\n",
2828 follow_child, detach_fork, retval);
2829 return retval;
2830 }
2831 }
2832
2833 /* Some target returned a fork event, but did not know how to follow it. */
2834 internal_error (__FILE__, __LINE__,
2835 _("could not find a target to follow fork"));
2836 }
2837
2838 void
2839 target_mourn_inferior (void)
2840 {
2841 struct target_ops *t;
2842
2843 for (t = current_target.beneath; t != NULL; t = t->beneath)
2844 {
2845 if (t->to_mourn_inferior != NULL)
2846 {
2847 t->to_mourn_inferior (t);
2848 if (targetdebug)
2849 fprintf_unfiltered (gdb_stdlog, "target_mourn_inferior ()\n");
2850
2851 /* We no longer need to keep handles on any of the object files.
2852 Make sure to release them to avoid unnecessarily locking any
2853 of them while we're not actually debugging. */
2854 bfd_cache_close_all ();
2855
2856 return;
2857 }
2858 }
2859
2860 internal_error (__FILE__, __LINE__,
2861 _("could not find a target to follow mourn inferior"));
2862 }
2863
2864 /* Look for a target which can describe architectural features, starting
2865 from TARGET. If we find one, return its description. */
2866
2867 const struct target_desc *
2868 target_read_description (struct target_ops *target)
2869 {
2870 struct target_ops *t;
2871
2872 for (t = target; t != NULL; t = t->beneath)
2873 if (t->to_read_description != NULL)
2874 {
2875 const struct target_desc *tdesc;
2876
2877 tdesc = t->to_read_description (t);
2878 if (tdesc)
2879 return tdesc;
2880 }
2881
2882 return NULL;
2883 }
2884
2885 /* The default implementation of to_search_memory.
2886 This implements a basic search of memory, reading target memory and
2887 performing the search here (as opposed to performing the search in on the
2888 target side with, for example, gdbserver). */
2889
2890 int
2891 simple_search_memory (struct target_ops *ops,
2892 CORE_ADDR start_addr, ULONGEST search_space_len,
2893 const gdb_byte *pattern, ULONGEST pattern_len,
2894 CORE_ADDR *found_addrp)
2895 {
2896 /* NOTE: also defined in find.c testcase. */
2897 #define SEARCH_CHUNK_SIZE 16000
2898 const unsigned chunk_size = SEARCH_CHUNK_SIZE;
2899 /* Buffer to hold memory contents for searching. */
2900 gdb_byte *search_buf;
2901 unsigned search_buf_size;
2902 struct cleanup *old_cleanups;
2903
2904 search_buf_size = chunk_size + pattern_len - 1;
2905
2906 /* No point in trying to allocate a buffer larger than the search space. */
2907 if (search_space_len < search_buf_size)
2908 search_buf_size = search_space_len;
2909
2910 search_buf = malloc (search_buf_size);
2911 if (search_buf == NULL)
2912 error (_("Unable to allocate memory to perform the search."));
2913 old_cleanups = make_cleanup (free_current_contents, &search_buf);
2914
2915 /* Prime the search buffer. */
2916
2917 if (target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2918 search_buf, start_addr, search_buf_size) != search_buf_size)
2919 {
2920 warning (_("Unable to access %s bytes of target "
2921 "memory at %s, halting search."),
2922 pulongest (search_buf_size), hex_string (start_addr));
2923 do_cleanups (old_cleanups);
2924 return -1;
2925 }
2926
2927 /* Perform the search.
2928
2929 The loop is kept simple by allocating [N + pattern-length - 1] bytes.
2930 When we've scanned N bytes we copy the trailing bytes to the start and
2931 read in another N bytes. */
2932
2933 while (search_space_len >= pattern_len)
2934 {
2935 gdb_byte *found_ptr;
2936 unsigned nr_search_bytes = min (search_space_len, search_buf_size);
2937
2938 found_ptr = memmem (search_buf, nr_search_bytes,
2939 pattern, pattern_len);
2940
2941 if (found_ptr != NULL)
2942 {
2943 CORE_ADDR found_addr = start_addr + (found_ptr - search_buf);
2944
2945 *found_addrp = found_addr;
2946 do_cleanups (old_cleanups);
2947 return 1;
2948 }
2949
2950 /* Not found in this chunk, skip to next chunk. */
2951
2952 /* Don't let search_space_len wrap here, it's unsigned. */
2953 if (search_space_len >= chunk_size)
2954 search_space_len -= chunk_size;
2955 else
2956 search_space_len = 0;
2957
2958 if (search_space_len >= pattern_len)
2959 {
2960 unsigned keep_len = search_buf_size - chunk_size;
2961 CORE_ADDR read_addr = start_addr + chunk_size + keep_len;
2962 int nr_to_read;
2963
2964 /* Copy the trailing part of the previous iteration to the front
2965 of the buffer for the next iteration. */
2966 gdb_assert (keep_len == pattern_len - 1);
2967 memcpy (search_buf, search_buf + chunk_size, keep_len);
2968
2969 nr_to_read = min (search_space_len - keep_len, chunk_size);
2970
2971 if (target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2972 search_buf + keep_len, read_addr,
2973 nr_to_read) != nr_to_read)
2974 {
2975 warning (_("Unable to access %s bytes of target "
2976 "memory at %s, halting search."),
2977 plongest (nr_to_read),
2978 hex_string (read_addr));
2979 do_cleanups (old_cleanups);
2980 return -1;
2981 }
2982
2983 start_addr += chunk_size;
2984 }
2985 }
2986
2987 /* Not found. */
2988
2989 do_cleanups (old_cleanups);
2990 return 0;
2991 }
2992
2993 /* Search SEARCH_SPACE_LEN bytes beginning at START_ADDR for the
2994 sequence of bytes in PATTERN with length PATTERN_LEN.
2995
2996 The result is 1 if found, 0 if not found, and -1 if there was an error
2997 requiring halting of the search (e.g. memory read error).
2998 If the pattern is found the address is recorded in FOUND_ADDRP. */
2999
3000 int
3001 target_search_memory (CORE_ADDR start_addr, ULONGEST search_space_len,
3002 const gdb_byte *pattern, ULONGEST pattern_len,
3003 CORE_ADDR *found_addrp)
3004 {
3005 struct target_ops *t;
3006 int found;
3007
3008 /* We don't use INHERIT to set current_target.to_search_memory,
3009 so we have to scan the target stack and handle targetdebug
3010 ourselves. */
3011
3012 if (targetdebug)
3013 fprintf_unfiltered (gdb_stdlog, "target_search_memory (%s, ...)\n",
3014 hex_string (start_addr));
3015
3016 for (t = current_target.beneath; t != NULL; t = t->beneath)
3017 if (t->to_search_memory != NULL)
3018 break;
3019
3020 if (t != NULL)
3021 {
3022 found = t->to_search_memory (t, start_addr, search_space_len,
3023 pattern, pattern_len, found_addrp);
3024 }
3025 else
3026 {
3027 /* If a special version of to_search_memory isn't available, use the
3028 simple version. */
3029 found = simple_search_memory (current_target.beneath,
3030 start_addr, search_space_len,
3031 pattern, pattern_len, found_addrp);
3032 }
3033
3034 if (targetdebug)
3035 fprintf_unfiltered (gdb_stdlog, " = %d\n", found);
3036
3037 return found;
3038 }
3039
3040 /* Look through the currently pushed targets. If none of them will
3041 be able to restart the currently running process, issue an error
3042 message. */
3043
3044 void
3045 target_require_runnable (void)
3046 {
3047 struct target_ops *t;
3048
3049 for (t = target_stack; t != NULL; t = t->beneath)
3050 {
3051 /* If this target knows how to create a new program, then
3052 assume we will still be able to after killing the current
3053 one. Either killing and mourning will not pop T, or else
3054 find_default_run_target will find it again. */
3055 if (t->to_create_inferior != NULL)
3056 return;
3057
3058 /* Do not worry about thread_stratum targets that can not
3059 create inferiors. Assume they will be pushed again if
3060 necessary, and continue to the process_stratum. */
3061 if (t->to_stratum == thread_stratum
3062 || t->to_stratum == arch_stratum)
3063 continue;
3064
3065 error (_("The \"%s\" target does not support \"run\". "
3066 "Try \"help target\" or \"continue\"."),
3067 t->to_shortname);
3068 }
3069
3070 /* This function is only called if the target is running. In that
3071 case there should have been a process_stratum target and it
3072 should either know how to create inferiors, or not... */
3073 internal_error (__FILE__, __LINE__, _("No targets found"));
3074 }
3075
3076 /* Look through the list of possible targets for a target that can
3077 execute a run or attach command without any other data. This is
3078 used to locate the default process stratum.
3079
3080 If DO_MESG is not NULL, the result is always valid (error() is
3081 called for errors); else, return NULL on error. */
3082
3083 static struct target_ops *
3084 find_default_run_target (char *do_mesg)
3085 {
3086 struct target_ops **t;
3087 struct target_ops *runable = NULL;
3088 int count;
3089
3090 count = 0;
3091
3092 for (t = target_structs; t < target_structs + target_struct_size;
3093 ++t)
3094 {
3095 if ((*t)->to_can_run && target_can_run (*t))
3096 {
3097 runable = *t;
3098 ++count;
3099 }
3100 }
3101
3102 if (count != 1)
3103 {
3104 if (do_mesg)
3105 error (_("Don't know how to %s. Try \"help target\"."), do_mesg);
3106 else
3107 return NULL;
3108 }
3109
3110 return runable;
3111 }
3112
3113 void
3114 find_default_attach (struct target_ops *ops, char *args, int from_tty)
3115 {
3116 struct target_ops *t;
3117
3118 t = find_default_run_target ("attach");
3119 (t->to_attach) (t, args, from_tty);
3120 return;
3121 }
3122
3123 void
3124 find_default_create_inferior (struct target_ops *ops,
3125 char *exec_file, char *allargs, char **env,
3126 int from_tty)
3127 {
3128 struct target_ops *t;
3129
3130 t = find_default_run_target ("run");
3131 (t->to_create_inferior) (t, exec_file, allargs, env, from_tty);
3132 return;
3133 }
3134
3135 static int
3136 find_default_can_async_p (void)
3137 {
3138 struct target_ops *t;
3139
3140 /* This may be called before the target is pushed on the stack;
3141 look for the default process stratum. If there's none, gdb isn't
3142 configured with a native debugger, and target remote isn't
3143 connected yet. */
3144 t = find_default_run_target (NULL);
3145 if (t && t->to_can_async_p)
3146 return (t->to_can_async_p) ();
3147 return 0;
3148 }
3149
3150 static int
3151 find_default_is_async_p (void)
3152 {
3153 struct target_ops *t;
3154
3155 /* This may be called before the target is pushed on the stack;
3156 look for the default process stratum. If there's none, gdb isn't
3157 configured with a native debugger, and target remote isn't
3158 connected yet. */
3159 t = find_default_run_target (NULL);
3160 if (t && t->to_is_async_p)
3161 return (t->to_is_async_p) ();
3162 return 0;
3163 }
3164
3165 static int
3166 find_default_supports_non_stop (void)
3167 {
3168 struct target_ops *t;
3169
3170 t = find_default_run_target (NULL);
3171 if (t && t->to_supports_non_stop)
3172 return (t->to_supports_non_stop) ();
3173 return 0;
3174 }
3175
3176 int
3177 target_supports_non_stop (void)
3178 {
3179 struct target_ops *t;
3180
3181 for (t = &current_target; t != NULL; t = t->beneath)
3182 if (t->to_supports_non_stop)
3183 return t->to_supports_non_stop ();
3184
3185 return 0;
3186 }
3187
3188 /* Implement the "info proc" command. */
3189
3190 int
3191 target_info_proc (char *args, enum info_proc_what what)
3192 {
3193 struct target_ops *t;
3194
3195 /* If we're already connected to something that can get us OS
3196 related data, use it. Otherwise, try using the native
3197 target. */
3198 if (current_target.to_stratum >= process_stratum)
3199 t = current_target.beneath;
3200 else
3201 t = find_default_run_target (NULL);
3202
3203 for (; t != NULL; t = t->beneath)
3204 {
3205 if (t->to_info_proc != NULL)
3206 {
3207 t->to_info_proc (t, args, what);
3208
3209 if (targetdebug)
3210 fprintf_unfiltered (gdb_stdlog,
3211 "target_info_proc (\"%s\", %d)\n", args, what);
3212
3213 return 1;
3214 }
3215 }
3216
3217 return 0;
3218 }
3219
3220 static int
3221 find_default_supports_disable_randomization (void)
3222 {
3223 struct target_ops *t;
3224
3225 t = find_default_run_target (NULL);
3226 if (t && t->to_supports_disable_randomization)
3227 return (t->to_supports_disable_randomization) ();
3228 return 0;
3229 }
3230
3231 int
3232 target_supports_disable_randomization (void)
3233 {
3234 struct target_ops *t;
3235
3236 for (t = &current_target; t != NULL; t = t->beneath)
3237 if (t->to_supports_disable_randomization)
3238 return t->to_supports_disable_randomization ();
3239
3240 return 0;
3241 }
3242
3243 char *
3244 target_get_osdata (const char *type)
3245 {
3246 struct target_ops *t;
3247
3248 /* If we're already connected to something that can get us OS
3249 related data, use it. Otherwise, try using the native
3250 target. */
3251 if (current_target.to_stratum >= process_stratum)
3252 t = current_target.beneath;
3253 else
3254 t = find_default_run_target ("get OS data");
3255
3256 if (!t)
3257 return NULL;
3258
3259 return target_read_stralloc (t, TARGET_OBJECT_OSDATA, type);
3260 }
3261
3262 /* Determine the current address space of thread PTID. */
3263
3264 struct address_space *
3265 target_thread_address_space (ptid_t ptid)
3266 {
3267 struct address_space *aspace;
3268 struct inferior *inf;
3269 struct target_ops *t;
3270
3271 for (t = current_target.beneath; t != NULL; t = t->beneath)
3272 {
3273 if (t->to_thread_address_space != NULL)
3274 {
3275 aspace = t->to_thread_address_space (t, ptid);
3276 gdb_assert (aspace);
3277
3278 if (targetdebug)
3279 fprintf_unfiltered (gdb_stdlog,
3280 "target_thread_address_space (%s) = %d\n",
3281 target_pid_to_str (ptid),
3282 address_space_num (aspace));
3283 return aspace;
3284 }
3285 }
3286
3287 /* Fall-back to the "main" address space of the inferior. */
3288 inf = find_inferior_pid (ptid_get_pid (ptid));
3289
3290 if (inf == NULL || inf->aspace == NULL)
3291 internal_error (__FILE__, __LINE__,
3292 _("Can't determine the current "
3293 "address space of thread %s\n"),
3294 target_pid_to_str (ptid));
3295
3296 return inf->aspace;
3297 }
3298
3299
3300 /* Target file operations. */
3301
3302 static struct target_ops *
3303 default_fileio_target (void)
3304 {
3305 /* If we're already connected to something that can perform
3306 file I/O, use it. Otherwise, try using the native target. */
3307 if (current_target.to_stratum >= process_stratum)
3308 return current_target.beneath;
3309 else
3310 return find_default_run_target ("file I/O");
3311 }
3312
3313 /* Open FILENAME on the target, using FLAGS and MODE. Return a
3314 target file descriptor, or -1 if an error occurs (and set
3315 *TARGET_ERRNO). */
3316 int
3317 target_fileio_open (const char *filename, int flags, int mode,
3318 int *target_errno)
3319 {
3320 struct target_ops *t;
3321
3322 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3323 {
3324 if (t->to_fileio_open != NULL)
3325 {
3326 int fd = t->to_fileio_open (filename, flags, mode, target_errno);
3327
3328 if (targetdebug)
3329 fprintf_unfiltered (gdb_stdlog,
3330 "target_fileio_open (%s,0x%x,0%o) = %d (%d)\n",
3331 filename, flags, mode,
3332 fd, fd != -1 ? 0 : *target_errno);
3333 return fd;
3334 }
3335 }
3336
3337 *target_errno = FILEIO_ENOSYS;
3338 return -1;
3339 }
3340
3341 /* Write up to LEN bytes from WRITE_BUF to FD on the target.
3342 Return the number of bytes written, or -1 if an error occurs
3343 (and set *TARGET_ERRNO). */
3344 int
3345 target_fileio_pwrite (int fd, const gdb_byte *write_buf, int len,
3346 ULONGEST offset, int *target_errno)
3347 {
3348 struct target_ops *t;
3349
3350 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3351 {
3352 if (t->to_fileio_pwrite != NULL)
3353 {
3354 int ret = t->to_fileio_pwrite (fd, write_buf, len, offset,
3355 target_errno);
3356
3357 if (targetdebug)
3358 fprintf_unfiltered (gdb_stdlog,
3359 "target_fileio_pwrite (%d,...,%d,%s) "
3360 "= %d (%d)\n",
3361 fd, len, pulongest (offset),
3362 ret, ret != -1 ? 0 : *target_errno);
3363 return ret;
3364 }
3365 }
3366
3367 *target_errno = FILEIO_ENOSYS;
3368 return -1;
3369 }
3370
3371 /* Read up to LEN bytes FD on the target into READ_BUF.
3372 Return the number of bytes read, or -1 if an error occurs
3373 (and set *TARGET_ERRNO). */
3374 int
3375 target_fileio_pread (int fd, gdb_byte *read_buf, int len,
3376 ULONGEST offset, int *target_errno)
3377 {
3378 struct target_ops *t;
3379
3380 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3381 {
3382 if (t->to_fileio_pread != NULL)
3383 {
3384 int ret = t->to_fileio_pread (fd, read_buf, len, offset,
3385 target_errno);
3386
3387 if (targetdebug)
3388 fprintf_unfiltered (gdb_stdlog,
3389 "target_fileio_pread (%d,...,%d,%s) "
3390 "= %d (%d)\n",
3391 fd, len, pulongest (offset),
3392 ret, ret != -1 ? 0 : *target_errno);
3393 return ret;
3394 }
3395 }
3396
3397 *target_errno = FILEIO_ENOSYS;
3398 return -1;
3399 }
3400
3401 /* Close FD on the target. Return 0, or -1 if an error occurs
3402 (and set *TARGET_ERRNO). */
3403 int
3404 target_fileio_close (int fd, int *target_errno)
3405 {
3406 struct target_ops *t;
3407
3408 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3409 {
3410 if (t->to_fileio_close != NULL)
3411 {
3412 int ret = t->to_fileio_close (fd, target_errno);
3413
3414 if (targetdebug)
3415 fprintf_unfiltered (gdb_stdlog,
3416 "target_fileio_close (%d) = %d (%d)\n",
3417 fd, ret, ret != -1 ? 0 : *target_errno);
3418 return ret;
3419 }
3420 }
3421
3422 *target_errno = FILEIO_ENOSYS;
3423 return -1;
3424 }
3425
3426 /* Unlink FILENAME on the target. Return 0, or -1 if an error
3427 occurs (and set *TARGET_ERRNO). */
3428 int
3429 target_fileio_unlink (const char *filename, int *target_errno)
3430 {
3431 struct target_ops *t;
3432
3433 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3434 {
3435 if (t->to_fileio_unlink != NULL)
3436 {
3437 int ret = t->to_fileio_unlink (filename, target_errno);
3438
3439 if (targetdebug)
3440 fprintf_unfiltered (gdb_stdlog,
3441 "target_fileio_unlink (%s) = %d (%d)\n",
3442 filename, ret, ret != -1 ? 0 : *target_errno);
3443 return ret;
3444 }
3445 }
3446
3447 *target_errno = FILEIO_ENOSYS;
3448 return -1;
3449 }
3450
3451 /* Read value of symbolic link FILENAME on the target. Return a
3452 null-terminated string allocated via xmalloc, or NULL if an error
3453 occurs (and set *TARGET_ERRNO). */
3454 char *
3455 target_fileio_readlink (const char *filename, int *target_errno)
3456 {
3457 struct target_ops *t;
3458
3459 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3460 {
3461 if (t->to_fileio_readlink != NULL)
3462 {
3463 char *ret = t->to_fileio_readlink (filename, target_errno);
3464
3465 if (targetdebug)
3466 fprintf_unfiltered (gdb_stdlog,
3467 "target_fileio_readlink (%s) = %s (%d)\n",
3468 filename, ret? ret : "(nil)",
3469 ret? 0 : *target_errno);
3470 return ret;
3471 }
3472 }
3473
3474 *target_errno = FILEIO_ENOSYS;
3475 return NULL;
3476 }
3477
3478 static void
3479 target_fileio_close_cleanup (void *opaque)
3480 {
3481 int fd = *(int *) opaque;
3482 int target_errno;
3483
3484 target_fileio_close (fd, &target_errno);
3485 }
3486
3487 /* Read target file FILENAME. Store the result in *BUF_P and
3488 return the size of the transferred data. PADDING additional bytes are
3489 available in *BUF_P. This is a helper function for
3490 target_fileio_read_alloc; see the declaration of that function for more
3491 information. */
3492
3493 static LONGEST
3494 target_fileio_read_alloc_1 (const char *filename,
3495 gdb_byte **buf_p, int padding)
3496 {
3497 struct cleanup *close_cleanup;
3498 size_t buf_alloc, buf_pos;
3499 gdb_byte *buf;
3500 LONGEST n;
3501 int fd;
3502 int target_errno;
3503
3504 fd = target_fileio_open (filename, FILEIO_O_RDONLY, 0700, &target_errno);
3505 if (fd == -1)
3506 return -1;
3507
3508 close_cleanup = make_cleanup (target_fileio_close_cleanup, &fd);
3509
3510 /* Start by reading up to 4K at a time. The target will throttle
3511 this number down if necessary. */
3512 buf_alloc = 4096;
3513 buf = xmalloc (buf_alloc);
3514 buf_pos = 0;
3515 while (1)
3516 {
3517 n = target_fileio_pread (fd, &buf[buf_pos],
3518 buf_alloc - buf_pos - padding, buf_pos,
3519 &target_errno);
3520 if (n < 0)
3521 {
3522 /* An error occurred. */
3523 do_cleanups (close_cleanup);
3524 xfree (buf);
3525 return -1;
3526 }
3527 else if (n == 0)
3528 {
3529 /* Read all there was. */
3530 do_cleanups (close_cleanup);
3531 if (buf_pos == 0)
3532 xfree (buf);
3533 else
3534 *buf_p = buf;
3535 return buf_pos;
3536 }
3537
3538 buf_pos += n;
3539
3540 /* If the buffer is filling up, expand it. */
3541 if (buf_alloc < buf_pos * 2)
3542 {
3543 buf_alloc *= 2;
3544 buf = xrealloc (buf, buf_alloc);
3545 }
3546
3547 QUIT;
3548 }
3549 }
3550
3551 /* Read target file FILENAME. Store the result in *BUF_P and return
3552 the size of the transferred data. See the declaration in "target.h"
3553 function for more information about the return value. */
3554
3555 LONGEST
3556 target_fileio_read_alloc (const char *filename, gdb_byte **buf_p)
3557 {
3558 return target_fileio_read_alloc_1 (filename, buf_p, 0);
3559 }
3560
3561 /* Read target file FILENAME. The result is NUL-terminated and
3562 returned as a string, allocated using xmalloc. If an error occurs
3563 or the transfer is unsupported, NULL is returned. Empty objects
3564 are returned as allocated but empty strings. A warning is issued
3565 if the result contains any embedded NUL bytes. */
3566
3567 char *
3568 target_fileio_read_stralloc (const char *filename)
3569 {
3570 gdb_byte *buffer;
3571 char *bufstr;
3572 LONGEST i, transferred;
3573
3574 transferred = target_fileio_read_alloc_1 (filename, &buffer, 1);
3575 bufstr = (char *) buffer;
3576
3577 if (transferred < 0)
3578 return NULL;
3579
3580 if (transferred == 0)
3581 return xstrdup ("");
3582
3583 bufstr[transferred] = 0;
3584
3585 /* Check for embedded NUL bytes; but allow trailing NULs. */
3586 for (i = strlen (bufstr); i < transferred; i++)
3587 if (bufstr[i] != 0)
3588 {
3589 warning (_("target file %s "
3590 "contained unexpected null characters"),
3591 filename);
3592 break;
3593 }
3594
3595 return bufstr;
3596 }
3597
3598
3599 static int
3600 default_region_ok_for_hw_watchpoint (CORE_ADDR addr, int len)
3601 {
3602 return (len <= gdbarch_ptr_bit (target_gdbarch ()) / TARGET_CHAR_BIT);
3603 }
3604
3605 static int
3606 default_watchpoint_addr_within_range (struct target_ops *target,
3607 CORE_ADDR addr,
3608 CORE_ADDR start, int length)
3609 {
3610 return addr >= start && addr < start + length;
3611 }
3612
3613 static struct gdbarch *
3614 default_thread_architecture (struct target_ops *ops, ptid_t ptid)
3615 {
3616 return target_gdbarch ();
3617 }
3618
3619 static int
3620 return_zero (void)
3621 {
3622 return 0;
3623 }
3624
3625 static int
3626 return_one (void)
3627 {
3628 return 1;
3629 }
3630
3631 static int
3632 return_minus_one (void)
3633 {
3634 return -1;
3635 }
3636
3637 /*
3638 * Find the next target down the stack from the specified target.
3639 */
3640
3641 struct target_ops *
3642 find_target_beneath (struct target_ops *t)
3643 {
3644 return t->beneath;
3645 }
3646
3647 \f
3648 /* The inferior process has died. Long live the inferior! */
3649
3650 void
3651 generic_mourn_inferior (void)
3652 {
3653 ptid_t ptid;
3654
3655 ptid = inferior_ptid;
3656 inferior_ptid = null_ptid;
3657
3658 /* Mark breakpoints uninserted in case something tries to delete a
3659 breakpoint while we delete the inferior's threads (which would
3660 fail, since the inferior is long gone). */
3661 mark_breakpoints_out ();
3662
3663 if (!ptid_equal (ptid, null_ptid))
3664 {
3665 int pid = ptid_get_pid (ptid);
3666 exit_inferior (pid);
3667 }
3668
3669 /* Note this wipes step-resume breakpoints, so needs to be done
3670 after exit_inferior, which ends up referencing the step-resume
3671 breakpoints through clear_thread_inferior_resources. */
3672 breakpoint_init_inferior (inf_exited);
3673
3674 registers_changed ();
3675
3676 reopen_exec_file ();
3677 reinit_frame_cache ();
3678
3679 if (deprecated_detach_hook)
3680 deprecated_detach_hook ();
3681 }
3682 \f
3683 /* Convert a normal process ID to a string. Returns the string in a
3684 static buffer. */
3685
3686 char *
3687 normal_pid_to_str (ptid_t ptid)
3688 {
3689 static char buf[32];
3690
3691 xsnprintf (buf, sizeof buf, "process %d", ptid_get_pid (ptid));
3692 return buf;
3693 }
3694
3695 static char *
3696 dummy_pid_to_str (struct target_ops *ops, ptid_t ptid)
3697 {
3698 return normal_pid_to_str (ptid);
3699 }
3700
3701 /* Error-catcher for target_find_memory_regions. */
3702 static int
3703 dummy_find_memory_regions (find_memory_region_ftype ignore1, void *ignore2)
3704 {
3705 error (_("Command not implemented for this target."));
3706 return 0;
3707 }
3708
3709 /* Error-catcher for target_make_corefile_notes. */
3710 static char *
3711 dummy_make_corefile_notes (bfd *ignore1, int *ignore2)
3712 {
3713 error (_("Command not implemented for this target."));
3714 return NULL;
3715 }
3716
3717 /* Error-catcher for target_get_bookmark. */
3718 static gdb_byte *
3719 dummy_get_bookmark (char *ignore1, int ignore2)
3720 {
3721 tcomplain ();
3722 return NULL;
3723 }
3724
3725 /* Error-catcher for target_goto_bookmark. */
3726 static void
3727 dummy_goto_bookmark (gdb_byte *ignore, int from_tty)
3728 {
3729 tcomplain ();
3730 }
3731
3732 /* Set up the handful of non-empty slots needed by the dummy target
3733 vector. */
3734
3735 static void
3736 init_dummy_target (void)
3737 {
3738 dummy_target.to_shortname = "None";
3739 dummy_target.to_longname = "None";
3740 dummy_target.to_doc = "";
3741 dummy_target.to_attach = find_default_attach;
3742 dummy_target.to_detach =
3743 (void (*)(struct target_ops *, char *, int))target_ignore;
3744 dummy_target.to_create_inferior = find_default_create_inferior;
3745 dummy_target.to_can_async_p = find_default_can_async_p;
3746 dummy_target.to_is_async_p = find_default_is_async_p;
3747 dummy_target.to_supports_non_stop = find_default_supports_non_stop;
3748 dummy_target.to_supports_disable_randomization
3749 = find_default_supports_disable_randomization;
3750 dummy_target.to_pid_to_str = dummy_pid_to_str;
3751 dummy_target.to_stratum = dummy_stratum;
3752 dummy_target.to_find_memory_regions = dummy_find_memory_regions;
3753 dummy_target.to_make_corefile_notes = dummy_make_corefile_notes;
3754 dummy_target.to_get_bookmark = dummy_get_bookmark;
3755 dummy_target.to_goto_bookmark = dummy_goto_bookmark;
3756 dummy_target.to_xfer_partial = default_xfer_partial;
3757 dummy_target.to_has_all_memory = (int (*) (struct target_ops *)) return_zero;
3758 dummy_target.to_has_memory = (int (*) (struct target_ops *)) return_zero;
3759 dummy_target.to_has_stack = (int (*) (struct target_ops *)) return_zero;
3760 dummy_target.to_has_registers = (int (*) (struct target_ops *)) return_zero;
3761 dummy_target.to_has_execution
3762 = (int (*) (struct target_ops *, ptid_t)) return_zero;
3763 dummy_target.to_stopped_by_watchpoint = return_zero;
3764 dummy_target.to_stopped_data_address =
3765 (int (*) (struct target_ops *, CORE_ADDR *)) return_zero;
3766 dummy_target.to_magic = OPS_MAGIC;
3767 }
3768 \f
3769 static void
3770 debug_to_open (char *args, int from_tty)
3771 {
3772 debug_target.to_open (args, from_tty);
3773
3774 fprintf_unfiltered (gdb_stdlog, "target_open (%s, %d)\n", args, from_tty);
3775 }
3776
3777 void
3778 target_close (struct target_ops *targ)
3779 {
3780 gdb_assert (!target_is_pushed (targ));
3781
3782 if (targ->to_xclose != NULL)
3783 targ->to_xclose (targ);
3784 else if (targ->to_close != NULL)
3785 targ->to_close ();
3786
3787 if (targetdebug)
3788 fprintf_unfiltered (gdb_stdlog, "target_close ()\n");
3789 }
3790
3791 void
3792 target_attach (char *args, int from_tty)
3793 {
3794 struct target_ops *t;
3795
3796 for (t = current_target.beneath; t != NULL; t = t->beneath)
3797 {
3798 if (t->to_attach != NULL)
3799 {
3800 t->to_attach (t, args, from_tty);
3801 if (targetdebug)
3802 fprintf_unfiltered (gdb_stdlog, "target_attach (%s, %d)\n",
3803 args, from_tty);
3804 return;
3805 }
3806 }
3807
3808 internal_error (__FILE__, __LINE__,
3809 _("could not find a target to attach"));
3810 }
3811
3812 int
3813 target_thread_alive (ptid_t ptid)
3814 {
3815 struct target_ops *t;
3816
3817 for (t = current_target.beneath; t != NULL; t = t->beneath)
3818 {
3819 if (t->to_thread_alive != NULL)
3820 {
3821 int retval;
3822
3823 retval = t->to_thread_alive (t, ptid);
3824 if (targetdebug)
3825 fprintf_unfiltered (gdb_stdlog, "target_thread_alive (%d) = %d\n",
3826 PIDGET (ptid), retval);
3827
3828 return retval;
3829 }
3830 }
3831
3832 return 0;
3833 }
3834
3835 void
3836 target_find_new_threads (void)
3837 {
3838 struct target_ops *t;
3839
3840 for (t = current_target.beneath; t != NULL; t = t->beneath)
3841 {
3842 if (t->to_find_new_threads != NULL)
3843 {
3844 t->to_find_new_threads (t);
3845 if (targetdebug)
3846 fprintf_unfiltered (gdb_stdlog, "target_find_new_threads ()\n");
3847
3848 return;
3849 }
3850 }
3851 }
3852
3853 void
3854 target_stop (ptid_t ptid)
3855 {
3856 if (!may_stop)
3857 {
3858 warning (_("May not interrupt or stop the target, ignoring attempt"));
3859 return;
3860 }
3861
3862 (*current_target.to_stop) (ptid);
3863 }
3864
3865 static void
3866 debug_to_post_attach (int pid)
3867 {
3868 debug_target.to_post_attach (pid);
3869
3870 fprintf_unfiltered (gdb_stdlog, "target_post_attach (%d)\n", pid);
3871 }
3872
3873 /* Concatenate ELEM to LIST, a comma separate list, and return the
3874 result. The LIST incoming argument is released. */
3875
3876 static char *
3877 str_comma_list_concat_elem (char *list, const char *elem)
3878 {
3879 if (list == NULL)
3880 return xstrdup (elem);
3881 else
3882 return reconcat (list, list, ", ", elem, (char *) NULL);
3883 }
3884
3885 /* Helper for target_options_to_string. If OPT is present in
3886 TARGET_OPTIONS, append the OPT_STR (string version of OPT) in RET.
3887 Returns the new resulting string. OPT is removed from
3888 TARGET_OPTIONS. */
3889
3890 static char *
3891 do_option (int *target_options, char *ret,
3892 int opt, char *opt_str)
3893 {
3894 if ((*target_options & opt) != 0)
3895 {
3896 ret = str_comma_list_concat_elem (ret, opt_str);
3897 *target_options &= ~opt;
3898 }
3899
3900 return ret;
3901 }
3902
3903 char *
3904 target_options_to_string (int target_options)
3905 {
3906 char *ret = NULL;
3907
3908 #define DO_TARG_OPTION(OPT) \
3909 ret = do_option (&target_options, ret, OPT, #OPT)
3910
3911 DO_TARG_OPTION (TARGET_WNOHANG);
3912
3913 if (target_options != 0)
3914 ret = str_comma_list_concat_elem (ret, "unknown???");
3915
3916 if (ret == NULL)
3917 ret = xstrdup ("");
3918 return ret;
3919 }
3920
3921 static void
3922 debug_print_register (const char * func,
3923 struct regcache *regcache, int regno)
3924 {
3925 struct gdbarch *gdbarch = get_regcache_arch (regcache);
3926
3927 fprintf_unfiltered (gdb_stdlog, "%s ", func);
3928 if (regno >= 0 && regno < gdbarch_num_regs (gdbarch)
3929 && gdbarch_register_name (gdbarch, regno) != NULL
3930 && gdbarch_register_name (gdbarch, regno)[0] != '\0')
3931 fprintf_unfiltered (gdb_stdlog, "(%s)",
3932 gdbarch_register_name (gdbarch, regno));
3933 else
3934 fprintf_unfiltered (gdb_stdlog, "(%d)", regno);
3935 if (regno >= 0 && regno < gdbarch_num_regs (gdbarch))
3936 {
3937 enum bfd_endian byte_order = gdbarch_byte_order (gdbarch);
3938 int i, size = register_size (gdbarch, regno);
3939 gdb_byte buf[MAX_REGISTER_SIZE];
3940
3941 regcache_raw_collect (regcache, regno, buf);
3942 fprintf_unfiltered (gdb_stdlog, " = ");
3943 for (i = 0; i < size; i++)
3944 {
3945 fprintf_unfiltered (gdb_stdlog, "%02x", buf[i]);
3946 }
3947 if (size <= sizeof (LONGEST))
3948 {
3949 ULONGEST val = extract_unsigned_integer (buf, size, byte_order);
3950
3951 fprintf_unfiltered (gdb_stdlog, " %s %s",
3952 core_addr_to_string_nz (val), plongest (val));
3953 }
3954 }
3955 fprintf_unfiltered (gdb_stdlog, "\n");
3956 }
3957
3958 void
3959 target_fetch_registers (struct regcache *regcache, int regno)
3960 {
3961 struct target_ops *t;
3962
3963 for (t = current_target.beneath; t != NULL; t = t->beneath)
3964 {
3965 if (t->to_fetch_registers != NULL)
3966 {
3967 t->to_fetch_registers (t, regcache, regno);
3968 if (targetdebug)
3969 debug_print_register ("target_fetch_registers", regcache, regno);
3970 return;
3971 }
3972 }
3973 }
3974
3975 void
3976 target_store_registers (struct regcache *regcache, int regno)
3977 {
3978 struct target_ops *t;
3979
3980 if (!may_write_registers)
3981 error (_("Writing to registers is not allowed (regno %d)"), regno);
3982
3983 for (t = current_target.beneath; t != NULL; t = t->beneath)
3984 {
3985 if (t->to_store_registers != NULL)
3986 {
3987 t->to_store_registers (t, regcache, regno);
3988 if (targetdebug)
3989 {
3990 debug_print_register ("target_store_registers", regcache, regno);
3991 }
3992 return;
3993 }
3994 }
3995
3996 noprocess ();
3997 }
3998
3999 int
4000 target_core_of_thread (ptid_t ptid)
4001 {
4002 struct target_ops *t;
4003
4004 for (t = current_target.beneath; t != NULL; t = t->beneath)
4005 {
4006 if (t->to_core_of_thread != NULL)
4007 {
4008 int retval = t->to_core_of_thread (t, ptid);
4009
4010 if (targetdebug)
4011 fprintf_unfiltered (gdb_stdlog,
4012 "target_core_of_thread (%d) = %d\n",
4013 PIDGET (ptid), retval);
4014 return retval;
4015 }
4016 }
4017
4018 return -1;
4019 }
4020
4021 int
4022 target_verify_memory (const gdb_byte *data, CORE_ADDR memaddr, ULONGEST size)
4023 {
4024 struct target_ops *t;
4025
4026 for (t = current_target.beneath; t != NULL; t = t->beneath)
4027 {
4028 if (t->to_verify_memory != NULL)
4029 {
4030 int retval = t->to_verify_memory (t, data, memaddr, size);
4031
4032 if (targetdebug)
4033 fprintf_unfiltered (gdb_stdlog,
4034 "target_verify_memory (%s, %s) = %d\n",
4035 paddress (target_gdbarch (), memaddr),
4036 pulongest (size),
4037 retval);
4038 return retval;
4039 }
4040 }
4041
4042 tcomplain ();
4043 }
4044
4045 /* The documentation for this function is in its prototype declaration in
4046 target.h. */
4047
4048 int
4049 target_insert_mask_watchpoint (CORE_ADDR addr, CORE_ADDR mask, int rw)
4050 {
4051 struct target_ops *t;
4052
4053 for (t = current_target.beneath; t != NULL; t = t->beneath)
4054 if (t->to_insert_mask_watchpoint != NULL)
4055 {
4056 int ret;
4057
4058 ret = t->to_insert_mask_watchpoint (t, addr, mask, rw);
4059
4060 if (targetdebug)
4061 fprintf_unfiltered (gdb_stdlog, "\
4062 target_insert_mask_watchpoint (%s, %s, %d) = %d\n",
4063 core_addr_to_string (addr),
4064 core_addr_to_string (mask), rw, ret);
4065
4066 return ret;
4067 }
4068
4069 return 1;
4070 }
4071
4072 /* The documentation for this function is in its prototype declaration in
4073 target.h. */
4074
4075 int
4076 target_remove_mask_watchpoint (CORE_ADDR addr, CORE_ADDR mask, int rw)
4077 {
4078 struct target_ops *t;
4079
4080 for (t = current_target.beneath; t != NULL; t = t->beneath)
4081 if (t->to_remove_mask_watchpoint != NULL)
4082 {
4083 int ret;
4084
4085 ret = t->to_remove_mask_watchpoint (t, addr, mask, rw);
4086
4087 if (targetdebug)
4088 fprintf_unfiltered (gdb_stdlog, "\
4089 target_remove_mask_watchpoint (%s, %s, %d) = %d\n",
4090 core_addr_to_string (addr),
4091 core_addr_to_string (mask), rw, ret);
4092
4093 return ret;
4094 }
4095
4096 return 1;
4097 }
4098
4099 /* The documentation for this function is in its prototype declaration
4100 in target.h. */
4101
4102 int
4103 target_masked_watch_num_registers (CORE_ADDR addr, CORE_ADDR mask)
4104 {
4105 struct target_ops *t;
4106
4107 for (t = current_target.beneath; t != NULL; t = t->beneath)
4108 if (t->to_masked_watch_num_registers != NULL)
4109 return t->to_masked_watch_num_registers (t, addr, mask);
4110
4111 return -1;
4112 }
4113
4114 /* The documentation for this function is in its prototype declaration
4115 in target.h. */
4116
4117 int
4118 target_ranged_break_num_registers (void)
4119 {
4120 struct target_ops *t;
4121
4122 for (t = current_target.beneath; t != NULL; t = t->beneath)
4123 if (t->to_ranged_break_num_registers != NULL)
4124 return t->to_ranged_break_num_registers (t);
4125
4126 return -1;
4127 }
4128
4129 /* See target.h. */
4130
4131 int
4132 target_supports_btrace (void)
4133 {
4134 struct target_ops *t;
4135
4136 for (t = current_target.beneath; t != NULL; t = t->beneath)
4137 if (t->to_supports_btrace != NULL)
4138 return t->to_supports_btrace ();
4139
4140 return 0;
4141 }
4142
4143 /* See target.h. */
4144
4145 struct btrace_target_info *
4146 target_enable_btrace (ptid_t ptid)
4147 {
4148 struct target_ops *t;
4149
4150 for (t = current_target.beneath; t != NULL; t = t->beneath)
4151 if (t->to_enable_btrace != NULL)
4152 return t->to_enable_btrace (ptid);
4153
4154 tcomplain ();
4155 return NULL;
4156 }
4157
4158 /* See target.h. */
4159
4160 void
4161 target_disable_btrace (struct btrace_target_info *btinfo)
4162 {
4163 struct target_ops *t;
4164
4165 for (t = current_target.beneath; t != NULL; t = t->beneath)
4166 if (t->to_disable_btrace != NULL)
4167 return t->to_disable_btrace (btinfo);
4168
4169 tcomplain ();
4170 }
4171
4172 /* See target.h. */
4173
4174 void
4175 target_teardown_btrace (struct btrace_target_info *btinfo)
4176 {
4177 struct target_ops *t;
4178
4179 for (t = current_target.beneath; t != NULL; t = t->beneath)
4180 if (t->to_teardown_btrace != NULL)
4181 return t->to_teardown_btrace (btinfo);
4182
4183 tcomplain ();
4184 }
4185
4186 /* See target.h. */
4187
4188 VEC (btrace_block_s) *
4189 target_read_btrace (struct btrace_target_info *btinfo,
4190 enum btrace_read_type type)
4191 {
4192 struct target_ops *t;
4193
4194 for (t = current_target.beneath; t != NULL; t = t->beneath)
4195 if (t->to_read_btrace != NULL)
4196 return t->to_read_btrace (btinfo, type);
4197
4198 tcomplain ();
4199 return NULL;
4200 }
4201
4202 /* See target.h. */
4203
4204 void
4205 target_stop_recording (void)
4206 {
4207 struct target_ops *t;
4208
4209 for (t = current_target.beneath; t != NULL; t = t->beneath)
4210 if (t->to_stop_recording != NULL)
4211 {
4212 t->to_stop_recording ();
4213 return;
4214 }
4215
4216 /* This is optional. */
4217 }
4218
4219 /* See target.h. */
4220
4221 void
4222 target_info_record (void)
4223 {
4224 struct target_ops *t;
4225
4226 for (t = current_target.beneath; t != NULL; t = t->beneath)
4227 if (t->to_info_record != NULL)
4228 {
4229 t->to_info_record ();
4230 return;
4231 }
4232
4233 tcomplain ();
4234 }
4235
4236 /* See target.h. */
4237
4238 void
4239 target_save_record (const char *filename)
4240 {
4241 struct target_ops *t;
4242
4243 for (t = current_target.beneath; t != NULL; t = t->beneath)
4244 if (t->to_save_record != NULL)
4245 {
4246 t->to_save_record (filename);
4247 return;
4248 }
4249
4250 tcomplain ();
4251 }
4252
4253 /* See target.h. */
4254
4255 int
4256 target_supports_delete_record (void)
4257 {
4258 struct target_ops *t;
4259
4260 for (t = current_target.beneath; t != NULL; t = t->beneath)
4261 if (t->to_delete_record != NULL)
4262 return 1;
4263
4264 return 0;
4265 }
4266
4267 /* See target.h. */
4268
4269 void
4270 target_delete_record (void)
4271 {
4272 struct target_ops *t;
4273
4274 for (t = current_target.beneath; t != NULL; t = t->beneath)
4275 if (t->to_delete_record != NULL)
4276 {
4277 t->to_delete_record ();
4278 return;
4279 }
4280
4281 tcomplain ();
4282 }
4283
4284 /* See target.h. */
4285
4286 int
4287 target_record_is_replaying (void)
4288 {
4289 struct target_ops *t;
4290
4291 for (t = current_target.beneath; t != NULL; t = t->beneath)
4292 if (t->to_record_is_replaying != NULL)
4293 return t->to_record_is_replaying ();
4294
4295 return 0;
4296 }
4297
4298 /* See target.h. */
4299
4300 void
4301 target_goto_record_begin (void)
4302 {
4303 struct target_ops *t;
4304
4305 for (t = current_target.beneath; t != NULL; t = t->beneath)
4306 if (t->to_goto_record_begin != NULL)
4307 {
4308 t->to_goto_record_begin ();
4309 return;
4310 }
4311
4312 tcomplain ();
4313 }
4314
4315 /* See target.h. */
4316
4317 void
4318 target_goto_record_end (void)
4319 {
4320 struct target_ops *t;
4321
4322 for (t = current_target.beneath; t != NULL; t = t->beneath)
4323 if (t->to_goto_record_end != NULL)
4324 {
4325 t->to_goto_record_end ();
4326 return;
4327 }
4328
4329 tcomplain ();
4330 }
4331
4332 /* See target.h. */
4333
4334 void
4335 target_goto_record (ULONGEST insn)
4336 {
4337 struct target_ops *t;
4338
4339 for (t = current_target.beneath; t != NULL; t = t->beneath)
4340 if (t->to_goto_record != NULL)
4341 {
4342 t->to_goto_record (insn);
4343 return;
4344 }
4345
4346 tcomplain ();
4347 }
4348
4349 /* See target.h. */
4350
4351 void
4352 target_insn_history (int size, int flags)
4353 {
4354 struct target_ops *t;
4355
4356 for (t = current_target.beneath; t != NULL; t = t->beneath)
4357 if (t->to_insn_history != NULL)
4358 {
4359 t->to_insn_history (size, flags);
4360 return;
4361 }
4362
4363 tcomplain ();
4364 }
4365
4366 /* See target.h. */
4367
4368 void
4369 target_insn_history_from (ULONGEST from, int size, int flags)
4370 {
4371 struct target_ops *t;
4372
4373 for (t = current_target.beneath; t != NULL; t = t->beneath)
4374 if (t->to_insn_history_from != NULL)
4375 {
4376 t->to_insn_history_from (from, size, flags);
4377 return;
4378 }
4379
4380 tcomplain ();
4381 }
4382
4383 /* See target.h. */
4384
4385 void
4386 target_insn_history_range (ULONGEST begin, ULONGEST end, int flags)
4387 {
4388 struct target_ops *t;
4389
4390 for (t = current_target.beneath; t != NULL; t = t->beneath)
4391 if (t->to_insn_history_range != NULL)
4392 {
4393 t->to_insn_history_range (begin, end, flags);
4394 return;
4395 }
4396
4397 tcomplain ();
4398 }
4399
4400 /* See target.h. */
4401
4402 void
4403 target_call_history (int size, int flags)
4404 {
4405 struct target_ops *t;
4406
4407 for (t = current_target.beneath; t != NULL; t = t->beneath)
4408 if (t->to_call_history != NULL)
4409 {
4410 t->to_call_history (size, flags);
4411 return;
4412 }
4413
4414 tcomplain ();
4415 }
4416
4417 /* See target.h. */
4418
4419 void
4420 target_call_history_from (ULONGEST begin, int size, int flags)
4421 {
4422 struct target_ops *t;
4423
4424 for (t = current_target.beneath; t != NULL; t = t->beneath)
4425 if (t->to_call_history_from != NULL)
4426 {
4427 t->to_call_history_from (begin, size, flags);
4428 return;
4429 }
4430
4431 tcomplain ();
4432 }
4433
4434 /* See target.h. */
4435
4436 void
4437 target_call_history_range (ULONGEST begin, ULONGEST end, int flags)
4438 {
4439 struct target_ops *t;
4440
4441 for (t = current_target.beneath; t != NULL; t = t->beneath)
4442 if (t->to_call_history_range != NULL)
4443 {
4444 t->to_call_history_range (begin, end, flags);
4445 return;
4446 }
4447
4448 tcomplain ();
4449 }
4450
4451 static void
4452 debug_to_prepare_to_store (struct regcache *regcache)
4453 {
4454 debug_target.to_prepare_to_store (regcache);
4455
4456 fprintf_unfiltered (gdb_stdlog, "target_prepare_to_store ()\n");
4457 }
4458
4459 static int
4460 deprecated_debug_xfer_memory (CORE_ADDR memaddr, bfd_byte *myaddr, int len,
4461 int write, struct mem_attrib *attrib,
4462 struct target_ops *target)
4463 {
4464 int retval;
4465
4466 retval = debug_target.deprecated_xfer_memory (memaddr, myaddr, len, write,
4467 attrib, target);
4468
4469 fprintf_unfiltered (gdb_stdlog,
4470 "target_xfer_memory (%s, xxx, %d, %s, xxx) = %d",
4471 paddress (target_gdbarch (), memaddr), len,
4472 write ? "write" : "read", retval);
4473
4474 if (retval > 0)
4475 {
4476 int i;
4477
4478 fputs_unfiltered (", bytes =", gdb_stdlog);
4479 for (i = 0; i < retval; i++)
4480 {
4481 if ((((intptr_t) &(myaddr[i])) & 0xf) == 0)
4482 {
4483 if (targetdebug < 2 && i > 0)
4484 {
4485 fprintf_unfiltered (gdb_stdlog, " ...");
4486 break;
4487 }
4488 fprintf_unfiltered (gdb_stdlog, "\n");
4489 }
4490
4491 fprintf_unfiltered (gdb_stdlog, " %02x", myaddr[i] & 0xff);
4492 }
4493 }
4494
4495 fputc_unfiltered ('\n', gdb_stdlog);
4496
4497 return retval;
4498 }
4499
4500 static void
4501 debug_to_files_info (struct target_ops *target)
4502 {
4503 debug_target.to_files_info (target);
4504
4505 fprintf_unfiltered (gdb_stdlog, "target_files_info (xxx)\n");
4506 }
4507
4508 static int
4509 debug_to_insert_breakpoint (struct gdbarch *gdbarch,
4510 struct bp_target_info *bp_tgt)
4511 {
4512 int retval;
4513
4514 retval = debug_target.to_insert_breakpoint (gdbarch, bp_tgt);
4515
4516 fprintf_unfiltered (gdb_stdlog,
4517 "target_insert_breakpoint (%s, xxx) = %ld\n",
4518 core_addr_to_string (bp_tgt->placed_address),
4519 (unsigned long) retval);
4520 return retval;
4521 }
4522
4523 static int
4524 debug_to_remove_breakpoint (struct gdbarch *gdbarch,
4525 struct bp_target_info *bp_tgt)
4526 {
4527 int retval;
4528
4529 retval = debug_target.to_remove_breakpoint (gdbarch, bp_tgt);
4530
4531 fprintf_unfiltered (gdb_stdlog,
4532 "target_remove_breakpoint (%s, xxx) = %ld\n",
4533 core_addr_to_string (bp_tgt->placed_address),
4534 (unsigned long) retval);
4535 return retval;
4536 }
4537
4538 static int
4539 debug_to_can_use_hw_breakpoint (int type, int cnt, int from_tty)
4540 {
4541 int retval;
4542
4543 retval = debug_target.to_can_use_hw_breakpoint (type, cnt, from_tty);
4544
4545 fprintf_unfiltered (gdb_stdlog,
4546 "target_can_use_hw_breakpoint (%ld, %ld, %ld) = %ld\n",
4547 (unsigned long) type,
4548 (unsigned long) cnt,
4549 (unsigned long) from_tty,
4550 (unsigned long) retval);
4551 return retval;
4552 }
4553
4554 static int
4555 debug_to_region_ok_for_hw_watchpoint (CORE_ADDR addr, int len)
4556 {
4557 CORE_ADDR retval;
4558
4559 retval = debug_target.to_region_ok_for_hw_watchpoint (addr, len);
4560
4561 fprintf_unfiltered (gdb_stdlog,
4562 "target_region_ok_for_hw_watchpoint (%s, %ld) = %s\n",
4563 core_addr_to_string (addr), (unsigned long) len,
4564 core_addr_to_string (retval));
4565 return retval;
4566 }
4567
4568 static int
4569 debug_to_can_accel_watchpoint_condition (CORE_ADDR addr, int len, int rw,
4570 struct expression *cond)
4571 {
4572 int retval;
4573
4574 retval = debug_target.to_can_accel_watchpoint_condition (addr, len,
4575 rw, cond);
4576
4577 fprintf_unfiltered (gdb_stdlog,
4578 "target_can_accel_watchpoint_condition "
4579 "(%s, %d, %d, %s) = %ld\n",
4580 core_addr_to_string (addr), len, rw,
4581 host_address_to_string (cond), (unsigned long) retval);
4582 return retval;
4583 }
4584
4585 static int
4586 debug_to_stopped_by_watchpoint (void)
4587 {
4588 int retval;
4589
4590 retval = debug_target.to_stopped_by_watchpoint ();
4591
4592 fprintf_unfiltered (gdb_stdlog,
4593 "target_stopped_by_watchpoint () = %ld\n",
4594 (unsigned long) retval);
4595 return retval;
4596 }
4597
4598 static int
4599 debug_to_stopped_data_address (struct target_ops *target, CORE_ADDR *addr)
4600 {
4601 int retval;
4602
4603 retval = debug_target.to_stopped_data_address (target, addr);
4604
4605 fprintf_unfiltered (gdb_stdlog,
4606 "target_stopped_data_address ([%s]) = %ld\n",
4607 core_addr_to_string (*addr),
4608 (unsigned long)retval);
4609 return retval;
4610 }
4611
4612 static int
4613 debug_to_watchpoint_addr_within_range (struct target_ops *target,
4614 CORE_ADDR addr,
4615 CORE_ADDR start, int length)
4616 {
4617 int retval;
4618
4619 retval = debug_target.to_watchpoint_addr_within_range (target, addr,
4620 start, length);
4621
4622 fprintf_filtered (gdb_stdlog,
4623 "target_watchpoint_addr_within_range (%s, %s, %d) = %d\n",
4624 core_addr_to_string (addr), core_addr_to_string (start),
4625 length, retval);
4626 return retval;
4627 }
4628
4629 static int
4630 debug_to_insert_hw_breakpoint (struct gdbarch *gdbarch,
4631 struct bp_target_info *bp_tgt)
4632 {
4633 int retval;
4634
4635 retval = debug_target.to_insert_hw_breakpoint (gdbarch, bp_tgt);
4636
4637 fprintf_unfiltered (gdb_stdlog,
4638 "target_insert_hw_breakpoint (%s, xxx) = %ld\n",
4639 core_addr_to_string (bp_tgt->placed_address),
4640 (unsigned long) retval);
4641 return retval;
4642 }
4643
4644 static int
4645 debug_to_remove_hw_breakpoint (struct gdbarch *gdbarch,
4646 struct bp_target_info *bp_tgt)
4647 {
4648 int retval;
4649
4650 retval = debug_target.to_remove_hw_breakpoint (gdbarch, bp_tgt);
4651
4652 fprintf_unfiltered (gdb_stdlog,
4653 "target_remove_hw_breakpoint (%s, xxx) = %ld\n",
4654 core_addr_to_string (bp_tgt->placed_address),
4655 (unsigned long) retval);
4656 return retval;
4657 }
4658
4659 static int
4660 debug_to_insert_watchpoint (CORE_ADDR addr, int len, int type,
4661 struct expression *cond)
4662 {
4663 int retval;
4664
4665 retval = debug_target.to_insert_watchpoint (addr, len, type, cond);
4666
4667 fprintf_unfiltered (gdb_stdlog,
4668 "target_insert_watchpoint (%s, %d, %d, %s) = %ld\n",
4669 core_addr_to_string (addr), len, type,
4670 host_address_to_string (cond), (unsigned long) retval);
4671 return retval;
4672 }
4673
4674 static int
4675 debug_to_remove_watchpoint (CORE_ADDR addr, int len, int type,
4676 struct expression *cond)
4677 {
4678 int retval;
4679
4680 retval = debug_target.to_remove_watchpoint (addr, len, type, cond);
4681
4682 fprintf_unfiltered (gdb_stdlog,
4683 "target_remove_watchpoint (%s, %d, %d, %s) = %ld\n",
4684 core_addr_to_string (addr), len, type,
4685 host_address_to_string (cond), (unsigned long) retval);
4686 return retval;
4687 }
4688
4689 static void
4690 debug_to_terminal_init (void)
4691 {
4692 debug_target.to_terminal_init ();
4693
4694 fprintf_unfiltered (gdb_stdlog, "target_terminal_init ()\n");
4695 }
4696
4697 static void
4698 debug_to_terminal_inferior (void)
4699 {
4700 debug_target.to_terminal_inferior ();
4701
4702 fprintf_unfiltered (gdb_stdlog, "target_terminal_inferior ()\n");
4703 }
4704
4705 static void
4706 debug_to_terminal_ours_for_output (void)
4707 {
4708 debug_target.to_terminal_ours_for_output ();
4709
4710 fprintf_unfiltered (gdb_stdlog, "target_terminal_ours_for_output ()\n");
4711 }
4712
4713 static void
4714 debug_to_terminal_ours (void)
4715 {
4716 debug_target.to_terminal_ours ();
4717
4718 fprintf_unfiltered (gdb_stdlog, "target_terminal_ours ()\n");
4719 }
4720
4721 static void
4722 debug_to_terminal_save_ours (void)
4723 {
4724 debug_target.to_terminal_save_ours ();
4725
4726 fprintf_unfiltered (gdb_stdlog, "target_terminal_save_ours ()\n");
4727 }
4728
4729 static void
4730 debug_to_terminal_info (const char *arg, int from_tty)
4731 {
4732 debug_target.to_terminal_info (arg, from_tty);
4733
4734 fprintf_unfiltered (gdb_stdlog, "target_terminal_info (%s, %d)\n", arg,
4735 from_tty);
4736 }
4737
4738 static void
4739 debug_to_load (char *args, int from_tty)
4740 {
4741 debug_target.to_load (args, from_tty);
4742
4743 fprintf_unfiltered (gdb_stdlog, "target_load (%s, %d)\n", args, from_tty);
4744 }
4745
4746 static void
4747 debug_to_post_startup_inferior (ptid_t ptid)
4748 {
4749 debug_target.to_post_startup_inferior (ptid);
4750
4751 fprintf_unfiltered (gdb_stdlog, "target_post_startup_inferior (%d)\n",
4752 PIDGET (ptid));
4753 }
4754
4755 static int
4756 debug_to_insert_fork_catchpoint (int pid)
4757 {
4758 int retval;
4759
4760 retval = debug_target.to_insert_fork_catchpoint (pid);
4761
4762 fprintf_unfiltered (gdb_stdlog, "target_insert_fork_catchpoint (%d) = %d\n",
4763 pid, retval);
4764
4765 return retval;
4766 }
4767
4768 static int
4769 debug_to_remove_fork_catchpoint (int pid)
4770 {
4771 int retval;
4772
4773 retval = debug_target.to_remove_fork_catchpoint (pid);
4774
4775 fprintf_unfiltered (gdb_stdlog, "target_remove_fork_catchpoint (%d) = %d\n",
4776 pid, retval);
4777
4778 return retval;
4779 }
4780
4781 static int
4782 debug_to_insert_vfork_catchpoint (int pid)
4783 {
4784 int retval;
4785
4786 retval = debug_target.to_insert_vfork_catchpoint (pid);
4787
4788 fprintf_unfiltered (gdb_stdlog, "target_insert_vfork_catchpoint (%d) = %d\n",
4789 pid, retval);
4790
4791 return retval;
4792 }
4793
4794 static int
4795 debug_to_remove_vfork_catchpoint (int pid)
4796 {
4797 int retval;
4798
4799 retval = debug_target.to_remove_vfork_catchpoint (pid);
4800
4801 fprintf_unfiltered (gdb_stdlog, "target_remove_vfork_catchpoint (%d) = %d\n",
4802 pid, retval);
4803
4804 return retval;
4805 }
4806
4807 static int
4808 debug_to_insert_exec_catchpoint (int pid)
4809 {
4810 int retval;
4811
4812 retval = debug_target.to_insert_exec_catchpoint (pid);
4813
4814 fprintf_unfiltered (gdb_stdlog, "target_insert_exec_catchpoint (%d) = %d\n",
4815 pid, retval);
4816
4817 return retval;
4818 }
4819
4820 static int
4821 debug_to_remove_exec_catchpoint (int pid)
4822 {
4823 int retval;
4824
4825 retval = debug_target.to_remove_exec_catchpoint (pid);
4826
4827 fprintf_unfiltered (gdb_stdlog, "target_remove_exec_catchpoint (%d) = %d\n",
4828 pid, retval);
4829
4830 return retval;
4831 }
4832
4833 static int
4834 debug_to_has_exited (int pid, int wait_status, int *exit_status)
4835 {
4836 int has_exited;
4837
4838 has_exited = debug_target.to_has_exited (pid, wait_status, exit_status);
4839
4840 fprintf_unfiltered (gdb_stdlog, "target_has_exited (%d, %d, %d) = %d\n",
4841 pid, wait_status, *exit_status, has_exited);
4842
4843 return has_exited;
4844 }
4845
4846 static int
4847 debug_to_can_run (void)
4848 {
4849 int retval;
4850
4851 retval = debug_target.to_can_run ();
4852
4853 fprintf_unfiltered (gdb_stdlog, "target_can_run () = %d\n", retval);
4854
4855 return retval;
4856 }
4857
4858 static struct gdbarch *
4859 debug_to_thread_architecture (struct target_ops *ops, ptid_t ptid)
4860 {
4861 struct gdbarch *retval;
4862
4863 retval = debug_target.to_thread_architecture (ops, ptid);
4864
4865 fprintf_unfiltered (gdb_stdlog,
4866 "target_thread_architecture (%s) = %s [%s]\n",
4867 target_pid_to_str (ptid),
4868 host_address_to_string (retval),
4869 gdbarch_bfd_arch_info (retval)->printable_name);
4870 return retval;
4871 }
4872
4873 static void
4874 debug_to_stop (ptid_t ptid)
4875 {
4876 debug_target.to_stop (ptid);
4877
4878 fprintf_unfiltered (gdb_stdlog, "target_stop (%s)\n",
4879 target_pid_to_str (ptid));
4880 }
4881
4882 static void
4883 debug_to_rcmd (char *command,
4884 struct ui_file *outbuf)
4885 {
4886 debug_target.to_rcmd (command, outbuf);
4887 fprintf_unfiltered (gdb_stdlog, "target_rcmd (%s, ...)\n", command);
4888 }
4889
4890 static char *
4891 debug_to_pid_to_exec_file (int pid)
4892 {
4893 char *exec_file;
4894
4895 exec_file = debug_target.to_pid_to_exec_file (pid);
4896
4897 fprintf_unfiltered (gdb_stdlog, "target_pid_to_exec_file (%d) = %s\n",
4898 pid, exec_file);
4899
4900 return exec_file;
4901 }
4902
4903 static void
4904 setup_target_debug (void)
4905 {
4906 memcpy (&debug_target, &current_target, sizeof debug_target);
4907
4908 current_target.to_open = debug_to_open;
4909 current_target.to_post_attach = debug_to_post_attach;
4910 current_target.to_prepare_to_store = debug_to_prepare_to_store;
4911 current_target.deprecated_xfer_memory = deprecated_debug_xfer_memory;
4912 current_target.to_files_info = debug_to_files_info;
4913 current_target.to_insert_breakpoint = debug_to_insert_breakpoint;
4914 current_target.to_remove_breakpoint = debug_to_remove_breakpoint;
4915 current_target.to_can_use_hw_breakpoint = debug_to_can_use_hw_breakpoint;
4916 current_target.to_insert_hw_breakpoint = debug_to_insert_hw_breakpoint;
4917 current_target.to_remove_hw_breakpoint = debug_to_remove_hw_breakpoint;
4918 current_target.to_insert_watchpoint = debug_to_insert_watchpoint;
4919 current_target.to_remove_watchpoint = debug_to_remove_watchpoint;
4920 current_target.to_stopped_by_watchpoint = debug_to_stopped_by_watchpoint;
4921 current_target.to_stopped_data_address = debug_to_stopped_data_address;
4922 current_target.to_watchpoint_addr_within_range
4923 = debug_to_watchpoint_addr_within_range;
4924 current_target.to_region_ok_for_hw_watchpoint
4925 = debug_to_region_ok_for_hw_watchpoint;
4926 current_target.to_can_accel_watchpoint_condition
4927 = debug_to_can_accel_watchpoint_condition;
4928 current_target.to_terminal_init = debug_to_terminal_init;
4929 current_target.to_terminal_inferior = debug_to_terminal_inferior;
4930 current_target.to_terminal_ours_for_output
4931 = debug_to_terminal_ours_for_output;
4932 current_target.to_terminal_ours = debug_to_terminal_ours;
4933 current_target.to_terminal_save_ours = debug_to_terminal_save_ours;
4934 current_target.to_terminal_info = debug_to_terminal_info;
4935 current_target.to_load = debug_to_load;
4936 current_target.to_post_startup_inferior = debug_to_post_startup_inferior;
4937 current_target.to_insert_fork_catchpoint = debug_to_insert_fork_catchpoint;
4938 current_target.to_remove_fork_catchpoint = debug_to_remove_fork_catchpoint;
4939 current_target.to_insert_vfork_catchpoint = debug_to_insert_vfork_catchpoint;
4940 current_target.to_remove_vfork_catchpoint = debug_to_remove_vfork_catchpoint;
4941 current_target.to_insert_exec_catchpoint = debug_to_insert_exec_catchpoint;
4942 current_target.to_remove_exec_catchpoint = debug_to_remove_exec_catchpoint;
4943 current_target.to_has_exited = debug_to_has_exited;
4944 current_target.to_can_run = debug_to_can_run;
4945 current_target.to_stop = debug_to_stop;
4946 current_target.to_rcmd = debug_to_rcmd;
4947 current_target.to_pid_to_exec_file = debug_to_pid_to_exec_file;
4948 current_target.to_thread_architecture = debug_to_thread_architecture;
4949 }
4950 \f
4951
4952 static char targ_desc[] =
4953 "Names of targets and files being debugged.\nShows the entire \
4954 stack of targets currently in use (including the exec-file,\n\
4955 core-file, and process, if any), as well as the symbol file name.";
4956
4957 static void
4958 do_monitor_command (char *cmd,
4959 int from_tty)
4960 {
4961 if ((current_target.to_rcmd
4962 == (void (*) (char *, struct ui_file *)) tcomplain)
4963 || (current_target.to_rcmd == debug_to_rcmd
4964 && (debug_target.to_rcmd
4965 == (void (*) (char *, struct ui_file *)) tcomplain)))
4966 error (_("\"monitor\" command not supported by this target."));
4967 target_rcmd (cmd, gdb_stdtarg);
4968 }
4969
4970 /* Print the name of each layers of our target stack. */
4971
4972 static void
4973 maintenance_print_target_stack (char *cmd, int from_tty)
4974 {
4975 struct target_ops *t;
4976
4977 printf_filtered (_("The current target stack is:\n"));
4978
4979 for (t = target_stack; t != NULL; t = t->beneath)
4980 {
4981 printf_filtered (" - %s (%s)\n", t->to_shortname, t->to_longname);
4982 }
4983 }
4984
4985 /* Controls if async mode is permitted. */
4986 int target_async_permitted = 0;
4987
4988 /* The set command writes to this variable. If the inferior is
4989 executing, target_async_permitted is *not* updated. */
4990 static int target_async_permitted_1 = 0;
4991
4992 static void
4993 set_target_async_command (char *args, int from_tty,
4994 struct cmd_list_element *c)
4995 {
4996 if (have_live_inferiors ())
4997 {
4998 target_async_permitted_1 = target_async_permitted;
4999 error (_("Cannot change this setting while the inferior is running."));
5000 }
5001
5002 target_async_permitted = target_async_permitted_1;
5003 }
5004
5005 static void
5006 show_target_async_command (struct ui_file *file, int from_tty,
5007 struct cmd_list_element *c,
5008 const char *value)
5009 {
5010 fprintf_filtered (file,
5011 _("Controlling the inferior in "
5012 "asynchronous mode is %s.\n"), value);
5013 }
5014
5015 /* Temporary copies of permission settings. */
5016
5017 static int may_write_registers_1 = 1;
5018 static int may_write_memory_1 = 1;
5019 static int may_insert_breakpoints_1 = 1;
5020 static int may_insert_tracepoints_1 = 1;
5021 static int may_insert_fast_tracepoints_1 = 1;
5022 static int may_stop_1 = 1;
5023
5024 /* Make the user-set values match the real values again. */
5025
5026 void
5027 update_target_permissions (void)
5028 {
5029 may_write_registers_1 = may_write_registers;
5030 may_write_memory_1 = may_write_memory;
5031 may_insert_breakpoints_1 = may_insert_breakpoints;
5032 may_insert_tracepoints_1 = may_insert_tracepoints;
5033 may_insert_fast_tracepoints_1 = may_insert_fast_tracepoints;
5034 may_stop_1 = may_stop;
5035 }
5036
5037 /* The one function handles (most of) the permission flags in the same
5038 way. */
5039
5040 static void
5041 set_target_permissions (char *args, int from_tty,
5042 struct cmd_list_element *c)
5043 {
5044 if (target_has_execution)
5045 {
5046 update_target_permissions ();
5047 error (_("Cannot change this setting while the inferior is running."));
5048 }
5049
5050 /* Make the real values match the user-changed values. */
5051 may_write_registers = may_write_registers_1;
5052 may_insert_breakpoints = may_insert_breakpoints_1;
5053 may_insert_tracepoints = may_insert_tracepoints_1;
5054 may_insert_fast_tracepoints = may_insert_fast_tracepoints_1;
5055 may_stop = may_stop_1;
5056 update_observer_mode ();
5057 }
5058
5059 /* Set memory write permission independently of observer mode. */
5060
5061 static void
5062 set_write_memory_permission (char *args, int from_tty,
5063 struct cmd_list_element *c)
5064 {
5065 /* Make the real values match the user-changed values. */
5066 may_write_memory = may_write_memory_1;
5067 update_observer_mode ();
5068 }
5069
5070
5071 void
5072 initialize_targets (void)
5073 {
5074 init_dummy_target ();
5075 push_target (&dummy_target);
5076
5077 add_info ("target", target_info, targ_desc);
5078 add_info ("files", target_info, targ_desc);
5079
5080 add_setshow_zuinteger_cmd ("target", class_maintenance, &targetdebug, _("\
5081 Set target debugging."), _("\
5082 Show target debugging."), _("\
5083 When non-zero, target debugging is enabled. Higher numbers are more\n\
5084 verbose. Changes do not take effect until the next \"run\" or \"target\"\n\
5085 command."),
5086 NULL,
5087 show_targetdebug,
5088 &setdebuglist, &showdebuglist);
5089
5090 add_setshow_boolean_cmd ("trust-readonly-sections", class_support,
5091 &trust_readonly, _("\
5092 Set mode for reading from readonly sections."), _("\
5093 Show mode for reading from readonly sections."), _("\
5094 When this mode is on, memory reads from readonly sections (such as .text)\n\
5095 will be read from the object file instead of from the target. This will\n\
5096 result in significant performance improvement for remote targets."),
5097 NULL,
5098 show_trust_readonly,
5099 &setlist, &showlist);
5100
5101 add_com ("monitor", class_obscure, do_monitor_command,
5102 _("Send a command to the remote monitor (remote targets only)."));
5103
5104 add_cmd ("target-stack", class_maintenance, maintenance_print_target_stack,
5105 _("Print the name of each layer of the internal target stack."),
5106 &maintenanceprintlist);
5107
5108 add_setshow_boolean_cmd ("target-async", no_class,
5109 &target_async_permitted_1, _("\
5110 Set whether gdb controls the inferior in asynchronous mode."), _("\
5111 Show whether gdb controls the inferior in asynchronous mode."), _("\
5112 Tells gdb whether to control the inferior in asynchronous mode."),
5113 set_target_async_command,
5114 show_target_async_command,
5115 &setlist,
5116 &showlist);
5117
5118 add_setshow_boolean_cmd ("stack-cache", class_support,
5119 &stack_cache_enabled_p_1, _("\
5120 Set cache use for stack access."), _("\
5121 Show cache use for stack access."), _("\
5122 When on, use the data cache for all stack access, regardless of any\n\
5123 configured memory regions. This improves remote performance significantly.\n\
5124 By default, caching for stack access is on."),
5125 set_stack_cache_enabled_p,
5126 show_stack_cache_enabled_p,
5127 &setlist, &showlist);
5128
5129 add_setshow_boolean_cmd ("may-write-registers", class_support,
5130 &may_write_registers_1, _("\
5131 Set permission to write into registers."), _("\
5132 Show permission to write into registers."), _("\
5133 When this permission is on, GDB may write into the target's registers.\n\
5134 Otherwise, any sort of write attempt will result in an error."),
5135 set_target_permissions, NULL,
5136 &setlist, &showlist);
5137
5138 add_setshow_boolean_cmd ("may-write-memory", class_support,
5139 &may_write_memory_1, _("\
5140 Set permission to write into target memory."), _("\
5141 Show permission to write into target memory."), _("\
5142 When this permission is on, GDB may write into the target's memory.\n\
5143 Otherwise, any sort of write attempt will result in an error."),
5144 set_write_memory_permission, NULL,
5145 &setlist, &showlist);
5146
5147 add_setshow_boolean_cmd ("may-insert-breakpoints", class_support,
5148 &may_insert_breakpoints_1, _("\
5149 Set permission to insert breakpoints in the target."), _("\
5150 Show permission to insert breakpoints in the target."), _("\
5151 When this permission is on, GDB may insert breakpoints in the program.\n\
5152 Otherwise, any sort of insertion attempt will result in an error."),
5153 set_target_permissions, NULL,
5154 &setlist, &showlist);
5155
5156 add_setshow_boolean_cmd ("may-insert-tracepoints", class_support,
5157 &may_insert_tracepoints_1, _("\
5158 Set permission to insert tracepoints in the target."), _("\
5159 Show permission to insert tracepoints in the target."), _("\
5160 When this permission is on, GDB may insert tracepoints in the program.\n\
5161 Otherwise, any sort of insertion attempt will result in an error."),
5162 set_target_permissions, NULL,
5163 &setlist, &showlist);
5164
5165 add_setshow_boolean_cmd ("may-insert-fast-tracepoints", class_support,
5166 &may_insert_fast_tracepoints_1, _("\
5167 Set permission to insert fast tracepoints in the target."), _("\
5168 Show permission to insert fast tracepoints in the target."), _("\
5169 When this permission is on, GDB may insert fast tracepoints.\n\
5170 Otherwise, any sort of insertion attempt will result in an error."),
5171 set_target_permissions, NULL,
5172 &setlist, &showlist);
5173
5174 add_setshow_boolean_cmd ("may-interrupt", class_support,
5175 &may_stop_1, _("\
5176 Set permission to interrupt or signal the target."), _("\
5177 Show permission to interrupt or signal the target."), _("\
5178 When this permission is on, GDB may interrupt/stop the target's execution.\n\
5179 Otherwise, any attempt to interrupt or stop will be ignored."),
5180 set_target_permissions, NULL,
5181 &setlist, &showlist);
5182
5183
5184 target_dcache = dcache_init ();
5185 }
This page took 0.202603 seconds and 4 git commands to generate.