Improve performance of large restore commands
[deliverable/binutils-gdb.git] / gdb / target.c
1 /* Select target systems and architectures at runtime for GDB.
2
3 Copyright (C) 1990-2013 Free Software Foundation, Inc.
4
5 Contributed by Cygnus Support.
6
7 This file is part of GDB.
8
9 This program is free software; you can redistribute it and/or modify
10 it under the terms of the GNU General Public License as published by
11 the Free Software Foundation; either version 3 of the License, or
12 (at your option) any later version.
13
14 This program is distributed in the hope that it will be useful,
15 but WITHOUT ANY WARRANTY; without even the implied warranty of
16 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
17 GNU General Public License for more details.
18
19 You should have received a copy of the GNU General Public License
20 along with this program. If not, see <http://www.gnu.org/licenses/>. */
21
22 #include "defs.h"
23 #include <errno.h>
24 #include "gdb_string.h"
25 #include "target.h"
26 #include "gdbcmd.h"
27 #include "symtab.h"
28 #include "inferior.h"
29 #include "bfd.h"
30 #include "symfile.h"
31 #include "objfiles.h"
32 #include "dcache.h"
33 #include <signal.h>
34 #include "regcache.h"
35 #include "gdb_assert.h"
36 #include "gdbcore.h"
37 #include "exceptions.h"
38 #include "target-descriptions.h"
39 #include "gdbthread.h"
40 #include "solib.h"
41 #include "exec.h"
42 #include "inline-frame.h"
43 #include "tracepoint.h"
44 #include "gdb/fileio.h"
45 #include "agent.h"
46
47 static void target_info (char *, int);
48
49 static void default_terminal_info (const char *, int);
50
51 static int default_watchpoint_addr_within_range (struct target_ops *,
52 CORE_ADDR, CORE_ADDR, int);
53
54 static int default_region_ok_for_hw_watchpoint (CORE_ADDR, int);
55
56 static void tcomplain (void) ATTRIBUTE_NORETURN;
57
58 static int nomemory (CORE_ADDR, char *, int, int, struct target_ops *);
59
60 static int return_zero (void);
61
62 static int return_one (void);
63
64 static int return_minus_one (void);
65
66 void target_ignore (void);
67
68 static void target_command (char *, int);
69
70 static struct target_ops *find_default_run_target (char *);
71
72 static LONGEST default_xfer_partial (struct target_ops *ops,
73 enum target_object object,
74 const char *annex, gdb_byte *readbuf,
75 const gdb_byte *writebuf,
76 ULONGEST offset, LONGEST len);
77
78 static LONGEST current_xfer_partial (struct target_ops *ops,
79 enum target_object object,
80 const char *annex, gdb_byte *readbuf,
81 const gdb_byte *writebuf,
82 ULONGEST offset, LONGEST len);
83
84 static struct gdbarch *default_thread_architecture (struct target_ops *ops,
85 ptid_t ptid);
86
87 static void init_dummy_target (void);
88
89 static struct target_ops debug_target;
90
91 static void debug_to_open (char *, int);
92
93 static void debug_to_prepare_to_store (struct regcache *);
94
95 static void debug_to_files_info (struct target_ops *);
96
97 static int debug_to_insert_breakpoint (struct gdbarch *,
98 struct bp_target_info *);
99
100 static int debug_to_remove_breakpoint (struct gdbarch *,
101 struct bp_target_info *);
102
103 static int debug_to_can_use_hw_breakpoint (int, int, int);
104
105 static int debug_to_insert_hw_breakpoint (struct gdbarch *,
106 struct bp_target_info *);
107
108 static int debug_to_remove_hw_breakpoint (struct gdbarch *,
109 struct bp_target_info *);
110
111 static int debug_to_insert_watchpoint (CORE_ADDR, int, int,
112 struct expression *);
113
114 static int debug_to_remove_watchpoint (CORE_ADDR, int, int,
115 struct expression *);
116
117 static int debug_to_stopped_by_watchpoint (void);
118
119 static int debug_to_stopped_data_address (struct target_ops *, CORE_ADDR *);
120
121 static int debug_to_watchpoint_addr_within_range (struct target_ops *,
122 CORE_ADDR, CORE_ADDR, int);
123
124 static int debug_to_region_ok_for_hw_watchpoint (CORE_ADDR, int);
125
126 static int debug_to_can_accel_watchpoint_condition (CORE_ADDR, int, int,
127 struct expression *);
128
129 static void debug_to_terminal_init (void);
130
131 static void debug_to_terminal_inferior (void);
132
133 static void debug_to_terminal_ours_for_output (void);
134
135 static void debug_to_terminal_save_ours (void);
136
137 static void debug_to_terminal_ours (void);
138
139 static void debug_to_load (char *, int);
140
141 static int debug_to_can_run (void);
142
143 static void debug_to_stop (ptid_t);
144
145 /* Pointer to array of target architecture structures; the size of the
146 array; the current index into the array; the allocated size of the
147 array. */
148 struct target_ops **target_structs;
149 unsigned target_struct_size;
150 unsigned target_struct_allocsize;
151 #define DEFAULT_ALLOCSIZE 10
152
153 /* The initial current target, so that there is always a semi-valid
154 current target. */
155
156 static struct target_ops dummy_target;
157
158 /* Top of target stack. */
159
160 static struct target_ops *target_stack;
161
162 /* The target structure we are currently using to talk to a process
163 or file or whatever "inferior" we have. */
164
165 struct target_ops current_target;
166
167 /* Command list for target. */
168
169 static struct cmd_list_element *targetlist = NULL;
170
171 /* Nonzero if we should trust readonly sections from the
172 executable when reading memory. */
173
174 static int trust_readonly = 0;
175
176 /* Nonzero if we should show true memory content including
177 memory breakpoint inserted by gdb. */
178
179 static int show_memory_breakpoints = 0;
180
181 /* These globals control whether GDB attempts to perform these
182 operations; they are useful for targets that need to prevent
183 inadvertant disruption, such as in non-stop mode. */
184
185 int may_write_registers = 1;
186
187 int may_write_memory = 1;
188
189 int may_insert_breakpoints = 1;
190
191 int may_insert_tracepoints = 1;
192
193 int may_insert_fast_tracepoints = 1;
194
195 int may_stop = 1;
196
197 /* Non-zero if we want to see trace of target level stuff. */
198
199 static unsigned int targetdebug = 0;
200 static void
201 show_targetdebug (struct ui_file *file, int from_tty,
202 struct cmd_list_element *c, const char *value)
203 {
204 fprintf_filtered (file, _("Target debugging is %s.\n"), value);
205 }
206
207 static void setup_target_debug (void);
208
209 /* The option sets this. */
210 static int stack_cache_enabled_p_1 = 1;
211 /* And set_stack_cache_enabled_p updates this.
212 The reason for the separation is so that we don't flush the cache for
213 on->on transitions. */
214 static int stack_cache_enabled_p = 1;
215
216 /* This is called *after* the stack-cache has been set.
217 Flush the cache for off->on and on->off transitions.
218 There's no real need to flush the cache for on->off transitions,
219 except cleanliness. */
220
221 static void
222 set_stack_cache_enabled_p (char *args, int from_tty,
223 struct cmd_list_element *c)
224 {
225 if (stack_cache_enabled_p != stack_cache_enabled_p_1)
226 target_dcache_invalidate ();
227
228 stack_cache_enabled_p = stack_cache_enabled_p_1;
229 }
230
231 static void
232 show_stack_cache_enabled_p (struct ui_file *file, int from_tty,
233 struct cmd_list_element *c, const char *value)
234 {
235 fprintf_filtered (file, _("Cache use for stack accesses is %s.\n"), value);
236 }
237
238 /* Cache of memory operations, to speed up remote access. */
239 static DCACHE *target_dcache;
240
241 /* Invalidate the target dcache. */
242
243 void
244 target_dcache_invalidate (void)
245 {
246 dcache_invalidate (target_dcache);
247 }
248
249 /* The user just typed 'target' without the name of a target. */
250
251 static void
252 target_command (char *arg, int from_tty)
253 {
254 fputs_filtered ("Argument required (target name). Try `help target'\n",
255 gdb_stdout);
256 }
257
258 /* Default target_has_* methods for process_stratum targets. */
259
260 int
261 default_child_has_all_memory (struct target_ops *ops)
262 {
263 /* If no inferior selected, then we can't read memory here. */
264 if (ptid_equal (inferior_ptid, null_ptid))
265 return 0;
266
267 return 1;
268 }
269
270 int
271 default_child_has_memory (struct target_ops *ops)
272 {
273 /* If no inferior selected, then we can't read memory here. */
274 if (ptid_equal (inferior_ptid, null_ptid))
275 return 0;
276
277 return 1;
278 }
279
280 int
281 default_child_has_stack (struct target_ops *ops)
282 {
283 /* If no inferior selected, there's no stack. */
284 if (ptid_equal (inferior_ptid, null_ptid))
285 return 0;
286
287 return 1;
288 }
289
290 int
291 default_child_has_registers (struct target_ops *ops)
292 {
293 /* Can't read registers from no inferior. */
294 if (ptid_equal (inferior_ptid, null_ptid))
295 return 0;
296
297 return 1;
298 }
299
300 int
301 default_child_has_execution (struct target_ops *ops, ptid_t the_ptid)
302 {
303 /* If there's no thread selected, then we can't make it run through
304 hoops. */
305 if (ptid_equal (the_ptid, null_ptid))
306 return 0;
307
308 return 1;
309 }
310
311
312 int
313 target_has_all_memory_1 (void)
314 {
315 struct target_ops *t;
316
317 for (t = current_target.beneath; t != NULL; t = t->beneath)
318 if (t->to_has_all_memory (t))
319 return 1;
320
321 return 0;
322 }
323
324 int
325 target_has_memory_1 (void)
326 {
327 struct target_ops *t;
328
329 for (t = current_target.beneath; t != NULL; t = t->beneath)
330 if (t->to_has_memory (t))
331 return 1;
332
333 return 0;
334 }
335
336 int
337 target_has_stack_1 (void)
338 {
339 struct target_ops *t;
340
341 for (t = current_target.beneath; t != NULL; t = t->beneath)
342 if (t->to_has_stack (t))
343 return 1;
344
345 return 0;
346 }
347
348 int
349 target_has_registers_1 (void)
350 {
351 struct target_ops *t;
352
353 for (t = current_target.beneath; t != NULL; t = t->beneath)
354 if (t->to_has_registers (t))
355 return 1;
356
357 return 0;
358 }
359
360 int
361 target_has_execution_1 (ptid_t the_ptid)
362 {
363 struct target_ops *t;
364
365 for (t = current_target.beneath; t != NULL; t = t->beneath)
366 if (t->to_has_execution (t, the_ptid))
367 return 1;
368
369 return 0;
370 }
371
372 int
373 target_has_execution_current (void)
374 {
375 return target_has_execution_1 (inferior_ptid);
376 }
377
378 /* Complete initialization of T. This ensures that various fields in
379 T are set, if needed by the target implementation. */
380
381 void
382 complete_target_initialization (struct target_ops *t)
383 {
384 /* Provide default values for all "must have" methods. */
385 if (t->to_xfer_partial == NULL)
386 t->to_xfer_partial = default_xfer_partial;
387
388 if (t->to_has_all_memory == NULL)
389 t->to_has_all_memory = (int (*) (struct target_ops *)) return_zero;
390
391 if (t->to_has_memory == NULL)
392 t->to_has_memory = (int (*) (struct target_ops *)) return_zero;
393
394 if (t->to_has_stack == NULL)
395 t->to_has_stack = (int (*) (struct target_ops *)) return_zero;
396
397 if (t->to_has_registers == NULL)
398 t->to_has_registers = (int (*) (struct target_ops *)) return_zero;
399
400 if (t->to_has_execution == NULL)
401 t->to_has_execution = (int (*) (struct target_ops *, ptid_t)) return_zero;
402 }
403
404 /* Add possible target architecture T to the list and add a new
405 command 'target T->to_shortname'. Set COMPLETER as the command's
406 completer if not NULL. */
407
408 void
409 add_target_with_completer (struct target_ops *t,
410 completer_ftype *completer)
411 {
412 struct cmd_list_element *c;
413
414 complete_target_initialization (t);
415
416 if (!target_structs)
417 {
418 target_struct_allocsize = DEFAULT_ALLOCSIZE;
419 target_structs = (struct target_ops **) xmalloc
420 (target_struct_allocsize * sizeof (*target_structs));
421 }
422 if (target_struct_size >= target_struct_allocsize)
423 {
424 target_struct_allocsize *= 2;
425 target_structs = (struct target_ops **)
426 xrealloc ((char *) target_structs,
427 target_struct_allocsize * sizeof (*target_structs));
428 }
429 target_structs[target_struct_size++] = t;
430
431 if (targetlist == NULL)
432 add_prefix_cmd ("target", class_run, target_command, _("\
433 Connect to a target machine or process.\n\
434 The first argument is the type or protocol of the target machine.\n\
435 Remaining arguments are interpreted by the target protocol. For more\n\
436 information on the arguments for a particular protocol, type\n\
437 `help target ' followed by the protocol name."),
438 &targetlist, "target ", 0, &cmdlist);
439 c = add_cmd (t->to_shortname, no_class, t->to_open, t->to_doc,
440 &targetlist);
441 if (completer != NULL)
442 set_cmd_completer (c, completer);
443 }
444
445 /* Add a possible target architecture to the list. */
446
447 void
448 add_target (struct target_ops *t)
449 {
450 add_target_with_completer (t, NULL);
451 }
452
453 /* See target.h. */
454
455 void
456 add_deprecated_target_alias (struct target_ops *t, char *alias)
457 {
458 struct cmd_list_element *c;
459 char *alt;
460
461 /* If we use add_alias_cmd, here, we do not get the deprecated warning,
462 see PR cli/15104. */
463 c = add_cmd (alias, no_class, t->to_open, t->to_doc, &targetlist);
464 alt = xstrprintf ("target %s", t->to_shortname);
465 deprecate_cmd (c, alt);
466 }
467
468 /* Stub functions */
469
470 void
471 target_ignore (void)
472 {
473 }
474
475 void
476 target_kill (void)
477 {
478 struct target_ops *t;
479
480 for (t = current_target.beneath; t != NULL; t = t->beneath)
481 if (t->to_kill != NULL)
482 {
483 if (targetdebug)
484 fprintf_unfiltered (gdb_stdlog, "target_kill ()\n");
485
486 t->to_kill (t);
487 return;
488 }
489
490 noprocess ();
491 }
492
493 void
494 target_load (char *arg, int from_tty)
495 {
496 target_dcache_invalidate ();
497 (*current_target.to_load) (arg, from_tty);
498 }
499
500 void
501 target_create_inferior (char *exec_file, char *args,
502 char **env, int from_tty)
503 {
504 struct target_ops *t;
505
506 for (t = current_target.beneath; t != NULL; t = t->beneath)
507 {
508 if (t->to_create_inferior != NULL)
509 {
510 t->to_create_inferior (t, exec_file, args, env, from_tty);
511 if (targetdebug)
512 fprintf_unfiltered (gdb_stdlog,
513 "target_create_inferior (%s, %s, xxx, %d)\n",
514 exec_file, args, from_tty);
515 return;
516 }
517 }
518
519 internal_error (__FILE__, __LINE__,
520 _("could not find a target to create inferior"));
521 }
522
523 void
524 target_terminal_inferior (void)
525 {
526 /* A background resume (``run&'') should leave GDB in control of the
527 terminal. Use target_can_async_p, not target_is_async_p, since at
528 this point the target is not async yet. However, if sync_execution
529 is not set, we know it will become async prior to resume. */
530 if (target_can_async_p () && !sync_execution)
531 return;
532
533 /* If GDB is resuming the inferior in the foreground, install
534 inferior's terminal modes. */
535 (*current_target.to_terminal_inferior) ();
536 }
537
538 static int
539 nomemory (CORE_ADDR memaddr, char *myaddr, int len, int write,
540 struct target_ops *t)
541 {
542 errno = EIO; /* Can't read/write this location. */
543 return 0; /* No bytes handled. */
544 }
545
546 static void
547 tcomplain (void)
548 {
549 error (_("You can't do that when your target is `%s'"),
550 current_target.to_shortname);
551 }
552
553 void
554 noprocess (void)
555 {
556 error (_("You can't do that without a process to debug."));
557 }
558
559 static void
560 default_terminal_info (const char *args, int from_tty)
561 {
562 printf_unfiltered (_("No saved terminal information.\n"));
563 }
564
565 /* A default implementation for the to_get_ada_task_ptid target method.
566
567 This function builds the PTID by using both LWP and TID as part of
568 the PTID lwp and tid elements. The pid used is the pid of the
569 inferior_ptid. */
570
571 static ptid_t
572 default_get_ada_task_ptid (long lwp, long tid)
573 {
574 return ptid_build (ptid_get_pid (inferior_ptid), lwp, tid);
575 }
576
577 static enum exec_direction_kind
578 default_execution_direction (void)
579 {
580 if (!target_can_execute_reverse)
581 return EXEC_FORWARD;
582 else if (!target_can_async_p ())
583 return EXEC_FORWARD;
584 else
585 gdb_assert_not_reached ("\
586 to_execution_direction must be implemented for reverse async");
587 }
588
589 /* Go through the target stack from top to bottom, copying over zero
590 entries in current_target, then filling in still empty entries. In
591 effect, we are doing class inheritance through the pushed target
592 vectors.
593
594 NOTE: cagney/2003-10-17: The problem with this inheritance, as it
595 is currently implemented, is that it discards any knowledge of
596 which target an inherited method originally belonged to.
597 Consequently, new new target methods should instead explicitly and
598 locally search the target stack for the target that can handle the
599 request. */
600
601 static void
602 update_current_target (void)
603 {
604 struct target_ops *t;
605
606 /* First, reset current's contents. */
607 memset (&current_target, 0, sizeof (current_target));
608
609 #define INHERIT(FIELD, TARGET) \
610 if (!current_target.FIELD) \
611 current_target.FIELD = (TARGET)->FIELD
612
613 for (t = target_stack; t; t = t->beneath)
614 {
615 INHERIT (to_shortname, t);
616 INHERIT (to_longname, t);
617 INHERIT (to_doc, t);
618 /* Do not inherit to_open. */
619 /* Do not inherit to_close. */
620 /* Do not inherit to_attach. */
621 INHERIT (to_post_attach, t);
622 INHERIT (to_attach_no_wait, t);
623 /* Do not inherit to_detach. */
624 /* Do not inherit to_disconnect. */
625 /* Do not inherit to_resume. */
626 /* Do not inherit to_wait. */
627 /* Do not inherit to_fetch_registers. */
628 /* Do not inherit to_store_registers. */
629 INHERIT (to_prepare_to_store, t);
630 INHERIT (deprecated_xfer_memory, t);
631 INHERIT (to_files_info, t);
632 INHERIT (to_insert_breakpoint, t);
633 INHERIT (to_remove_breakpoint, t);
634 INHERIT (to_can_use_hw_breakpoint, t);
635 INHERIT (to_insert_hw_breakpoint, t);
636 INHERIT (to_remove_hw_breakpoint, t);
637 /* Do not inherit to_ranged_break_num_registers. */
638 INHERIT (to_insert_watchpoint, t);
639 INHERIT (to_remove_watchpoint, t);
640 /* Do not inherit to_insert_mask_watchpoint. */
641 /* Do not inherit to_remove_mask_watchpoint. */
642 INHERIT (to_stopped_data_address, t);
643 INHERIT (to_have_steppable_watchpoint, t);
644 INHERIT (to_have_continuable_watchpoint, t);
645 INHERIT (to_stopped_by_watchpoint, t);
646 INHERIT (to_watchpoint_addr_within_range, t);
647 INHERIT (to_region_ok_for_hw_watchpoint, t);
648 INHERIT (to_can_accel_watchpoint_condition, t);
649 /* Do not inherit to_masked_watch_num_registers. */
650 INHERIT (to_terminal_init, t);
651 INHERIT (to_terminal_inferior, t);
652 INHERIT (to_terminal_ours_for_output, t);
653 INHERIT (to_terminal_ours, t);
654 INHERIT (to_terminal_save_ours, t);
655 INHERIT (to_terminal_info, t);
656 /* Do not inherit to_kill. */
657 INHERIT (to_load, t);
658 /* Do no inherit to_create_inferior. */
659 INHERIT (to_post_startup_inferior, t);
660 INHERIT (to_insert_fork_catchpoint, t);
661 INHERIT (to_remove_fork_catchpoint, t);
662 INHERIT (to_insert_vfork_catchpoint, t);
663 INHERIT (to_remove_vfork_catchpoint, t);
664 /* Do not inherit to_follow_fork. */
665 INHERIT (to_insert_exec_catchpoint, t);
666 INHERIT (to_remove_exec_catchpoint, t);
667 INHERIT (to_set_syscall_catchpoint, t);
668 INHERIT (to_has_exited, t);
669 /* Do not inherit to_mourn_inferior. */
670 INHERIT (to_can_run, t);
671 /* Do not inherit to_pass_signals. */
672 /* Do not inherit to_program_signals. */
673 /* Do not inherit to_thread_alive. */
674 /* Do not inherit to_find_new_threads. */
675 /* Do not inherit to_pid_to_str. */
676 INHERIT (to_extra_thread_info, t);
677 INHERIT (to_thread_name, t);
678 INHERIT (to_stop, t);
679 /* Do not inherit to_xfer_partial. */
680 INHERIT (to_rcmd, t);
681 INHERIT (to_pid_to_exec_file, t);
682 INHERIT (to_log_command, t);
683 INHERIT (to_stratum, t);
684 /* Do not inherit to_has_all_memory. */
685 /* Do not inherit to_has_memory. */
686 /* Do not inherit to_has_stack. */
687 /* Do not inherit to_has_registers. */
688 /* Do not inherit to_has_execution. */
689 INHERIT (to_has_thread_control, t);
690 INHERIT (to_can_async_p, t);
691 INHERIT (to_is_async_p, t);
692 INHERIT (to_async, t);
693 INHERIT (to_find_memory_regions, t);
694 INHERIT (to_make_corefile_notes, t);
695 INHERIT (to_get_bookmark, t);
696 INHERIT (to_goto_bookmark, t);
697 /* Do not inherit to_get_thread_local_address. */
698 INHERIT (to_can_execute_reverse, t);
699 INHERIT (to_execution_direction, t);
700 INHERIT (to_thread_architecture, t);
701 /* Do not inherit to_read_description. */
702 INHERIT (to_get_ada_task_ptid, t);
703 /* Do not inherit to_search_memory. */
704 INHERIT (to_supports_multi_process, t);
705 INHERIT (to_supports_enable_disable_tracepoint, t);
706 INHERIT (to_supports_string_tracing, t);
707 INHERIT (to_trace_init, t);
708 INHERIT (to_download_tracepoint, t);
709 INHERIT (to_can_download_tracepoint, t);
710 INHERIT (to_download_trace_state_variable, t);
711 INHERIT (to_enable_tracepoint, t);
712 INHERIT (to_disable_tracepoint, t);
713 INHERIT (to_trace_set_readonly_regions, t);
714 INHERIT (to_trace_start, t);
715 INHERIT (to_get_trace_status, t);
716 INHERIT (to_get_tracepoint_status, t);
717 INHERIT (to_trace_stop, t);
718 INHERIT (to_trace_find, t);
719 INHERIT (to_get_trace_state_variable_value, t);
720 INHERIT (to_save_trace_data, t);
721 INHERIT (to_upload_tracepoints, t);
722 INHERIT (to_upload_trace_state_variables, t);
723 INHERIT (to_get_raw_trace_data, t);
724 INHERIT (to_get_min_fast_tracepoint_insn_len, t);
725 INHERIT (to_set_disconnected_tracing, t);
726 INHERIT (to_set_circular_trace_buffer, t);
727 INHERIT (to_set_trace_buffer_size, t);
728 INHERIT (to_set_trace_notes, t);
729 INHERIT (to_get_tib_address, t);
730 INHERIT (to_set_permissions, t);
731 INHERIT (to_static_tracepoint_marker_at, t);
732 INHERIT (to_static_tracepoint_markers_by_strid, t);
733 INHERIT (to_traceframe_info, t);
734 INHERIT (to_use_agent, t);
735 INHERIT (to_can_use_agent, t);
736 INHERIT (to_augmented_libraries_svr4_read, t);
737 INHERIT (to_magic, t);
738 INHERIT (to_supports_evaluation_of_breakpoint_conditions, t);
739 INHERIT (to_can_run_breakpoint_commands, t);
740 /* Do not inherit to_memory_map. */
741 /* Do not inherit to_flash_erase. */
742 /* Do not inherit to_flash_done. */
743 }
744 #undef INHERIT
745
746 /* Clean up a target struct so it no longer has any zero pointers in
747 it. Some entries are defaulted to a method that print an error,
748 others are hard-wired to a standard recursive default. */
749
750 #define de_fault(field, value) \
751 if (!current_target.field) \
752 current_target.field = value
753
754 de_fault (to_open,
755 (void (*) (char *, int))
756 tcomplain);
757 de_fault (to_close,
758 (void (*) (void))
759 target_ignore);
760 de_fault (to_post_attach,
761 (void (*) (int))
762 target_ignore);
763 de_fault (to_prepare_to_store,
764 (void (*) (struct regcache *))
765 noprocess);
766 de_fault (deprecated_xfer_memory,
767 (int (*) (CORE_ADDR, gdb_byte *, int, int,
768 struct mem_attrib *, struct target_ops *))
769 nomemory);
770 de_fault (to_files_info,
771 (void (*) (struct target_ops *))
772 target_ignore);
773 de_fault (to_insert_breakpoint,
774 memory_insert_breakpoint);
775 de_fault (to_remove_breakpoint,
776 memory_remove_breakpoint);
777 de_fault (to_can_use_hw_breakpoint,
778 (int (*) (int, int, int))
779 return_zero);
780 de_fault (to_insert_hw_breakpoint,
781 (int (*) (struct gdbarch *, struct bp_target_info *))
782 return_minus_one);
783 de_fault (to_remove_hw_breakpoint,
784 (int (*) (struct gdbarch *, struct bp_target_info *))
785 return_minus_one);
786 de_fault (to_insert_watchpoint,
787 (int (*) (CORE_ADDR, int, int, struct expression *))
788 return_minus_one);
789 de_fault (to_remove_watchpoint,
790 (int (*) (CORE_ADDR, int, int, struct expression *))
791 return_minus_one);
792 de_fault (to_stopped_by_watchpoint,
793 (int (*) (void))
794 return_zero);
795 de_fault (to_stopped_data_address,
796 (int (*) (struct target_ops *, CORE_ADDR *))
797 return_zero);
798 de_fault (to_watchpoint_addr_within_range,
799 default_watchpoint_addr_within_range);
800 de_fault (to_region_ok_for_hw_watchpoint,
801 default_region_ok_for_hw_watchpoint);
802 de_fault (to_can_accel_watchpoint_condition,
803 (int (*) (CORE_ADDR, int, int, struct expression *))
804 return_zero);
805 de_fault (to_terminal_init,
806 (void (*) (void))
807 target_ignore);
808 de_fault (to_terminal_inferior,
809 (void (*) (void))
810 target_ignore);
811 de_fault (to_terminal_ours_for_output,
812 (void (*) (void))
813 target_ignore);
814 de_fault (to_terminal_ours,
815 (void (*) (void))
816 target_ignore);
817 de_fault (to_terminal_save_ours,
818 (void (*) (void))
819 target_ignore);
820 de_fault (to_terminal_info,
821 default_terminal_info);
822 de_fault (to_load,
823 (void (*) (char *, int))
824 tcomplain);
825 de_fault (to_post_startup_inferior,
826 (void (*) (ptid_t))
827 target_ignore);
828 de_fault (to_insert_fork_catchpoint,
829 (int (*) (int))
830 return_one);
831 de_fault (to_remove_fork_catchpoint,
832 (int (*) (int))
833 return_one);
834 de_fault (to_insert_vfork_catchpoint,
835 (int (*) (int))
836 return_one);
837 de_fault (to_remove_vfork_catchpoint,
838 (int (*) (int))
839 return_one);
840 de_fault (to_insert_exec_catchpoint,
841 (int (*) (int))
842 return_one);
843 de_fault (to_remove_exec_catchpoint,
844 (int (*) (int))
845 return_one);
846 de_fault (to_set_syscall_catchpoint,
847 (int (*) (int, int, int, int, int *))
848 return_one);
849 de_fault (to_has_exited,
850 (int (*) (int, int, int *))
851 return_zero);
852 de_fault (to_can_run,
853 return_zero);
854 de_fault (to_extra_thread_info,
855 (char *(*) (struct thread_info *))
856 return_zero);
857 de_fault (to_thread_name,
858 (char *(*) (struct thread_info *))
859 return_zero);
860 de_fault (to_stop,
861 (void (*) (ptid_t))
862 target_ignore);
863 current_target.to_xfer_partial = current_xfer_partial;
864 de_fault (to_rcmd,
865 (void (*) (char *, struct ui_file *))
866 tcomplain);
867 de_fault (to_pid_to_exec_file,
868 (char *(*) (int))
869 return_zero);
870 de_fault (to_async,
871 (void (*) (void (*) (enum inferior_event_type, void*), void*))
872 tcomplain);
873 de_fault (to_thread_architecture,
874 default_thread_architecture);
875 current_target.to_read_description = NULL;
876 de_fault (to_get_ada_task_ptid,
877 (ptid_t (*) (long, long))
878 default_get_ada_task_ptid);
879 de_fault (to_supports_multi_process,
880 (int (*) (void))
881 return_zero);
882 de_fault (to_supports_enable_disable_tracepoint,
883 (int (*) (void))
884 return_zero);
885 de_fault (to_supports_string_tracing,
886 (int (*) (void))
887 return_zero);
888 de_fault (to_trace_init,
889 (void (*) (void))
890 tcomplain);
891 de_fault (to_download_tracepoint,
892 (void (*) (struct bp_location *))
893 tcomplain);
894 de_fault (to_can_download_tracepoint,
895 (int (*) (void))
896 return_zero);
897 de_fault (to_download_trace_state_variable,
898 (void (*) (struct trace_state_variable *))
899 tcomplain);
900 de_fault (to_enable_tracepoint,
901 (void (*) (struct bp_location *))
902 tcomplain);
903 de_fault (to_disable_tracepoint,
904 (void (*) (struct bp_location *))
905 tcomplain);
906 de_fault (to_trace_set_readonly_regions,
907 (void (*) (void))
908 tcomplain);
909 de_fault (to_trace_start,
910 (void (*) (void))
911 tcomplain);
912 de_fault (to_get_trace_status,
913 (int (*) (struct trace_status *))
914 return_minus_one);
915 de_fault (to_get_tracepoint_status,
916 (void (*) (struct breakpoint *, struct uploaded_tp *))
917 tcomplain);
918 de_fault (to_trace_stop,
919 (void (*) (void))
920 tcomplain);
921 de_fault (to_trace_find,
922 (int (*) (enum trace_find_type, int, CORE_ADDR, CORE_ADDR, int *))
923 return_minus_one);
924 de_fault (to_get_trace_state_variable_value,
925 (int (*) (int, LONGEST *))
926 return_zero);
927 de_fault (to_save_trace_data,
928 (int (*) (const char *))
929 tcomplain);
930 de_fault (to_upload_tracepoints,
931 (int (*) (struct uploaded_tp **))
932 return_zero);
933 de_fault (to_upload_trace_state_variables,
934 (int (*) (struct uploaded_tsv **))
935 return_zero);
936 de_fault (to_get_raw_trace_data,
937 (LONGEST (*) (gdb_byte *, ULONGEST, LONGEST))
938 tcomplain);
939 de_fault (to_get_min_fast_tracepoint_insn_len,
940 (int (*) (void))
941 return_minus_one);
942 de_fault (to_set_disconnected_tracing,
943 (void (*) (int))
944 target_ignore);
945 de_fault (to_set_circular_trace_buffer,
946 (void (*) (int))
947 target_ignore);
948 de_fault (to_set_trace_buffer_size,
949 (void (*) (LONGEST))
950 target_ignore);
951 de_fault (to_set_trace_notes,
952 (int (*) (const char *, const char *, const char *))
953 return_zero);
954 de_fault (to_get_tib_address,
955 (int (*) (ptid_t, CORE_ADDR *))
956 tcomplain);
957 de_fault (to_set_permissions,
958 (void (*) (void))
959 target_ignore);
960 de_fault (to_static_tracepoint_marker_at,
961 (int (*) (CORE_ADDR, struct static_tracepoint_marker *))
962 return_zero);
963 de_fault (to_static_tracepoint_markers_by_strid,
964 (VEC(static_tracepoint_marker_p) * (*) (const char *))
965 tcomplain);
966 de_fault (to_traceframe_info,
967 (struct traceframe_info * (*) (void))
968 return_zero);
969 de_fault (to_supports_evaluation_of_breakpoint_conditions,
970 (int (*) (void))
971 return_zero);
972 de_fault (to_can_run_breakpoint_commands,
973 (int (*) (void))
974 return_zero);
975 de_fault (to_use_agent,
976 (int (*) (int))
977 tcomplain);
978 de_fault (to_can_use_agent,
979 (int (*) (void))
980 return_zero);
981 de_fault (to_augmented_libraries_svr4_read,
982 (int (*) (void))
983 return_zero);
984 de_fault (to_execution_direction, default_execution_direction);
985
986 #undef de_fault
987
988 /* Finally, position the target-stack beneath the squashed
989 "current_target". That way code looking for a non-inherited
990 target method can quickly and simply find it. */
991 current_target.beneath = target_stack;
992
993 if (targetdebug)
994 setup_target_debug ();
995 }
996
997 /* Push a new target type into the stack of the existing target accessors,
998 possibly superseding some of the existing accessors.
999
1000 Rather than allow an empty stack, we always have the dummy target at
1001 the bottom stratum, so we can call the function vectors without
1002 checking them. */
1003
1004 void
1005 push_target (struct target_ops *t)
1006 {
1007 struct target_ops **cur;
1008
1009 /* Check magic number. If wrong, it probably means someone changed
1010 the struct definition, but not all the places that initialize one. */
1011 if (t->to_magic != OPS_MAGIC)
1012 {
1013 fprintf_unfiltered (gdb_stderr,
1014 "Magic number of %s target struct wrong\n",
1015 t->to_shortname);
1016 internal_error (__FILE__, __LINE__,
1017 _("failed internal consistency check"));
1018 }
1019
1020 /* Find the proper stratum to install this target in. */
1021 for (cur = &target_stack; (*cur) != NULL; cur = &(*cur)->beneath)
1022 {
1023 if ((int) (t->to_stratum) >= (int) (*cur)->to_stratum)
1024 break;
1025 }
1026
1027 /* If there's already targets at this stratum, remove them. */
1028 /* FIXME: cagney/2003-10-15: I think this should be popping all
1029 targets to CUR, and not just those at this stratum level. */
1030 while ((*cur) != NULL && t->to_stratum == (*cur)->to_stratum)
1031 {
1032 /* There's already something at this stratum level. Close it,
1033 and un-hook it from the stack. */
1034 struct target_ops *tmp = (*cur);
1035
1036 (*cur) = (*cur)->beneath;
1037 tmp->beneath = NULL;
1038 target_close (tmp);
1039 }
1040
1041 /* We have removed all targets in our stratum, now add the new one. */
1042 t->beneath = (*cur);
1043 (*cur) = t;
1044
1045 update_current_target ();
1046 }
1047
1048 /* Remove a target_ops vector from the stack, wherever it may be.
1049 Return how many times it was removed (0 or 1). */
1050
1051 int
1052 unpush_target (struct target_ops *t)
1053 {
1054 struct target_ops **cur;
1055 struct target_ops *tmp;
1056
1057 if (t->to_stratum == dummy_stratum)
1058 internal_error (__FILE__, __LINE__,
1059 _("Attempt to unpush the dummy target"));
1060
1061 /* Look for the specified target. Note that we assume that a target
1062 can only occur once in the target stack. */
1063
1064 for (cur = &target_stack; (*cur) != NULL; cur = &(*cur)->beneath)
1065 {
1066 if ((*cur) == t)
1067 break;
1068 }
1069
1070 /* If we don't find target_ops, quit. Only open targets should be
1071 closed. */
1072 if ((*cur) == NULL)
1073 return 0;
1074
1075 /* Unchain the target. */
1076 tmp = (*cur);
1077 (*cur) = (*cur)->beneath;
1078 tmp->beneath = NULL;
1079
1080 update_current_target ();
1081
1082 /* Finally close the target. Note we do this after unchaining, so
1083 any target method calls from within the target_close
1084 implementation don't end up in T anymore. */
1085 target_close (t);
1086
1087 return 1;
1088 }
1089
1090 void
1091 pop_all_targets_above (enum strata above_stratum)
1092 {
1093 while ((int) (current_target.to_stratum) > (int) above_stratum)
1094 {
1095 if (!unpush_target (target_stack))
1096 {
1097 fprintf_unfiltered (gdb_stderr,
1098 "pop_all_targets couldn't find target %s\n",
1099 target_stack->to_shortname);
1100 internal_error (__FILE__, __LINE__,
1101 _("failed internal consistency check"));
1102 break;
1103 }
1104 }
1105 }
1106
1107 void
1108 pop_all_targets (void)
1109 {
1110 pop_all_targets_above (dummy_stratum);
1111 }
1112
1113 /* Return 1 if T is now pushed in the target stack. Return 0 otherwise. */
1114
1115 int
1116 target_is_pushed (struct target_ops *t)
1117 {
1118 struct target_ops **cur;
1119
1120 /* Check magic number. If wrong, it probably means someone changed
1121 the struct definition, but not all the places that initialize one. */
1122 if (t->to_magic != OPS_MAGIC)
1123 {
1124 fprintf_unfiltered (gdb_stderr,
1125 "Magic number of %s target struct wrong\n",
1126 t->to_shortname);
1127 internal_error (__FILE__, __LINE__,
1128 _("failed internal consistency check"));
1129 }
1130
1131 for (cur = &target_stack; (*cur) != NULL; cur = &(*cur)->beneath)
1132 if (*cur == t)
1133 return 1;
1134
1135 return 0;
1136 }
1137
1138 /* Using the objfile specified in OBJFILE, find the address for the
1139 current thread's thread-local storage with offset OFFSET. */
1140 CORE_ADDR
1141 target_translate_tls_address (struct objfile *objfile, CORE_ADDR offset)
1142 {
1143 volatile CORE_ADDR addr = 0;
1144 struct target_ops *target;
1145
1146 for (target = current_target.beneath;
1147 target != NULL;
1148 target = target->beneath)
1149 {
1150 if (target->to_get_thread_local_address != NULL)
1151 break;
1152 }
1153
1154 if (target != NULL
1155 && gdbarch_fetch_tls_load_module_address_p (target_gdbarch ()))
1156 {
1157 ptid_t ptid = inferior_ptid;
1158 volatile struct gdb_exception ex;
1159
1160 TRY_CATCH (ex, RETURN_MASK_ALL)
1161 {
1162 CORE_ADDR lm_addr;
1163
1164 /* Fetch the load module address for this objfile. */
1165 lm_addr = gdbarch_fetch_tls_load_module_address (target_gdbarch (),
1166 objfile);
1167 /* If it's 0, throw the appropriate exception. */
1168 if (lm_addr == 0)
1169 throw_error (TLS_LOAD_MODULE_NOT_FOUND_ERROR,
1170 _("TLS load module not found"));
1171
1172 addr = target->to_get_thread_local_address (target, ptid,
1173 lm_addr, offset);
1174 }
1175 /* If an error occurred, print TLS related messages here. Otherwise,
1176 throw the error to some higher catcher. */
1177 if (ex.reason < 0)
1178 {
1179 int objfile_is_library = (objfile->flags & OBJF_SHARED);
1180
1181 switch (ex.error)
1182 {
1183 case TLS_NO_LIBRARY_SUPPORT_ERROR:
1184 error (_("Cannot find thread-local variables "
1185 "in this thread library."));
1186 break;
1187 case TLS_LOAD_MODULE_NOT_FOUND_ERROR:
1188 if (objfile_is_library)
1189 error (_("Cannot find shared library `%s' in dynamic"
1190 " linker's load module list"), objfile_name (objfile));
1191 else
1192 error (_("Cannot find executable file `%s' in dynamic"
1193 " linker's load module list"), objfile_name (objfile));
1194 break;
1195 case TLS_NOT_ALLOCATED_YET_ERROR:
1196 if (objfile_is_library)
1197 error (_("The inferior has not yet allocated storage for"
1198 " thread-local variables in\n"
1199 "the shared library `%s'\n"
1200 "for %s"),
1201 objfile_name (objfile), target_pid_to_str (ptid));
1202 else
1203 error (_("The inferior has not yet allocated storage for"
1204 " thread-local variables in\n"
1205 "the executable `%s'\n"
1206 "for %s"),
1207 objfile_name (objfile), target_pid_to_str (ptid));
1208 break;
1209 case TLS_GENERIC_ERROR:
1210 if (objfile_is_library)
1211 error (_("Cannot find thread-local storage for %s, "
1212 "shared library %s:\n%s"),
1213 target_pid_to_str (ptid),
1214 objfile_name (objfile), ex.message);
1215 else
1216 error (_("Cannot find thread-local storage for %s, "
1217 "executable file %s:\n%s"),
1218 target_pid_to_str (ptid),
1219 objfile_name (objfile), ex.message);
1220 break;
1221 default:
1222 throw_exception (ex);
1223 break;
1224 }
1225 }
1226 }
1227 /* It wouldn't be wrong here to try a gdbarch method, too; finding
1228 TLS is an ABI-specific thing. But we don't do that yet. */
1229 else
1230 error (_("Cannot find thread-local variables on this target"));
1231
1232 return addr;
1233 }
1234
1235 const char *
1236 target_xfer_error_to_string (enum target_xfer_error err)
1237 {
1238 #define CASE(X) case X: return #X
1239 switch (err)
1240 {
1241 CASE(TARGET_XFER_E_IO);
1242 CASE(TARGET_XFER_E_UNAVAILABLE);
1243 default:
1244 return "<unknown>";
1245 }
1246 #undef CASE
1247 };
1248
1249
1250 #undef MIN
1251 #define MIN(A, B) (((A) <= (B)) ? (A) : (B))
1252
1253 /* target_read_string -- read a null terminated string, up to LEN bytes,
1254 from MEMADDR in target. Set *ERRNOP to the errno code, or 0 if successful.
1255 Set *STRING to a pointer to malloc'd memory containing the data; the caller
1256 is responsible for freeing it. Return the number of bytes successfully
1257 read. */
1258
1259 int
1260 target_read_string (CORE_ADDR memaddr, char **string, int len, int *errnop)
1261 {
1262 int tlen, offset, i;
1263 gdb_byte buf[4];
1264 int errcode = 0;
1265 char *buffer;
1266 int buffer_allocated;
1267 char *bufptr;
1268 unsigned int nbytes_read = 0;
1269
1270 gdb_assert (string);
1271
1272 /* Small for testing. */
1273 buffer_allocated = 4;
1274 buffer = xmalloc (buffer_allocated);
1275 bufptr = buffer;
1276
1277 while (len > 0)
1278 {
1279 tlen = MIN (len, 4 - (memaddr & 3));
1280 offset = memaddr & 3;
1281
1282 errcode = target_read_memory (memaddr & ~3, buf, sizeof buf);
1283 if (errcode != 0)
1284 {
1285 /* The transfer request might have crossed the boundary to an
1286 unallocated region of memory. Retry the transfer, requesting
1287 a single byte. */
1288 tlen = 1;
1289 offset = 0;
1290 errcode = target_read_memory (memaddr, buf, 1);
1291 if (errcode != 0)
1292 goto done;
1293 }
1294
1295 if (bufptr - buffer + tlen > buffer_allocated)
1296 {
1297 unsigned int bytes;
1298
1299 bytes = bufptr - buffer;
1300 buffer_allocated *= 2;
1301 buffer = xrealloc (buffer, buffer_allocated);
1302 bufptr = buffer + bytes;
1303 }
1304
1305 for (i = 0; i < tlen; i++)
1306 {
1307 *bufptr++ = buf[i + offset];
1308 if (buf[i + offset] == '\000')
1309 {
1310 nbytes_read += i + 1;
1311 goto done;
1312 }
1313 }
1314
1315 memaddr += tlen;
1316 len -= tlen;
1317 nbytes_read += tlen;
1318 }
1319 done:
1320 *string = buffer;
1321 if (errnop != NULL)
1322 *errnop = errcode;
1323 return nbytes_read;
1324 }
1325
1326 struct target_section_table *
1327 target_get_section_table (struct target_ops *target)
1328 {
1329 struct target_ops *t;
1330
1331 if (targetdebug)
1332 fprintf_unfiltered (gdb_stdlog, "target_get_section_table ()\n");
1333
1334 for (t = target; t != NULL; t = t->beneath)
1335 if (t->to_get_section_table != NULL)
1336 return (*t->to_get_section_table) (t);
1337
1338 return NULL;
1339 }
1340
1341 /* Find a section containing ADDR. */
1342
1343 struct target_section *
1344 target_section_by_addr (struct target_ops *target, CORE_ADDR addr)
1345 {
1346 struct target_section_table *table = target_get_section_table (target);
1347 struct target_section *secp;
1348
1349 if (table == NULL)
1350 return NULL;
1351
1352 for (secp = table->sections; secp < table->sections_end; secp++)
1353 {
1354 if (addr >= secp->addr && addr < secp->endaddr)
1355 return secp;
1356 }
1357 return NULL;
1358 }
1359
1360 /* Read memory from the live target, even if currently inspecting a
1361 traceframe. The return is the same as that of target_read. */
1362
1363 static LONGEST
1364 target_read_live_memory (enum target_object object,
1365 ULONGEST memaddr, gdb_byte *myaddr, LONGEST len)
1366 {
1367 LONGEST ret;
1368 struct cleanup *cleanup;
1369
1370 /* Switch momentarily out of tfind mode so to access live memory.
1371 Note that this must not clear global state, such as the frame
1372 cache, which must still remain valid for the previous traceframe.
1373 We may be _building_ the frame cache at this point. */
1374 cleanup = make_cleanup_restore_traceframe_number ();
1375 set_traceframe_number (-1);
1376
1377 ret = target_read (current_target.beneath, object, NULL,
1378 myaddr, memaddr, len);
1379
1380 do_cleanups (cleanup);
1381 return ret;
1382 }
1383
1384 /* Using the set of read-only target sections of OPS, read live
1385 read-only memory. Note that the actual reads start from the
1386 top-most target again.
1387
1388 For interface/parameters/return description see target.h,
1389 to_xfer_partial. */
1390
1391 static LONGEST
1392 memory_xfer_live_readonly_partial (struct target_ops *ops,
1393 enum target_object object,
1394 gdb_byte *readbuf, ULONGEST memaddr,
1395 LONGEST len)
1396 {
1397 struct target_section *secp;
1398 struct target_section_table *table;
1399
1400 secp = target_section_by_addr (ops, memaddr);
1401 if (secp != NULL
1402 && (bfd_get_section_flags (secp->the_bfd_section->owner,
1403 secp->the_bfd_section)
1404 & SEC_READONLY))
1405 {
1406 struct target_section *p;
1407 ULONGEST memend = memaddr + len;
1408
1409 table = target_get_section_table (ops);
1410
1411 for (p = table->sections; p < table->sections_end; p++)
1412 {
1413 if (memaddr >= p->addr)
1414 {
1415 if (memend <= p->endaddr)
1416 {
1417 /* Entire transfer is within this section. */
1418 return target_read_live_memory (object, memaddr,
1419 readbuf, len);
1420 }
1421 else if (memaddr >= p->endaddr)
1422 {
1423 /* This section ends before the transfer starts. */
1424 continue;
1425 }
1426 else
1427 {
1428 /* This section overlaps the transfer. Just do half. */
1429 len = p->endaddr - memaddr;
1430 return target_read_live_memory (object, memaddr,
1431 readbuf, len);
1432 }
1433 }
1434 }
1435 }
1436
1437 return 0;
1438 }
1439
1440 /* Perform a partial memory transfer.
1441 For docs see target.h, to_xfer_partial. */
1442
1443 static LONGEST
1444 memory_xfer_partial_1 (struct target_ops *ops, enum target_object object,
1445 void *readbuf, const void *writebuf, ULONGEST memaddr,
1446 LONGEST len)
1447 {
1448 LONGEST res;
1449 int reg_len;
1450 struct mem_region *region;
1451 struct inferior *inf;
1452
1453 /* For accesses to unmapped overlay sections, read directly from
1454 files. Must do this first, as MEMADDR may need adjustment. */
1455 if (readbuf != NULL && overlay_debugging)
1456 {
1457 struct obj_section *section = find_pc_overlay (memaddr);
1458
1459 if (pc_in_unmapped_range (memaddr, section))
1460 {
1461 struct target_section_table *table
1462 = target_get_section_table (ops);
1463 const char *section_name = section->the_bfd_section->name;
1464
1465 memaddr = overlay_mapped_address (memaddr, section);
1466 return section_table_xfer_memory_partial (readbuf, writebuf,
1467 memaddr, len,
1468 table->sections,
1469 table->sections_end,
1470 section_name);
1471 }
1472 }
1473
1474 /* Try the executable files, if "trust-readonly-sections" is set. */
1475 if (readbuf != NULL && trust_readonly)
1476 {
1477 struct target_section *secp;
1478 struct target_section_table *table;
1479
1480 secp = target_section_by_addr (ops, memaddr);
1481 if (secp != NULL
1482 && (bfd_get_section_flags (secp->the_bfd_section->owner,
1483 secp->the_bfd_section)
1484 & SEC_READONLY))
1485 {
1486 table = target_get_section_table (ops);
1487 return section_table_xfer_memory_partial (readbuf, writebuf,
1488 memaddr, len,
1489 table->sections,
1490 table->sections_end,
1491 NULL);
1492 }
1493 }
1494
1495 /* If reading unavailable memory in the context of traceframes, and
1496 this address falls within a read-only section, fallback to
1497 reading from live memory. */
1498 if (readbuf != NULL && get_traceframe_number () != -1)
1499 {
1500 VEC(mem_range_s) *available;
1501
1502 /* If we fail to get the set of available memory, then the
1503 target does not support querying traceframe info, and so we
1504 attempt reading from the traceframe anyway (assuming the
1505 target implements the old QTro packet then). */
1506 if (traceframe_available_memory (&available, memaddr, len))
1507 {
1508 struct cleanup *old_chain;
1509
1510 old_chain = make_cleanup (VEC_cleanup(mem_range_s), &available);
1511
1512 if (VEC_empty (mem_range_s, available)
1513 || VEC_index (mem_range_s, available, 0)->start != memaddr)
1514 {
1515 /* Don't read into the traceframe's available
1516 memory. */
1517 if (!VEC_empty (mem_range_s, available))
1518 {
1519 LONGEST oldlen = len;
1520
1521 len = VEC_index (mem_range_s, available, 0)->start - memaddr;
1522 gdb_assert (len <= oldlen);
1523 }
1524
1525 do_cleanups (old_chain);
1526
1527 /* This goes through the topmost target again. */
1528 res = memory_xfer_live_readonly_partial (ops, object,
1529 readbuf, memaddr, len);
1530 if (res > 0)
1531 return res;
1532
1533 /* No use trying further, we know some memory starting
1534 at MEMADDR isn't available. */
1535 return TARGET_XFER_E_UNAVAILABLE;
1536 }
1537
1538 /* Don't try to read more than how much is available, in
1539 case the target implements the deprecated QTro packet to
1540 cater for older GDBs (the target's knowledge of read-only
1541 sections may be outdated by now). */
1542 len = VEC_index (mem_range_s, available, 0)->length;
1543
1544 do_cleanups (old_chain);
1545 }
1546 }
1547
1548 /* Try GDB's internal data cache. */
1549 region = lookup_mem_region (memaddr);
1550 /* region->hi == 0 means there's no upper bound. */
1551 if (memaddr + len < region->hi || region->hi == 0)
1552 reg_len = len;
1553 else
1554 reg_len = region->hi - memaddr;
1555
1556 switch (region->attrib.mode)
1557 {
1558 case MEM_RO:
1559 if (writebuf != NULL)
1560 return -1;
1561 break;
1562
1563 case MEM_WO:
1564 if (readbuf != NULL)
1565 return -1;
1566 break;
1567
1568 case MEM_FLASH:
1569 /* We only support writing to flash during "load" for now. */
1570 if (writebuf != NULL)
1571 error (_("Writing to flash memory forbidden in this context"));
1572 break;
1573
1574 case MEM_NONE:
1575 return -1;
1576 }
1577
1578 if (!ptid_equal (inferior_ptid, null_ptid))
1579 inf = find_inferior_pid (ptid_get_pid (inferior_ptid));
1580 else
1581 inf = NULL;
1582
1583 if (inf != NULL
1584 /* The dcache reads whole cache lines; that doesn't play well
1585 with reading from a trace buffer, because reading outside of
1586 the collected memory range fails. */
1587 && get_traceframe_number () == -1
1588 && (region->attrib.cache
1589 || (stack_cache_enabled_p && object == TARGET_OBJECT_STACK_MEMORY)))
1590 {
1591 if (readbuf != NULL)
1592 res = dcache_xfer_memory (ops, target_dcache, memaddr, readbuf,
1593 reg_len, 0);
1594 else
1595 /* FIXME drow/2006-08-09: If we're going to preserve const
1596 correctness dcache_xfer_memory should take readbuf and
1597 writebuf. */
1598 res = dcache_xfer_memory (ops, target_dcache, memaddr,
1599 (void *) writebuf,
1600 reg_len, 1);
1601 if (res <= 0)
1602 return -1;
1603 else
1604 return res;
1605 }
1606
1607 /* If none of those methods found the memory we wanted, fall back
1608 to a target partial transfer. Normally a single call to
1609 to_xfer_partial is enough; if it doesn't recognize an object
1610 it will call the to_xfer_partial of the next target down.
1611 But for memory this won't do. Memory is the only target
1612 object which can be read from more than one valid target.
1613 A core file, for instance, could have some of memory but
1614 delegate other bits to the target below it. So, we must
1615 manually try all targets. */
1616
1617 do
1618 {
1619 res = ops->to_xfer_partial (ops, TARGET_OBJECT_MEMORY, NULL,
1620 readbuf, writebuf, memaddr, reg_len);
1621 if (res > 0)
1622 break;
1623
1624 /* We want to continue past core files to executables, but not
1625 past a running target's memory. */
1626 if (ops->to_has_all_memory (ops))
1627 break;
1628
1629 ops = ops->beneath;
1630 }
1631 while (ops != NULL);
1632
1633 /* Make sure the cache gets updated no matter what - if we are writing
1634 to the stack. Even if this write is not tagged as such, we still need
1635 to update the cache. */
1636
1637 if (res > 0
1638 && inf != NULL
1639 && writebuf != NULL
1640 && !region->attrib.cache
1641 && stack_cache_enabled_p
1642 && object != TARGET_OBJECT_STACK_MEMORY)
1643 {
1644 dcache_update (target_dcache, memaddr, (void *) writebuf, res);
1645 }
1646
1647 /* If we still haven't got anything, return the last error. We
1648 give up. */
1649 return res;
1650 }
1651
1652 /* Perform a partial memory transfer. For docs see target.h,
1653 to_xfer_partial. */
1654
1655 static LONGEST
1656 memory_xfer_partial (struct target_ops *ops, enum target_object object,
1657 void *readbuf, const void *writebuf, ULONGEST memaddr,
1658 LONGEST len)
1659 {
1660 int res;
1661
1662 /* Zero length requests are ok and require no work. */
1663 if (len == 0)
1664 return 0;
1665
1666 /* Fill in READBUF with breakpoint shadows, or WRITEBUF with
1667 breakpoint insns, thus hiding out from higher layers whether
1668 there are software breakpoints inserted in the code stream. */
1669 if (readbuf != NULL)
1670 {
1671 res = memory_xfer_partial_1 (ops, object, readbuf, NULL, memaddr, len);
1672
1673 if (res > 0 && !show_memory_breakpoints)
1674 breakpoint_xfer_memory (readbuf, NULL, NULL, memaddr, res);
1675 }
1676 else
1677 {
1678 void *buf;
1679 struct cleanup *old_chain;
1680
1681 /* A large write request is likely to be partially satisfied
1682 by memory_xfer_partial_1. We will continually malloc
1683 and free a copy of the entire write request for breakpoint
1684 shadow handling even though we only end up writing a small
1685 subset of it. Cap writes to 4KB to mitigate this. */
1686 len = min (4096, len);
1687
1688 buf = xmalloc (len);
1689 old_chain = make_cleanup (xfree, buf);
1690 memcpy (buf, writebuf, len);
1691
1692 breakpoint_xfer_memory (NULL, buf, writebuf, memaddr, len);
1693 res = memory_xfer_partial_1 (ops, object, NULL, buf, memaddr, len);
1694
1695 do_cleanups (old_chain);
1696 }
1697
1698 return res;
1699 }
1700
1701 static void
1702 restore_show_memory_breakpoints (void *arg)
1703 {
1704 show_memory_breakpoints = (uintptr_t) arg;
1705 }
1706
1707 struct cleanup *
1708 make_show_memory_breakpoints_cleanup (int show)
1709 {
1710 int current = show_memory_breakpoints;
1711
1712 show_memory_breakpoints = show;
1713 return make_cleanup (restore_show_memory_breakpoints,
1714 (void *) (uintptr_t) current);
1715 }
1716
1717 /* For docs see target.h, to_xfer_partial. */
1718
1719 LONGEST
1720 target_xfer_partial (struct target_ops *ops,
1721 enum target_object object, const char *annex,
1722 void *readbuf, const void *writebuf,
1723 ULONGEST offset, LONGEST len)
1724 {
1725 LONGEST retval;
1726
1727 gdb_assert (ops->to_xfer_partial != NULL);
1728
1729 if (writebuf && !may_write_memory)
1730 error (_("Writing to memory is not allowed (addr %s, len %s)"),
1731 core_addr_to_string_nz (offset), plongest (len));
1732
1733 /* If this is a memory transfer, let the memory-specific code
1734 have a look at it instead. Memory transfers are more
1735 complicated. */
1736 if (object == TARGET_OBJECT_MEMORY || object == TARGET_OBJECT_STACK_MEMORY)
1737 retval = memory_xfer_partial (ops, object, readbuf,
1738 writebuf, offset, len);
1739 else
1740 {
1741 enum target_object raw_object = object;
1742
1743 /* If this is a raw memory transfer, request the normal
1744 memory object from other layers. */
1745 if (raw_object == TARGET_OBJECT_RAW_MEMORY)
1746 raw_object = TARGET_OBJECT_MEMORY;
1747
1748 retval = ops->to_xfer_partial (ops, raw_object, annex, readbuf,
1749 writebuf, offset, len);
1750 }
1751
1752 if (targetdebug)
1753 {
1754 const unsigned char *myaddr = NULL;
1755
1756 fprintf_unfiltered (gdb_stdlog,
1757 "%s:target_xfer_partial "
1758 "(%d, %s, %s, %s, %s, %s) = %s",
1759 ops->to_shortname,
1760 (int) object,
1761 (annex ? annex : "(null)"),
1762 host_address_to_string (readbuf),
1763 host_address_to_string (writebuf),
1764 core_addr_to_string_nz (offset),
1765 plongest (len), plongest (retval));
1766
1767 if (readbuf)
1768 myaddr = readbuf;
1769 if (writebuf)
1770 myaddr = writebuf;
1771 if (retval > 0 && myaddr != NULL)
1772 {
1773 int i;
1774
1775 fputs_unfiltered (", bytes =", gdb_stdlog);
1776 for (i = 0; i < retval; i++)
1777 {
1778 if ((((intptr_t) &(myaddr[i])) & 0xf) == 0)
1779 {
1780 if (targetdebug < 2 && i > 0)
1781 {
1782 fprintf_unfiltered (gdb_stdlog, " ...");
1783 break;
1784 }
1785 fprintf_unfiltered (gdb_stdlog, "\n");
1786 }
1787
1788 fprintf_unfiltered (gdb_stdlog, " %02x", myaddr[i] & 0xff);
1789 }
1790 }
1791
1792 fputc_unfiltered ('\n', gdb_stdlog);
1793 }
1794 return retval;
1795 }
1796
1797 /* Read LEN bytes of target memory at address MEMADDR, placing the
1798 results in GDB's memory at MYADDR. Returns either 0 for success or
1799 a target_xfer_error value if any error occurs.
1800
1801 If an error occurs, no guarantee is made about the contents of the data at
1802 MYADDR. In particular, the caller should not depend upon partial reads
1803 filling the buffer with good data. There is no way for the caller to know
1804 how much good data might have been transfered anyway. Callers that can
1805 deal with partial reads should call target_read (which will retry until
1806 it makes no progress, and then return how much was transferred). */
1807
1808 int
1809 target_read_memory (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1810 {
1811 /* Dispatch to the topmost target, not the flattened current_target.
1812 Memory accesses check target->to_has_(all_)memory, and the
1813 flattened target doesn't inherit those. */
1814 if (target_read (current_target.beneath, TARGET_OBJECT_MEMORY, NULL,
1815 myaddr, memaddr, len) == len)
1816 return 0;
1817 else
1818 return TARGET_XFER_E_IO;
1819 }
1820
1821 /* Like target_read_memory, but specify explicitly that this is a read from
1822 the target's stack. This may trigger different cache behavior. */
1823
1824 int
1825 target_read_stack (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1826 {
1827 /* Dispatch to the topmost target, not the flattened current_target.
1828 Memory accesses check target->to_has_(all_)memory, and the
1829 flattened target doesn't inherit those. */
1830
1831 if (target_read (current_target.beneath, TARGET_OBJECT_STACK_MEMORY, NULL,
1832 myaddr, memaddr, len) == len)
1833 return 0;
1834 else
1835 return TARGET_XFER_E_IO;
1836 }
1837
1838 /* Write LEN bytes from MYADDR to target memory at address MEMADDR.
1839 Returns either 0 for success or a target_xfer_error value if any
1840 error occurs. If an error occurs, no guarantee is made about how
1841 much data got written. Callers that can deal with partial writes
1842 should call target_write. */
1843
1844 int
1845 target_write_memory (CORE_ADDR memaddr, const gdb_byte *myaddr, ssize_t len)
1846 {
1847 /* Dispatch to the topmost target, not the flattened current_target.
1848 Memory accesses check target->to_has_(all_)memory, and the
1849 flattened target doesn't inherit those. */
1850 if (target_write (current_target.beneath, TARGET_OBJECT_MEMORY, NULL,
1851 myaddr, memaddr, len) == len)
1852 return 0;
1853 else
1854 return TARGET_XFER_E_IO;
1855 }
1856
1857 /* Write LEN bytes from MYADDR to target raw memory at address
1858 MEMADDR. Returns either 0 for success or a target_xfer_error value
1859 if any error occurs. If an error occurs, no guarantee is made
1860 about how much data got written. Callers that can deal with
1861 partial writes should call target_write. */
1862
1863 int
1864 target_write_raw_memory (CORE_ADDR memaddr, const gdb_byte *myaddr, ssize_t len)
1865 {
1866 /* Dispatch to the topmost target, not the flattened current_target.
1867 Memory accesses check target->to_has_(all_)memory, and the
1868 flattened target doesn't inherit those. */
1869 if (target_write (current_target.beneath, TARGET_OBJECT_RAW_MEMORY, NULL,
1870 myaddr, memaddr, len) == len)
1871 return 0;
1872 else
1873 return TARGET_XFER_E_IO;
1874 }
1875
1876 /* Fetch the target's memory map. */
1877
1878 VEC(mem_region_s) *
1879 target_memory_map (void)
1880 {
1881 VEC(mem_region_s) *result;
1882 struct mem_region *last_one, *this_one;
1883 int ix;
1884 struct target_ops *t;
1885
1886 if (targetdebug)
1887 fprintf_unfiltered (gdb_stdlog, "target_memory_map ()\n");
1888
1889 for (t = current_target.beneath; t != NULL; t = t->beneath)
1890 if (t->to_memory_map != NULL)
1891 break;
1892
1893 if (t == NULL)
1894 return NULL;
1895
1896 result = t->to_memory_map (t);
1897 if (result == NULL)
1898 return NULL;
1899
1900 qsort (VEC_address (mem_region_s, result),
1901 VEC_length (mem_region_s, result),
1902 sizeof (struct mem_region), mem_region_cmp);
1903
1904 /* Check that regions do not overlap. Simultaneously assign
1905 a numbering for the "mem" commands to use to refer to
1906 each region. */
1907 last_one = NULL;
1908 for (ix = 0; VEC_iterate (mem_region_s, result, ix, this_one); ix++)
1909 {
1910 this_one->number = ix;
1911
1912 if (last_one && last_one->hi > this_one->lo)
1913 {
1914 warning (_("Overlapping regions in memory map: ignoring"));
1915 VEC_free (mem_region_s, result);
1916 return NULL;
1917 }
1918 last_one = this_one;
1919 }
1920
1921 return result;
1922 }
1923
1924 void
1925 target_flash_erase (ULONGEST address, LONGEST length)
1926 {
1927 struct target_ops *t;
1928
1929 for (t = current_target.beneath; t != NULL; t = t->beneath)
1930 if (t->to_flash_erase != NULL)
1931 {
1932 if (targetdebug)
1933 fprintf_unfiltered (gdb_stdlog, "target_flash_erase (%s, %s)\n",
1934 hex_string (address), phex (length, 0));
1935 t->to_flash_erase (t, address, length);
1936 return;
1937 }
1938
1939 tcomplain ();
1940 }
1941
1942 void
1943 target_flash_done (void)
1944 {
1945 struct target_ops *t;
1946
1947 for (t = current_target.beneath; t != NULL; t = t->beneath)
1948 if (t->to_flash_done != NULL)
1949 {
1950 if (targetdebug)
1951 fprintf_unfiltered (gdb_stdlog, "target_flash_done\n");
1952 t->to_flash_done (t);
1953 return;
1954 }
1955
1956 tcomplain ();
1957 }
1958
1959 static void
1960 show_trust_readonly (struct ui_file *file, int from_tty,
1961 struct cmd_list_element *c, const char *value)
1962 {
1963 fprintf_filtered (file,
1964 _("Mode for reading from readonly sections is %s.\n"),
1965 value);
1966 }
1967
1968 /* More generic transfers. */
1969
1970 static LONGEST
1971 default_xfer_partial (struct target_ops *ops, enum target_object object,
1972 const char *annex, gdb_byte *readbuf,
1973 const gdb_byte *writebuf, ULONGEST offset, LONGEST len)
1974 {
1975 if (object == TARGET_OBJECT_MEMORY
1976 && ops->deprecated_xfer_memory != NULL)
1977 /* If available, fall back to the target's
1978 "deprecated_xfer_memory" method. */
1979 {
1980 int xfered = -1;
1981
1982 errno = 0;
1983 if (writebuf != NULL)
1984 {
1985 void *buffer = xmalloc (len);
1986 struct cleanup *cleanup = make_cleanup (xfree, buffer);
1987
1988 memcpy (buffer, writebuf, len);
1989 xfered = ops->deprecated_xfer_memory (offset, buffer, len,
1990 1/*write*/, NULL, ops);
1991 do_cleanups (cleanup);
1992 }
1993 if (readbuf != NULL)
1994 xfered = ops->deprecated_xfer_memory (offset, readbuf, len,
1995 0/*read*/, NULL, ops);
1996 if (xfered > 0)
1997 return xfered;
1998 else if (xfered == 0 && errno == 0)
1999 /* "deprecated_xfer_memory" uses 0, cross checked against
2000 ERRNO as one indication of an error. */
2001 return 0;
2002 else
2003 return -1;
2004 }
2005 else if (ops->beneath != NULL)
2006 return ops->beneath->to_xfer_partial (ops->beneath, object, annex,
2007 readbuf, writebuf, offset, len);
2008 else
2009 return -1;
2010 }
2011
2012 /* The xfer_partial handler for the topmost target. Unlike the default,
2013 it does not need to handle memory specially; it just passes all
2014 requests down the stack. */
2015
2016 static LONGEST
2017 current_xfer_partial (struct target_ops *ops, enum target_object object,
2018 const char *annex, gdb_byte *readbuf,
2019 const gdb_byte *writebuf, ULONGEST offset, LONGEST len)
2020 {
2021 if (ops->beneath != NULL)
2022 return ops->beneath->to_xfer_partial (ops->beneath, object, annex,
2023 readbuf, writebuf, offset, len);
2024 else
2025 return -1;
2026 }
2027
2028 /* Target vector read/write partial wrapper functions. */
2029
2030 static LONGEST
2031 target_read_partial (struct target_ops *ops,
2032 enum target_object object,
2033 const char *annex, gdb_byte *buf,
2034 ULONGEST offset, LONGEST len)
2035 {
2036 return target_xfer_partial (ops, object, annex, buf, NULL, offset, len);
2037 }
2038
2039 static LONGEST
2040 target_write_partial (struct target_ops *ops,
2041 enum target_object object,
2042 const char *annex, const gdb_byte *buf,
2043 ULONGEST offset, LONGEST len)
2044 {
2045 return target_xfer_partial (ops, object, annex, NULL, buf, offset, len);
2046 }
2047
2048 /* Wrappers to perform the full transfer. */
2049
2050 /* For docs on target_read see target.h. */
2051
2052 LONGEST
2053 target_read (struct target_ops *ops,
2054 enum target_object object,
2055 const char *annex, gdb_byte *buf,
2056 ULONGEST offset, LONGEST len)
2057 {
2058 LONGEST xfered = 0;
2059
2060 while (xfered < len)
2061 {
2062 LONGEST xfer = target_read_partial (ops, object, annex,
2063 (gdb_byte *) buf + xfered,
2064 offset + xfered, len - xfered);
2065
2066 /* Call an observer, notifying them of the xfer progress? */
2067 if (xfer == 0)
2068 return xfered;
2069 if (xfer < 0)
2070 return -1;
2071 xfered += xfer;
2072 QUIT;
2073 }
2074 return len;
2075 }
2076
2077 /* Assuming that the entire [begin, end) range of memory cannot be
2078 read, try to read whatever subrange is possible to read.
2079
2080 The function returns, in RESULT, either zero or one memory block.
2081 If there's a readable subrange at the beginning, it is completely
2082 read and returned. Any further readable subrange will not be read.
2083 Otherwise, if there's a readable subrange at the end, it will be
2084 completely read and returned. Any readable subranges before it
2085 (obviously, not starting at the beginning), will be ignored. In
2086 other cases -- either no readable subrange, or readable subrange(s)
2087 that is neither at the beginning, or end, nothing is returned.
2088
2089 The purpose of this function is to handle a read across a boundary
2090 of accessible memory in a case when memory map is not available.
2091 The above restrictions are fine for this case, but will give
2092 incorrect results if the memory is 'patchy'. However, supporting
2093 'patchy' memory would require trying to read every single byte,
2094 and it seems unacceptable solution. Explicit memory map is
2095 recommended for this case -- and target_read_memory_robust will
2096 take care of reading multiple ranges then. */
2097
2098 static void
2099 read_whatever_is_readable (struct target_ops *ops,
2100 ULONGEST begin, ULONGEST end,
2101 VEC(memory_read_result_s) **result)
2102 {
2103 gdb_byte *buf = xmalloc (end - begin);
2104 ULONGEST current_begin = begin;
2105 ULONGEST current_end = end;
2106 int forward;
2107 memory_read_result_s r;
2108
2109 /* If we previously failed to read 1 byte, nothing can be done here. */
2110 if (end - begin <= 1)
2111 {
2112 xfree (buf);
2113 return;
2114 }
2115
2116 /* Check that either first or the last byte is readable, and give up
2117 if not. This heuristic is meant to permit reading accessible memory
2118 at the boundary of accessible region. */
2119 if (target_read_partial (ops, TARGET_OBJECT_MEMORY, NULL,
2120 buf, begin, 1) == 1)
2121 {
2122 forward = 1;
2123 ++current_begin;
2124 }
2125 else if (target_read_partial (ops, TARGET_OBJECT_MEMORY, NULL,
2126 buf + (end-begin) - 1, end - 1, 1) == 1)
2127 {
2128 forward = 0;
2129 --current_end;
2130 }
2131 else
2132 {
2133 xfree (buf);
2134 return;
2135 }
2136
2137 /* Loop invariant is that the [current_begin, current_end) was previously
2138 found to be not readable as a whole.
2139
2140 Note loop condition -- if the range has 1 byte, we can't divide the range
2141 so there's no point trying further. */
2142 while (current_end - current_begin > 1)
2143 {
2144 ULONGEST first_half_begin, first_half_end;
2145 ULONGEST second_half_begin, second_half_end;
2146 LONGEST xfer;
2147 ULONGEST middle = current_begin + (current_end - current_begin)/2;
2148
2149 if (forward)
2150 {
2151 first_half_begin = current_begin;
2152 first_half_end = middle;
2153 second_half_begin = middle;
2154 second_half_end = current_end;
2155 }
2156 else
2157 {
2158 first_half_begin = middle;
2159 first_half_end = current_end;
2160 second_half_begin = current_begin;
2161 second_half_end = middle;
2162 }
2163
2164 xfer = target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2165 buf + (first_half_begin - begin),
2166 first_half_begin,
2167 first_half_end - first_half_begin);
2168
2169 if (xfer == first_half_end - first_half_begin)
2170 {
2171 /* This half reads up fine. So, the error must be in the
2172 other half. */
2173 current_begin = second_half_begin;
2174 current_end = second_half_end;
2175 }
2176 else
2177 {
2178 /* This half is not readable. Because we've tried one byte, we
2179 know some part of this half if actually redable. Go to the next
2180 iteration to divide again and try to read.
2181
2182 We don't handle the other half, because this function only tries
2183 to read a single readable subrange. */
2184 current_begin = first_half_begin;
2185 current_end = first_half_end;
2186 }
2187 }
2188
2189 if (forward)
2190 {
2191 /* The [begin, current_begin) range has been read. */
2192 r.begin = begin;
2193 r.end = current_begin;
2194 r.data = buf;
2195 }
2196 else
2197 {
2198 /* The [current_end, end) range has been read. */
2199 LONGEST rlen = end - current_end;
2200
2201 r.data = xmalloc (rlen);
2202 memcpy (r.data, buf + current_end - begin, rlen);
2203 r.begin = current_end;
2204 r.end = end;
2205 xfree (buf);
2206 }
2207 VEC_safe_push(memory_read_result_s, (*result), &r);
2208 }
2209
2210 void
2211 free_memory_read_result_vector (void *x)
2212 {
2213 VEC(memory_read_result_s) *v = x;
2214 memory_read_result_s *current;
2215 int ix;
2216
2217 for (ix = 0; VEC_iterate (memory_read_result_s, v, ix, current); ++ix)
2218 {
2219 xfree (current->data);
2220 }
2221 VEC_free (memory_read_result_s, v);
2222 }
2223
2224 VEC(memory_read_result_s) *
2225 read_memory_robust (struct target_ops *ops, ULONGEST offset, LONGEST len)
2226 {
2227 VEC(memory_read_result_s) *result = 0;
2228
2229 LONGEST xfered = 0;
2230 while (xfered < len)
2231 {
2232 struct mem_region *region = lookup_mem_region (offset + xfered);
2233 LONGEST rlen;
2234
2235 /* If there is no explicit region, a fake one should be created. */
2236 gdb_assert (region);
2237
2238 if (region->hi == 0)
2239 rlen = len - xfered;
2240 else
2241 rlen = region->hi - offset;
2242
2243 if (region->attrib.mode == MEM_NONE || region->attrib.mode == MEM_WO)
2244 {
2245 /* Cannot read this region. Note that we can end up here only
2246 if the region is explicitly marked inaccessible, or
2247 'inaccessible-by-default' is in effect. */
2248 xfered += rlen;
2249 }
2250 else
2251 {
2252 LONGEST to_read = min (len - xfered, rlen);
2253 gdb_byte *buffer = (gdb_byte *)xmalloc (to_read);
2254
2255 LONGEST xfer = target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2256 (gdb_byte *) buffer,
2257 offset + xfered, to_read);
2258 /* Call an observer, notifying them of the xfer progress? */
2259 if (xfer <= 0)
2260 {
2261 /* Got an error reading full chunk. See if maybe we can read
2262 some subrange. */
2263 xfree (buffer);
2264 read_whatever_is_readable (ops, offset + xfered,
2265 offset + xfered + to_read, &result);
2266 xfered += to_read;
2267 }
2268 else
2269 {
2270 struct memory_read_result r;
2271 r.data = buffer;
2272 r.begin = offset + xfered;
2273 r.end = r.begin + xfer;
2274 VEC_safe_push (memory_read_result_s, result, &r);
2275 xfered += xfer;
2276 }
2277 QUIT;
2278 }
2279 }
2280 return result;
2281 }
2282
2283
2284 /* An alternative to target_write with progress callbacks. */
2285
2286 LONGEST
2287 target_write_with_progress (struct target_ops *ops,
2288 enum target_object object,
2289 const char *annex, const gdb_byte *buf,
2290 ULONGEST offset, LONGEST len,
2291 void (*progress) (ULONGEST, void *), void *baton)
2292 {
2293 LONGEST xfered = 0;
2294
2295 /* Give the progress callback a chance to set up. */
2296 if (progress)
2297 (*progress) (0, baton);
2298
2299 while (xfered < len)
2300 {
2301 LONGEST xfer = target_write_partial (ops, object, annex,
2302 (gdb_byte *) buf + xfered,
2303 offset + xfered, len - xfered);
2304
2305 if (xfer == 0)
2306 return xfered;
2307 if (xfer < 0)
2308 return -1;
2309
2310 if (progress)
2311 (*progress) (xfer, baton);
2312
2313 xfered += xfer;
2314 QUIT;
2315 }
2316 return len;
2317 }
2318
2319 /* For docs on target_write see target.h. */
2320
2321 LONGEST
2322 target_write (struct target_ops *ops,
2323 enum target_object object,
2324 const char *annex, const gdb_byte *buf,
2325 ULONGEST offset, LONGEST len)
2326 {
2327 return target_write_with_progress (ops, object, annex, buf, offset, len,
2328 NULL, NULL);
2329 }
2330
2331 /* Read OBJECT/ANNEX using OPS. Store the result in *BUF_P and return
2332 the size of the transferred data. PADDING additional bytes are
2333 available in *BUF_P. This is a helper function for
2334 target_read_alloc; see the declaration of that function for more
2335 information. */
2336
2337 static LONGEST
2338 target_read_alloc_1 (struct target_ops *ops, enum target_object object,
2339 const char *annex, gdb_byte **buf_p, int padding)
2340 {
2341 size_t buf_alloc, buf_pos;
2342 gdb_byte *buf;
2343 LONGEST n;
2344
2345 /* This function does not have a length parameter; it reads the
2346 entire OBJECT). Also, it doesn't support objects fetched partly
2347 from one target and partly from another (in a different stratum,
2348 e.g. a core file and an executable). Both reasons make it
2349 unsuitable for reading memory. */
2350 gdb_assert (object != TARGET_OBJECT_MEMORY);
2351
2352 /* Start by reading up to 4K at a time. The target will throttle
2353 this number down if necessary. */
2354 buf_alloc = 4096;
2355 buf = xmalloc (buf_alloc);
2356 buf_pos = 0;
2357 while (1)
2358 {
2359 n = target_read_partial (ops, object, annex, &buf[buf_pos],
2360 buf_pos, buf_alloc - buf_pos - padding);
2361 if (n < 0)
2362 {
2363 /* An error occurred. */
2364 xfree (buf);
2365 return -1;
2366 }
2367 else if (n == 0)
2368 {
2369 /* Read all there was. */
2370 if (buf_pos == 0)
2371 xfree (buf);
2372 else
2373 *buf_p = buf;
2374 return buf_pos;
2375 }
2376
2377 buf_pos += n;
2378
2379 /* If the buffer is filling up, expand it. */
2380 if (buf_alloc < buf_pos * 2)
2381 {
2382 buf_alloc *= 2;
2383 buf = xrealloc (buf, buf_alloc);
2384 }
2385
2386 QUIT;
2387 }
2388 }
2389
2390 /* Read OBJECT/ANNEX using OPS. Store the result in *BUF_P and return
2391 the size of the transferred data. See the declaration in "target.h"
2392 function for more information about the return value. */
2393
2394 LONGEST
2395 target_read_alloc (struct target_ops *ops, enum target_object object,
2396 const char *annex, gdb_byte **buf_p)
2397 {
2398 return target_read_alloc_1 (ops, object, annex, buf_p, 0);
2399 }
2400
2401 /* Read OBJECT/ANNEX using OPS. The result is NUL-terminated and
2402 returned as a string, allocated using xmalloc. If an error occurs
2403 or the transfer is unsupported, NULL is returned. Empty objects
2404 are returned as allocated but empty strings. A warning is issued
2405 if the result contains any embedded NUL bytes. */
2406
2407 char *
2408 target_read_stralloc (struct target_ops *ops, enum target_object object,
2409 const char *annex)
2410 {
2411 gdb_byte *buffer;
2412 char *bufstr;
2413 LONGEST i, transferred;
2414
2415 transferred = target_read_alloc_1 (ops, object, annex, &buffer, 1);
2416 bufstr = (char *) buffer;
2417
2418 if (transferred < 0)
2419 return NULL;
2420
2421 if (transferred == 0)
2422 return xstrdup ("");
2423
2424 bufstr[transferred] = 0;
2425
2426 /* Check for embedded NUL bytes; but allow trailing NULs. */
2427 for (i = strlen (bufstr); i < transferred; i++)
2428 if (bufstr[i] != 0)
2429 {
2430 warning (_("target object %d, annex %s, "
2431 "contained unexpected null characters"),
2432 (int) object, annex ? annex : "(none)");
2433 break;
2434 }
2435
2436 return bufstr;
2437 }
2438
2439 /* Memory transfer methods. */
2440
2441 void
2442 get_target_memory (struct target_ops *ops, CORE_ADDR addr, gdb_byte *buf,
2443 LONGEST len)
2444 {
2445 /* This method is used to read from an alternate, non-current
2446 target. This read must bypass the overlay support (as symbols
2447 don't match this target), and GDB's internal cache (wrong cache
2448 for this target). */
2449 if (target_read (ops, TARGET_OBJECT_RAW_MEMORY, NULL, buf, addr, len)
2450 != len)
2451 memory_error (TARGET_XFER_E_IO, addr);
2452 }
2453
2454 ULONGEST
2455 get_target_memory_unsigned (struct target_ops *ops, CORE_ADDR addr,
2456 int len, enum bfd_endian byte_order)
2457 {
2458 gdb_byte buf[sizeof (ULONGEST)];
2459
2460 gdb_assert (len <= sizeof (buf));
2461 get_target_memory (ops, addr, buf, len);
2462 return extract_unsigned_integer (buf, len, byte_order);
2463 }
2464
2465 int
2466 target_insert_breakpoint (struct gdbarch *gdbarch,
2467 struct bp_target_info *bp_tgt)
2468 {
2469 if (!may_insert_breakpoints)
2470 {
2471 warning (_("May not insert breakpoints"));
2472 return 1;
2473 }
2474
2475 return (*current_target.to_insert_breakpoint) (gdbarch, bp_tgt);
2476 }
2477
2478 int
2479 target_remove_breakpoint (struct gdbarch *gdbarch,
2480 struct bp_target_info *bp_tgt)
2481 {
2482 /* This is kind of a weird case to handle, but the permission might
2483 have been changed after breakpoints were inserted - in which case
2484 we should just take the user literally and assume that any
2485 breakpoints should be left in place. */
2486 if (!may_insert_breakpoints)
2487 {
2488 warning (_("May not remove breakpoints"));
2489 return 1;
2490 }
2491
2492 return (*current_target.to_remove_breakpoint) (gdbarch, bp_tgt);
2493 }
2494
2495 static void
2496 target_info (char *args, int from_tty)
2497 {
2498 struct target_ops *t;
2499 int has_all_mem = 0;
2500
2501 if (symfile_objfile != NULL)
2502 printf_unfiltered (_("Symbols from \"%s\".\n"),
2503 objfile_name (symfile_objfile));
2504
2505 for (t = target_stack; t != NULL; t = t->beneath)
2506 {
2507 if (!(*t->to_has_memory) (t))
2508 continue;
2509
2510 if ((int) (t->to_stratum) <= (int) dummy_stratum)
2511 continue;
2512 if (has_all_mem)
2513 printf_unfiltered (_("\tWhile running this, "
2514 "GDB does not access memory from...\n"));
2515 printf_unfiltered ("%s:\n", t->to_longname);
2516 (t->to_files_info) (t);
2517 has_all_mem = (*t->to_has_all_memory) (t);
2518 }
2519 }
2520
2521 /* This function is called before any new inferior is created, e.g.
2522 by running a program, attaching, or connecting to a target.
2523 It cleans up any state from previous invocations which might
2524 change between runs. This is a subset of what target_preopen
2525 resets (things which might change between targets). */
2526
2527 void
2528 target_pre_inferior (int from_tty)
2529 {
2530 /* Clear out solib state. Otherwise the solib state of the previous
2531 inferior might have survived and is entirely wrong for the new
2532 target. This has been observed on GNU/Linux using glibc 2.3. How
2533 to reproduce:
2534
2535 bash$ ./foo&
2536 [1] 4711
2537 bash$ ./foo&
2538 [1] 4712
2539 bash$ gdb ./foo
2540 [...]
2541 (gdb) attach 4711
2542 (gdb) detach
2543 (gdb) attach 4712
2544 Cannot access memory at address 0xdeadbeef
2545 */
2546
2547 /* In some OSs, the shared library list is the same/global/shared
2548 across inferiors. If code is shared between processes, so are
2549 memory regions and features. */
2550 if (!gdbarch_has_global_solist (target_gdbarch ()))
2551 {
2552 no_shared_libraries (NULL, from_tty);
2553
2554 invalidate_target_mem_regions ();
2555
2556 target_clear_description ();
2557 }
2558
2559 agent_capability_invalidate ();
2560 }
2561
2562 /* Callback for iterate_over_inferiors. Gets rid of the given
2563 inferior. */
2564
2565 static int
2566 dispose_inferior (struct inferior *inf, void *args)
2567 {
2568 struct thread_info *thread;
2569
2570 thread = any_thread_of_process (inf->pid);
2571 if (thread)
2572 {
2573 switch_to_thread (thread->ptid);
2574
2575 /* Core inferiors actually should be detached, not killed. */
2576 if (target_has_execution)
2577 target_kill ();
2578 else
2579 target_detach (NULL, 0);
2580 }
2581
2582 return 0;
2583 }
2584
2585 /* This is to be called by the open routine before it does
2586 anything. */
2587
2588 void
2589 target_preopen (int from_tty)
2590 {
2591 dont_repeat ();
2592
2593 if (have_inferiors ())
2594 {
2595 if (!from_tty
2596 || !have_live_inferiors ()
2597 || query (_("A program is being debugged already. Kill it? ")))
2598 iterate_over_inferiors (dispose_inferior, NULL);
2599 else
2600 error (_("Program not killed."));
2601 }
2602
2603 /* Calling target_kill may remove the target from the stack. But if
2604 it doesn't (which seems like a win for UDI), remove it now. */
2605 /* Leave the exec target, though. The user may be switching from a
2606 live process to a core of the same program. */
2607 pop_all_targets_above (file_stratum);
2608
2609 target_pre_inferior (from_tty);
2610 }
2611
2612 /* Detach a target after doing deferred register stores. */
2613
2614 void
2615 target_detach (char *args, int from_tty)
2616 {
2617 struct target_ops* t;
2618
2619 if (gdbarch_has_global_breakpoints (target_gdbarch ()))
2620 /* Don't remove global breakpoints here. They're removed on
2621 disconnection from the target. */
2622 ;
2623 else
2624 /* If we're in breakpoints-always-inserted mode, have to remove
2625 them before detaching. */
2626 remove_breakpoints_pid (ptid_get_pid (inferior_ptid));
2627
2628 prepare_for_detach ();
2629
2630 for (t = current_target.beneath; t != NULL; t = t->beneath)
2631 {
2632 if (t->to_detach != NULL)
2633 {
2634 t->to_detach (t, args, from_tty);
2635 if (targetdebug)
2636 fprintf_unfiltered (gdb_stdlog, "target_detach (%s, %d)\n",
2637 args, from_tty);
2638 return;
2639 }
2640 }
2641
2642 internal_error (__FILE__, __LINE__, _("could not find a target to detach"));
2643 }
2644
2645 void
2646 target_disconnect (char *args, int from_tty)
2647 {
2648 struct target_ops *t;
2649
2650 /* If we're in breakpoints-always-inserted mode or if breakpoints
2651 are global across processes, we have to remove them before
2652 disconnecting. */
2653 remove_breakpoints ();
2654
2655 for (t = current_target.beneath; t != NULL; t = t->beneath)
2656 if (t->to_disconnect != NULL)
2657 {
2658 if (targetdebug)
2659 fprintf_unfiltered (gdb_stdlog, "target_disconnect (%s, %d)\n",
2660 args, from_tty);
2661 t->to_disconnect (t, args, from_tty);
2662 return;
2663 }
2664
2665 tcomplain ();
2666 }
2667
2668 ptid_t
2669 target_wait (ptid_t ptid, struct target_waitstatus *status, int options)
2670 {
2671 struct target_ops *t;
2672
2673 for (t = current_target.beneath; t != NULL; t = t->beneath)
2674 {
2675 if (t->to_wait != NULL)
2676 {
2677 ptid_t retval = (*t->to_wait) (t, ptid, status, options);
2678
2679 if (targetdebug)
2680 {
2681 char *status_string;
2682 char *options_string;
2683
2684 status_string = target_waitstatus_to_string (status);
2685 options_string = target_options_to_string (options);
2686 fprintf_unfiltered (gdb_stdlog,
2687 "target_wait (%d, status, options={%s})"
2688 " = %d, %s\n",
2689 ptid_get_pid (ptid), options_string,
2690 ptid_get_pid (retval), status_string);
2691 xfree (status_string);
2692 xfree (options_string);
2693 }
2694
2695 return retval;
2696 }
2697 }
2698
2699 noprocess ();
2700 }
2701
2702 char *
2703 target_pid_to_str (ptid_t ptid)
2704 {
2705 struct target_ops *t;
2706
2707 for (t = current_target.beneath; t != NULL; t = t->beneath)
2708 {
2709 if (t->to_pid_to_str != NULL)
2710 return (*t->to_pid_to_str) (t, ptid);
2711 }
2712
2713 return normal_pid_to_str (ptid);
2714 }
2715
2716 char *
2717 target_thread_name (struct thread_info *info)
2718 {
2719 struct target_ops *t;
2720
2721 for (t = current_target.beneath; t != NULL; t = t->beneath)
2722 {
2723 if (t->to_thread_name != NULL)
2724 return (*t->to_thread_name) (info);
2725 }
2726
2727 return NULL;
2728 }
2729
2730 void
2731 target_resume (ptid_t ptid, int step, enum gdb_signal signal)
2732 {
2733 struct target_ops *t;
2734
2735 target_dcache_invalidate ();
2736
2737 for (t = current_target.beneath; t != NULL; t = t->beneath)
2738 {
2739 if (t->to_resume != NULL)
2740 {
2741 t->to_resume (t, ptid, step, signal);
2742 if (targetdebug)
2743 fprintf_unfiltered (gdb_stdlog, "target_resume (%d, %s, %s)\n",
2744 ptid_get_pid (ptid),
2745 step ? "step" : "continue",
2746 gdb_signal_to_name (signal));
2747
2748 registers_changed_ptid (ptid);
2749 set_executing (ptid, 1);
2750 set_running (ptid, 1);
2751 clear_inline_frame_state (ptid);
2752 return;
2753 }
2754 }
2755
2756 noprocess ();
2757 }
2758
2759 void
2760 target_pass_signals (int numsigs, unsigned char *pass_signals)
2761 {
2762 struct target_ops *t;
2763
2764 for (t = current_target.beneath; t != NULL; t = t->beneath)
2765 {
2766 if (t->to_pass_signals != NULL)
2767 {
2768 if (targetdebug)
2769 {
2770 int i;
2771
2772 fprintf_unfiltered (gdb_stdlog, "target_pass_signals (%d, {",
2773 numsigs);
2774
2775 for (i = 0; i < numsigs; i++)
2776 if (pass_signals[i])
2777 fprintf_unfiltered (gdb_stdlog, " %s",
2778 gdb_signal_to_name (i));
2779
2780 fprintf_unfiltered (gdb_stdlog, " })\n");
2781 }
2782
2783 (*t->to_pass_signals) (numsigs, pass_signals);
2784 return;
2785 }
2786 }
2787 }
2788
2789 void
2790 target_program_signals (int numsigs, unsigned char *program_signals)
2791 {
2792 struct target_ops *t;
2793
2794 for (t = current_target.beneath; t != NULL; t = t->beneath)
2795 {
2796 if (t->to_program_signals != NULL)
2797 {
2798 if (targetdebug)
2799 {
2800 int i;
2801
2802 fprintf_unfiltered (gdb_stdlog, "target_program_signals (%d, {",
2803 numsigs);
2804
2805 for (i = 0; i < numsigs; i++)
2806 if (program_signals[i])
2807 fprintf_unfiltered (gdb_stdlog, " %s",
2808 gdb_signal_to_name (i));
2809
2810 fprintf_unfiltered (gdb_stdlog, " })\n");
2811 }
2812
2813 (*t->to_program_signals) (numsigs, program_signals);
2814 return;
2815 }
2816 }
2817 }
2818
2819 /* Look through the list of possible targets for a target that can
2820 follow forks. */
2821
2822 int
2823 target_follow_fork (int follow_child, int detach_fork)
2824 {
2825 struct target_ops *t;
2826
2827 for (t = current_target.beneath; t != NULL; t = t->beneath)
2828 {
2829 if (t->to_follow_fork != NULL)
2830 {
2831 int retval = t->to_follow_fork (t, follow_child, detach_fork);
2832
2833 if (targetdebug)
2834 fprintf_unfiltered (gdb_stdlog,
2835 "target_follow_fork (%d, %d) = %d\n",
2836 follow_child, detach_fork, retval);
2837 return retval;
2838 }
2839 }
2840
2841 /* Some target returned a fork event, but did not know how to follow it. */
2842 internal_error (__FILE__, __LINE__,
2843 _("could not find a target to follow fork"));
2844 }
2845
2846 void
2847 target_mourn_inferior (void)
2848 {
2849 struct target_ops *t;
2850
2851 for (t = current_target.beneath; t != NULL; t = t->beneath)
2852 {
2853 if (t->to_mourn_inferior != NULL)
2854 {
2855 t->to_mourn_inferior (t);
2856 if (targetdebug)
2857 fprintf_unfiltered (gdb_stdlog, "target_mourn_inferior ()\n");
2858
2859 /* We no longer need to keep handles on any of the object files.
2860 Make sure to release them to avoid unnecessarily locking any
2861 of them while we're not actually debugging. */
2862 bfd_cache_close_all ();
2863
2864 return;
2865 }
2866 }
2867
2868 internal_error (__FILE__, __LINE__,
2869 _("could not find a target to follow mourn inferior"));
2870 }
2871
2872 /* Look for a target which can describe architectural features, starting
2873 from TARGET. If we find one, return its description. */
2874
2875 const struct target_desc *
2876 target_read_description (struct target_ops *target)
2877 {
2878 struct target_ops *t;
2879
2880 for (t = target; t != NULL; t = t->beneath)
2881 if (t->to_read_description != NULL)
2882 {
2883 const struct target_desc *tdesc;
2884
2885 tdesc = t->to_read_description (t);
2886 if (tdesc)
2887 return tdesc;
2888 }
2889
2890 return NULL;
2891 }
2892
2893 /* The default implementation of to_search_memory.
2894 This implements a basic search of memory, reading target memory and
2895 performing the search here (as opposed to performing the search in on the
2896 target side with, for example, gdbserver). */
2897
2898 int
2899 simple_search_memory (struct target_ops *ops,
2900 CORE_ADDR start_addr, ULONGEST search_space_len,
2901 const gdb_byte *pattern, ULONGEST pattern_len,
2902 CORE_ADDR *found_addrp)
2903 {
2904 /* NOTE: also defined in find.c testcase. */
2905 #define SEARCH_CHUNK_SIZE 16000
2906 const unsigned chunk_size = SEARCH_CHUNK_SIZE;
2907 /* Buffer to hold memory contents for searching. */
2908 gdb_byte *search_buf;
2909 unsigned search_buf_size;
2910 struct cleanup *old_cleanups;
2911
2912 search_buf_size = chunk_size + pattern_len - 1;
2913
2914 /* No point in trying to allocate a buffer larger than the search space. */
2915 if (search_space_len < search_buf_size)
2916 search_buf_size = search_space_len;
2917
2918 search_buf = malloc (search_buf_size);
2919 if (search_buf == NULL)
2920 error (_("Unable to allocate memory to perform the search."));
2921 old_cleanups = make_cleanup (free_current_contents, &search_buf);
2922
2923 /* Prime the search buffer. */
2924
2925 if (target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2926 search_buf, start_addr, search_buf_size) != search_buf_size)
2927 {
2928 warning (_("Unable to access %s bytes of target "
2929 "memory at %s, halting search."),
2930 pulongest (search_buf_size), hex_string (start_addr));
2931 do_cleanups (old_cleanups);
2932 return -1;
2933 }
2934
2935 /* Perform the search.
2936
2937 The loop is kept simple by allocating [N + pattern-length - 1] bytes.
2938 When we've scanned N bytes we copy the trailing bytes to the start and
2939 read in another N bytes. */
2940
2941 while (search_space_len >= pattern_len)
2942 {
2943 gdb_byte *found_ptr;
2944 unsigned nr_search_bytes = min (search_space_len, search_buf_size);
2945
2946 found_ptr = memmem (search_buf, nr_search_bytes,
2947 pattern, pattern_len);
2948
2949 if (found_ptr != NULL)
2950 {
2951 CORE_ADDR found_addr = start_addr + (found_ptr - search_buf);
2952
2953 *found_addrp = found_addr;
2954 do_cleanups (old_cleanups);
2955 return 1;
2956 }
2957
2958 /* Not found in this chunk, skip to next chunk. */
2959
2960 /* Don't let search_space_len wrap here, it's unsigned. */
2961 if (search_space_len >= chunk_size)
2962 search_space_len -= chunk_size;
2963 else
2964 search_space_len = 0;
2965
2966 if (search_space_len >= pattern_len)
2967 {
2968 unsigned keep_len = search_buf_size - chunk_size;
2969 CORE_ADDR read_addr = start_addr + chunk_size + keep_len;
2970 int nr_to_read;
2971
2972 /* Copy the trailing part of the previous iteration to the front
2973 of the buffer for the next iteration. */
2974 gdb_assert (keep_len == pattern_len - 1);
2975 memcpy (search_buf, search_buf + chunk_size, keep_len);
2976
2977 nr_to_read = min (search_space_len - keep_len, chunk_size);
2978
2979 if (target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2980 search_buf + keep_len, read_addr,
2981 nr_to_read) != nr_to_read)
2982 {
2983 warning (_("Unable to access %s bytes of target "
2984 "memory at %s, halting search."),
2985 plongest (nr_to_read),
2986 hex_string (read_addr));
2987 do_cleanups (old_cleanups);
2988 return -1;
2989 }
2990
2991 start_addr += chunk_size;
2992 }
2993 }
2994
2995 /* Not found. */
2996
2997 do_cleanups (old_cleanups);
2998 return 0;
2999 }
3000
3001 /* Search SEARCH_SPACE_LEN bytes beginning at START_ADDR for the
3002 sequence of bytes in PATTERN with length PATTERN_LEN.
3003
3004 The result is 1 if found, 0 if not found, and -1 if there was an error
3005 requiring halting of the search (e.g. memory read error).
3006 If the pattern is found the address is recorded in FOUND_ADDRP. */
3007
3008 int
3009 target_search_memory (CORE_ADDR start_addr, ULONGEST search_space_len,
3010 const gdb_byte *pattern, ULONGEST pattern_len,
3011 CORE_ADDR *found_addrp)
3012 {
3013 struct target_ops *t;
3014 int found;
3015
3016 /* We don't use INHERIT to set current_target.to_search_memory,
3017 so we have to scan the target stack and handle targetdebug
3018 ourselves. */
3019
3020 if (targetdebug)
3021 fprintf_unfiltered (gdb_stdlog, "target_search_memory (%s, ...)\n",
3022 hex_string (start_addr));
3023
3024 for (t = current_target.beneath; t != NULL; t = t->beneath)
3025 if (t->to_search_memory != NULL)
3026 break;
3027
3028 if (t != NULL)
3029 {
3030 found = t->to_search_memory (t, start_addr, search_space_len,
3031 pattern, pattern_len, found_addrp);
3032 }
3033 else
3034 {
3035 /* If a special version of to_search_memory isn't available, use the
3036 simple version. */
3037 found = simple_search_memory (current_target.beneath,
3038 start_addr, search_space_len,
3039 pattern, pattern_len, found_addrp);
3040 }
3041
3042 if (targetdebug)
3043 fprintf_unfiltered (gdb_stdlog, " = %d\n", found);
3044
3045 return found;
3046 }
3047
3048 /* Look through the currently pushed targets. If none of them will
3049 be able to restart the currently running process, issue an error
3050 message. */
3051
3052 void
3053 target_require_runnable (void)
3054 {
3055 struct target_ops *t;
3056
3057 for (t = target_stack; t != NULL; t = t->beneath)
3058 {
3059 /* If this target knows how to create a new program, then
3060 assume we will still be able to after killing the current
3061 one. Either killing and mourning will not pop T, or else
3062 find_default_run_target will find it again. */
3063 if (t->to_create_inferior != NULL)
3064 return;
3065
3066 /* Do not worry about thread_stratum targets that can not
3067 create inferiors. Assume they will be pushed again if
3068 necessary, and continue to the process_stratum. */
3069 if (t->to_stratum == thread_stratum
3070 || t->to_stratum == arch_stratum)
3071 continue;
3072
3073 error (_("The \"%s\" target does not support \"run\". "
3074 "Try \"help target\" or \"continue\"."),
3075 t->to_shortname);
3076 }
3077
3078 /* This function is only called if the target is running. In that
3079 case there should have been a process_stratum target and it
3080 should either know how to create inferiors, or not... */
3081 internal_error (__FILE__, __LINE__, _("No targets found"));
3082 }
3083
3084 /* Look through the list of possible targets for a target that can
3085 execute a run or attach command without any other data. This is
3086 used to locate the default process stratum.
3087
3088 If DO_MESG is not NULL, the result is always valid (error() is
3089 called for errors); else, return NULL on error. */
3090
3091 static struct target_ops *
3092 find_default_run_target (char *do_mesg)
3093 {
3094 struct target_ops **t;
3095 struct target_ops *runable = NULL;
3096 int count;
3097
3098 count = 0;
3099
3100 for (t = target_structs; t < target_structs + target_struct_size;
3101 ++t)
3102 {
3103 if ((*t)->to_can_run && target_can_run (*t))
3104 {
3105 runable = *t;
3106 ++count;
3107 }
3108 }
3109
3110 if (count != 1)
3111 {
3112 if (do_mesg)
3113 error (_("Don't know how to %s. Try \"help target\"."), do_mesg);
3114 else
3115 return NULL;
3116 }
3117
3118 return runable;
3119 }
3120
3121 void
3122 find_default_attach (struct target_ops *ops, char *args, int from_tty)
3123 {
3124 struct target_ops *t;
3125
3126 t = find_default_run_target ("attach");
3127 (t->to_attach) (t, args, from_tty);
3128 return;
3129 }
3130
3131 void
3132 find_default_create_inferior (struct target_ops *ops,
3133 char *exec_file, char *allargs, char **env,
3134 int from_tty)
3135 {
3136 struct target_ops *t;
3137
3138 t = find_default_run_target ("run");
3139 (t->to_create_inferior) (t, exec_file, allargs, env, from_tty);
3140 return;
3141 }
3142
3143 static int
3144 find_default_can_async_p (void)
3145 {
3146 struct target_ops *t;
3147
3148 /* This may be called before the target is pushed on the stack;
3149 look for the default process stratum. If there's none, gdb isn't
3150 configured with a native debugger, and target remote isn't
3151 connected yet. */
3152 t = find_default_run_target (NULL);
3153 if (t && t->to_can_async_p)
3154 return (t->to_can_async_p) ();
3155 return 0;
3156 }
3157
3158 static int
3159 find_default_is_async_p (void)
3160 {
3161 struct target_ops *t;
3162
3163 /* This may be called before the target is pushed on the stack;
3164 look for the default process stratum. If there's none, gdb isn't
3165 configured with a native debugger, and target remote isn't
3166 connected yet. */
3167 t = find_default_run_target (NULL);
3168 if (t && t->to_is_async_p)
3169 return (t->to_is_async_p) ();
3170 return 0;
3171 }
3172
3173 static int
3174 find_default_supports_non_stop (void)
3175 {
3176 struct target_ops *t;
3177
3178 t = find_default_run_target (NULL);
3179 if (t && t->to_supports_non_stop)
3180 return (t->to_supports_non_stop) ();
3181 return 0;
3182 }
3183
3184 int
3185 target_supports_non_stop (void)
3186 {
3187 struct target_ops *t;
3188
3189 for (t = &current_target; t != NULL; t = t->beneath)
3190 if (t->to_supports_non_stop)
3191 return t->to_supports_non_stop ();
3192
3193 return 0;
3194 }
3195
3196 /* Implement the "info proc" command. */
3197
3198 int
3199 target_info_proc (char *args, enum info_proc_what what)
3200 {
3201 struct target_ops *t;
3202
3203 /* If we're already connected to something that can get us OS
3204 related data, use it. Otherwise, try using the native
3205 target. */
3206 if (current_target.to_stratum >= process_stratum)
3207 t = current_target.beneath;
3208 else
3209 t = find_default_run_target (NULL);
3210
3211 for (; t != NULL; t = t->beneath)
3212 {
3213 if (t->to_info_proc != NULL)
3214 {
3215 t->to_info_proc (t, args, what);
3216
3217 if (targetdebug)
3218 fprintf_unfiltered (gdb_stdlog,
3219 "target_info_proc (\"%s\", %d)\n", args, what);
3220
3221 return 1;
3222 }
3223 }
3224
3225 return 0;
3226 }
3227
3228 static int
3229 find_default_supports_disable_randomization (void)
3230 {
3231 struct target_ops *t;
3232
3233 t = find_default_run_target (NULL);
3234 if (t && t->to_supports_disable_randomization)
3235 return (t->to_supports_disable_randomization) ();
3236 return 0;
3237 }
3238
3239 int
3240 target_supports_disable_randomization (void)
3241 {
3242 struct target_ops *t;
3243
3244 for (t = &current_target; t != NULL; t = t->beneath)
3245 if (t->to_supports_disable_randomization)
3246 return t->to_supports_disable_randomization ();
3247
3248 return 0;
3249 }
3250
3251 char *
3252 target_get_osdata (const char *type)
3253 {
3254 struct target_ops *t;
3255
3256 /* If we're already connected to something that can get us OS
3257 related data, use it. Otherwise, try using the native
3258 target. */
3259 if (current_target.to_stratum >= process_stratum)
3260 t = current_target.beneath;
3261 else
3262 t = find_default_run_target ("get OS data");
3263
3264 if (!t)
3265 return NULL;
3266
3267 return target_read_stralloc (t, TARGET_OBJECT_OSDATA, type);
3268 }
3269
3270 /* Determine the current address space of thread PTID. */
3271
3272 struct address_space *
3273 target_thread_address_space (ptid_t ptid)
3274 {
3275 struct address_space *aspace;
3276 struct inferior *inf;
3277 struct target_ops *t;
3278
3279 for (t = current_target.beneath; t != NULL; t = t->beneath)
3280 {
3281 if (t->to_thread_address_space != NULL)
3282 {
3283 aspace = t->to_thread_address_space (t, ptid);
3284 gdb_assert (aspace);
3285
3286 if (targetdebug)
3287 fprintf_unfiltered (gdb_stdlog,
3288 "target_thread_address_space (%s) = %d\n",
3289 target_pid_to_str (ptid),
3290 address_space_num (aspace));
3291 return aspace;
3292 }
3293 }
3294
3295 /* Fall-back to the "main" address space of the inferior. */
3296 inf = find_inferior_pid (ptid_get_pid (ptid));
3297
3298 if (inf == NULL || inf->aspace == NULL)
3299 internal_error (__FILE__, __LINE__,
3300 _("Can't determine the current "
3301 "address space of thread %s\n"),
3302 target_pid_to_str (ptid));
3303
3304 return inf->aspace;
3305 }
3306
3307
3308 /* Target file operations. */
3309
3310 static struct target_ops *
3311 default_fileio_target (void)
3312 {
3313 /* If we're already connected to something that can perform
3314 file I/O, use it. Otherwise, try using the native target. */
3315 if (current_target.to_stratum >= process_stratum)
3316 return current_target.beneath;
3317 else
3318 return find_default_run_target ("file I/O");
3319 }
3320
3321 /* Open FILENAME on the target, using FLAGS and MODE. Return a
3322 target file descriptor, or -1 if an error occurs (and set
3323 *TARGET_ERRNO). */
3324 int
3325 target_fileio_open (const char *filename, int flags, int mode,
3326 int *target_errno)
3327 {
3328 struct target_ops *t;
3329
3330 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3331 {
3332 if (t->to_fileio_open != NULL)
3333 {
3334 int fd = t->to_fileio_open (filename, flags, mode, target_errno);
3335
3336 if (targetdebug)
3337 fprintf_unfiltered (gdb_stdlog,
3338 "target_fileio_open (%s,0x%x,0%o) = %d (%d)\n",
3339 filename, flags, mode,
3340 fd, fd != -1 ? 0 : *target_errno);
3341 return fd;
3342 }
3343 }
3344
3345 *target_errno = FILEIO_ENOSYS;
3346 return -1;
3347 }
3348
3349 /* Write up to LEN bytes from WRITE_BUF to FD on the target.
3350 Return the number of bytes written, or -1 if an error occurs
3351 (and set *TARGET_ERRNO). */
3352 int
3353 target_fileio_pwrite (int fd, const gdb_byte *write_buf, int len,
3354 ULONGEST offset, int *target_errno)
3355 {
3356 struct target_ops *t;
3357
3358 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3359 {
3360 if (t->to_fileio_pwrite != NULL)
3361 {
3362 int ret = t->to_fileio_pwrite (fd, write_buf, len, offset,
3363 target_errno);
3364
3365 if (targetdebug)
3366 fprintf_unfiltered (gdb_stdlog,
3367 "target_fileio_pwrite (%d,...,%d,%s) "
3368 "= %d (%d)\n",
3369 fd, len, pulongest (offset),
3370 ret, ret != -1 ? 0 : *target_errno);
3371 return ret;
3372 }
3373 }
3374
3375 *target_errno = FILEIO_ENOSYS;
3376 return -1;
3377 }
3378
3379 /* Read up to LEN bytes FD on the target into READ_BUF.
3380 Return the number of bytes read, or -1 if an error occurs
3381 (and set *TARGET_ERRNO). */
3382 int
3383 target_fileio_pread (int fd, gdb_byte *read_buf, int len,
3384 ULONGEST offset, int *target_errno)
3385 {
3386 struct target_ops *t;
3387
3388 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3389 {
3390 if (t->to_fileio_pread != NULL)
3391 {
3392 int ret = t->to_fileio_pread (fd, read_buf, len, offset,
3393 target_errno);
3394
3395 if (targetdebug)
3396 fprintf_unfiltered (gdb_stdlog,
3397 "target_fileio_pread (%d,...,%d,%s) "
3398 "= %d (%d)\n",
3399 fd, len, pulongest (offset),
3400 ret, ret != -1 ? 0 : *target_errno);
3401 return ret;
3402 }
3403 }
3404
3405 *target_errno = FILEIO_ENOSYS;
3406 return -1;
3407 }
3408
3409 /* Close FD on the target. Return 0, or -1 if an error occurs
3410 (and set *TARGET_ERRNO). */
3411 int
3412 target_fileio_close (int fd, int *target_errno)
3413 {
3414 struct target_ops *t;
3415
3416 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3417 {
3418 if (t->to_fileio_close != NULL)
3419 {
3420 int ret = t->to_fileio_close (fd, target_errno);
3421
3422 if (targetdebug)
3423 fprintf_unfiltered (gdb_stdlog,
3424 "target_fileio_close (%d) = %d (%d)\n",
3425 fd, ret, ret != -1 ? 0 : *target_errno);
3426 return ret;
3427 }
3428 }
3429
3430 *target_errno = FILEIO_ENOSYS;
3431 return -1;
3432 }
3433
3434 /* Unlink FILENAME on the target. Return 0, or -1 if an error
3435 occurs (and set *TARGET_ERRNO). */
3436 int
3437 target_fileio_unlink (const char *filename, int *target_errno)
3438 {
3439 struct target_ops *t;
3440
3441 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3442 {
3443 if (t->to_fileio_unlink != NULL)
3444 {
3445 int ret = t->to_fileio_unlink (filename, target_errno);
3446
3447 if (targetdebug)
3448 fprintf_unfiltered (gdb_stdlog,
3449 "target_fileio_unlink (%s) = %d (%d)\n",
3450 filename, ret, ret != -1 ? 0 : *target_errno);
3451 return ret;
3452 }
3453 }
3454
3455 *target_errno = FILEIO_ENOSYS;
3456 return -1;
3457 }
3458
3459 /* Read value of symbolic link FILENAME on the target. Return a
3460 null-terminated string allocated via xmalloc, or NULL if an error
3461 occurs (and set *TARGET_ERRNO). */
3462 char *
3463 target_fileio_readlink (const char *filename, int *target_errno)
3464 {
3465 struct target_ops *t;
3466
3467 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3468 {
3469 if (t->to_fileio_readlink != NULL)
3470 {
3471 char *ret = t->to_fileio_readlink (filename, target_errno);
3472
3473 if (targetdebug)
3474 fprintf_unfiltered (gdb_stdlog,
3475 "target_fileio_readlink (%s) = %s (%d)\n",
3476 filename, ret? ret : "(nil)",
3477 ret? 0 : *target_errno);
3478 return ret;
3479 }
3480 }
3481
3482 *target_errno = FILEIO_ENOSYS;
3483 return NULL;
3484 }
3485
3486 static void
3487 target_fileio_close_cleanup (void *opaque)
3488 {
3489 int fd = *(int *) opaque;
3490 int target_errno;
3491
3492 target_fileio_close (fd, &target_errno);
3493 }
3494
3495 /* Read target file FILENAME. Store the result in *BUF_P and
3496 return the size of the transferred data. PADDING additional bytes are
3497 available in *BUF_P. This is a helper function for
3498 target_fileio_read_alloc; see the declaration of that function for more
3499 information. */
3500
3501 static LONGEST
3502 target_fileio_read_alloc_1 (const char *filename,
3503 gdb_byte **buf_p, int padding)
3504 {
3505 struct cleanup *close_cleanup;
3506 size_t buf_alloc, buf_pos;
3507 gdb_byte *buf;
3508 LONGEST n;
3509 int fd;
3510 int target_errno;
3511
3512 fd = target_fileio_open (filename, FILEIO_O_RDONLY, 0700, &target_errno);
3513 if (fd == -1)
3514 return -1;
3515
3516 close_cleanup = make_cleanup (target_fileio_close_cleanup, &fd);
3517
3518 /* Start by reading up to 4K at a time. The target will throttle
3519 this number down if necessary. */
3520 buf_alloc = 4096;
3521 buf = xmalloc (buf_alloc);
3522 buf_pos = 0;
3523 while (1)
3524 {
3525 n = target_fileio_pread (fd, &buf[buf_pos],
3526 buf_alloc - buf_pos - padding, buf_pos,
3527 &target_errno);
3528 if (n < 0)
3529 {
3530 /* An error occurred. */
3531 do_cleanups (close_cleanup);
3532 xfree (buf);
3533 return -1;
3534 }
3535 else if (n == 0)
3536 {
3537 /* Read all there was. */
3538 do_cleanups (close_cleanup);
3539 if (buf_pos == 0)
3540 xfree (buf);
3541 else
3542 *buf_p = buf;
3543 return buf_pos;
3544 }
3545
3546 buf_pos += n;
3547
3548 /* If the buffer is filling up, expand it. */
3549 if (buf_alloc < buf_pos * 2)
3550 {
3551 buf_alloc *= 2;
3552 buf = xrealloc (buf, buf_alloc);
3553 }
3554
3555 QUIT;
3556 }
3557 }
3558
3559 /* Read target file FILENAME. Store the result in *BUF_P and return
3560 the size of the transferred data. See the declaration in "target.h"
3561 function for more information about the return value. */
3562
3563 LONGEST
3564 target_fileio_read_alloc (const char *filename, gdb_byte **buf_p)
3565 {
3566 return target_fileio_read_alloc_1 (filename, buf_p, 0);
3567 }
3568
3569 /* Read target file FILENAME. The result is NUL-terminated and
3570 returned as a string, allocated using xmalloc. If an error occurs
3571 or the transfer is unsupported, NULL is returned. Empty objects
3572 are returned as allocated but empty strings. A warning is issued
3573 if the result contains any embedded NUL bytes. */
3574
3575 char *
3576 target_fileio_read_stralloc (const char *filename)
3577 {
3578 gdb_byte *buffer;
3579 char *bufstr;
3580 LONGEST i, transferred;
3581
3582 transferred = target_fileio_read_alloc_1 (filename, &buffer, 1);
3583 bufstr = (char *) buffer;
3584
3585 if (transferred < 0)
3586 return NULL;
3587
3588 if (transferred == 0)
3589 return xstrdup ("");
3590
3591 bufstr[transferred] = 0;
3592
3593 /* Check for embedded NUL bytes; but allow trailing NULs. */
3594 for (i = strlen (bufstr); i < transferred; i++)
3595 if (bufstr[i] != 0)
3596 {
3597 warning (_("target file %s "
3598 "contained unexpected null characters"),
3599 filename);
3600 break;
3601 }
3602
3603 return bufstr;
3604 }
3605
3606
3607 static int
3608 default_region_ok_for_hw_watchpoint (CORE_ADDR addr, int len)
3609 {
3610 return (len <= gdbarch_ptr_bit (target_gdbarch ()) / TARGET_CHAR_BIT);
3611 }
3612
3613 static int
3614 default_watchpoint_addr_within_range (struct target_ops *target,
3615 CORE_ADDR addr,
3616 CORE_ADDR start, int length)
3617 {
3618 return addr >= start && addr < start + length;
3619 }
3620
3621 static struct gdbarch *
3622 default_thread_architecture (struct target_ops *ops, ptid_t ptid)
3623 {
3624 return target_gdbarch ();
3625 }
3626
3627 static int
3628 return_zero (void)
3629 {
3630 return 0;
3631 }
3632
3633 static int
3634 return_one (void)
3635 {
3636 return 1;
3637 }
3638
3639 static int
3640 return_minus_one (void)
3641 {
3642 return -1;
3643 }
3644
3645 /*
3646 * Find the next target down the stack from the specified target.
3647 */
3648
3649 struct target_ops *
3650 find_target_beneath (struct target_ops *t)
3651 {
3652 return t->beneath;
3653 }
3654
3655 \f
3656 /* The inferior process has died. Long live the inferior! */
3657
3658 void
3659 generic_mourn_inferior (void)
3660 {
3661 ptid_t ptid;
3662
3663 ptid = inferior_ptid;
3664 inferior_ptid = null_ptid;
3665
3666 /* Mark breakpoints uninserted in case something tries to delete a
3667 breakpoint while we delete the inferior's threads (which would
3668 fail, since the inferior is long gone). */
3669 mark_breakpoints_out ();
3670
3671 if (!ptid_equal (ptid, null_ptid))
3672 {
3673 int pid = ptid_get_pid (ptid);
3674 exit_inferior (pid);
3675 }
3676
3677 /* Note this wipes step-resume breakpoints, so needs to be done
3678 after exit_inferior, which ends up referencing the step-resume
3679 breakpoints through clear_thread_inferior_resources. */
3680 breakpoint_init_inferior (inf_exited);
3681
3682 registers_changed ();
3683
3684 reopen_exec_file ();
3685 reinit_frame_cache ();
3686
3687 if (deprecated_detach_hook)
3688 deprecated_detach_hook ();
3689 }
3690 \f
3691 /* Convert a normal process ID to a string. Returns the string in a
3692 static buffer. */
3693
3694 char *
3695 normal_pid_to_str (ptid_t ptid)
3696 {
3697 static char buf[32];
3698
3699 xsnprintf (buf, sizeof buf, "process %d", ptid_get_pid (ptid));
3700 return buf;
3701 }
3702
3703 static char *
3704 dummy_pid_to_str (struct target_ops *ops, ptid_t ptid)
3705 {
3706 return normal_pid_to_str (ptid);
3707 }
3708
3709 /* Error-catcher for target_find_memory_regions. */
3710 static int
3711 dummy_find_memory_regions (find_memory_region_ftype ignore1, void *ignore2)
3712 {
3713 error (_("Command not implemented for this target."));
3714 return 0;
3715 }
3716
3717 /* Error-catcher for target_make_corefile_notes. */
3718 static char *
3719 dummy_make_corefile_notes (bfd *ignore1, int *ignore2)
3720 {
3721 error (_("Command not implemented for this target."));
3722 return NULL;
3723 }
3724
3725 /* Error-catcher for target_get_bookmark. */
3726 static gdb_byte *
3727 dummy_get_bookmark (char *ignore1, int ignore2)
3728 {
3729 tcomplain ();
3730 return NULL;
3731 }
3732
3733 /* Error-catcher for target_goto_bookmark. */
3734 static void
3735 dummy_goto_bookmark (gdb_byte *ignore, int from_tty)
3736 {
3737 tcomplain ();
3738 }
3739
3740 /* Set up the handful of non-empty slots needed by the dummy target
3741 vector. */
3742
3743 static void
3744 init_dummy_target (void)
3745 {
3746 dummy_target.to_shortname = "None";
3747 dummy_target.to_longname = "None";
3748 dummy_target.to_doc = "";
3749 dummy_target.to_attach = find_default_attach;
3750 dummy_target.to_detach =
3751 (void (*)(struct target_ops *, char *, int))target_ignore;
3752 dummy_target.to_create_inferior = find_default_create_inferior;
3753 dummy_target.to_can_async_p = find_default_can_async_p;
3754 dummy_target.to_is_async_p = find_default_is_async_p;
3755 dummy_target.to_supports_non_stop = find_default_supports_non_stop;
3756 dummy_target.to_supports_disable_randomization
3757 = find_default_supports_disable_randomization;
3758 dummy_target.to_pid_to_str = dummy_pid_to_str;
3759 dummy_target.to_stratum = dummy_stratum;
3760 dummy_target.to_find_memory_regions = dummy_find_memory_regions;
3761 dummy_target.to_make_corefile_notes = dummy_make_corefile_notes;
3762 dummy_target.to_get_bookmark = dummy_get_bookmark;
3763 dummy_target.to_goto_bookmark = dummy_goto_bookmark;
3764 dummy_target.to_xfer_partial = default_xfer_partial;
3765 dummy_target.to_has_all_memory = (int (*) (struct target_ops *)) return_zero;
3766 dummy_target.to_has_memory = (int (*) (struct target_ops *)) return_zero;
3767 dummy_target.to_has_stack = (int (*) (struct target_ops *)) return_zero;
3768 dummy_target.to_has_registers = (int (*) (struct target_ops *)) return_zero;
3769 dummy_target.to_has_execution
3770 = (int (*) (struct target_ops *, ptid_t)) return_zero;
3771 dummy_target.to_stopped_by_watchpoint = return_zero;
3772 dummy_target.to_stopped_data_address =
3773 (int (*) (struct target_ops *, CORE_ADDR *)) return_zero;
3774 dummy_target.to_magic = OPS_MAGIC;
3775 }
3776 \f
3777 static void
3778 debug_to_open (char *args, int from_tty)
3779 {
3780 debug_target.to_open (args, from_tty);
3781
3782 fprintf_unfiltered (gdb_stdlog, "target_open (%s, %d)\n", args, from_tty);
3783 }
3784
3785 void
3786 target_close (struct target_ops *targ)
3787 {
3788 gdb_assert (!target_is_pushed (targ));
3789
3790 if (targ->to_xclose != NULL)
3791 targ->to_xclose (targ);
3792 else if (targ->to_close != NULL)
3793 targ->to_close ();
3794
3795 if (targetdebug)
3796 fprintf_unfiltered (gdb_stdlog, "target_close ()\n");
3797 }
3798
3799 void
3800 target_attach (char *args, int from_tty)
3801 {
3802 struct target_ops *t;
3803
3804 for (t = current_target.beneath; t != NULL; t = t->beneath)
3805 {
3806 if (t->to_attach != NULL)
3807 {
3808 t->to_attach (t, args, from_tty);
3809 if (targetdebug)
3810 fprintf_unfiltered (gdb_stdlog, "target_attach (%s, %d)\n",
3811 args, from_tty);
3812 return;
3813 }
3814 }
3815
3816 internal_error (__FILE__, __LINE__,
3817 _("could not find a target to attach"));
3818 }
3819
3820 int
3821 target_thread_alive (ptid_t ptid)
3822 {
3823 struct target_ops *t;
3824
3825 for (t = current_target.beneath; t != NULL; t = t->beneath)
3826 {
3827 if (t->to_thread_alive != NULL)
3828 {
3829 int retval;
3830
3831 retval = t->to_thread_alive (t, ptid);
3832 if (targetdebug)
3833 fprintf_unfiltered (gdb_stdlog, "target_thread_alive (%d) = %d\n",
3834 ptid_get_pid (ptid), retval);
3835
3836 return retval;
3837 }
3838 }
3839
3840 return 0;
3841 }
3842
3843 void
3844 target_find_new_threads (void)
3845 {
3846 struct target_ops *t;
3847
3848 for (t = current_target.beneath; t != NULL; t = t->beneath)
3849 {
3850 if (t->to_find_new_threads != NULL)
3851 {
3852 t->to_find_new_threads (t);
3853 if (targetdebug)
3854 fprintf_unfiltered (gdb_stdlog, "target_find_new_threads ()\n");
3855
3856 return;
3857 }
3858 }
3859 }
3860
3861 void
3862 target_stop (ptid_t ptid)
3863 {
3864 if (!may_stop)
3865 {
3866 warning (_("May not interrupt or stop the target, ignoring attempt"));
3867 return;
3868 }
3869
3870 (*current_target.to_stop) (ptid);
3871 }
3872
3873 static void
3874 debug_to_post_attach (int pid)
3875 {
3876 debug_target.to_post_attach (pid);
3877
3878 fprintf_unfiltered (gdb_stdlog, "target_post_attach (%d)\n", pid);
3879 }
3880
3881 /* Concatenate ELEM to LIST, a comma separate list, and return the
3882 result. The LIST incoming argument is released. */
3883
3884 static char *
3885 str_comma_list_concat_elem (char *list, const char *elem)
3886 {
3887 if (list == NULL)
3888 return xstrdup (elem);
3889 else
3890 return reconcat (list, list, ", ", elem, (char *) NULL);
3891 }
3892
3893 /* Helper for target_options_to_string. If OPT is present in
3894 TARGET_OPTIONS, append the OPT_STR (string version of OPT) in RET.
3895 Returns the new resulting string. OPT is removed from
3896 TARGET_OPTIONS. */
3897
3898 static char *
3899 do_option (int *target_options, char *ret,
3900 int opt, char *opt_str)
3901 {
3902 if ((*target_options & opt) != 0)
3903 {
3904 ret = str_comma_list_concat_elem (ret, opt_str);
3905 *target_options &= ~opt;
3906 }
3907
3908 return ret;
3909 }
3910
3911 char *
3912 target_options_to_string (int target_options)
3913 {
3914 char *ret = NULL;
3915
3916 #define DO_TARG_OPTION(OPT) \
3917 ret = do_option (&target_options, ret, OPT, #OPT)
3918
3919 DO_TARG_OPTION (TARGET_WNOHANG);
3920
3921 if (target_options != 0)
3922 ret = str_comma_list_concat_elem (ret, "unknown???");
3923
3924 if (ret == NULL)
3925 ret = xstrdup ("");
3926 return ret;
3927 }
3928
3929 static void
3930 debug_print_register (const char * func,
3931 struct regcache *regcache, int regno)
3932 {
3933 struct gdbarch *gdbarch = get_regcache_arch (regcache);
3934
3935 fprintf_unfiltered (gdb_stdlog, "%s ", func);
3936 if (regno >= 0 && regno < gdbarch_num_regs (gdbarch)
3937 && gdbarch_register_name (gdbarch, regno) != NULL
3938 && gdbarch_register_name (gdbarch, regno)[0] != '\0')
3939 fprintf_unfiltered (gdb_stdlog, "(%s)",
3940 gdbarch_register_name (gdbarch, regno));
3941 else
3942 fprintf_unfiltered (gdb_stdlog, "(%d)", regno);
3943 if (regno >= 0 && regno < gdbarch_num_regs (gdbarch))
3944 {
3945 enum bfd_endian byte_order = gdbarch_byte_order (gdbarch);
3946 int i, size = register_size (gdbarch, regno);
3947 gdb_byte buf[MAX_REGISTER_SIZE];
3948
3949 regcache_raw_collect (regcache, regno, buf);
3950 fprintf_unfiltered (gdb_stdlog, " = ");
3951 for (i = 0; i < size; i++)
3952 {
3953 fprintf_unfiltered (gdb_stdlog, "%02x", buf[i]);
3954 }
3955 if (size <= sizeof (LONGEST))
3956 {
3957 ULONGEST val = extract_unsigned_integer (buf, size, byte_order);
3958
3959 fprintf_unfiltered (gdb_stdlog, " %s %s",
3960 core_addr_to_string_nz (val), plongest (val));
3961 }
3962 }
3963 fprintf_unfiltered (gdb_stdlog, "\n");
3964 }
3965
3966 void
3967 target_fetch_registers (struct regcache *regcache, int regno)
3968 {
3969 struct target_ops *t;
3970
3971 for (t = current_target.beneath; t != NULL; t = t->beneath)
3972 {
3973 if (t->to_fetch_registers != NULL)
3974 {
3975 t->to_fetch_registers (t, regcache, regno);
3976 if (targetdebug)
3977 debug_print_register ("target_fetch_registers", regcache, regno);
3978 return;
3979 }
3980 }
3981 }
3982
3983 void
3984 target_store_registers (struct regcache *regcache, int regno)
3985 {
3986 struct target_ops *t;
3987
3988 if (!may_write_registers)
3989 error (_("Writing to registers is not allowed (regno %d)"), regno);
3990
3991 for (t = current_target.beneath; t != NULL; t = t->beneath)
3992 {
3993 if (t->to_store_registers != NULL)
3994 {
3995 t->to_store_registers (t, regcache, regno);
3996 if (targetdebug)
3997 {
3998 debug_print_register ("target_store_registers", regcache, regno);
3999 }
4000 return;
4001 }
4002 }
4003
4004 noprocess ();
4005 }
4006
4007 int
4008 target_core_of_thread (ptid_t ptid)
4009 {
4010 struct target_ops *t;
4011
4012 for (t = current_target.beneath; t != NULL; t = t->beneath)
4013 {
4014 if (t->to_core_of_thread != NULL)
4015 {
4016 int retval = t->to_core_of_thread (t, ptid);
4017
4018 if (targetdebug)
4019 fprintf_unfiltered (gdb_stdlog,
4020 "target_core_of_thread (%d) = %d\n",
4021 ptid_get_pid (ptid), retval);
4022 return retval;
4023 }
4024 }
4025
4026 return -1;
4027 }
4028
4029 int
4030 target_verify_memory (const gdb_byte *data, CORE_ADDR memaddr, ULONGEST size)
4031 {
4032 struct target_ops *t;
4033
4034 for (t = current_target.beneath; t != NULL; t = t->beneath)
4035 {
4036 if (t->to_verify_memory != NULL)
4037 {
4038 int retval = t->to_verify_memory (t, data, memaddr, size);
4039
4040 if (targetdebug)
4041 fprintf_unfiltered (gdb_stdlog,
4042 "target_verify_memory (%s, %s) = %d\n",
4043 paddress (target_gdbarch (), memaddr),
4044 pulongest (size),
4045 retval);
4046 return retval;
4047 }
4048 }
4049
4050 tcomplain ();
4051 }
4052
4053 /* The documentation for this function is in its prototype declaration in
4054 target.h. */
4055
4056 int
4057 target_insert_mask_watchpoint (CORE_ADDR addr, CORE_ADDR mask, int rw)
4058 {
4059 struct target_ops *t;
4060
4061 for (t = current_target.beneath; t != NULL; t = t->beneath)
4062 if (t->to_insert_mask_watchpoint != NULL)
4063 {
4064 int ret;
4065
4066 ret = t->to_insert_mask_watchpoint (t, addr, mask, rw);
4067
4068 if (targetdebug)
4069 fprintf_unfiltered (gdb_stdlog, "\
4070 target_insert_mask_watchpoint (%s, %s, %d) = %d\n",
4071 core_addr_to_string (addr),
4072 core_addr_to_string (mask), rw, ret);
4073
4074 return ret;
4075 }
4076
4077 return 1;
4078 }
4079
4080 /* The documentation for this function is in its prototype declaration in
4081 target.h. */
4082
4083 int
4084 target_remove_mask_watchpoint (CORE_ADDR addr, CORE_ADDR mask, int rw)
4085 {
4086 struct target_ops *t;
4087
4088 for (t = current_target.beneath; t != NULL; t = t->beneath)
4089 if (t->to_remove_mask_watchpoint != NULL)
4090 {
4091 int ret;
4092
4093 ret = t->to_remove_mask_watchpoint (t, addr, mask, rw);
4094
4095 if (targetdebug)
4096 fprintf_unfiltered (gdb_stdlog, "\
4097 target_remove_mask_watchpoint (%s, %s, %d) = %d\n",
4098 core_addr_to_string (addr),
4099 core_addr_to_string (mask), rw, ret);
4100
4101 return ret;
4102 }
4103
4104 return 1;
4105 }
4106
4107 /* The documentation for this function is in its prototype declaration
4108 in target.h. */
4109
4110 int
4111 target_masked_watch_num_registers (CORE_ADDR addr, CORE_ADDR mask)
4112 {
4113 struct target_ops *t;
4114
4115 for (t = current_target.beneath; t != NULL; t = t->beneath)
4116 if (t->to_masked_watch_num_registers != NULL)
4117 return t->to_masked_watch_num_registers (t, addr, mask);
4118
4119 return -1;
4120 }
4121
4122 /* The documentation for this function is in its prototype declaration
4123 in target.h. */
4124
4125 int
4126 target_ranged_break_num_registers (void)
4127 {
4128 struct target_ops *t;
4129
4130 for (t = current_target.beneath; t != NULL; t = t->beneath)
4131 if (t->to_ranged_break_num_registers != NULL)
4132 return t->to_ranged_break_num_registers (t);
4133
4134 return -1;
4135 }
4136
4137 /* See target.h. */
4138
4139 int
4140 target_supports_btrace (void)
4141 {
4142 struct target_ops *t;
4143
4144 for (t = current_target.beneath; t != NULL; t = t->beneath)
4145 if (t->to_supports_btrace != NULL)
4146 return t->to_supports_btrace ();
4147
4148 return 0;
4149 }
4150
4151 /* See target.h. */
4152
4153 struct btrace_target_info *
4154 target_enable_btrace (ptid_t ptid)
4155 {
4156 struct target_ops *t;
4157
4158 for (t = current_target.beneath; t != NULL; t = t->beneath)
4159 if (t->to_enable_btrace != NULL)
4160 return t->to_enable_btrace (ptid);
4161
4162 tcomplain ();
4163 return NULL;
4164 }
4165
4166 /* See target.h. */
4167
4168 void
4169 target_disable_btrace (struct btrace_target_info *btinfo)
4170 {
4171 struct target_ops *t;
4172
4173 for (t = current_target.beneath; t != NULL; t = t->beneath)
4174 if (t->to_disable_btrace != NULL)
4175 {
4176 t->to_disable_btrace (btinfo);
4177 return;
4178 }
4179
4180 tcomplain ();
4181 }
4182
4183 /* See target.h. */
4184
4185 void
4186 target_teardown_btrace (struct btrace_target_info *btinfo)
4187 {
4188 struct target_ops *t;
4189
4190 for (t = current_target.beneath; t != NULL; t = t->beneath)
4191 if (t->to_teardown_btrace != NULL)
4192 {
4193 t->to_teardown_btrace (btinfo);
4194 return;
4195 }
4196
4197 tcomplain ();
4198 }
4199
4200 /* See target.h. */
4201
4202 VEC (btrace_block_s) *
4203 target_read_btrace (struct btrace_target_info *btinfo,
4204 enum btrace_read_type type)
4205 {
4206 struct target_ops *t;
4207
4208 for (t = current_target.beneath; t != NULL; t = t->beneath)
4209 if (t->to_read_btrace != NULL)
4210 return t->to_read_btrace (btinfo, type);
4211
4212 tcomplain ();
4213 return NULL;
4214 }
4215
4216 /* See target.h. */
4217
4218 void
4219 target_stop_recording (void)
4220 {
4221 struct target_ops *t;
4222
4223 for (t = current_target.beneath; t != NULL; t = t->beneath)
4224 if (t->to_stop_recording != NULL)
4225 {
4226 t->to_stop_recording ();
4227 return;
4228 }
4229
4230 /* This is optional. */
4231 }
4232
4233 /* See target.h. */
4234
4235 void
4236 target_info_record (void)
4237 {
4238 struct target_ops *t;
4239
4240 for (t = current_target.beneath; t != NULL; t = t->beneath)
4241 if (t->to_info_record != NULL)
4242 {
4243 t->to_info_record ();
4244 return;
4245 }
4246
4247 tcomplain ();
4248 }
4249
4250 /* See target.h. */
4251
4252 void
4253 target_save_record (const char *filename)
4254 {
4255 struct target_ops *t;
4256
4257 for (t = current_target.beneath; t != NULL; t = t->beneath)
4258 if (t->to_save_record != NULL)
4259 {
4260 t->to_save_record (filename);
4261 return;
4262 }
4263
4264 tcomplain ();
4265 }
4266
4267 /* See target.h. */
4268
4269 int
4270 target_supports_delete_record (void)
4271 {
4272 struct target_ops *t;
4273
4274 for (t = current_target.beneath; t != NULL; t = t->beneath)
4275 if (t->to_delete_record != NULL)
4276 return 1;
4277
4278 return 0;
4279 }
4280
4281 /* See target.h. */
4282
4283 void
4284 target_delete_record (void)
4285 {
4286 struct target_ops *t;
4287
4288 for (t = current_target.beneath; t != NULL; t = t->beneath)
4289 if (t->to_delete_record != NULL)
4290 {
4291 t->to_delete_record ();
4292 return;
4293 }
4294
4295 tcomplain ();
4296 }
4297
4298 /* See target.h. */
4299
4300 int
4301 target_record_is_replaying (void)
4302 {
4303 struct target_ops *t;
4304
4305 for (t = current_target.beneath; t != NULL; t = t->beneath)
4306 if (t->to_record_is_replaying != NULL)
4307 return t->to_record_is_replaying ();
4308
4309 return 0;
4310 }
4311
4312 /* See target.h. */
4313
4314 void
4315 target_goto_record_begin (void)
4316 {
4317 struct target_ops *t;
4318
4319 for (t = current_target.beneath; t != NULL; t = t->beneath)
4320 if (t->to_goto_record_begin != NULL)
4321 {
4322 t->to_goto_record_begin ();
4323 return;
4324 }
4325
4326 tcomplain ();
4327 }
4328
4329 /* See target.h. */
4330
4331 void
4332 target_goto_record_end (void)
4333 {
4334 struct target_ops *t;
4335
4336 for (t = current_target.beneath; t != NULL; t = t->beneath)
4337 if (t->to_goto_record_end != NULL)
4338 {
4339 t->to_goto_record_end ();
4340 return;
4341 }
4342
4343 tcomplain ();
4344 }
4345
4346 /* See target.h. */
4347
4348 void
4349 target_goto_record (ULONGEST insn)
4350 {
4351 struct target_ops *t;
4352
4353 for (t = current_target.beneath; t != NULL; t = t->beneath)
4354 if (t->to_goto_record != NULL)
4355 {
4356 t->to_goto_record (insn);
4357 return;
4358 }
4359
4360 tcomplain ();
4361 }
4362
4363 /* See target.h. */
4364
4365 void
4366 target_insn_history (int size, int flags)
4367 {
4368 struct target_ops *t;
4369
4370 for (t = current_target.beneath; t != NULL; t = t->beneath)
4371 if (t->to_insn_history != NULL)
4372 {
4373 t->to_insn_history (size, flags);
4374 return;
4375 }
4376
4377 tcomplain ();
4378 }
4379
4380 /* See target.h. */
4381
4382 void
4383 target_insn_history_from (ULONGEST from, int size, int flags)
4384 {
4385 struct target_ops *t;
4386
4387 for (t = current_target.beneath; t != NULL; t = t->beneath)
4388 if (t->to_insn_history_from != NULL)
4389 {
4390 t->to_insn_history_from (from, size, flags);
4391 return;
4392 }
4393
4394 tcomplain ();
4395 }
4396
4397 /* See target.h. */
4398
4399 void
4400 target_insn_history_range (ULONGEST begin, ULONGEST end, int flags)
4401 {
4402 struct target_ops *t;
4403
4404 for (t = current_target.beneath; t != NULL; t = t->beneath)
4405 if (t->to_insn_history_range != NULL)
4406 {
4407 t->to_insn_history_range (begin, end, flags);
4408 return;
4409 }
4410
4411 tcomplain ();
4412 }
4413
4414 /* See target.h. */
4415
4416 void
4417 target_call_history (int size, int flags)
4418 {
4419 struct target_ops *t;
4420
4421 for (t = current_target.beneath; t != NULL; t = t->beneath)
4422 if (t->to_call_history != NULL)
4423 {
4424 t->to_call_history (size, flags);
4425 return;
4426 }
4427
4428 tcomplain ();
4429 }
4430
4431 /* See target.h. */
4432
4433 void
4434 target_call_history_from (ULONGEST begin, int size, int flags)
4435 {
4436 struct target_ops *t;
4437
4438 for (t = current_target.beneath; t != NULL; t = t->beneath)
4439 if (t->to_call_history_from != NULL)
4440 {
4441 t->to_call_history_from (begin, size, flags);
4442 return;
4443 }
4444
4445 tcomplain ();
4446 }
4447
4448 /* See target.h. */
4449
4450 void
4451 target_call_history_range (ULONGEST begin, ULONGEST end, int flags)
4452 {
4453 struct target_ops *t;
4454
4455 for (t = current_target.beneath; t != NULL; t = t->beneath)
4456 if (t->to_call_history_range != NULL)
4457 {
4458 t->to_call_history_range (begin, end, flags);
4459 return;
4460 }
4461
4462 tcomplain ();
4463 }
4464
4465 static void
4466 debug_to_prepare_to_store (struct regcache *regcache)
4467 {
4468 debug_target.to_prepare_to_store (regcache);
4469
4470 fprintf_unfiltered (gdb_stdlog, "target_prepare_to_store ()\n");
4471 }
4472
4473 static int
4474 deprecated_debug_xfer_memory (CORE_ADDR memaddr, bfd_byte *myaddr, int len,
4475 int write, struct mem_attrib *attrib,
4476 struct target_ops *target)
4477 {
4478 int retval;
4479
4480 retval = debug_target.deprecated_xfer_memory (memaddr, myaddr, len, write,
4481 attrib, target);
4482
4483 fprintf_unfiltered (gdb_stdlog,
4484 "target_xfer_memory (%s, xxx, %d, %s, xxx) = %d",
4485 paddress (target_gdbarch (), memaddr), len,
4486 write ? "write" : "read", retval);
4487
4488 if (retval > 0)
4489 {
4490 int i;
4491
4492 fputs_unfiltered (", bytes =", gdb_stdlog);
4493 for (i = 0; i < retval; i++)
4494 {
4495 if ((((intptr_t) &(myaddr[i])) & 0xf) == 0)
4496 {
4497 if (targetdebug < 2 && i > 0)
4498 {
4499 fprintf_unfiltered (gdb_stdlog, " ...");
4500 break;
4501 }
4502 fprintf_unfiltered (gdb_stdlog, "\n");
4503 }
4504
4505 fprintf_unfiltered (gdb_stdlog, " %02x", myaddr[i] & 0xff);
4506 }
4507 }
4508
4509 fputc_unfiltered ('\n', gdb_stdlog);
4510
4511 return retval;
4512 }
4513
4514 static void
4515 debug_to_files_info (struct target_ops *target)
4516 {
4517 debug_target.to_files_info (target);
4518
4519 fprintf_unfiltered (gdb_stdlog, "target_files_info (xxx)\n");
4520 }
4521
4522 static int
4523 debug_to_insert_breakpoint (struct gdbarch *gdbarch,
4524 struct bp_target_info *bp_tgt)
4525 {
4526 int retval;
4527
4528 retval = debug_target.to_insert_breakpoint (gdbarch, bp_tgt);
4529
4530 fprintf_unfiltered (gdb_stdlog,
4531 "target_insert_breakpoint (%s, xxx) = %ld\n",
4532 core_addr_to_string (bp_tgt->placed_address),
4533 (unsigned long) retval);
4534 return retval;
4535 }
4536
4537 static int
4538 debug_to_remove_breakpoint (struct gdbarch *gdbarch,
4539 struct bp_target_info *bp_tgt)
4540 {
4541 int retval;
4542
4543 retval = debug_target.to_remove_breakpoint (gdbarch, bp_tgt);
4544
4545 fprintf_unfiltered (gdb_stdlog,
4546 "target_remove_breakpoint (%s, xxx) = %ld\n",
4547 core_addr_to_string (bp_tgt->placed_address),
4548 (unsigned long) retval);
4549 return retval;
4550 }
4551
4552 static int
4553 debug_to_can_use_hw_breakpoint (int type, int cnt, int from_tty)
4554 {
4555 int retval;
4556
4557 retval = debug_target.to_can_use_hw_breakpoint (type, cnt, from_tty);
4558
4559 fprintf_unfiltered (gdb_stdlog,
4560 "target_can_use_hw_breakpoint (%ld, %ld, %ld) = %ld\n",
4561 (unsigned long) type,
4562 (unsigned long) cnt,
4563 (unsigned long) from_tty,
4564 (unsigned long) retval);
4565 return retval;
4566 }
4567
4568 static int
4569 debug_to_region_ok_for_hw_watchpoint (CORE_ADDR addr, int len)
4570 {
4571 CORE_ADDR retval;
4572
4573 retval = debug_target.to_region_ok_for_hw_watchpoint (addr, len);
4574
4575 fprintf_unfiltered (gdb_stdlog,
4576 "target_region_ok_for_hw_watchpoint (%s, %ld) = %s\n",
4577 core_addr_to_string (addr), (unsigned long) len,
4578 core_addr_to_string (retval));
4579 return retval;
4580 }
4581
4582 static int
4583 debug_to_can_accel_watchpoint_condition (CORE_ADDR addr, int len, int rw,
4584 struct expression *cond)
4585 {
4586 int retval;
4587
4588 retval = debug_target.to_can_accel_watchpoint_condition (addr, len,
4589 rw, cond);
4590
4591 fprintf_unfiltered (gdb_stdlog,
4592 "target_can_accel_watchpoint_condition "
4593 "(%s, %d, %d, %s) = %ld\n",
4594 core_addr_to_string (addr), len, rw,
4595 host_address_to_string (cond), (unsigned long) retval);
4596 return retval;
4597 }
4598
4599 static int
4600 debug_to_stopped_by_watchpoint (void)
4601 {
4602 int retval;
4603
4604 retval = debug_target.to_stopped_by_watchpoint ();
4605
4606 fprintf_unfiltered (gdb_stdlog,
4607 "target_stopped_by_watchpoint () = %ld\n",
4608 (unsigned long) retval);
4609 return retval;
4610 }
4611
4612 static int
4613 debug_to_stopped_data_address (struct target_ops *target, CORE_ADDR *addr)
4614 {
4615 int retval;
4616
4617 retval = debug_target.to_stopped_data_address (target, addr);
4618
4619 fprintf_unfiltered (gdb_stdlog,
4620 "target_stopped_data_address ([%s]) = %ld\n",
4621 core_addr_to_string (*addr),
4622 (unsigned long)retval);
4623 return retval;
4624 }
4625
4626 static int
4627 debug_to_watchpoint_addr_within_range (struct target_ops *target,
4628 CORE_ADDR addr,
4629 CORE_ADDR start, int length)
4630 {
4631 int retval;
4632
4633 retval = debug_target.to_watchpoint_addr_within_range (target, addr,
4634 start, length);
4635
4636 fprintf_filtered (gdb_stdlog,
4637 "target_watchpoint_addr_within_range (%s, %s, %d) = %d\n",
4638 core_addr_to_string (addr), core_addr_to_string (start),
4639 length, retval);
4640 return retval;
4641 }
4642
4643 static int
4644 debug_to_insert_hw_breakpoint (struct gdbarch *gdbarch,
4645 struct bp_target_info *bp_tgt)
4646 {
4647 int retval;
4648
4649 retval = debug_target.to_insert_hw_breakpoint (gdbarch, bp_tgt);
4650
4651 fprintf_unfiltered (gdb_stdlog,
4652 "target_insert_hw_breakpoint (%s, xxx) = %ld\n",
4653 core_addr_to_string (bp_tgt->placed_address),
4654 (unsigned long) retval);
4655 return retval;
4656 }
4657
4658 static int
4659 debug_to_remove_hw_breakpoint (struct gdbarch *gdbarch,
4660 struct bp_target_info *bp_tgt)
4661 {
4662 int retval;
4663
4664 retval = debug_target.to_remove_hw_breakpoint (gdbarch, bp_tgt);
4665
4666 fprintf_unfiltered (gdb_stdlog,
4667 "target_remove_hw_breakpoint (%s, xxx) = %ld\n",
4668 core_addr_to_string (bp_tgt->placed_address),
4669 (unsigned long) retval);
4670 return retval;
4671 }
4672
4673 static int
4674 debug_to_insert_watchpoint (CORE_ADDR addr, int len, int type,
4675 struct expression *cond)
4676 {
4677 int retval;
4678
4679 retval = debug_target.to_insert_watchpoint (addr, len, type, cond);
4680
4681 fprintf_unfiltered (gdb_stdlog,
4682 "target_insert_watchpoint (%s, %d, %d, %s) = %ld\n",
4683 core_addr_to_string (addr), len, type,
4684 host_address_to_string (cond), (unsigned long) retval);
4685 return retval;
4686 }
4687
4688 static int
4689 debug_to_remove_watchpoint (CORE_ADDR addr, int len, int type,
4690 struct expression *cond)
4691 {
4692 int retval;
4693
4694 retval = debug_target.to_remove_watchpoint (addr, len, type, cond);
4695
4696 fprintf_unfiltered (gdb_stdlog,
4697 "target_remove_watchpoint (%s, %d, %d, %s) = %ld\n",
4698 core_addr_to_string (addr), len, type,
4699 host_address_to_string (cond), (unsigned long) retval);
4700 return retval;
4701 }
4702
4703 static void
4704 debug_to_terminal_init (void)
4705 {
4706 debug_target.to_terminal_init ();
4707
4708 fprintf_unfiltered (gdb_stdlog, "target_terminal_init ()\n");
4709 }
4710
4711 static void
4712 debug_to_terminal_inferior (void)
4713 {
4714 debug_target.to_terminal_inferior ();
4715
4716 fprintf_unfiltered (gdb_stdlog, "target_terminal_inferior ()\n");
4717 }
4718
4719 static void
4720 debug_to_terminal_ours_for_output (void)
4721 {
4722 debug_target.to_terminal_ours_for_output ();
4723
4724 fprintf_unfiltered (gdb_stdlog, "target_terminal_ours_for_output ()\n");
4725 }
4726
4727 static void
4728 debug_to_terminal_ours (void)
4729 {
4730 debug_target.to_terminal_ours ();
4731
4732 fprintf_unfiltered (gdb_stdlog, "target_terminal_ours ()\n");
4733 }
4734
4735 static void
4736 debug_to_terminal_save_ours (void)
4737 {
4738 debug_target.to_terminal_save_ours ();
4739
4740 fprintf_unfiltered (gdb_stdlog, "target_terminal_save_ours ()\n");
4741 }
4742
4743 static void
4744 debug_to_terminal_info (const char *arg, int from_tty)
4745 {
4746 debug_target.to_terminal_info (arg, from_tty);
4747
4748 fprintf_unfiltered (gdb_stdlog, "target_terminal_info (%s, %d)\n", arg,
4749 from_tty);
4750 }
4751
4752 static void
4753 debug_to_load (char *args, int from_tty)
4754 {
4755 debug_target.to_load (args, from_tty);
4756
4757 fprintf_unfiltered (gdb_stdlog, "target_load (%s, %d)\n", args, from_tty);
4758 }
4759
4760 static void
4761 debug_to_post_startup_inferior (ptid_t ptid)
4762 {
4763 debug_target.to_post_startup_inferior (ptid);
4764
4765 fprintf_unfiltered (gdb_stdlog, "target_post_startup_inferior (%d)\n",
4766 ptid_get_pid (ptid));
4767 }
4768
4769 static int
4770 debug_to_insert_fork_catchpoint (int pid)
4771 {
4772 int retval;
4773
4774 retval = debug_target.to_insert_fork_catchpoint (pid);
4775
4776 fprintf_unfiltered (gdb_stdlog, "target_insert_fork_catchpoint (%d) = %d\n",
4777 pid, retval);
4778
4779 return retval;
4780 }
4781
4782 static int
4783 debug_to_remove_fork_catchpoint (int pid)
4784 {
4785 int retval;
4786
4787 retval = debug_target.to_remove_fork_catchpoint (pid);
4788
4789 fprintf_unfiltered (gdb_stdlog, "target_remove_fork_catchpoint (%d) = %d\n",
4790 pid, retval);
4791
4792 return retval;
4793 }
4794
4795 static int
4796 debug_to_insert_vfork_catchpoint (int pid)
4797 {
4798 int retval;
4799
4800 retval = debug_target.to_insert_vfork_catchpoint (pid);
4801
4802 fprintf_unfiltered (gdb_stdlog, "target_insert_vfork_catchpoint (%d) = %d\n",
4803 pid, retval);
4804
4805 return retval;
4806 }
4807
4808 static int
4809 debug_to_remove_vfork_catchpoint (int pid)
4810 {
4811 int retval;
4812
4813 retval = debug_target.to_remove_vfork_catchpoint (pid);
4814
4815 fprintf_unfiltered (gdb_stdlog, "target_remove_vfork_catchpoint (%d) = %d\n",
4816 pid, retval);
4817
4818 return retval;
4819 }
4820
4821 static int
4822 debug_to_insert_exec_catchpoint (int pid)
4823 {
4824 int retval;
4825
4826 retval = debug_target.to_insert_exec_catchpoint (pid);
4827
4828 fprintf_unfiltered (gdb_stdlog, "target_insert_exec_catchpoint (%d) = %d\n",
4829 pid, retval);
4830
4831 return retval;
4832 }
4833
4834 static int
4835 debug_to_remove_exec_catchpoint (int pid)
4836 {
4837 int retval;
4838
4839 retval = debug_target.to_remove_exec_catchpoint (pid);
4840
4841 fprintf_unfiltered (gdb_stdlog, "target_remove_exec_catchpoint (%d) = %d\n",
4842 pid, retval);
4843
4844 return retval;
4845 }
4846
4847 static int
4848 debug_to_has_exited (int pid, int wait_status, int *exit_status)
4849 {
4850 int has_exited;
4851
4852 has_exited = debug_target.to_has_exited (pid, wait_status, exit_status);
4853
4854 fprintf_unfiltered (gdb_stdlog, "target_has_exited (%d, %d, %d) = %d\n",
4855 pid, wait_status, *exit_status, has_exited);
4856
4857 return has_exited;
4858 }
4859
4860 static int
4861 debug_to_can_run (void)
4862 {
4863 int retval;
4864
4865 retval = debug_target.to_can_run ();
4866
4867 fprintf_unfiltered (gdb_stdlog, "target_can_run () = %d\n", retval);
4868
4869 return retval;
4870 }
4871
4872 static struct gdbarch *
4873 debug_to_thread_architecture (struct target_ops *ops, ptid_t ptid)
4874 {
4875 struct gdbarch *retval;
4876
4877 retval = debug_target.to_thread_architecture (ops, ptid);
4878
4879 fprintf_unfiltered (gdb_stdlog,
4880 "target_thread_architecture (%s) = %s [%s]\n",
4881 target_pid_to_str (ptid),
4882 host_address_to_string (retval),
4883 gdbarch_bfd_arch_info (retval)->printable_name);
4884 return retval;
4885 }
4886
4887 static void
4888 debug_to_stop (ptid_t ptid)
4889 {
4890 debug_target.to_stop (ptid);
4891
4892 fprintf_unfiltered (gdb_stdlog, "target_stop (%s)\n",
4893 target_pid_to_str (ptid));
4894 }
4895
4896 static void
4897 debug_to_rcmd (char *command,
4898 struct ui_file *outbuf)
4899 {
4900 debug_target.to_rcmd (command, outbuf);
4901 fprintf_unfiltered (gdb_stdlog, "target_rcmd (%s, ...)\n", command);
4902 }
4903
4904 static char *
4905 debug_to_pid_to_exec_file (int pid)
4906 {
4907 char *exec_file;
4908
4909 exec_file = debug_target.to_pid_to_exec_file (pid);
4910
4911 fprintf_unfiltered (gdb_stdlog, "target_pid_to_exec_file (%d) = %s\n",
4912 pid, exec_file);
4913
4914 return exec_file;
4915 }
4916
4917 static void
4918 setup_target_debug (void)
4919 {
4920 memcpy (&debug_target, &current_target, sizeof debug_target);
4921
4922 current_target.to_open = debug_to_open;
4923 current_target.to_post_attach = debug_to_post_attach;
4924 current_target.to_prepare_to_store = debug_to_prepare_to_store;
4925 current_target.deprecated_xfer_memory = deprecated_debug_xfer_memory;
4926 current_target.to_files_info = debug_to_files_info;
4927 current_target.to_insert_breakpoint = debug_to_insert_breakpoint;
4928 current_target.to_remove_breakpoint = debug_to_remove_breakpoint;
4929 current_target.to_can_use_hw_breakpoint = debug_to_can_use_hw_breakpoint;
4930 current_target.to_insert_hw_breakpoint = debug_to_insert_hw_breakpoint;
4931 current_target.to_remove_hw_breakpoint = debug_to_remove_hw_breakpoint;
4932 current_target.to_insert_watchpoint = debug_to_insert_watchpoint;
4933 current_target.to_remove_watchpoint = debug_to_remove_watchpoint;
4934 current_target.to_stopped_by_watchpoint = debug_to_stopped_by_watchpoint;
4935 current_target.to_stopped_data_address = debug_to_stopped_data_address;
4936 current_target.to_watchpoint_addr_within_range
4937 = debug_to_watchpoint_addr_within_range;
4938 current_target.to_region_ok_for_hw_watchpoint
4939 = debug_to_region_ok_for_hw_watchpoint;
4940 current_target.to_can_accel_watchpoint_condition
4941 = debug_to_can_accel_watchpoint_condition;
4942 current_target.to_terminal_init = debug_to_terminal_init;
4943 current_target.to_terminal_inferior = debug_to_terminal_inferior;
4944 current_target.to_terminal_ours_for_output
4945 = debug_to_terminal_ours_for_output;
4946 current_target.to_terminal_ours = debug_to_terminal_ours;
4947 current_target.to_terminal_save_ours = debug_to_terminal_save_ours;
4948 current_target.to_terminal_info = debug_to_terminal_info;
4949 current_target.to_load = debug_to_load;
4950 current_target.to_post_startup_inferior = debug_to_post_startup_inferior;
4951 current_target.to_insert_fork_catchpoint = debug_to_insert_fork_catchpoint;
4952 current_target.to_remove_fork_catchpoint = debug_to_remove_fork_catchpoint;
4953 current_target.to_insert_vfork_catchpoint = debug_to_insert_vfork_catchpoint;
4954 current_target.to_remove_vfork_catchpoint = debug_to_remove_vfork_catchpoint;
4955 current_target.to_insert_exec_catchpoint = debug_to_insert_exec_catchpoint;
4956 current_target.to_remove_exec_catchpoint = debug_to_remove_exec_catchpoint;
4957 current_target.to_has_exited = debug_to_has_exited;
4958 current_target.to_can_run = debug_to_can_run;
4959 current_target.to_stop = debug_to_stop;
4960 current_target.to_rcmd = debug_to_rcmd;
4961 current_target.to_pid_to_exec_file = debug_to_pid_to_exec_file;
4962 current_target.to_thread_architecture = debug_to_thread_architecture;
4963 }
4964 \f
4965
4966 static char targ_desc[] =
4967 "Names of targets and files being debugged.\nShows the entire \
4968 stack of targets currently in use (including the exec-file,\n\
4969 core-file, and process, if any), as well as the symbol file name.";
4970
4971 static void
4972 do_monitor_command (char *cmd,
4973 int from_tty)
4974 {
4975 if ((current_target.to_rcmd
4976 == (void (*) (char *, struct ui_file *)) tcomplain)
4977 || (current_target.to_rcmd == debug_to_rcmd
4978 && (debug_target.to_rcmd
4979 == (void (*) (char *, struct ui_file *)) tcomplain)))
4980 error (_("\"monitor\" command not supported by this target."));
4981 target_rcmd (cmd, gdb_stdtarg);
4982 }
4983
4984 /* Print the name of each layers of our target stack. */
4985
4986 static void
4987 maintenance_print_target_stack (char *cmd, int from_tty)
4988 {
4989 struct target_ops *t;
4990
4991 printf_filtered (_("The current target stack is:\n"));
4992
4993 for (t = target_stack; t != NULL; t = t->beneath)
4994 {
4995 printf_filtered (" - %s (%s)\n", t->to_shortname, t->to_longname);
4996 }
4997 }
4998
4999 /* Controls if async mode is permitted. */
5000 int target_async_permitted = 0;
5001
5002 /* The set command writes to this variable. If the inferior is
5003 executing, target_async_permitted is *not* updated. */
5004 static int target_async_permitted_1 = 0;
5005
5006 static void
5007 set_target_async_command (char *args, int from_tty,
5008 struct cmd_list_element *c)
5009 {
5010 if (have_live_inferiors ())
5011 {
5012 target_async_permitted_1 = target_async_permitted;
5013 error (_("Cannot change this setting while the inferior is running."));
5014 }
5015
5016 target_async_permitted = target_async_permitted_1;
5017 }
5018
5019 static void
5020 show_target_async_command (struct ui_file *file, int from_tty,
5021 struct cmd_list_element *c,
5022 const char *value)
5023 {
5024 fprintf_filtered (file,
5025 _("Controlling the inferior in "
5026 "asynchronous mode is %s.\n"), value);
5027 }
5028
5029 /* Temporary copies of permission settings. */
5030
5031 static int may_write_registers_1 = 1;
5032 static int may_write_memory_1 = 1;
5033 static int may_insert_breakpoints_1 = 1;
5034 static int may_insert_tracepoints_1 = 1;
5035 static int may_insert_fast_tracepoints_1 = 1;
5036 static int may_stop_1 = 1;
5037
5038 /* Make the user-set values match the real values again. */
5039
5040 void
5041 update_target_permissions (void)
5042 {
5043 may_write_registers_1 = may_write_registers;
5044 may_write_memory_1 = may_write_memory;
5045 may_insert_breakpoints_1 = may_insert_breakpoints;
5046 may_insert_tracepoints_1 = may_insert_tracepoints;
5047 may_insert_fast_tracepoints_1 = may_insert_fast_tracepoints;
5048 may_stop_1 = may_stop;
5049 }
5050
5051 /* The one function handles (most of) the permission flags in the same
5052 way. */
5053
5054 static void
5055 set_target_permissions (char *args, int from_tty,
5056 struct cmd_list_element *c)
5057 {
5058 if (target_has_execution)
5059 {
5060 update_target_permissions ();
5061 error (_("Cannot change this setting while the inferior is running."));
5062 }
5063
5064 /* Make the real values match the user-changed values. */
5065 may_write_registers = may_write_registers_1;
5066 may_insert_breakpoints = may_insert_breakpoints_1;
5067 may_insert_tracepoints = may_insert_tracepoints_1;
5068 may_insert_fast_tracepoints = may_insert_fast_tracepoints_1;
5069 may_stop = may_stop_1;
5070 update_observer_mode ();
5071 }
5072
5073 /* Set memory write permission independently of observer mode. */
5074
5075 static void
5076 set_write_memory_permission (char *args, int from_tty,
5077 struct cmd_list_element *c)
5078 {
5079 /* Make the real values match the user-changed values. */
5080 may_write_memory = may_write_memory_1;
5081 update_observer_mode ();
5082 }
5083
5084
5085 void
5086 initialize_targets (void)
5087 {
5088 init_dummy_target ();
5089 push_target (&dummy_target);
5090
5091 add_info ("target", target_info, targ_desc);
5092 add_info ("files", target_info, targ_desc);
5093
5094 add_setshow_zuinteger_cmd ("target", class_maintenance, &targetdebug, _("\
5095 Set target debugging."), _("\
5096 Show target debugging."), _("\
5097 When non-zero, target debugging is enabled. Higher numbers are more\n\
5098 verbose. Changes do not take effect until the next \"run\" or \"target\"\n\
5099 command."),
5100 NULL,
5101 show_targetdebug,
5102 &setdebuglist, &showdebuglist);
5103
5104 add_setshow_boolean_cmd ("trust-readonly-sections", class_support,
5105 &trust_readonly, _("\
5106 Set mode for reading from readonly sections."), _("\
5107 Show mode for reading from readonly sections."), _("\
5108 When this mode is on, memory reads from readonly sections (such as .text)\n\
5109 will be read from the object file instead of from the target. This will\n\
5110 result in significant performance improvement for remote targets."),
5111 NULL,
5112 show_trust_readonly,
5113 &setlist, &showlist);
5114
5115 add_com ("monitor", class_obscure, do_monitor_command,
5116 _("Send a command to the remote monitor (remote targets only)."));
5117
5118 add_cmd ("target-stack", class_maintenance, maintenance_print_target_stack,
5119 _("Print the name of each layer of the internal target stack."),
5120 &maintenanceprintlist);
5121
5122 add_setshow_boolean_cmd ("target-async", no_class,
5123 &target_async_permitted_1, _("\
5124 Set whether gdb controls the inferior in asynchronous mode."), _("\
5125 Show whether gdb controls the inferior in asynchronous mode."), _("\
5126 Tells gdb whether to control the inferior in asynchronous mode."),
5127 set_target_async_command,
5128 show_target_async_command,
5129 &setlist,
5130 &showlist);
5131
5132 add_setshow_boolean_cmd ("stack-cache", class_support,
5133 &stack_cache_enabled_p_1, _("\
5134 Set cache use for stack access."), _("\
5135 Show cache use for stack access."), _("\
5136 When on, use the data cache for all stack access, regardless of any\n\
5137 configured memory regions. This improves remote performance significantly.\n\
5138 By default, caching for stack access is on."),
5139 set_stack_cache_enabled_p,
5140 show_stack_cache_enabled_p,
5141 &setlist, &showlist);
5142
5143 add_setshow_boolean_cmd ("may-write-registers", class_support,
5144 &may_write_registers_1, _("\
5145 Set permission to write into registers."), _("\
5146 Show permission to write into registers."), _("\
5147 When this permission is on, GDB may write into the target's registers.\n\
5148 Otherwise, any sort of write attempt will result in an error."),
5149 set_target_permissions, NULL,
5150 &setlist, &showlist);
5151
5152 add_setshow_boolean_cmd ("may-write-memory", class_support,
5153 &may_write_memory_1, _("\
5154 Set permission to write into target memory."), _("\
5155 Show permission to write into target memory."), _("\
5156 When this permission is on, GDB may write into the target's memory.\n\
5157 Otherwise, any sort of write attempt will result in an error."),
5158 set_write_memory_permission, NULL,
5159 &setlist, &showlist);
5160
5161 add_setshow_boolean_cmd ("may-insert-breakpoints", class_support,
5162 &may_insert_breakpoints_1, _("\
5163 Set permission to insert breakpoints in the target."), _("\
5164 Show permission to insert breakpoints in the target."), _("\
5165 When this permission is on, GDB may insert breakpoints in the program.\n\
5166 Otherwise, any sort of insertion attempt will result in an error."),
5167 set_target_permissions, NULL,
5168 &setlist, &showlist);
5169
5170 add_setshow_boolean_cmd ("may-insert-tracepoints", class_support,
5171 &may_insert_tracepoints_1, _("\
5172 Set permission to insert tracepoints in the target."), _("\
5173 Show permission to insert tracepoints in the target."), _("\
5174 When this permission is on, GDB may insert tracepoints in the program.\n\
5175 Otherwise, any sort of insertion attempt will result in an error."),
5176 set_target_permissions, NULL,
5177 &setlist, &showlist);
5178
5179 add_setshow_boolean_cmd ("may-insert-fast-tracepoints", class_support,
5180 &may_insert_fast_tracepoints_1, _("\
5181 Set permission to insert fast tracepoints in the target."), _("\
5182 Show permission to insert fast tracepoints in the target."), _("\
5183 When this permission is on, GDB may insert fast tracepoints.\n\
5184 Otherwise, any sort of insertion attempt will result in an error."),
5185 set_target_permissions, NULL,
5186 &setlist, &showlist);
5187
5188 add_setshow_boolean_cmd ("may-interrupt", class_support,
5189 &may_stop_1, _("\
5190 Set permission to interrupt or signal the target."), _("\
5191 Show permission to interrupt or signal the target."), _("\
5192 When this permission is on, GDB may interrupt/stop the target's execution.\n\
5193 Otherwise, any attempt to interrupt or stop will be ignored."),
5194 set_target_permissions, NULL,
5195 &setlist, &showlist);
5196
5197
5198 target_dcache = dcache_init ();
5199 }
This page took 0.130024 seconds and 5 git commands to generate.