Remove range_s VEC
[deliverable/binutils-gdb.git] / gdb / value.c
1 /* Low level packing and unpacking of values for GDB, the GNU Debugger.
2
3 Copyright (C) 1986-2018 Free Software Foundation, Inc.
4
5 This file is part of GDB.
6
7 This program is free software; you can redistribute it and/or modify
8 it under the terms of the GNU General Public License as published by
9 the Free Software Foundation; either version 3 of the License, or
10 (at your option) any later version.
11
12 This program is distributed in the hope that it will be useful,
13 but WITHOUT ANY WARRANTY; without even the implied warranty of
14 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
15 GNU General Public License for more details.
16
17 You should have received a copy of the GNU General Public License
18 along with this program. If not, see <http://www.gnu.org/licenses/>. */
19
20 #include "defs.h"
21 #include "arch-utils.h"
22 #include "symtab.h"
23 #include "gdbtypes.h"
24 #include "value.h"
25 #include "gdbcore.h"
26 #include "command.h"
27 #include "gdbcmd.h"
28 #include "target.h"
29 #include "language.h"
30 #include "demangle.h"
31 #include "regcache.h"
32 #include "block.h"
33 #include "target-float.h"
34 #include "objfiles.h"
35 #include "valprint.h"
36 #include "cli/cli-decode.h"
37 #include "extension.h"
38 #include <ctype.h>
39 #include "tracepoint.h"
40 #include "cp-abi.h"
41 #include "user-regs.h"
42 #include <algorithm>
43 #include "completer.h"
44
45 /* Definition of a user function. */
46 struct internal_function
47 {
48 /* The name of the function. It is a bit odd to have this in the
49 function itself -- the user might use a differently-named
50 convenience variable to hold the function. */
51 char *name;
52
53 /* The handler. */
54 internal_function_fn handler;
55
56 /* User data for the handler. */
57 void *cookie;
58 };
59
60 /* Defines an [OFFSET, OFFSET + LENGTH) range. */
61
62 struct range
63 {
64 /* Lowest offset in the range. */
65 LONGEST offset;
66
67 /* Length of the range. */
68 LONGEST length;
69
70 /* Returns true if THIS is strictly less than OTHER, useful for
71 searching. We keep ranges sorted by offset and coalesce
72 overlapping and contiguous ranges, so this just compares the
73 starting offset. */
74
75 bool operator< (const range &other) const
76 {
77 return offset < other.offset;
78 }
79 };
80
81 /* Returns true if the ranges defined by [offset1, offset1+len1) and
82 [offset2, offset2+len2) overlap. */
83
84 static int
85 ranges_overlap (LONGEST offset1, LONGEST len1,
86 LONGEST offset2, LONGEST len2)
87 {
88 ULONGEST h, l;
89
90 l = std::max (offset1, offset2);
91 h = std::min (offset1 + len1, offset2 + len2);
92 return (l < h);
93 }
94
95 /* Returns true if RANGES contains any range that overlaps [OFFSET,
96 OFFSET+LENGTH). */
97
98 static int
99 ranges_contain (const std::vector<range> &ranges, LONGEST offset,
100 LONGEST length)
101 {
102 range what;
103
104 what.offset = offset;
105 what.length = length;
106
107 /* We keep ranges sorted by offset and coalesce overlapping and
108 contiguous ranges, so to check if a range list contains a given
109 range, we can do a binary search for the position the given range
110 would be inserted if we only considered the starting OFFSET of
111 ranges. We call that position I. Since we also have LENGTH to
112 care for (this is a range afterall), we need to check if the
113 _previous_ range overlaps the I range. E.g.,
114
115 R
116 |---|
117 |---| |---| |------| ... |--|
118 0 1 2 N
119
120 I=1
121
122 In the case above, the binary search would return `I=1', meaning,
123 this OFFSET should be inserted at position 1, and the current
124 position 1 should be pushed further (and before 2). But, `0'
125 overlaps with R.
126
127 Then we need to check if the I range overlaps the I range itself.
128 E.g.,
129
130 R
131 |---|
132 |---| |---| |-------| ... |--|
133 0 1 2 N
134
135 I=1
136 */
137
138
139 auto i = std::lower_bound (ranges.begin (), ranges.end (), what);
140
141 if (i > ranges.begin ())
142 {
143 const struct range &bef = *(i - 1);
144
145 if (ranges_overlap (bef.offset, bef.length, offset, length))
146 return 1;
147 }
148
149 if (i < ranges.end ())
150 {
151 const struct range &r = *i;
152
153 if (ranges_overlap (r.offset, r.length, offset, length))
154 return 1;
155 }
156
157 return 0;
158 }
159
160 static struct cmd_list_element *functionlist;
161
162 /* Note that the fields in this structure are arranged to save a bit
163 of memory. */
164
165 struct value
166 {
167 explicit value (struct type *type_)
168 : modifiable (1),
169 lazy (1),
170 initialized (1),
171 stack (0),
172 type (type_),
173 enclosing_type (type_)
174 {
175 location.address = 0;
176 }
177
178 ~value ()
179 {
180 if (VALUE_LVAL (this) == lval_computed)
181 {
182 const struct lval_funcs *funcs = location.computed.funcs;
183
184 if (funcs->free_closure)
185 funcs->free_closure (this);
186 }
187 else if (VALUE_LVAL (this) == lval_xcallable)
188 delete location.xm_worker;
189
190 xfree (contents);
191 }
192
193 DISABLE_COPY_AND_ASSIGN (value);
194
195 /* Type of value; either not an lval, or one of the various
196 different possible kinds of lval. */
197 enum lval_type lval = not_lval;
198
199 /* Is it modifiable? Only relevant if lval != not_lval. */
200 unsigned int modifiable : 1;
201
202 /* If zero, contents of this value are in the contents field. If
203 nonzero, contents are in inferior. If the lval field is lval_memory,
204 the contents are in inferior memory at location.address plus offset.
205 The lval field may also be lval_register.
206
207 WARNING: This field is used by the code which handles watchpoints
208 (see breakpoint.c) to decide whether a particular value can be
209 watched by hardware watchpoints. If the lazy flag is set for
210 some member of a value chain, it is assumed that this member of
211 the chain doesn't need to be watched as part of watching the
212 value itself. This is how GDB avoids watching the entire struct
213 or array when the user wants to watch a single struct member or
214 array element. If you ever change the way lazy flag is set and
215 reset, be sure to consider this use as well! */
216 unsigned int lazy : 1;
217
218 /* If value is a variable, is it initialized or not. */
219 unsigned int initialized : 1;
220
221 /* If value is from the stack. If this is set, read_stack will be
222 used instead of read_memory to enable extra caching. */
223 unsigned int stack : 1;
224
225 /* Location of value (if lval). */
226 union
227 {
228 /* If lval == lval_memory, this is the address in the inferior */
229 CORE_ADDR address;
230
231 /*If lval == lval_register, the value is from a register. */
232 struct
233 {
234 /* Register number. */
235 int regnum;
236 /* Frame ID of "next" frame to which a register value is relative.
237 If the register value is found relative to frame F, then the
238 frame id of F->next will be stored in next_frame_id. */
239 struct frame_id next_frame_id;
240 } reg;
241
242 /* Pointer to internal variable. */
243 struct internalvar *internalvar;
244
245 /* Pointer to xmethod worker. */
246 struct xmethod_worker *xm_worker;
247
248 /* If lval == lval_computed, this is a set of function pointers
249 to use to access and describe the value, and a closure pointer
250 for them to use. */
251 struct
252 {
253 /* Functions to call. */
254 const struct lval_funcs *funcs;
255
256 /* Closure for those functions to use. */
257 void *closure;
258 } computed;
259 } location;
260
261 /* Describes offset of a value within lval of a structure in target
262 addressable memory units. Note also the member embedded_offset
263 below. */
264 LONGEST offset = 0;
265
266 /* Only used for bitfields; number of bits contained in them. */
267 LONGEST bitsize = 0;
268
269 /* Only used for bitfields; position of start of field. For
270 gdbarch_bits_big_endian=0 targets, it is the position of the LSB. For
271 gdbarch_bits_big_endian=1 targets, it is the position of the MSB. */
272 LONGEST bitpos = 0;
273
274 /* The number of references to this value. When a value is created,
275 the value chain holds a reference, so REFERENCE_COUNT is 1. If
276 release_value is called, this value is removed from the chain but
277 the caller of release_value now has a reference to this value.
278 The caller must arrange for a call to value_free later. */
279 int reference_count = 1;
280
281 /* Only used for bitfields; the containing value. This allows a
282 single read from the target when displaying multiple
283 bitfields. */
284 value_ref_ptr parent;
285
286 /* Type of the value. */
287 struct type *type;
288
289 /* If a value represents a C++ object, then the `type' field gives
290 the object's compile-time type. If the object actually belongs
291 to some class derived from `type', perhaps with other base
292 classes and additional members, then `type' is just a subobject
293 of the real thing, and the full object is probably larger than
294 `type' would suggest.
295
296 If `type' is a dynamic class (i.e. one with a vtable), then GDB
297 can actually determine the object's run-time type by looking at
298 the run-time type information in the vtable. When this
299 information is available, we may elect to read in the entire
300 object, for several reasons:
301
302 - When printing the value, the user would probably rather see the
303 full object, not just the limited portion apparent from the
304 compile-time type.
305
306 - If `type' has virtual base classes, then even printing `type'
307 alone may require reaching outside the `type' portion of the
308 object to wherever the virtual base class has been stored.
309
310 When we store the entire object, `enclosing_type' is the run-time
311 type -- the complete object -- and `embedded_offset' is the
312 offset of `type' within that larger type, in target addressable memory
313 units. The value_contents() macro takes `embedded_offset' into account,
314 so most GDB code continues to see the `type' portion of the value, just
315 as the inferior would.
316
317 If `type' is a pointer to an object, then `enclosing_type' is a
318 pointer to the object's run-time type, and `pointed_to_offset' is
319 the offset in target addressable memory units from the full object
320 to the pointed-to object -- that is, the value `embedded_offset' would
321 have if we followed the pointer and fetched the complete object.
322 (I don't really see the point. Why not just determine the
323 run-time type when you indirect, and avoid the special case? The
324 contents don't matter until you indirect anyway.)
325
326 If we're not doing anything fancy, `enclosing_type' is equal to
327 `type', and `embedded_offset' is zero, so everything works
328 normally. */
329 struct type *enclosing_type;
330 LONGEST embedded_offset = 0;
331 LONGEST pointed_to_offset = 0;
332
333 /* Actual contents of the value. Target byte-order. NULL or not
334 valid if lazy is nonzero. */
335 gdb_byte *contents = nullptr;
336
337 /* Unavailable ranges in CONTENTS. We mark unavailable ranges,
338 rather than available, since the common and default case is for a
339 value to be available. This is filled in at value read time.
340 The unavailable ranges are tracked in bits. Note that a contents
341 bit that has been optimized out doesn't really exist in the
342 program, so it can't be marked unavailable either. */
343 std::vector<range> unavailable;
344
345 /* Likewise, but for optimized out contents (a chunk of the value of
346 a variable that does not actually exist in the program). If LVAL
347 is lval_register, this is a register ($pc, $sp, etc., never a
348 program variable) that has not been saved in the frame. Not
349 saved registers and optimized-out program variables values are
350 treated pretty much the same, except not-saved registers have a
351 different string representation and related error strings. */
352 std::vector<range> optimized_out;
353 };
354
355 /* See value.h. */
356
357 struct gdbarch *
358 get_value_arch (const struct value *value)
359 {
360 return get_type_arch (value_type (value));
361 }
362
363 int
364 value_bits_available (const struct value *value, LONGEST offset, LONGEST length)
365 {
366 gdb_assert (!value->lazy);
367
368 return !ranges_contain (value->unavailable, offset, length);
369 }
370
371 int
372 value_bytes_available (const struct value *value,
373 LONGEST offset, LONGEST length)
374 {
375 return value_bits_available (value,
376 offset * TARGET_CHAR_BIT,
377 length * TARGET_CHAR_BIT);
378 }
379
380 int
381 value_bits_any_optimized_out (const struct value *value, int bit_offset, int bit_length)
382 {
383 gdb_assert (!value->lazy);
384
385 return ranges_contain (value->optimized_out, bit_offset, bit_length);
386 }
387
388 int
389 value_entirely_available (struct value *value)
390 {
391 /* We can only tell whether the whole value is available when we try
392 to read it. */
393 if (value->lazy)
394 value_fetch_lazy (value);
395
396 if (value->unavailable.empty ())
397 return 1;
398 return 0;
399 }
400
401 /* Returns true if VALUE is entirely covered by RANGES. If the value
402 is lazy, it'll be read now. Note that RANGE is a pointer to
403 pointer because reading the value might change *RANGE. */
404
405 static int
406 value_entirely_covered_by_range_vector (struct value *value,
407 const std::vector<range> &ranges)
408 {
409 /* We can only tell whether the whole value is optimized out /
410 unavailable when we try to read it. */
411 if (value->lazy)
412 value_fetch_lazy (value);
413
414 if (ranges.size () == 1)
415 {
416 const struct range &t = ranges[0];
417
418 if (t.offset == 0
419 && t.length == (TARGET_CHAR_BIT
420 * TYPE_LENGTH (value_enclosing_type (value))))
421 return 1;
422 }
423
424 return 0;
425 }
426
427 int
428 value_entirely_unavailable (struct value *value)
429 {
430 return value_entirely_covered_by_range_vector (value, value->unavailable);
431 }
432
433 int
434 value_entirely_optimized_out (struct value *value)
435 {
436 return value_entirely_covered_by_range_vector (value, value->optimized_out);
437 }
438
439 /* Insert into the vector pointed to by VECTORP the bit range starting of
440 OFFSET bits, and extending for the next LENGTH bits. */
441
442 static void
443 insert_into_bit_range_vector (std::vector<range> *vectorp,
444 LONGEST offset, LONGEST length)
445 {
446 range newr;
447
448 /* Insert the range sorted. If there's overlap or the new range
449 would be contiguous with an existing range, merge. */
450
451 newr.offset = offset;
452 newr.length = length;
453
454 /* Do a binary search for the position the given range would be
455 inserted if we only considered the starting OFFSET of ranges.
456 Call that position I. Since we also have LENGTH to care for
457 (this is a range afterall), we need to check if the _previous_
458 range overlaps the I range. E.g., calling R the new range:
459
460 #1 - overlaps with previous
461
462 R
463 |-...-|
464 |---| |---| |------| ... |--|
465 0 1 2 N
466
467 I=1
468
469 In the case #1 above, the binary search would return `I=1',
470 meaning, this OFFSET should be inserted at position 1, and the
471 current position 1 should be pushed further (and become 2). But,
472 note that `0' overlaps with R, so we want to merge them.
473
474 A similar consideration needs to be taken if the new range would
475 be contiguous with the previous range:
476
477 #2 - contiguous with previous
478
479 R
480 |-...-|
481 |--| |---| |------| ... |--|
482 0 1 2 N
483
484 I=1
485
486 If there's no overlap with the previous range, as in:
487
488 #3 - not overlapping and not contiguous
489
490 R
491 |-...-|
492 |--| |---| |------| ... |--|
493 0 1 2 N
494
495 I=1
496
497 or if I is 0:
498
499 #4 - R is the range with lowest offset
500
501 R
502 |-...-|
503 |--| |---| |------| ... |--|
504 0 1 2 N
505
506 I=0
507
508 ... we just push the new range to I.
509
510 All the 4 cases above need to consider that the new range may
511 also overlap several of the ranges that follow, or that R may be
512 contiguous with the following range, and merge. E.g.,
513
514 #5 - overlapping following ranges
515
516 R
517 |------------------------|
518 |--| |---| |------| ... |--|
519 0 1 2 N
520
521 I=0
522
523 or:
524
525 R
526 |-------|
527 |--| |---| |------| ... |--|
528 0 1 2 N
529
530 I=1
531
532 */
533
534 auto i = std::lower_bound (vectorp->begin (), vectorp->end (), newr);
535 if (i > vectorp->begin ())
536 {
537 struct range &bef = *(i - 1);
538
539 if (ranges_overlap (bef.offset, bef.length, offset, length))
540 {
541 /* #1 */
542 ULONGEST l = std::min (bef.offset, offset);
543 ULONGEST h = std::max (bef.offset + bef.length, offset + length);
544
545 bef.offset = l;
546 bef.length = h - l;
547 i--;
548 }
549 else if (offset == bef.offset + bef.length)
550 {
551 /* #2 */
552 bef.length += length;
553 i--;
554 }
555 else
556 {
557 /* #3 */
558 i = vectorp->insert (i, newr);
559 }
560 }
561 else
562 {
563 /* #4 */
564 i = vectorp->insert (i, newr);
565 }
566
567 /* Check whether the ranges following the one we've just added or
568 touched can be folded in (#5 above). */
569 if (i != vectorp->end () && i + 1 < vectorp->end ())
570 {
571 int removed = 0;
572 auto next = i + 1;
573
574 /* Get the range we just touched. */
575 struct range &t = *i;
576 removed = 0;
577
578 i = next;
579 for (; i < vectorp->end (); i++)
580 {
581 struct range &r = *i;
582 if (r.offset <= t.offset + t.length)
583 {
584 ULONGEST l, h;
585
586 l = std::min (t.offset, r.offset);
587 h = std::max (t.offset + t.length, r.offset + r.length);
588
589 t.offset = l;
590 t.length = h - l;
591
592 removed++;
593 }
594 else
595 {
596 /* If we couldn't merge this one, we won't be able to
597 merge following ones either, since the ranges are
598 always sorted by OFFSET. */
599 break;
600 }
601 }
602
603 if (removed != 0)
604 vectorp->erase (next, next + removed);
605 }
606 }
607
608 void
609 mark_value_bits_unavailable (struct value *value,
610 LONGEST offset, LONGEST length)
611 {
612 insert_into_bit_range_vector (&value->unavailable, offset, length);
613 }
614
615 void
616 mark_value_bytes_unavailable (struct value *value,
617 LONGEST offset, LONGEST length)
618 {
619 mark_value_bits_unavailable (value,
620 offset * TARGET_CHAR_BIT,
621 length * TARGET_CHAR_BIT);
622 }
623
624 /* Find the first range in RANGES that overlaps the range defined by
625 OFFSET and LENGTH, starting at element POS in the RANGES vector,
626 Returns the index into RANGES where such overlapping range was
627 found, or -1 if none was found. */
628
629 static int
630 find_first_range_overlap (const std::vector<range> *ranges, int pos,
631 LONGEST offset, LONGEST length)
632 {
633 int i;
634
635 for (i = pos; i < ranges->size (); i++)
636 {
637 const range &r = (*ranges)[i];
638 if (ranges_overlap (r.offset, r.length, offset, length))
639 return i;
640 }
641
642 return -1;
643 }
644
645 /* Compare LENGTH_BITS of memory at PTR1 + OFFSET1_BITS with the memory at
646 PTR2 + OFFSET2_BITS. Return 0 if the memory is the same, otherwise
647 return non-zero.
648
649 It must always be the case that:
650 OFFSET1_BITS % TARGET_CHAR_BIT == OFFSET2_BITS % TARGET_CHAR_BIT
651
652 It is assumed that memory can be accessed from:
653 PTR + (OFFSET_BITS / TARGET_CHAR_BIT)
654 to:
655 PTR + ((OFFSET_BITS + LENGTH_BITS + TARGET_CHAR_BIT - 1)
656 / TARGET_CHAR_BIT) */
657 static int
658 memcmp_with_bit_offsets (const gdb_byte *ptr1, size_t offset1_bits,
659 const gdb_byte *ptr2, size_t offset2_bits,
660 size_t length_bits)
661 {
662 gdb_assert (offset1_bits % TARGET_CHAR_BIT
663 == offset2_bits % TARGET_CHAR_BIT);
664
665 if (offset1_bits % TARGET_CHAR_BIT != 0)
666 {
667 size_t bits;
668 gdb_byte mask, b1, b2;
669
670 /* The offset from the base pointers PTR1 and PTR2 is not a complete
671 number of bytes. A number of bits up to either the next exact
672 byte boundary, or LENGTH_BITS (which ever is sooner) will be
673 compared. */
674 bits = TARGET_CHAR_BIT - offset1_bits % TARGET_CHAR_BIT;
675 gdb_assert (bits < sizeof (mask) * TARGET_CHAR_BIT);
676 mask = (1 << bits) - 1;
677
678 if (length_bits < bits)
679 {
680 mask &= ~(gdb_byte) ((1 << (bits - length_bits)) - 1);
681 bits = length_bits;
682 }
683
684 /* Now load the two bytes and mask off the bits we care about. */
685 b1 = *(ptr1 + offset1_bits / TARGET_CHAR_BIT) & mask;
686 b2 = *(ptr2 + offset2_bits / TARGET_CHAR_BIT) & mask;
687
688 if (b1 != b2)
689 return 1;
690
691 /* Now update the length and offsets to take account of the bits
692 we've just compared. */
693 length_bits -= bits;
694 offset1_bits += bits;
695 offset2_bits += bits;
696 }
697
698 if (length_bits % TARGET_CHAR_BIT != 0)
699 {
700 size_t bits;
701 size_t o1, o2;
702 gdb_byte mask, b1, b2;
703
704 /* The length is not an exact number of bytes. After the previous
705 IF.. block then the offsets are byte aligned, or the
706 length is zero (in which case this code is not reached). Compare
707 a number of bits at the end of the region, starting from an exact
708 byte boundary. */
709 bits = length_bits % TARGET_CHAR_BIT;
710 o1 = offset1_bits + length_bits - bits;
711 o2 = offset2_bits + length_bits - bits;
712
713 gdb_assert (bits < sizeof (mask) * TARGET_CHAR_BIT);
714 mask = ((1 << bits) - 1) << (TARGET_CHAR_BIT - bits);
715
716 gdb_assert (o1 % TARGET_CHAR_BIT == 0);
717 gdb_assert (o2 % TARGET_CHAR_BIT == 0);
718
719 b1 = *(ptr1 + o1 / TARGET_CHAR_BIT) & mask;
720 b2 = *(ptr2 + o2 / TARGET_CHAR_BIT) & mask;
721
722 if (b1 != b2)
723 return 1;
724
725 length_bits -= bits;
726 }
727
728 if (length_bits > 0)
729 {
730 /* We've now taken care of any stray "bits" at the start, or end of
731 the region to compare, the remainder can be covered with a simple
732 memcmp. */
733 gdb_assert (offset1_bits % TARGET_CHAR_BIT == 0);
734 gdb_assert (offset2_bits % TARGET_CHAR_BIT == 0);
735 gdb_assert (length_bits % TARGET_CHAR_BIT == 0);
736
737 return memcmp (ptr1 + offset1_bits / TARGET_CHAR_BIT,
738 ptr2 + offset2_bits / TARGET_CHAR_BIT,
739 length_bits / TARGET_CHAR_BIT);
740 }
741
742 /* Length is zero, regions match. */
743 return 0;
744 }
745
746 /* Helper struct for find_first_range_overlap_and_match and
747 value_contents_bits_eq. Keep track of which slot of a given ranges
748 vector have we last looked at. */
749
750 struct ranges_and_idx
751 {
752 /* The ranges. */
753 const std::vector<range> *ranges;
754
755 /* The range we've last found in RANGES. Given ranges are sorted,
756 we can start the next lookup here. */
757 int idx;
758 };
759
760 /* Helper function for value_contents_bits_eq. Compare LENGTH bits of
761 RP1's ranges starting at OFFSET1 bits with LENGTH bits of RP2's
762 ranges starting at OFFSET2 bits. Return true if the ranges match
763 and fill in *L and *H with the overlapping window relative to
764 (both) OFFSET1 or OFFSET2. */
765
766 static int
767 find_first_range_overlap_and_match (struct ranges_and_idx *rp1,
768 struct ranges_and_idx *rp2,
769 LONGEST offset1, LONGEST offset2,
770 LONGEST length, ULONGEST *l, ULONGEST *h)
771 {
772 rp1->idx = find_first_range_overlap (rp1->ranges, rp1->idx,
773 offset1, length);
774 rp2->idx = find_first_range_overlap (rp2->ranges, rp2->idx,
775 offset2, length);
776
777 if (rp1->idx == -1 && rp2->idx == -1)
778 {
779 *l = length;
780 *h = length;
781 return 1;
782 }
783 else if (rp1->idx == -1 || rp2->idx == -1)
784 return 0;
785 else
786 {
787 const range *r1, *r2;
788 ULONGEST l1, h1;
789 ULONGEST l2, h2;
790
791 r1 = &(*rp1->ranges)[rp1->idx];
792 r2 = &(*rp2->ranges)[rp2->idx];
793
794 /* Get the unavailable windows intersected by the incoming
795 ranges. The first and last ranges that overlap the argument
796 range may be wider than said incoming arguments ranges. */
797 l1 = std::max (offset1, r1->offset);
798 h1 = std::min (offset1 + length, r1->offset + r1->length);
799
800 l2 = std::max (offset2, r2->offset);
801 h2 = std::min (offset2 + length, offset2 + r2->length);
802
803 /* Make them relative to the respective start offsets, so we can
804 compare them for equality. */
805 l1 -= offset1;
806 h1 -= offset1;
807
808 l2 -= offset2;
809 h2 -= offset2;
810
811 /* Different ranges, no match. */
812 if (l1 != l2 || h1 != h2)
813 return 0;
814
815 *h = h1;
816 *l = l1;
817 return 1;
818 }
819 }
820
821 /* Helper function for value_contents_eq. The only difference is that
822 this function is bit rather than byte based.
823
824 Compare LENGTH bits of VAL1's contents starting at OFFSET1 bits
825 with LENGTH bits of VAL2's contents starting at OFFSET2 bits.
826 Return true if the available bits match. */
827
828 static bool
829 value_contents_bits_eq (const struct value *val1, int offset1,
830 const struct value *val2, int offset2,
831 int length)
832 {
833 /* Each array element corresponds to a ranges source (unavailable,
834 optimized out). '1' is for VAL1, '2' for VAL2. */
835 struct ranges_and_idx rp1[2], rp2[2];
836
837 /* See function description in value.h. */
838 gdb_assert (!val1->lazy && !val2->lazy);
839
840 /* We shouldn't be trying to compare past the end of the values. */
841 gdb_assert (offset1 + length
842 <= TYPE_LENGTH (val1->enclosing_type) * TARGET_CHAR_BIT);
843 gdb_assert (offset2 + length
844 <= TYPE_LENGTH (val2->enclosing_type) * TARGET_CHAR_BIT);
845
846 memset (&rp1, 0, sizeof (rp1));
847 memset (&rp2, 0, sizeof (rp2));
848 rp1[0].ranges = &val1->unavailable;
849 rp2[0].ranges = &val2->unavailable;
850 rp1[1].ranges = &val1->optimized_out;
851 rp2[1].ranges = &val2->optimized_out;
852
853 while (length > 0)
854 {
855 ULONGEST l = 0, h = 0; /* init for gcc -Wall */
856 int i;
857
858 for (i = 0; i < 2; i++)
859 {
860 ULONGEST l_tmp, h_tmp;
861
862 /* The contents only match equal if the invalid/unavailable
863 contents ranges match as well. */
864 if (!find_first_range_overlap_and_match (&rp1[i], &rp2[i],
865 offset1, offset2, length,
866 &l_tmp, &h_tmp))
867 return false;
868
869 /* We're interested in the lowest/first range found. */
870 if (i == 0 || l_tmp < l)
871 {
872 l = l_tmp;
873 h = h_tmp;
874 }
875 }
876
877 /* Compare the available/valid contents. */
878 if (memcmp_with_bit_offsets (val1->contents, offset1,
879 val2->contents, offset2, l) != 0)
880 return false;
881
882 length -= h;
883 offset1 += h;
884 offset2 += h;
885 }
886
887 return true;
888 }
889
890 bool
891 value_contents_eq (const struct value *val1, LONGEST offset1,
892 const struct value *val2, LONGEST offset2,
893 LONGEST length)
894 {
895 return value_contents_bits_eq (val1, offset1 * TARGET_CHAR_BIT,
896 val2, offset2 * TARGET_CHAR_BIT,
897 length * TARGET_CHAR_BIT);
898 }
899
900
901 /* The value-history records all the values printed by print commands
902 during this session. */
903
904 static std::vector<value_ref_ptr> value_history;
905
906 \f
907 /* List of all value objects currently allocated
908 (except for those released by calls to release_value)
909 This is so they can be freed after each command. */
910
911 static std::vector<value_ref_ptr> all_values;
912
913 /* Allocate a lazy value for type TYPE. Its actual content is
914 "lazily" allocated too: the content field of the return value is
915 NULL; it will be allocated when it is fetched from the target. */
916
917 struct value *
918 allocate_value_lazy (struct type *type)
919 {
920 struct value *val;
921
922 /* Call check_typedef on our type to make sure that, if TYPE
923 is a TYPE_CODE_TYPEDEF, its length is set to the length
924 of the target type instead of zero. However, we do not
925 replace the typedef type by the target type, because we want
926 to keep the typedef in order to be able to set the VAL's type
927 description correctly. */
928 check_typedef (type);
929
930 val = new struct value (type);
931
932 /* Values start out on the all_values chain. */
933 all_values.emplace_back (val);
934
935 return val;
936 }
937
938 /* The maximum size, in bytes, that GDB will try to allocate for a value.
939 The initial value of 64k was not selected for any specific reason, it is
940 just a reasonable starting point. */
941
942 static int max_value_size = 65536; /* 64k bytes */
943
944 /* It is critical that the MAX_VALUE_SIZE is at least as big as the size of
945 LONGEST, otherwise GDB will not be able to parse integer values from the
946 CLI; for example if the MAX_VALUE_SIZE could be set to 1 then GDB would
947 be unable to parse "set max-value-size 2".
948
949 As we want a consistent GDB experience across hosts with different sizes
950 of LONGEST, this arbitrary minimum value was selected, so long as this
951 is bigger than LONGEST on all GDB supported hosts we're fine. */
952
953 #define MIN_VALUE_FOR_MAX_VALUE_SIZE 16
954 gdb_static_assert (sizeof (LONGEST) <= MIN_VALUE_FOR_MAX_VALUE_SIZE);
955
956 /* Implement the "set max-value-size" command. */
957
958 static void
959 set_max_value_size (const char *args, int from_tty,
960 struct cmd_list_element *c)
961 {
962 gdb_assert (max_value_size == -1 || max_value_size >= 0);
963
964 if (max_value_size > -1 && max_value_size < MIN_VALUE_FOR_MAX_VALUE_SIZE)
965 {
966 max_value_size = MIN_VALUE_FOR_MAX_VALUE_SIZE;
967 error (_("max-value-size set too low, increasing to %d bytes"),
968 max_value_size);
969 }
970 }
971
972 /* Implement the "show max-value-size" command. */
973
974 static void
975 show_max_value_size (struct ui_file *file, int from_tty,
976 struct cmd_list_element *c, const char *value)
977 {
978 if (max_value_size == -1)
979 fprintf_filtered (file, _("Maximum value size is unlimited.\n"));
980 else
981 fprintf_filtered (file, _("Maximum value size is %d bytes.\n"),
982 max_value_size);
983 }
984
985 /* Called before we attempt to allocate or reallocate a buffer for the
986 contents of a value. TYPE is the type of the value for which we are
987 allocating the buffer. If the buffer is too large (based on the user
988 controllable setting) then throw an error. If this function returns
989 then we should attempt to allocate the buffer. */
990
991 static void
992 check_type_length_before_alloc (const struct type *type)
993 {
994 unsigned int length = TYPE_LENGTH (type);
995
996 if (max_value_size > -1 && length > max_value_size)
997 {
998 if (TYPE_NAME (type) != NULL)
999 error (_("value of type `%s' requires %u bytes, which is more "
1000 "than max-value-size"), TYPE_NAME (type), length);
1001 else
1002 error (_("value requires %u bytes, which is more than "
1003 "max-value-size"), length);
1004 }
1005 }
1006
1007 /* Allocate the contents of VAL if it has not been allocated yet. */
1008
1009 static void
1010 allocate_value_contents (struct value *val)
1011 {
1012 if (!val->contents)
1013 {
1014 check_type_length_before_alloc (val->enclosing_type);
1015 val->contents
1016 = (gdb_byte *) xzalloc (TYPE_LENGTH (val->enclosing_type));
1017 }
1018 }
1019
1020 /* Allocate a value and its contents for type TYPE. */
1021
1022 struct value *
1023 allocate_value (struct type *type)
1024 {
1025 struct value *val = allocate_value_lazy (type);
1026
1027 allocate_value_contents (val);
1028 val->lazy = 0;
1029 return val;
1030 }
1031
1032 /* Allocate a value that has the correct length
1033 for COUNT repetitions of type TYPE. */
1034
1035 struct value *
1036 allocate_repeat_value (struct type *type, int count)
1037 {
1038 int low_bound = current_language->string_lower_bound; /* ??? */
1039 /* FIXME-type-allocation: need a way to free this type when we are
1040 done with it. */
1041 struct type *array_type
1042 = lookup_array_range_type (type, low_bound, count + low_bound - 1);
1043
1044 return allocate_value (array_type);
1045 }
1046
1047 struct value *
1048 allocate_computed_value (struct type *type,
1049 const struct lval_funcs *funcs,
1050 void *closure)
1051 {
1052 struct value *v = allocate_value_lazy (type);
1053
1054 VALUE_LVAL (v) = lval_computed;
1055 v->location.computed.funcs = funcs;
1056 v->location.computed.closure = closure;
1057
1058 return v;
1059 }
1060
1061 /* Allocate NOT_LVAL value for type TYPE being OPTIMIZED_OUT. */
1062
1063 struct value *
1064 allocate_optimized_out_value (struct type *type)
1065 {
1066 struct value *retval = allocate_value_lazy (type);
1067
1068 mark_value_bytes_optimized_out (retval, 0, TYPE_LENGTH (type));
1069 set_value_lazy (retval, 0);
1070 return retval;
1071 }
1072
1073 /* Accessor methods. */
1074
1075 struct type *
1076 value_type (const struct value *value)
1077 {
1078 return value->type;
1079 }
1080 void
1081 deprecated_set_value_type (struct value *value, struct type *type)
1082 {
1083 value->type = type;
1084 }
1085
1086 LONGEST
1087 value_offset (const struct value *value)
1088 {
1089 return value->offset;
1090 }
1091 void
1092 set_value_offset (struct value *value, LONGEST offset)
1093 {
1094 value->offset = offset;
1095 }
1096
1097 LONGEST
1098 value_bitpos (const struct value *value)
1099 {
1100 return value->bitpos;
1101 }
1102 void
1103 set_value_bitpos (struct value *value, LONGEST bit)
1104 {
1105 value->bitpos = bit;
1106 }
1107
1108 LONGEST
1109 value_bitsize (const struct value *value)
1110 {
1111 return value->bitsize;
1112 }
1113 void
1114 set_value_bitsize (struct value *value, LONGEST bit)
1115 {
1116 value->bitsize = bit;
1117 }
1118
1119 struct value *
1120 value_parent (const struct value *value)
1121 {
1122 return value->parent.get ();
1123 }
1124
1125 /* See value.h. */
1126
1127 void
1128 set_value_parent (struct value *value, struct value *parent)
1129 {
1130 value->parent = value_ref_ptr (value_incref (parent));
1131 }
1132
1133 gdb_byte *
1134 value_contents_raw (struct value *value)
1135 {
1136 struct gdbarch *arch = get_value_arch (value);
1137 int unit_size = gdbarch_addressable_memory_unit_size (arch);
1138
1139 allocate_value_contents (value);
1140 return value->contents + value->embedded_offset * unit_size;
1141 }
1142
1143 gdb_byte *
1144 value_contents_all_raw (struct value *value)
1145 {
1146 allocate_value_contents (value);
1147 return value->contents;
1148 }
1149
1150 struct type *
1151 value_enclosing_type (const struct value *value)
1152 {
1153 return value->enclosing_type;
1154 }
1155
1156 /* Look at value.h for description. */
1157
1158 struct type *
1159 value_actual_type (struct value *value, int resolve_simple_types,
1160 int *real_type_found)
1161 {
1162 struct value_print_options opts;
1163 struct type *result;
1164
1165 get_user_print_options (&opts);
1166
1167 if (real_type_found)
1168 *real_type_found = 0;
1169 result = value_type (value);
1170 if (opts.objectprint)
1171 {
1172 /* If result's target type is TYPE_CODE_STRUCT, proceed to
1173 fetch its rtti type. */
1174 if ((TYPE_CODE (result) == TYPE_CODE_PTR || TYPE_IS_REFERENCE (result))
1175 && TYPE_CODE (check_typedef (TYPE_TARGET_TYPE (result)))
1176 == TYPE_CODE_STRUCT
1177 && !value_optimized_out (value))
1178 {
1179 struct type *real_type;
1180
1181 real_type = value_rtti_indirect_type (value, NULL, NULL, NULL);
1182 if (real_type)
1183 {
1184 if (real_type_found)
1185 *real_type_found = 1;
1186 result = real_type;
1187 }
1188 }
1189 else if (resolve_simple_types)
1190 {
1191 if (real_type_found)
1192 *real_type_found = 1;
1193 result = value_enclosing_type (value);
1194 }
1195 }
1196
1197 return result;
1198 }
1199
1200 void
1201 error_value_optimized_out (void)
1202 {
1203 error (_("value has been optimized out"));
1204 }
1205
1206 static void
1207 require_not_optimized_out (const struct value *value)
1208 {
1209 if (!value->optimized_out.empty ())
1210 {
1211 if (value->lval == lval_register)
1212 error (_("register has not been saved in frame"));
1213 else
1214 error_value_optimized_out ();
1215 }
1216 }
1217
1218 static void
1219 require_available (const struct value *value)
1220 {
1221 if (!value->unavailable.empty ())
1222 throw_error (NOT_AVAILABLE_ERROR, _("value is not available"));
1223 }
1224
1225 const gdb_byte *
1226 value_contents_for_printing (struct value *value)
1227 {
1228 if (value->lazy)
1229 value_fetch_lazy (value);
1230 return value->contents;
1231 }
1232
1233 const gdb_byte *
1234 value_contents_for_printing_const (const struct value *value)
1235 {
1236 gdb_assert (!value->lazy);
1237 return value->contents;
1238 }
1239
1240 const gdb_byte *
1241 value_contents_all (struct value *value)
1242 {
1243 const gdb_byte *result = value_contents_for_printing (value);
1244 require_not_optimized_out (value);
1245 require_available (value);
1246 return result;
1247 }
1248
1249 /* Copy ranges in SRC_RANGE that overlap [SRC_BIT_OFFSET,
1250 SRC_BIT_OFFSET+BIT_LENGTH) ranges into *DST_RANGE, adjusted. */
1251
1252 static void
1253 ranges_copy_adjusted (std::vector<range> *dst_range, int dst_bit_offset,
1254 const std::vector<range> &src_range, int src_bit_offset,
1255 int bit_length)
1256 {
1257 for (const range &r : src_range)
1258 {
1259 ULONGEST h, l;
1260
1261 l = std::max (r.offset, (LONGEST) src_bit_offset);
1262 h = std::min (r.offset + r.length,
1263 (LONGEST) src_bit_offset + bit_length);
1264
1265 if (l < h)
1266 insert_into_bit_range_vector (dst_range,
1267 dst_bit_offset + (l - src_bit_offset),
1268 h - l);
1269 }
1270 }
1271
1272 /* Copy the ranges metadata in SRC that overlaps [SRC_BIT_OFFSET,
1273 SRC_BIT_OFFSET+BIT_LENGTH) into DST, adjusted. */
1274
1275 static void
1276 value_ranges_copy_adjusted (struct value *dst, int dst_bit_offset,
1277 const struct value *src, int src_bit_offset,
1278 int bit_length)
1279 {
1280 ranges_copy_adjusted (&dst->unavailable, dst_bit_offset,
1281 src->unavailable, src_bit_offset,
1282 bit_length);
1283 ranges_copy_adjusted (&dst->optimized_out, dst_bit_offset,
1284 src->optimized_out, src_bit_offset,
1285 bit_length);
1286 }
1287
1288 /* Copy LENGTH target addressable memory units of SRC value's (all) contents
1289 (value_contents_all) starting at SRC_OFFSET, into DST value's (all)
1290 contents, starting at DST_OFFSET. If unavailable contents are
1291 being copied from SRC, the corresponding DST contents are marked
1292 unavailable accordingly. Neither DST nor SRC may be lazy
1293 values.
1294
1295 It is assumed the contents of DST in the [DST_OFFSET,
1296 DST_OFFSET+LENGTH) range are wholly available. */
1297
1298 void
1299 value_contents_copy_raw (struct value *dst, LONGEST dst_offset,
1300 struct value *src, LONGEST src_offset, LONGEST length)
1301 {
1302 LONGEST src_bit_offset, dst_bit_offset, bit_length;
1303 struct gdbarch *arch = get_value_arch (src);
1304 int unit_size = gdbarch_addressable_memory_unit_size (arch);
1305
1306 /* A lazy DST would make that this copy operation useless, since as
1307 soon as DST's contents were un-lazied (by a later value_contents
1308 call, say), the contents would be overwritten. A lazy SRC would
1309 mean we'd be copying garbage. */
1310 gdb_assert (!dst->lazy && !src->lazy);
1311
1312 /* The overwritten DST range gets unavailability ORed in, not
1313 replaced. Make sure to remember to implement replacing if it
1314 turns out actually necessary. */
1315 gdb_assert (value_bytes_available (dst, dst_offset, length));
1316 gdb_assert (!value_bits_any_optimized_out (dst,
1317 TARGET_CHAR_BIT * dst_offset,
1318 TARGET_CHAR_BIT * length));
1319
1320 /* Copy the data. */
1321 memcpy (value_contents_all_raw (dst) + dst_offset * unit_size,
1322 value_contents_all_raw (src) + src_offset * unit_size,
1323 length * unit_size);
1324
1325 /* Copy the meta-data, adjusted. */
1326 src_bit_offset = src_offset * unit_size * HOST_CHAR_BIT;
1327 dst_bit_offset = dst_offset * unit_size * HOST_CHAR_BIT;
1328 bit_length = length * unit_size * HOST_CHAR_BIT;
1329
1330 value_ranges_copy_adjusted (dst, dst_bit_offset,
1331 src, src_bit_offset,
1332 bit_length);
1333 }
1334
1335 /* Copy LENGTH bytes of SRC value's (all) contents
1336 (value_contents_all) starting at SRC_OFFSET byte, into DST value's
1337 (all) contents, starting at DST_OFFSET. If unavailable contents
1338 are being copied from SRC, the corresponding DST contents are
1339 marked unavailable accordingly. DST must not be lazy. If SRC is
1340 lazy, it will be fetched now.
1341
1342 It is assumed the contents of DST in the [DST_OFFSET,
1343 DST_OFFSET+LENGTH) range are wholly available. */
1344
1345 void
1346 value_contents_copy (struct value *dst, LONGEST dst_offset,
1347 struct value *src, LONGEST src_offset, LONGEST length)
1348 {
1349 if (src->lazy)
1350 value_fetch_lazy (src);
1351
1352 value_contents_copy_raw (dst, dst_offset, src, src_offset, length);
1353 }
1354
1355 int
1356 value_lazy (const struct value *value)
1357 {
1358 return value->lazy;
1359 }
1360
1361 void
1362 set_value_lazy (struct value *value, int val)
1363 {
1364 value->lazy = val;
1365 }
1366
1367 int
1368 value_stack (const struct value *value)
1369 {
1370 return value->stack;
1371 }
1372
1373 void
1374 set_value_stack (struct value *value, int val)
1375 {
1376 value->stack = val;
1377 }
1378
1379 const gdb_byte *
1380 value_contents (struct value *value)
1381 {
1382 const gdb_byte *result = value_contents_writeable (value);
1383 require_not_optimized_out (value);
1384 require_available (value);
1385 return result;
1386 }
1387
1388 gdb_byte *
1389 value_contents_writeable (struct value *value)
1390 {
1391 if (value->lazy)
1392 value_fetch_lazy (value);
1393 return value_contents_raw (value);
1394 }
1395
1396 int
1397 value_optimized_out (struct value *value)
1398 {
1399 /* We can only know if a value is optimized out once we have tried to
1400 fetch it. */
1401 if (value->optimized_out.empty () && value->lazy)
1402 {
1403 TRY
1404 {
1405 value_fetch_lazy (value);
1406 }
1407 CATCH (ex, RETURN_MASK_ERROR)
1408 {
1409 /* Fall back to checking value->optimized_out. */
1410 }
1411 END_CATCH
1412 }
1413
1414 return !value->optimized_out.empty ();
1415 }
1416
1417 /* Mark contents of VALUE as optimized out, starting at OFFSET bytes, and
1418 the following LENGTH bytes. */
1419
1420 void
1421 mark_value_bytes_optimized_out (struct value *value, int offset, int length)
1422 {
1423 mark_value_bits_optimized_out (value,
1424 offset * TARGET_CHAR_BIT,
1425 length * TARGET_CHAR_BIT);
1426 }
1427
1428 /* See value.h. */
1429
1430 void
1431 mark_value_bits_optimized_out (struct value *value,
1432 LONGEST offset, LONGEST length)
1433 {
1434 insert_into_bit_range_vector (&value->optimized_out, offset, length);
1435 }
1436
1437 int
1438 value_bits_synthetic_pointer (const struct value *value,
1439 LONGEST offset, LONGEST length)
1440 {
1441 if (value->lval != lval_computed
1442 || !value->location.computed.funcs->check_synthetic_pointer)
1443 return 0;
1444 return value->location.computed.funcs->check_synthetic_pointer (value,
1445 offset,
1446 length);
1447 }
1448
1449 LONGEST
1450 value_embedded_offset (const struct value *value)
1451 {
1452 return value->embedded_offset;
1453 }
1454
1455 void
1456 set_value_embedded_offset (struct value *value, LONGEST val)
1457 {
1458 value->embedded_offset = val;
1459 }
1460
1461 LONGEST
1462 value_pointed_to_offset (const struct value *value)
1463 {
1464 return value->pointed_to_offset;
1465 }
1466
1467 void
1468 set_value_pointed_to_offset (struct value *value, LONGEST val)
1469 {
1470 value->pointed_to_offset = val;
1471 }
1472
1473 const struct lval_funcs *
1474 value_computed_funcs (const struct value *v)
1475 {
1476 gdb_assert (value_lval_const (v) == lval_computed);
1477
1478 return v->location.computed.funcs;
1479 }
1480
1481 void *
1482 value_computed_closure (const struct value *v)
1483 {
1484 gdb_assert (v->lval == lval_computed);
1485
1486 return v->location.computed.closure;
1487 }
1488
1489 enum lval_type *
1490 deprecated_value_lval_hack (struct value *value)
1491 {
1492 return &value->lval;
1493 }
1494
1495 enum lval_type
1496 value_lval_const (const struct value *value)
1497 {
1498 return value->lval;
1499 }
1500
1501 CORE_ADDR
1502 value_address (const struct value *value)
1503 {
1504 if (value->lval != lval_memory)
1505 return 0;
1506 if (value->parent != NULL)
1507 return value_address (value->parent.get ()) + value->offset;
1508 if (NULL != TYPE_DATA_LOCATION (value_type (value)))
1509 {
1510 gdb_assert (PROP_CONST == TYPE_DATA_LOCATION_KIND (value_type (value)));
1511 return TYPE_DATA_LOCATION_ADDR (value_type (value));
1512 }
1513
1514 return value->location.address + value->offset;
1515 }
1516
1517 CORE_ADDR
1518 value_raw_address (const struct value *value)
1519 {
1520 if (value->lval != lval_memory)
1521 return 0;
1522 return value->location.address;
1523 }
1524
1525 void
1526 set_value_address (struct value *value, CORE_ADDR addr)
1527 {
1528 gdb_assert (value->lval == lval_memory);
1529 value->location.address = addr;
1530 }
1531
1532 struct internalvar **
1533 deprecated_value_internalvar_hack (struct value *value)
1534 {
1535 return &value->location.internalvar;
1536 }
1537
1538 struct frame_id *
1539 deprecated_value_next_frame_id_hack (struct value *value)
1540 {
1541 gdb_assert (value->lval == lval_register);
1542 return &value->location.reg.next_frame_id;
1543 }
1544
1545 int *
1546 deprecated_value_regnum_hack (struct value *value)
1547 {
1548 gdb_assert (value->lval == lval_register);
1549 return &value->location.reg.regnum;
1550 }
1551
1552 int
1553 deprecated_value_modifiable (const struct value *value)
1554 {
1555 return value->modifiable;
1556 }
1557 \f
1558 /* Return a mark in the value chain. All values allocated after the
1559 mark is obtained (except for those released) are subject to being freed
1560 if a subsequent value_free_to_mark is passed the mark. */
1561 struct value *
1562 value_mark (void)
1563 {
1564 if (all_values.empty ())
1565 return nullptr;
1566 return all_values.back ().get ();
1567 }
1568
1569 /* Take a reference to VAL. VAL will not be deallocated until all
1570 references are released. */
1571
1572 struct value *
1573 value_incref (struct value *val)
1574 {
1575 val->reference_count++;
1576 return val;
1577 }
1578
1579 /* Release a reference to VAL, which was acquired with value_incref.
1580 This function is also called to deallocate values from the value
1581 chain. */
1582
1583 void
1584 value_decref (struct value *val)
1585 {
1586 if (val != nullptr)
1587 {
1588 gdb_assert (val->reference_count > 0);
1589 val->reference_count--;
1590 if (val->reference_count == 0)
1591 delete val;
1592 }
1593 }
1594
1595 /* Free all values allocated since MARK was obtained by value_mark
1596 (except for those released). */
1597 void
1598 value_free_to_mark (const struct value *mark)
1599 {
1600 auto iter = std::find (all_values.begin (), all_values.end (), mark);
1601 if (iter == all_values.end ())
1602 all_values.clear ();
1603 else
1604 all_values.erase (iter + 1, all_values.end ());
1605 }
1606
1607 /* Remove VAL from the chain all_values
1608 so it will not be freed automatically. */
1609
1610 value_ref_ptr
1611 release_value (struct value *val)
1612 {
1613 struct value *v;
1614
1615 if (val == nullptr)
1616 return value_ref_ptr ();
1617
1618 std::vector<value_ref_ptr>::reverse_iterator iter;
1619 for (iter = all_values.rbegin (); iter != all_values.rend (); ++iter)
1620 {
1621 if (*iter == val)
1622 {
1623 value_ref_ptr result = *iter;
1624 all_values.erase (iter.base () - 1);
1625 return result;
1626 }
1627 }
1628
1629 /* We must always return an owned reference. Normally this happens
1630 because we transfer the reference from the value chain, but in
1631 this case the value was not on the chain. */
1632 return value_ref_ptr (value_incref (val));
1633 }
1634
1635 /* See value.h. */
1636
1637 std::vector<value_ref_ptr>
1638 value_release_to_mark (const struct value *mark)
1639 {
1640 std::vector<value_ref_ptr> result;
1641
1642 auto iter = std::find (all_values.begin (), all_values.end (), mark);
1643 if (iter == all_values.end ())
1644 std::swap (result, all_values);
1645 else
1646 {
1647 std::move (iter + 1, all_values.end (), std::back_inserter (result));
1648 all_values.erase (iter + 1, all_values.end ());
1649 }
1650 std::reverse (result.begin (), result.end ());
1651 return result;
1652 }
1653
1654 /* Return a copy of the value ARG.
1655 It contains the same contents, for same memory address,
1656 but it's a different block of storage. */
1657
1658 struct value *
1659 value_copy (struct value *arg)
1660 {
1661 struct type *encl_type = value_enclosing_type (arg);
1662 struct value *val;
1663
1664 if (value_lazy (arg))
1665 val = allocate_value_lazy (encl_type);
1666 else
1667 val = allocate_value (encl_type);
1668 val->type = arg->type;
1669 VALUE_LVAL (val) = VALUE_LVAL (arg);
1670 val->location = arg->location;
1671 val->offset = arg->offset;
1672 val->bitpos = arg->bitpos;
1673 val->bitsize = arg->bitsize;
1674 val->lazy = arg->lazy;
1675 val->embedded_offset = value_embedded_offset (arg);
1676 val->pointed_to_offset = arg->pointed_to_offset;
1677 val->modifiable = arg->modifiable;
1678 if (!value_lazy (val))
1679 {
1680 memcpy (value_contents_all_raw (val), value_contents_all_raw (arg),
1681 TYPE_LENGTH (value_enclosing_type (arg)));
1682
1683 }
1684 val->unavailable = arg->unavailable;
1685 val->optimized_out = arg->optimized_out;
1686 val->parent = arg->parent;
1687 if (VALUE_LVAL (val) == lval_computed)
1688 {
1689 const struct lval_funcs *funcs = val->location.computed.funcs;
1690
1691 if (funcs->copy_closure)
1692 val->location.computed.closure = funcs->copy_closure (val);
1693 }
1694 return val;
1695 }
1696
1697 /* Return a "const" and/or "volatile" qualified version of the value V.
1698 If CNST is true, then the returned value will be qualified with
1699 "const".
1700 if VOLTL is true, then the returned value will be qualified with
1701 "volatile". */
1702
1703 struct value *
1704 make_cv_value (int cnst, int voltl, struct value *v)
1705 {
1706 struct type *val_type = value_type (v);
1707 struct type *enclosing_type = value_enclosing_type (v);
1708 struct value *cv_val = value_copy (v);
1709
1710 deprecated_set_value_type (cv_val,
1711 make_cv_type (cnst, voltl, val_type, NULL));
1712 set_value_enclosing_type (cv_val,
1713 make_cv_type (cnst, voltl, enclosing_type, NULL));
1714
1715 return cv_val;
1716 }
1717
1718 /* Return a version of ARG that is non-lvalue. */
1719
1720 struct value *
1721 value_non_lval (struct value *arg)
1722 {
1723 if (VALUE_LVAL (arg) != not_lval)
1724 {
1725 struct type *enc_type = value_enclosing_type (arg);
1726 struct value *val = allocate_value (enc_type);
1727
1728 memcpy (value_contents_all_raw (val), value_contents_all (arg),
1729 TYPE_LENGTH (enc_type));
1730 val->type = arg->type;
1731 set_value_embedded_offset (val, value_embedded_offset (arg));
1732 set_value_pointed_to_offset (val, value_pointed_to_offset (arg));
1733 return val;
1734 }
1735 return arg;
1736 }
1737
1738 /* Write contents of V at ADDR and set its lval type to be LVAL_MEMORY. */
1739
1740 void
1741 value_force_lval (struct value *v, CORE_ADDR addr)
1742 {
1743 gdb_assert (VALUE_LVAL (v) == not_lval);
1744
1745 write_memory (addr, value_contents_raw (v), TYPE_LENGTH (value_type (v)));
1746 v->lval = lval_memory;
1747 v->location.address = addr;
1748 }
1749
1750 void
1751 set_value_component_location (struct value *component,
1752 const struct value *whole)
1753 {
1754 struct type *type;
1755
1756 gdb_assert (whole->lval != lval_xcallable);
1757
1758 if (whole->lval == lval_internalvar)
1759 VALUE_LVAL (component) = lval_internalvar_component;
1760 else
1761 VALUE_LVAL (component) = whole->lval;
1762
1763 component->location = whole->location;
1764 if (whole->lval == lval_computed)
1765 {
1766 const struct lval_funcs *funcs = whole->location.computed.funcs;
1767
1768 if (funcs->copy_closure)
1769 component->location.computed.closure = funcs->copy_closure (whole);
1770 }
1771
1772 /* If type has a dynamic resolved location property
1773 update it's value address. */
1774 type = value_type (whole);
1775 if (NULL != TYPE_DATA_LOCATION (type)
1776 && TYPE_DATA_LOCATION_KIND (type) == PROP_CONST)
1777 set_value_address (component, TYPE_DATA_LOCATION_ADDR (type));
1778 }
1779
1780 /* Access to the value history. */
1781
1782 /* Record a new value in the value history.
1783 Returns the absolute history index of the entry. */
1784
1785 int
1786 record_latest_value (struct value *val)
1787 {
1788 int i;
1789
1790 /* We don't want this value to have anything to do with the inferior anymore.
1791 In particular, "set $1 = 50" should not affect the variable from which
1792 the value was taken, and fast watchpoints should be able to assume that
1793 a value on the value history never changes. */
1794 if (value_lazy (val))
1795 value_fetch_lazy (val);
1796 /* We preserve VALUE_LVAL so that the user can find out where it was fetched
1797 from. This is a bit dubious, because then *&$1 does not just return $1
1798 but the current contents of that location. c'est la vie... */
1799 val->modifiable = 0;
1800
1801 value_history.push_back (release_value (val));
1802
1803 return value_history.size ();
1804 }
1805
1806 /* Return a copy of the value in the history with sequence number NUM. */
1807
1808 struct value *
1809 access_value_history (int num)
1810 {
1811 int i;
1812 int absnum = num;
1813
1814 if (absnum <= 0)
1815 absnum += value_history.size ();
1816
1817 if (absnum <= 0)
1818 {
1819 if (num == 0)
1820 error (_("The history is empty."));
1821 else if (num == 1)
1822 error (_("There is only one value in the history."));
1823 else
1824 error (_("History does not go back to $$%d."), -num);
1825 }
1826 if (absnum > value_history.size ())
1827 error (_("History has not yet reached $%d."), absnum);
1828
1829 absnum--;
1830
1831 return value_copy (value_history[absnum].get ());
1832 }
1833
1834 static void
1835 show_values (const char *num_exp, int from_tty)
1836 {
1837 int i;
1838 struct value *val;
1839 static int num = 1;
1840
1841 if (num_exp)
1842 {
1843 /* "show values +" should print from the stored position.
1844 "show values <exp>" should print around value number <exp>. */
1845 if (num_exp[0] != '+' || num_exp[1] != '\0')
1846 num = parse_and_eval_long (num_exp) - 5;
1847 }
1848 else
1849 {
1850 /* "show values" means print the last 10 values. */
1851 num = value_history.size () - 9;
1852 }
1853
1854 if (num <= 0)
1855 num = 1;
1856
1857 for (i = num; i < num + 10 && i <= value_history.size (); i++)
1858 {
1859 struct value_print_options opts;
1860
1861 val = access_value_history (i);
1862 printf_filtered (("$%d = "), i);
1863 get_user_print_options (&opts);
1864 value_print (val, gdb_stdout, &opts);
1865 printf_filtered (("\n"));
1866 }
1867
1868 /* The next "show values +" should start after what we just printed. */
1869 num += 10;
1870
1871 /* Hitting just return after this command should do the same thing as
1872 "show values +". If num_exp is null, this is unnecessary, since
1873 "show values +" is not useful after "show values". */
1874 if (from_tty && num_exp)
1875 set_repeat_arguments ("+");
1876 }
1877 \f
1878 enum internalvar_kind
1879 {
1880 /* The internal variable is empty. */
1881 INTERNALVAR_VOID,
1882
1883 /* The value of the internal variable is provided directly as
1884 a GDB value object. */
1885 INTERNALVAR_VALUE,
1886
1887 /* A fresh value is computed via a call-back routine on every
1888 access to the internal variable. */
1889 INTERNALVAR_MAKE_VALUE,
1890
1891 /* The internal variable holds a GDB internal convenience function. */
1892 INTERNALVAR_FUNCTION,
1893
1894 /* The variable holds an integer value. */
1895 INTERNALVAR_INTEGER,
1896
1897 /* The variable holds a GDB-provided string. */
1898 INTERNALVAR_STRING,
1899 };
1900
1901 union internalvar_data
1902 {
1903 /* A value object used with INTERNALVAR_VALUE. */
1904 struct value *value;
1905
1906 /* The call-back routine used with INTERNALVAR_MAKE_VALUE. */
1907 struct
1908 {
1909 /* The functions to call. */
1910 const struct internalvar_funcs *functions;
1911
1912 /* The function's user-data. */
1913 void *data;
1914 } make_value;
1915
1916 /* The internal function used with INTERNALVAR_FUNCTION. */
1917 struct
1918 {
1919 struct internal_function *function;
1920 /* True if this is the canonical name for the function. */
1921 int canonical;
1922 } fn;
1923
1924 /* An integer value used with INTERNALVAR_INTEGER. */
1925 struct
1926 {
1927 /* If type is non-NULL, it will be used as the type to generate
1928 a value for this internal variable. If type is NULL, a default
1929 integer type for the architecture is used. */
1930 struct type *type;
1931 LONGEST val;
1932 } integer;
1933
1934 /* A string value used with INTERNALVAR_STRING. */
1935 char *string;
1936 };
1937
1938 /* Internal variables. These are variables within the debugger
1939 that hold values assigned by debugger commands.
1940 The user refers to them with a '$' prefix
1941 that does not appear in the variable names stored internally. */
1942
1943 struct internalvar
1944 {
1945 struct internalvar *next;
1946 char *name;
1947
1948 /* We support various different kinds of content of an internal variable.
1949 enum internalvar_kind specifies the kind, and union internalvar_data
1950 provides the data associated with this particular kind. */
1951
1952 enum internalvar_kind kind;
1953
1954 union internalvar_data u;
1955 };
1956
1957 static struct internalvar *internalvars;
1958
1959 /* If the variable does not already exist create it and give it the
1960 value given. If no value is given then the default is zero. */
1961 static void
1962 init_if_undefined_command (const char* args, int from_tty)
1963 {
1964 struct internalvar* intvar;
1965
1966 /* Parse the expression - this is taken from set_command(). */
1967 expression_up expr = parse_expression (args);
1968
1969 /* Validate the expression.
1970 Was the expression an assignment?
1971 Or even an expression at all? */
1972 if (expr->nelts == 0 || expr->elts[0].opcode != BINOP_ASSIGN)
1973 error (_("Init-if-undefined requires an assignment expression."));
1974
1975 /* Extract the variable from the parsed expression.
1976 In the case of an assign the lvalue will be in elts[1] and elts[2]. */
1977 if (expr->elts[1].opcode != OP_INTERNALVAR)
1978 error (_("The first parameter to init-if-undefined "
1979 "should be a GDB variable."));
1980 intvar = expr->elts[2].internalvar;
1981
1982 /* Only evaluate the expression if the lvalue is void.
1983 This may still fail if the expresssion is invalid. */
1984 if (intvar->kind == INTERNALVAR_VOID)
1985 evaluate_expression (expr.get ());
1986 }
1987
1988
1989 /* Look up an internal variable with name NAME. NAME should not
1990 normally include a dollar sign.
1991
1992 If the specified internal variable does not exist,
1993 the return value is NULL. */
1994
1995 struct internalvar *
1996 lookup_only_internalvar (const char *name)
1997 {
1998 struct internalvar *var;
1999
2000 for (var = internalvars; var; var = var->next)
2001 if (strcmp (var->name, name) == 0)
2002 return var;
2003
2004 return NULL;
2005 }
2006
2007 /* Complete NAME by comparing it to the names of internal
2008 variables. */
2009
2010 void
2011 complete_internalvar (completion_tracker &tracker, const char *name)
2012 {
2013 struct internalvar *var;
2014 int len;
2015
2016 len = strlen (name);
2017
2018 for (var = internalvars; var; var = var->next)
2019 if (strncmp (var->name, name, len) == 0)
2020 {
2021 gdb::unique_xmalloc_ptr<char> copy (xstrdup (var->name));
2022
2023 tracker.add_completion (std::move (copy));
2024 }
2025 }
2026
2027 /* Create an internal variable with name NAME and with a void value.
2028 NAME should not normally include a dollar sign. */
2029
2030 struct internalvar *
2031 create_internalvar (const char *name)
2032 {
2033 struct internalvar *var = XNEW (struct internalvar);
2034
2035 var->name = concat (name, (char *)NULL);
2036 var->kind = INTERNALVAR_VOID;
2037 var->next = internalvars;
2038 internalvars = var;
2039 return var;
2040 }
2041
2042 /* Create an internal variable with name NAME and register FUN as the
2043 function that value_of_internalvar uses to create a value whenever
2044 this variable is referenced. NAME should not normally include a
2045 dollar sign. DATA is passed uninterpreted to FUN when it is
2046 called. CLEANUP, if not NULL, is called when the internal variable
2047 is destroyed. It is passed DATA as its only argument. */
2048
2049 struct internalvar *
2050 create_internalvar_type_lazy (const char *name,
2051 const struct internalvar_funcs *funcs,
2052 void *data)
2053 {
2054 struct internalvar *var = create_internalvar (name);
2055
2056 var->kind = INTERNALVAR_MAKE_VALUE;
2057 var->u.make_value.functions = funcs;
2058 var->u.make_value.data = data;
2059 return var;
2060 }
2061
2062 /* See documentation in value.h. */
2063
2064 int
2065 compile_internalvar_to_ax (struct internalvar *var,
2066 struct agent_expr *expr,
2067 struct axs_value *value)
2068 {
2069 if (var->kind != INTERNALVAR_MAKE_VALUE
2070 || var->u.make_value.functions->compile_to_ax == NULL)
2071 return 0;
2072
2073 var->u.make_value.functions->compile_to_ax (var, expr, value,
2074 var->u.make_value.data);
2075 return 1;
2076 }
2077
2078 /* Look up an internal variable with name NAME. NAME should not
2079 normally include a dollar sign.
2080
2081 If the specified internal variable does not exist,
2082 one is created, with a void value. */
2083
2084 struct internalvar *
2085 lookup_internalvar (const char *name)
2086 {
2087 struct internalvar *var;
2088
2089 var = lookup_only_internalvar (name);
2090 if (var)
2091 return var;
2092
2093 return create_internalvar (name);
2094 }
2095
2096 /* Return current value of internal variable VAR. For variables that
2097 are not inherently typed, use a value type appropriate for GDBARCH. */
2098
2099 struct value *
2100 value_of_internalvar (struct gdbarch *gdbarch, struct internalvar *var)
2101 {
2102 struct value *val;
2103 struct trace_state_variable *tsv;
2104
2105 /* If there is a trace state variable of the same name, assume that
2106 is what we really want to see. */
2107 tsv = find_trace_state_variable (var->name);
2108 if (tsv)
2109 {
2110 tsv->value_known = target_get_trace_state_variable_value (tsv->number,
2111 &(tsv->value));
2112 if (tsv->value_known)
2113 val = value_from_longest (builtin_type (gdbarch)->builtin_int64,
2114 tsv->value);
2115 else
2116 val = allocate_value (builtin_type (gdbarch)->builtin_void);
2117 return val;
2118 }
2119
2120 switch (var->kind)
2121 {
2122 case INTERNALVAR_VOID:
2123 val = allocate_value (builtin_type (gdbarch)->builtin_void);
2124 break;
2125
2126 case INTERNALVAR_FUNCTION:
2127 val = allocate_value (builtin_type (gdbarch)->internal_fn);
2128 break;
2129
2130 case INTERNALVAR_INTEGER:
2131 if (!var->u.integer.type)
2132 val = value_from_longest (builtin_type (gdbarch)->builtin_int,
2133 var->u.integer.val);
2134 else
2135 val = value_from_longest (var->u.integer.type, var->u.integer.val);
2136 break;
2137
2138 case INTERNALVAR_STRING:
2139 val = value_cstring (var->u.string, strlen (var->u.string),
2140 builtin_type (gdbarch)->builtin_char);
2141 break;
2142
2143 case INTERNALVAR_VALUE:
2144 val = value_copy (var->u.value);
2145 if (value_lazy (val))
2146 value_fetch_lazy (val);
2147 break;
2148
2149 case INTERNALVAR_MAKE_VALUE:
2150 val = (*var->u.make_value.functions->make_value) (gdbarch, var,
2151 var->u.make_value.data);
2152 break;
2153
2154 default:
2155 internal_error (__FILE__, __LINE__, _("bad kind"));
2156 }
2157
2158 /* Change the VALUE_LVAL to lval_internalvar so that future operations
2159 on this value go back to affect the original internal variable.
2160
2161 Do not do this for INTERNALVAR_MAKE_VALUE variables, as those have
2162 no underlying modifyable state in the internal variable.
2163
2164 Likewise, if the variable's value is a computed lvalue, we want
2165 references to it to produce another computed lvalue, where
2166 references and assignments actually operate through the
2167 computed value's functions.
2168
2169 This means that internal variables with computed values
2170 behave a little differently from other internal variables:
2171 assignments to them don't just replace the previous value
2172 altogether. At the moment, this seems like the behavior we
2173 want. */
2174
2175 if (var->kind != INTERNALVAR_MAKE_VALUE
2176 && val->lval != lval_computed)
2177 {
2178 VALUE_LVAL (val) = lval_internalvar;
2179 VALUE_INTERNALVAR (val) = var;
2180 }
2181
2182 return val;
2183 }
2184
2185 int
2186 get_internalvar_integer (struct internalvar *var, LONGEST *result)
2187 {
2188 if (var->kind == INTERNALVAR_INTEGER)
2189 {
2190 *result = var->u.integer.val;
2191 return 1;
2192 }
2193
2194 if (var->kind == INTERNALVAR_VALUE)
2195 {
2196 struct type *type = check_typedef (value_type (var->u.value));
2197
2198 if (TYPE_CODE (type) == TYPE_CODE_INT)
2199 {
2200 *result = value_as_long (var->u.value);
2201 return 1;
2202 }
2203 }
2204
2205 return 0;
2206 }
2207
2208 static int
2209 get_internalvar_function (struct internalvar *var,
2210 struct internal_function **result)
2211 {
2212 switch (var->kind)
2213 {
2214 case INTERNALVAR_FUNCTION:
2215 *result = var->u.fn.function;
2216 return 1;
2217
2218 default:
2219 return 0;
2220 }
2221 }
2222
2223 void
2224 set_internalvar_component (struct internalvar *var,
2225 LONGEST offset, LONGEST bitpos,
2226 LONGEST bitsize, struct value *newval)
2227 {
2228 gdb_byte *addr;
2229 struct gdbarch *arch;
2230 int unit_size;
2231
2232 switch (var->kind)
2233 {
2234 case INTERNALVAR_VALUE:
2235 addr = value_contents_writeable (var->u.value);
2236 arch = get_value_arch (var->u.value);
2237 unit_size = gdbarch_addressable_memory_unit_size (arch);
2238
2239 if (bitsize)
2240 modify_field (value_type (var->u.value), addr + offset,
2241 value_as_long (newval), bitpos, bitsize);
2242 else
2243 memcpy (addr + offset * unit_size, value_contents (newval),
2244 TYPE_LENGTH (value_type (newval)));
2245 break;
2246
2247 default:
2248 /* We can never get a component of any other kind. */
2249 internal_error (__FILE__, __LINE__, _("set_internalvar_component"));
2250 }
2251 }
2252
2253 void
2254 set_internalvar (struct internalvar *var, struct value *val)
2255 {
2256 enum internalvar_kind new_kind;
2257 union internalvar_data new_data = { 0 };
2258
2259 if (var->kind == INTERNALVAR_FUNCTION && var->u.fn.canonical)
2260 error (_("Cannot overwrite convenience function %s"), var->name);
2261
2262 /* Prepare new contents. */
2263 switch (TYPE_CODE (check_typedef (value_type (val))))
2264 {
2265 case TYPE_CODE_VOID:
2266 new_kind = INTERNALVAR_VOID;
2267 break;
2268
2269 case TYPE_CODE_INTERNAL_FUNCTION:
2270 gdb_assert (VALUE_LVAL (val) == lval_internalvar);
2271 new_kind = INTERNALVAR_FUNCTION;
2272 get_internalvar_function (VALUE_INTERNALVAR (val),
2273 &new_data.fn.function);
2274 /* Copies created here are never canonical. */
2275 break;
2276
2277 default:
2278 new_kind = INTERNALVAR_VALUE;
2279 new_data.value = value_copy (val);
2280 new_data.value->modifiable = 1;
2281
2282 /* Force the value to be fetched from the target now, to avoid problems
2283 later when this internalvar is referenced and the target is gone or
2284 has changed. */
2285 if (value_lazy (new_data.value))
2286 value_fetch_lazy (new_data.value);
2287
2288 /* Release the value from the value chain to prevent it from being
2289 deleted by free_all_values. From here on this function should not
2290 call error () until new_data is installed into the var->u to avoid
2291 leaking memory. */
2292 release_value (new_data.value).release ();
2293
2294 /* Internal variables which are created from values with a dynamic
2295 location don't need the location property of the origin anymore.
2296 The resolved dynamic location is used prior then any other address
2297 when accessing the value.
2298 If we keep it, we would still refer to the origin value.
2299 Remove the location property in case it exist. */
2300 remove_dyn_prop (DYN_PROP_DATA_LOCATION, value_type (new_data.value));
2301
2302 break;
2303 }
2304
2305 /* Clean up old contents. */
2306 clear_internalvar (var);
2307
2308 /* Switch over. */
2309 var->kind = new_kind;
2310 var->u = new_data;
2311 /* End code which must not call error(). */
2312 }
2313
2314 void
2315 set_internalvar_integer (struct internalvar *var, LONGEST l)
2316 {
2317 /* Clean up old contents. */
2318 clear_internalvar (var);
2319
2320 var->kind = INTERNALVAR_INTEGER;
2321 var->u.integer.type = NULL;
2322 var->u.integer.val = l;
2323 }
2324
2325 void
2326 set_internalvar_string (struct internalvar *var, const char *string)
2327 {
2328 /* Clean up old contents. */
2329 clear_internalvar (var);
2330
2331 var->kind = INTERNALVAR_STRING;
2332 var->u.string = xstrdup (string);
2333 }
2334
2335 static void
2336 set_internalvar_function (struct internalvar *var, struct internal_function *f)
2337 {
2338 /* Clean up old contents. */
2339 clear_internalvar (var);
2340
2341 var->kind = INTERNALVAR_FUNCTION;
2342 var->u.fn.function = f;
2343 var->u.fn.canonical = 1;
2344 /* Variables installed here are always the canonical version. */
2345 }
2346
2347 void
2348 clear_internalvar (struct internalvar *var)
2349 {
2350 /* Clean up old contents. */
2351 switch (var->kind)
2352 {
2353 case INTERNALVAR_VALUE:
2354 value_decref (var->u.value);
2355 break;
2356
2357 case INTERNALVAR_STRING:
2358 xfree (var->u.string);
2359 break;
2360
2361 case INTERNALVAR_MAKE_VALUE:
2362 if (var->u.make_value.functions->destroy != NULL)
2363 var->u.make_value.functions->destroy (var->u.make_value.data);
2364 break;
2365
2366 default:
2367 break;
2368 }
2369
2370 /* Reset to void kind. */
2371 var->kind = INTERNALVAR_VOID;
2372 }
2373
2374 char *
2375 internalvar_name (const struct internalvar *var)
2376 {
2377 return var->name;
2378 }
2379
2380 static struct internal_function *
2381 create_internal_function (const char *name,
2382 internal_function_fn handler, void *cookie)
2383 {
2384 struct internal_function *ifn = XNEW (struct internal_function);
2385
2386 ifn->name = xstrdup (name);
2387 ifn->handler = handler;
2388 ifn->cookie = cookie;
2389 return ifn;
2390 }
2391
2392 char *
2393 value_internal_function_name (struct value *val)
2394 {
2395 struct internal_function *ifn;
2396 int result;
2397
2398 gdb_assert (VALUE_LVAL (val) == lval_internalvar);
2399 result = get_internalvar_function (VALUE_INTERNALVAR (val), &ifn);
2400 gdb_assert (result);
2401
2402 return ifn->name;
2403 }
2404
2405 struct value *
2406 call_internal_function (struct gdbarch *gdbarch,
2407 const struct language_defn *language,
2408 struct value *func, int argc, struct value **argv)
2409 {
2410 struct internal_function *ifn;
2411 int result;
2412
2413 gdb_assert (VALUE_LVAL (func) == lval_internalvar);
2414 result = get_internalvar_function (VALUE_INTERNALVAR (func), &ifn);
2415 gdb_assert (result);
2416
2417 return (*ifn->handler) (gdbarch, language, ifn->cookie, argc, argv);
2418 }
2419
2420 /* The 'function' command. This does nothing -- it is just a
2421 placeholder to let "help function NAME" work. This is also used as
2422 the implementation of the sub-command that is created when
2423 registering an internal function. */
2424 static void
2425 function_command (const char *command, int from_tty)
2426 {
2427 /* Do nothing. */
2428 }
2429
2430 /* Clean up if an internal function's command is destroyed. */
2431 static void
2432 function_destroyer (struct cmd_list_element *self, void *ignore)
2433 {
2434 xfree ((char *) self->name);
2435 xfree ((char *) self->doc);
2436 }
2437
2438 /* Add a new internal function. NAME is the name of the function; DOC
2439 is a documentation string describing the function. HANDLER is
2440 called when the function is invoked. COOKIE is an arbitrary
2441 pointer which is passed to HANDLER and is intended for "user
2442 data". */
2443 void
2444 add_internal_function (const char *name, const char *doc,
2445 internal_function_fn handler, void *cookie)
2446 {
2447 struct cmd_list_element *cmd;
2448 struct internal_function *ifn;
2449 struct internalvar *var = lookup_internalvar (name);
2450
2451 ifn = create_internal_function (name, handler, cookie);
2452 set_internalvar_function (var, ifn);
2453
2454 cmd = add_cmd (xstrdup (name), no_class, function_command, (char *) doc,
2455 &functionlist);
2456 cmd->destroyer = function_destroyer;
2457 }
2458
2459 /* Update VALUE before discarding OBJFILE. COPIED_TYPES is used to
2460 prevent cycles / duplicates. */
2461
2462 void
2463 preserve_one_value (struct value *value, struct objfile *objfile,
2464 htab_t copied_types)
2465 {
2466 if (TYPE_OBJFILE (value->type) == objfile)
2467 value->type = copy_type_recursive (objfile, value->type, copied_types);
2468
2469 if (TYPE_OBJFILE (value->enclosing_type) == objfile)
2470 value->enclosing_type = copy_type_recursive (objfile,
2471 value->enclosing_type,
2472 copied_types);
2473 }
2474
2475 /* Likewise for internal variable VAR. */
2476
2477 static void
2478 preserve_one_internalvar (struct internalvar *var, struct objfile *objfile,
2479 htab_t copied_types)
2480 {
2481 switch (var->kind)
2482 {
2483 case INTERNALVAR_INTEGER:
2484 if (var->u.integer.type && TYPE_OBJFILE (var->u.integer.type) == objfile)
2485 var->u.integer.type
2486 = copy_type_recursive (objfile, var->u.integer.type, copied_types);
2487 break;
2488
2489 case INTERNALVAR_VALUE:
2490 preserve_one_value (var->u.value, objfile, copied_types);
2491 break;
2492 }
2493 }
2494
2495 /* Update the internal variables and value history when OBJFILE is
2496 discarded; we must copy the types out of the objfile. New global types
2497 will be created for every convenience variable which currently points to
2498 this objfile's types, and the convenience variables will be adjusted to
2499 use the new global types. */
2500
2501 void
2502 preserve_values (struct objfile *objfile)
2503 {
2504 htab_t copied_types;
2505 struct internalvar *var;
2506 int i;
2507
2508 /* Create the hash table. We allocate on the objfile's obstack, since
2509 it is soon to be deleted. */
2510 copied_types = create_copied_types_hash (objfile);
2511
2512 for (const value_ref_ptr &item : value_history)
2513 preserve_one_value (item.get (), objfile, copied_types);
2514
2515 for (var = internalvars; var; var = var->next)
2516 preserve_one_internalvar (var, objfile, copied_types);
2517
2518 preserve_ext_lang_values (objfile, copied_types);
2519
2520 htab_delete (copied_types);
2521 }
2522
2523 static void
2524 show_convenience (const char *ignore, int from_tty)
2525 {
2526 struct gdbarch *gdbarch = get_current_arch ();
2527 struct internalvar *var;
2528 int varseen = 0;
2529 struct value_print_options opts;
2530
2531 get_user_print_options (&opts);
2532 for (var = internalvars; var; var = var->next)
2533 {
2534
2535 if (!varseen)
2536 {
2537 varseen = 1;
2538 }
2539 printf_filtered (("$%s = "), var->name);
2540
2541 TRY
2542 {
2543 struct value *val;
2544
2545 val = value_of_internalvar (gdbarch, var);
2546 value_print (val, gdb_stdout, &opts);
2547 }
2548 CATCH (ex, RETURN_MASK_ERROR)
2549 {
2550 fprintf_filtered (gdb_stdout, _("<error: %s>"), ex.message);
2551 }
2552 END_CATCH
2553
2554 printf_filtered (("\n"));
2555 }
2556 if (!varseen)
2557 {
2558 /* This text does not mention convenience functions on purpose.
2559 The user can't create them except via Python, and if Python support
2560 is installed this message will never be printed ($_streq will
2561 exist). */
2562 printf_unfiltered (_("No debugger convenience variables now defined.\n"
2563 "Convenience variables have "
2564 "names starting with \"$\";\n"
2565 "use \"set\" as in \"set "
2566 "$foo = 5\" to define them.\n"));
2567 }
2568 }
2569 \f
2570
2571 /* See value.h. */
2572
2573 struct value *
2574 value_from_xmethod (xmethod_worker_up &&worker)
2575 {
2576 struct value *v;
2577
2578 v = allocate_value (builtin_type (target_gdbarch ())->xmethod);
2579 v->lval = lval_xcallable;
2580 v->location.xm_worker = worker.release ();
2581 v->modifiable = 0;
2582
2583 return v;
2584 }
2585
2586 /* Return the type of the result of TYPE_CODE_XMETHOD value METHOD. */
2587
2588 struct type *
2589 result_type_of_xmethod (struct value *method, int argc, struct value **argv)
2590 {
2591 gdb_assert (TYPE_CODE (value_type (method)) == TYPE_CODE_XMETHOD
2592 && method->lval == lval_xcallable && argc > 0);
2593
2594 return method->location.xm_worker->get_result_type
2595 (argv[0], argv + 1, argc - 1);
2596 }
2597
2598 /* Call the xmethod corresponding to the TYPE_CODE_XMETHOD value METHOD. */
2599
2600 struct value *
2601 call_xmethod (struct value *method, int argc, struct value **argv)
2602 {
2603 gdb_assert (TYPE_CODE (value_type (method)) == TYPE_CODE_XMETHOD
2604 && method->lval == lval_xcallable && argc > 0);
2605
2606 return method->location.xm_worker->invoke (argv[0], argv + 1, argc - 1);
2607 }
2608 \f
2609 /* Extract a value as a C number (either long or double).
2610 Knows how to convert fixed values to double, or
2611 floating values to long.
2612 Does not deallocate the value. */
2613
2614 LONGEST
2615 value_as_long (struct value *val)
2616 {
2617 /* This coerces arrays and functions, which is necessary (e.g.
2618 in disassemble_command). It also dereferences references, which
2619 I suspect is the most logical thing to do. */
2620 val = coerce_array (val);
2621 return unpack_long (value_type (val), value_contents (val));
2622 }
2623
2624 /* Extract a value as a C pointer. Does not deallocate the value.
2625 Note that val's type may not actually be a pointer; value_as_long
2626 handles all the cases. */
2627 CORE_ADDR
2628 value_as_address (struct value *val)
2629 {
2630 struct gdbarch *gdbarch = get_type_arch (value_type (val));
2631
2632 /* Assume a CORE_ADDR can fit in a LONGEST (for now). Not sure
2633 whether we want this to be true eventually. */
2634 #if 0
2635 /* gdbarch_addr_bits_remove is wrong if we are being called for a
2636 non-address (e.g. argument to "signal", "info break", etc.), or
2637 for pointers to char, in which the low bits *are* significant. */
2638 return gdbarch_addr_bits_remove (gdbarch, value_as_long (val));
2639 #else
2640
2641 /* There are several targets (IA-64, PowerPC, and others) which
2642 don't represent pointers to functions as simply the address of
2643 the function's entry point. For example, on the IA-64, a
2644 function pointer points to a two-word descriptor, generated by
2645 the linker, which contains the function's entry point, and the
2646 value the IA-64 "global pointer" register should have --- to
2647 support position-independent code. The linker generates
2648 descriptors only for those functions whose addresses are taken.
2649
2650 On such targets, it's difficult for GDB to convert an arbitrary
2651 function address into a function pointer; it has to either find
2652 an existing descriptor for that function, or call malloc and
2653 build its own. On some targets, it is impossible for GDB to
2654 build a descriptor at all: the descriptor must contain a jump
2655 instruction; data memory cannot be executed; and code memory
2656 cannot be modified.
2657
2658 Upon entry to this function, if VAL is a value of type `function'
2659 (that is, TYPE_CODE (VALUE_TYPE (val)) == TYPE_CODE_FUNC), then
2660 value_address (val) is the address of the function. This is what
2661 you'll get if you evaluate an expression like `main'. The call
2662 to COERCE_ARRAY below actually does all the usual unary
2663 conversions, which includes converting values of type `function'
2664 to `pointer to function'. This is the challenging conversion
2665 discussed above. Then, `unpack_long' will convert that pointer
2666 back into an address.
2667
2668 So, suppose the user types `disassemble foo' on an architecture
2669 with a strange function pointer representation, on which GDB
2670 cannot build its own descriptors, and suppose further that `foo'
2671 has no linker-built descriptor. The address->pointer conversion
2672 will signal an error and prevent the command from running, even
2673 though the next step would have been to convert the pointer
2674 directly back into the same address.
2675
2676 The following shortcut avoids this whole mess. If VAL is a
2677 function, just return its address directly. */
2678 if (TYPE_CODE (value_type (val)) == TYPE_CODE_FUNC
2679 || TYPE_CODE (value_type (val)) == TYPE_CODE_METHOD)
2680 return value_address (val);
2681
2682 val = coerce_array (val);
2683
2684 /* Some architectures (e.g. Harvard), map instruction and data
2685 addresses onto a single large unified address space. For
2686 instance: An architecture may consider a large integer in the
2687 range 0x10000000 .. 0x1000ffff to already represent a data
2688 addresses (hence not need a pointer to address conversion) while
2689 a small integer would still need to be converted integer to
2690 pointer to address. Just assume such architectures handle all
2691 integer conversions in a single function. */
2692
2693 /* JimB writes:
2694
2695 I think INTEGER_TO_ADDRESS is a good idea as proposed --- but we
2696 must admonish GDB hackers to make sure its behavior matches the
2697 compiler's, whenever possible.
2698
2699 In general, I think GDB should evaluate expressions the same way
2700 the compiler does. When the user copies an expression out of
2701 their source code and hands it to a `print' command, they should
2702 get the same value the compiler would have computed. Any
2703 deviation from this rule can cause major confusion and annoyance,
2704 and needs to be justified carefully. In other words, GDB doesn't
2705 really have the freedom to do these conversions in clever and
2706 useful ways.
2707
2708 AndrewC pointed out that users aren't complaining about how GDB
2709 casts integers to pointers; they are complaining that they can't
2710 take an address from a disassembly listing and give it to `x/i'.
2711 This is certainly important.
2712
2713 Adding an architecture method like integer_to_address() certainly
2714 makes it possible for GDB to "get it right" in all circumstances
2715 --- the target has complete control over how things get done, so
2716 people can Do The Right Thing for their target without breaking
2717 anyone else. The standard doesn't specify how integers get
2718 converted to pointers; usually, the ABI doesn't either, but
2719 ABI-specific code is a more reasonable place to handle it. */
2720
2721 if (TYPE_CODE (value_type (val)) != TYPE_CODE_PTR
2722 && !TYPE_IS_REFERENCE (value_type (val))
2723 && gdbarch_integer_to_address_p (gdbarch))
2724 return gdbarch_integer_to_address (gdbarch, value_type (val),
2725 value_contents (val));
2726
2727 return unpack_long (value_type (val), value_contents (val));
2728 #endif
2729 }
2730 \f
2731 /* Unpack raw data (copied from debugee, target byte order) at VALADDR
2732 as a long, or as a double, assuming the raw data is described
2733 by type TYPE. Knows how to convert different sizes of values
2734 and can convert between fixed and floating point. We don't assume
2735 any alignment for the raw data. Return value is in host byte order.
2736
2737 If you want functions and arrays to be coerced to pointers, and
2738 references to be dereferenced, call value_as_long() instead.
2739
2740 C++: It is assumed that the front-end has taken care of
2741 all matters concerning pointers to members. A pointer
2742 to member which reaches here is considered to be equivalent
2743 to an INT (or some size). After all, it is only an offset. */
2744
2745 LONGEST
2746 unpack_long (struct type *type, const gdb_byte *valaddr)
2747 {
2748 enum bfd_endian byte_order = gdbarch_byte_order (get_type_arch (type));
2749 enum type_code code = TYPE_CODE (type);
2750 int len = TYPE_LENGTH (type);
2751 int nosign = TYPE_UNSIGNED (type);
2752
2753 switch (code)
2754 {
2755 case TYPE_CODE_TYPEDEF:
2756 return unpack_long (check_typedef (type), valaddr);
2757 case TYPE_CODE_ENUM:
2758 case TYPE_CODE_FLAGS:
2759 case TYPE_CODE_BOOL:
2760 case TYPE_CODE_INT:
2761 case TYPE_CODE_CHAR:
2762 case TYPE_CODE_RANGE:
2763 case TYPE_CODE_MEMBERPTR:
2764 if (nosign)
2765 return extract_unsigned_integer (valaddr, len, byte_order);
2766 else
2767 return extract_signed_integer (valaddr, len, byte_order);
2768
2769 case TYPE_CODE_FLT:
2770 case TYPE_CODE_DECFLOAT:
2771 return target_float_to_longest (valaddr, type);
2772
2773 case TYPE_CODE_PTR:
2774 case TYPE_CODE_REF:
2775 case TYPE_CODE_RVALUE_REF:
2776 /* Assume a CORE_ADDR can fit in a LONGEST (for now). Not sure
2777 whether we want this to be true eventually. */
2778 return extract_typed_address (valaddr, type);
2779
2780 default:
2781 error (_("Value can't be converted to integer."));
2782 }
2783 return 0; /* Placate lint. */
2784 }
2785
2786 /* Unpack raw data (copied from debugee, target byte order) at VALADDR
2787 as a CORE_ADDR, assuming the raw data is described by type TYPE.
2788 We don't assume any alignment for the raw data. Return value is in
2789 host byte order.
2790
2791 If you want functions and arrays to be coerced to pointers, and
2792 references to be dereferenced, call value_as_address() instead.
2793
2794 C++: It is assumed that the front-end has taken care of
2795 all matters concerning pointers to members. A pointer
2796 to member which reaches here is considered to be equivalent
2797 to an INT (or some size). After all, it is only an offset. */
2798
2799 CORE_ADDR
2800 unpack_pointer (struct type *type, const gdb_byte *valaddr)
2801 {
2802 /* Assume a CORE_ADDR can fit in a LONGEST (for now). Not sure
2803 whether we want this to be true eventually. */
2804 return unpack_long (type, valaddr);
2805 }
2806
2807 bool
2808 is_floating_value (struct value *val)
2809 {
2810 struct type *type = check_typedef (value_type (val));
2811
2812 if (is_floating_type (type))
2813 {
2814 if (!target_float_is_valid (value_contents (val), type))
2815 error (_("Invalid floating value found in program."));
2816 return true;
2817 }
2818
2819 return false;
2820 }
2821
2822 \f
2823 /* Get the value of the FIELDNO'th field (which must be static) of
2824 TYPE. */
2825
2826 struct value *
2827 value_static_field (struct type *type, int fieldno)
2828 {
2829 struct value *retval;
2830
2831 switch (TYPE_FIELD_LOC_KIND (type, fieldno))
2832 {
2833 case FIELD_LOC_KIND_PHYSADDR:
2834 retval = value_at_lazy (TYPE_FIELD_TYPE (type, fieldno),
2835 TYPE_FIELD_STATIC_PHYSADDR (type, fieldno));
2836 break;
2837 case FIELD_LOC_KIND_PHYSNAME:
2838 {
2839 const char *phys_name = TYPE_FIELD_STATIC_PHYSNAME (type, fieldno);
2840 /* TYPE_FIELD_NAME (type, fieldno); */
2841 struct block_symbol sym = lookup_symbol (phys_name, 0, VAR_DOMAIN, 0);
2842
2843 if (sym.symbol == NULL)
2844 {
2845 /* With some compilers, e.g. HP aCC, static data members are
2846 reported as non-debuggable symbols. */
2847 struct bound_minimal_symbol msym
2848 = lookup_minimal_symbol (phys_name, NULL, NULL);
2849 struct type *field_type = TYPE_FIELD_TYPE (type, fieldno);
2850
2851 if (!msym.minsym)
2852 retval = allocate_optimized_out_value (field_type);
2853 else
2854 retval = value_at_lazy (field_type, BMSYMBOL_VALUE_ADDRESS (msym));
2855 }
2856 else
2857 retval = value_of_variable (sym.symbol, sym.block);
2858 break;
2859 }
2860 default:
2861 gdb_assert_not_reached ("unexpected field location kind");
2862 }
2863
2864 return retval;
2865 }
2866
2867 /* Change the enclosing type of a value object VAL to NEW_ENCL_TYPE.
2868 You have to be careful here, since the size of the data area for the value
2869 is set by the length of the enclosing type. So if NEW_ENCL_TYPE is bigger
2870 than the old enclosing type, you have to allocate more space for the
2871 data. */
2872
2873 void
2874 set_value_enclosing_type (struct value *val, struct type *new_encl_type)
2875 {
2876 if (TYPE_LENGTH (new_encl_type) > TYPE_LENGTH (value_enclosing_type (val)))
2877 {
2878 check_type_length_before_alloc (new_encl_type);
2879 val->contents
2880 = (gdb_byte *) xrealloc (val->contents, TYPE_LENGTH (new_encl_type));
2881 }
2882
2883 val->enclosing_type = new_encl_type;
2884 }
2885
2886 /* Given a value ARG1 (offset by OFFSET bytes)
2887 of a struct or union type ARG_TYPE,
2888 extract and return the value of one of its (non-static) fields.
2889 FIELDNO says which field. */
2890
2891 struct value *
2892 value_primitive_field (struct value *arg1, LONGEST offset,
2893 int fieldno, struct type *arg_type)
2894 {
2895 struct value *v;
2896 struct type *type;
2897 struct gdbarch *arch = get_value_arch (arg1);
2898 int unit_size = gdbarch_addressable_memory_unit_size (arch);
2899
2900 arg_type = check_typedef (arg_type);
2901 type = TYPE_FIELD_TYPE (arg_type, fieldno);
2902
2903 /* Call check_typedef on our type to make sure that, if TYPE
2904 is a TYPE_CODE_TYPEDEF, its length is set to the length
2905 of the target type instead of zero. However, we do not
2906 replace the typedef type by the target type, because we want
2907 to keep the typedef in order to be able to print the type
2908 description correctly. */
2909 check_typedef (type);
2910
2911 if (TYPE_FIELD_BITSIZE (arg_type, fieldno))
2912 {
2913 /* Handle packed fields.
2914
2915 Create a new value for the bitfield, with bitpos and bitsize
2916 set. If possible, arrange offset and bitpos so that we can
2917 do a single aligned read of the size of the containing type.
2918 Otherwise, adjust offset to the byte containing the first
2919 bit. Assume that the address, offset, and embedded offset
2920 are sufficiently aligned. */
2921
2922 LONGEST bitpos = TYPE_FIELD_BITPOS (arg_type, fieldno);
2923 LONGEST container_bitsize = TYPE_LENGTH (type) * 8;
2924
2925 v = allocate_value_lazy (type);
2926 v->bitsize = TYPE_FIELD_BITSIZE (arg_type, fieldno);
2927 if ((bitpos % container_bitsize) + v->bitsize <= container_bitsize
2928 && TYPE_LENGTH (type) <= (int) sizeof (LONGEST))
2929 v->bitpos = bitpos % container_bitsize;
2930 else
2931 v->bitpos = bitpos % 8;
2932 v->offset = (value_embedded_offset (arg1)
2933 + offset
2934 + (bitpos - v->bitpos) / 8);
2935 set_value_parent (v, arg1);
2936 if (!value_lazy (arg1))
2937 value_fetch_lazy (v);
2938 }
2939 else if (fieldno < TYPE_N_BASECLASSES (arg_type))
2940 {
2941 /* This field is actually a base subobject, so preserve the
2942 entire object's contents for later references to virtual
2943 bases, etc. */
2944 LONGEST boffset;
2945
2946 /* Lazy register values with offsets are not supported. */
2947 if (VALUE_LVAL (arg1) == lval_register && value_lazy (arg1))
2948 value_fetch_lazy (arg1);
2949
2950 /* We special case virtual inheritance here because this
2951 requires access to the contents, which we would rather avoid
2952 for references to ordinary fields of unavailable values. */
2953 if (BASETYPE_VIA_VIRTUAL (arg_type, fieldno))
2954 boffset = baseclass_offset (arg_type, fieldno,
2955 value_contents (arg1),
2956 value_embedded_offset (arg1),
2957 value_address (arg1),
2958 arg1);
2959 else
2960 boffset = TYPE_FIELD_BITPOS (arg_type, fieldno) / 8;
2961
2962 if (value_lazy (arg1))
2963 v = allocate_value_lazy (value_enclosing_type (arg1));
2964 else
2965 {
2966 v = allocate_value (value_enclosing_type (arg1));
2967 value_contents_copy_raw (v, 0, arg1, 0,
2968 TYPE_LENGTH (value_enclosing_type (arg1)));
2969 }
2970 v->type = type;
2971 v->offset = value_offset (arg1);
2972 v->embedded_offset = offset + value_embedded_offset (arg1) + boffset;
2973 }
2974 else if (NULL != TYPE_DATA_LOCATION (type))
2975 {
2976 /* Field is a dynamic data member. */
2977
2978 gdb_assert (0 == offset);
2979 /* We expect an already resolved data location. */
2980 gdb_assert (PROP_CONST == TYPE_DATA_LOCATION_KIND (type));
2981 /* For dynamic data types defer memory allocation
2982 until we actual access the value. */
2983 v = allocate_value_lazy (type);
2984 }
2985 else
2986 {
2987 /* Plain old data member */
2988 offset += (TYPE_FIELD_BITPOS (arg_type, fieldno)
2989 / (HOST_CHAR_BIT * unit_size));
2990
2991 /* Lazy register values with offsets are not supported. */
2992 if (VALUE_LVAL (arg1) == lval_register && value_lazy (arg1))
2993 value_fetch_lazy (arg1);
2994
2995 if (value_lazy (arg1))
2996 v = allocate_value_lazy (type);
2997 else
2998 {
2999 v = allocate_value (type);
3000 value_contents_copy_raw (v, value_embedded_offset (v),
3001 arg1, value_embedded_offset (arg1) + offset,
3002 type_length_units (type));
3003 }
3004 v->offset = (value_offset (arg1) + offset
3005 + value_embedded_offset (arg1));
3006 }
3007 set_value_component_location (v, arg1);
3008 return v;
3009 }
3010
3011 /* Given a value ARG1 of a struct or union type,
3012 extract and return the value of one of its (non-static) fields.
3013 FIELDNO says which field. */
3014
3015 struct value *
3016 value_field (struct value *arg1, int fieldno)
3017 {
3018 return value_primitive_field (arg1, 0, fieldno, value_type (arg1));
3019 }
3020
3021 /* Return a non-virtual function as a value.
3022 F is the list of member functions which contains the desired method.
3023 J is an index into F which provides the desired method.
3024
3025 We only use the symbol for its address, so be happy with either a
3026 full symbol or a minimal symbol. */
3027
3028 struct value *
3029 value_fn_field (struct value **arg1p, struct fn_field *f,
3030 int j, struct type *type,
3031 LONGEST offset)
3032 {
3033 struct value *v;
3034 struct type *ftype = TYPE_FN_FIELD_TYPE (f, j);
3035 const char *physname = TYPE_FN_FIELD_PHYSNAME (f, j);
3036 struct symbol *sym;
3037 struct bound_minimal_symbol msym;
3038
3039 sym = lookup_symbol (physname, 0, VAR_DOMAIN, 0).symbol;
3040 if (sym != NULL)
3041 {
3042 memset (&msym, 0, sizeof (msym));
3043 }
3044 else
3045 {
3046 gdb_assert (sym == NULL);
3047 msym = lookup_bound_minimal_symbol (physname);
3048 if (msym.minsym == NULL)
3049 return NULL;
3050 }
3051
3052 v = allocate_value (ftype);
3053 VALUE_LVAL (v) = lval_memory;
3054 if (sym)
3055 {
3056 set_value_address (v, BLOCK_START (SYMBOL_BLOCK_VALUE (sym)));
3057 }
3058 else
3059 {
3060 /* The minimal symbol might point to a function descriptor;
3061 resolve it to the actual code address instead. */
3062 struct objfile *objfile = msym.objfile;
3063 struct gdbarch *gdbarch = get_objfile_arch (objfile);
3064
3065 set_value_address (v,
3066 gdbarch_convert_from_func_ptr_addr
3067 (gdbarch, BMSYMBOL_VALUE_ADDRESS (msym), &current_target));
3068 }
3069
3070 if (arg1p)
3071 {
3072 if (type != value_type (*arg1p))
3073 *arg1p = value_ind (value_cast (lookup_pointer_type (type),
3074 value_addr (*arg1p)));
3075
3076 /* Move the `this' pointer according to the offset.
3077 VALUE_OFFSET (*arg1p) += offset; */
3078 }
3079
3080 return v;
3081 }
3082
3083 \f
3084
3085 /* Unpack a bitfield of the specified FIELD_TYPE, from the object at
3086 VALADDR, and store the result in *RESULT.
3087 The bitfield starts at BITPOS bits and contains BITSIZE bits; if
3088 BITSIZE is zero, then the length is taken from FIELD_TYPE.
3089
3090 Extracting bits depends on endianness of the machine. Compute the
3091 number of least significant bits to discard. For big endian machines,
3092 we compute the total number of bits in the anonymous object, subtract
3093 off the bit count from the MSB of the object to the MSB of the
3094 bitfield, then the size of the bitfield, which leaves the LSB discard
3095 count. For little endian machines, the discard count is simply the
3096 number of bits from the LSB of the anonymous object to the LSB of the
3097 bitfield.
3098
3099 If the field is signed, we also do sign extension. */
3100
3101 static LONGEST
3102 unpack_bits_as_long (struct type *field_type, const gdb_byte *valaddr,
3103 LONGEST bitpos, LONGEST bitsize)
3104 {
3105 enum bfd_endian byte_order = gdbarch_byte_order (get_type_arch (field_type));
3106 ULONGEST val;
3107 ULONGEST valmask;
3108 int lsbcount;
3109 LONGEST bytes_read;
3110 LONGEST read_offset;
3111
3112 /* Read the minimum number of bytes required; there may not be
3113 enough bytes to read an entire ULONGEST. */
3114 field_type = check_typedef (field_type);
3115 if (bitsize)
3116 bytes_read = ((bitpos % 8) + bitsize + 7) / 8;
3117 else
3118 {
3119 bytes_read = TYPE_LENGTH (field_type);
3120 bitsize = 8 * bytes_read;
3121 }
3122
3123 read_offset = bitpos / 8;
3124
3125 val = extract_unsigned_integer (valaddr + read_offset,
3126 bytes_read, byte_order);
3127
3128 /* Extract bits. See comment above. */
3129
3130 if (gdbarch_bits_big_endian (get_type_arch (field_type)))
3131 lsbcount = (bytes_read * 8 - bitpos % 8 - bitsize);
3132 else
3133 lsbcount = (bitpos % 8);
3134 val >>= lsbcount;
3135
3136 /* If the field does not entirely fill a LONGEST, then zero the sign bits.
3137 If the field is signed, and is negative, then sign extend. */
3138
3139 if (bitsize < 8 * (int) sizeof (val))
3140 {
3141 valmask = (((ULONGEST) 1) << bitsize) - 1;
3142 val &= valmask;
3143 if (!TYPE_UNSIGNED (field_type))
3144 {
3145 if (val & (valmask ^ (valmask >> 1)))
3146 {
3147 val |= ~valmask;
3148 }
3149 }
3150 }
3151
3152 return val;
3153 }
3154
3155 /* Unpack a field FIELDNO of the specified TYPE, from the object at
3156 VALADDR + EMBEDDED_OFFSET. VALADDR points to the contents of
3157 ORIGINAL_VALUE, which must not be NULL. See
3158 unpack_value_bits_as_long for more details. */
3159
3160 int
3161 unpack_value_field_as_long (struct type *type, const gdb_byte *valaddr,
3162 LONGEST embedded_offset, int fieldno,
3163 const struct value *val, LONGEST *result)
3164 {
3165 int bitpos = TYPE_FIELD_BITPOS (type, fieldno);
3166 int bitsize = TYPE_FIELD_BITSIZE (type, fieldno);
3167 struct type *field_type = TYPE_FIELD_TYPE (type, fieldno);
3168 int bit_offset;
3169
3170 gdb_assert (val != NULL);
3171
3172 bit_offset = embedded_offset * TARGET_CHAR_BIT + bitpos;
3173 if (value_bits_any_optimized_out (val, bit_offset, bitsize)
3174 || !value_bits_available (val, bit_offset, bitsize))
3175 return 0;
3176
3177 *result = unpack_bits_as_long (field_type, valaddr + embedded_offset,
3178 bitpos, bitsize);
3179 return 1;
3180 }
3181
3182 /* Unpack a field FIELDNO of the specified TYPE, from the anonymous
3183 object at VALADDR. See unpack_bits_as_long for more details. */
3184
3185 LONGEST
3186 unpack_field_as_long (struct type *type, const gdb_byte *valaddr, int fieldno)
3187 {
3188 int bitpos = TYPE_FIELD_BITPOS (type, fieldno);
3189 int bitsize = TYPE_FIELD_BITSIZE (type, fieldno);
3190 struct type *field_type = TYPE_FIELD_TYPE (type, fieldno);
3191
3192 return unpack_bits_as_long (field_type, valaddr, bitpos, bitsize);
3193 }
3194
3195 /* Unpack a bitfield of BITSIZE bits found at BITPOS in the object at
3196 VALADDR + EMBEDDEDOFFSET that has the type of DEST_VAL and store
3197 the contents in DEST_VAL, zero or sign extending if the type of
3198 DEST_VAL is wider than BITSIZE. VALADDR points to the contents of
3199 VAL. If the VAL's contents required to extract the bitfield from
3200 are unavailable/optimized out, DEST_VAL is correspondingly
3201 marked unavailable/optimized out. */
3202
3203 void
3204 unpack_value_bitfield (struct value *dest_val,
3205 LONGEST bitpos, LONGEST bitsize,
3206 const gdb_byte *valaddr, LONGEST embedded_offset,
3207 const struct value *val)
3208 {
3209 enum bfd_endian byte_order;
3210 int src_bit_offset;
3211 int dst_bit_offset;
3212 struct type *field_type = value_type (dest_val);
3213
3214 byte_order = gdbarch_byte_order (get_type_arch (field_type));
3215
3216 /* First, unpack and sign extend the bitfield as if it was wholly
3217 valid. Optimized out/unavailable bits are read as zero, but
3218 that's OK, as they'll end up marked below. If the VAL is
3219 wholly-invalid we may have skipped allocating its contents,
3220 though. See allocate_optimized_out_value. */
3221 if (valaddr != NULL)
3222 {
3223 LONGEST num;
3224
3225 num = unpack_bits_as_long (field_type, valaddr + embedded_offset,
3226 bitpos, bitsize);
3227 store_signed_integer (value_contents_raw (dest_val),
3228 TYPE_LENGTH (field_type), byte_order, num);
3229 }
3230
3231 /* Now copy the optimized out / unavailability ranges to the right
3232 bits. */
3233 src_bit_offset = embedded_offset * TARGET_CHAR_BIT + bitpos;
3234 if (byte_order == BFD_ENDIAN_BIG)
3235 dst_bit_offset = TYPE_LENGTH (field_type) * TARGET_CHAR_BIT - bitsize;
3236 else
3237 dst_bit_offset = 0;
3238 value_ranges_copy_adjusted (dest_val, dst_bit_offset,
3239 val, src_bit_offset, bitsize);
3240 }
3241
3242 /* Return a new value with type TYPE, which is FIELDNO field of the
3243 object at VALADDR + EMBEDDEDOFFSET. VALADDR points to the contents
3244 of VAL. If the VAL's contents required to extract the bitfield
3245 from are unavailable/optimized out, the new value is
3246 correspondingly marked unavailable/optimized out. */
3247
3248 struct value *
3249 value_field_bitfield (struct type *type, int fieldno,
3250 const gdb_byte *valaddr,
3251 LONGEST embedded_offset, const struct value *val)
3252 {
3253 int bitpos = TYPE_FIELD_BITPOS (type, fieldno);
3254 int bitsize = TYPE_FIELD_BITSIZE (type, fieldno);
3255 struct value *res_val = allocate_value (TYPE_FIELD_TYPE (type, fieldno));
3256
3257 unpack_value_bitfield (res_val, bitpos, bitsize,
3258 valaddr, embedded_offset, val);
3259
3260 return res_val;
3261 }
3262
3263 /* Modify the value of a bitfield. ADDR points to a block of memory in
3264 target byte order; the bitfield starts in the byte pointed to. FIELDVAL
3265 is the desired value of the field, in host byte order. BITPOS and BITSIZE
3266 indicate which bits (in target bit order) comprise the bitfield.
3267 Requires 0 < BITSIZE <= lbits, 0 <= BITPOS % 8 + BITSIZE <= lbits, and
3268 0 <= BITPOS, where lbits is the size of a LONGEST in bits. */
3269
3270 void
3271 modify_field (struct type *type, gdb_byte *addr,
3272 LONGEST fieldval, LONGEST bitpos, LONGEST bitsize)
3273 {
3274 enum bfd_endian byte_order = gdbarch_byte_order (get_type_arch (type));
3275 ULONGEST oword;
3276 ULONGEST mask = (ULONGEST) -1 >> (8 * sizeof (ULONGEST) - bitsize);
3277 LONGEST bytesize;
3278
3279 /* Normalize BITPOS. */
3280 addr += bitpos / 8;
3281 bitpos %= 8;
3282
3283 /* If a negative fieldval fits in the field in question, chop
3284 off the sign extension bits. */
3285 if ((~fieldval & ~(mask >> 1)) == 0)
3286 fieldval &= mask;
3287
3288 /* Warn if value is too big to fit in the field in question. */
3289 if (0 != (fieldval & ~mask))
3290 {
3291 /* FIXME: would like to include fieldval in the message, but
3292 we don't have a sprintf_longest. */
3293 warning (_("Value does not fit in %s bits."), plongest (bitsize));
3294
3295 /* Truncate it, otherwise adjoining fields may be corrupted. */
3296 fieldval &= mask;
3297 }
3298
3299 /* Ensure no bytes outside of the modified ones get accessed as it may cause
3300 false valgrind reports. */
3301
3302 bytesize = (bitpos + bitsize + 7) / 8;
3303 oword = extract_unsigned_integer (addr, bytesize, byte_order);
3304
3305 /* Shifting for bit field depends on endianness of the target machine. */
3306 if (gdbarch_bits_big_endian (get_type_arch (type)))
3307 bitpos = bytesize * 8 - bitpos - bitsize;
3308
3309 oword &= ~(mask << bitpos);
3310 oword |= fieldval << bitpos;
3311
3312 store_unsigned_integer (addr, bytesize, byte_order, oword);
3313 }
3314 \f
3315 /* Pack NUM into BUF using a target format of TYPE. */
3316
3317 void
3318 pack_long (gdb_byte *buf, struct type *type, LONGEST num)
3319 {
3320 enum bfd_endian byte_order = gdbarch_byte_order (get_type_arch (type));
3321 LONGEST len;
3322
3323 type = check_typedef (type);
3324 len = TYPE_LENGTH (type);
3325
3326 switch (TYPE_CODE (type))
3327 {
3328 case TYPE_CODE_INT:
3329 case TYPE_CODE_CHAR:
3330 case TYPE_CODE_ENUM:
3331 case TYPE_CODE_FLAGS:
3332 case TYPE_CODE_BOOL:
3333 case TYPE_CODE_RANGE:
3334 case TYPE_CODE_MEMBERPTR:
3335 store_signed_integer (buf, len, byte_order, num);
3336 break;
3337
3338 case TYPE_CODE_REF:
3339 case TYPE_CODE_RVALUE_REF:
3340 case TYPE_CODE_PTR:
3341 store_typed_address (buf, type, (CORE_ADDR) num);
3342 break;
3343
3344 case TYPE_CODE_FLT:
3345 case TYPE_CODE_DECFLOAT:
3346 target_float_from_longest (buf, type, num);
3347 break;
3348
3349 default:
3350 error (_("Unexpected type (%d) encountered for integer constant."),
3351 TYPE_CODE (type));
3352 }
3353 }
3354
3355
3356 /* Pack NUM into BUF using a target format of TYPE. */
3357
3358 static void
3359 pack_unsigned_long (gdb_byte *buf, struct type *type, ULONGEST num)
3360 {
3361 LONGEST len;
3362 enum bfd_endian byte_order;
3363
3364 type = check_typedef (type);
3365 len = TYPE_LENGTH (type);
3366 byte_order = gdbarch_byte_order (get_type_arch (type));
3367
3368 switch (TYPE_CODE (type))
3369 {
3370 case TYPE_CODE_INT:
3371 case TYPE_CODE_CHAR:
3372 case TYPE_CODE_ENUM:
3373 case TYPE_CODE_FLAGS:
3374 case TYPE_CODE_BOOL:
3375 case TYPE_CODE_RANGE:
3376 case TYPE_CODE_MEMBERPTR:
3377 store_unsigned_integer (buf, len, byte_order, num);
3378 break;
3379
3380 case TYPE_CODE_REF:
3381 case TYPE_CODE_RVALUE_REF:
3382 case TYPE_CODE_PTR:
3383 store_typed_address (buf, type, (CORE_ADDR) num);
3384 break;
3385
3386 case TYPE_CODE_FLT:
3387 case TYPE_CODE_DECFLOAT:
3388 target_float_from_ulongest (buf, type, num);
3389 break;
3390
3391 default:
3392 error (_("Unexpected type (%d) encountered "
3393 "for unsigned integer constant."),
3394 TYPE_CODE (type));
3395 }
3396 }
3397
3398
3399 /* Convert C numbers into newly allocated values. */
3400
3401 struct value *
3402 value_from_longest (struct type *type, LONGEST num)
3403 {
3404 struct value *val = allocate_value (type);
3405
3406 pack_long (value_contents_raw (val), type, num);
3407 return val;
3408 }
3409
3410
3411 /* Convert C unsigned numbers into newly allocated values. */
3412
3413 struct value *
3414 value_from_ulongest (struct type *type, ULONGEST num)
3415 {
3416 struct value *val = allocate_value (type);
3417
3418 pack_unsigned_long (value_contents_raw (val), type, num);
3419
3420 return val;
3421 }
3422
3423
3424 /* Create a value representing a pointer of type TYPE to the address
3425 ADDR. */
3426
3427 struct value *
3428 value_from_pointer (struct type *type, CORE_ADDR addr)
3429 {
3430 struct value *val = allocate_value (type);
3431
3432 store_typed_address (value_contents_raw (val),
3433 check_typedef (type), addr);
3434 return val;
3435 }
3436
3437
3438 /* Create a value of type TYPE whose contents come from VALADDR, if it
3439 is non-null, and whose memory address (in the inferior) is
3440 ADDRESS. The type of the created value may differ from the passed
3441 type TYPE. Make sure to retrieve values new type after this call.
3442 Note that TYPE is not passed through resolve_dynamic_type; this is
3443 a special API intended for use only by Ada. */
3444
3445 struct value *
3446 value_from_contents_and_address_unresolved (struct type *type,
3447 const gdb_byte *valaddr,
3448 CORE_ADDR address)
3449 {
3450 struct value *v;
3451
3452 if (valaddr == NULL)
3453 v = allocate_value_lazy (type);
3454 else
3455 v = value_from_contents (type, valaddr);
3456 VALUE_LVAL (v) = lval_memory;
3457 set_value_address (v, address);
3458 return v;
3459 }
3460
3461 /* Create a value of type TYPE whose contents come from VALADDR, if it
3462 is non-null, and whose memory address (in the inferior) is
3463 ADDRESS. The type of the created value may differ from the passed
3464 type TYPE. Make sure to retrieve values new type after this call. */
3465
3466 struct value *
3467 value_from_contents_and_address (struct type *type,
3468 const gdb_byte *valaddr,
3469 CORE_ADDR address)
3470 {
3471 struct type *resolved_type = resolve_dynamic_type (type, valaddr, address);
3472 struct type *resolved_type_no_typedef = check_typedef (resolved_type);
3473 struct value *v;
3474
3475 if (valaddr == NULL)
3476 v = allocate_value_lazy (resolved_type);
3477 else
3478 v = value_from_contents (resolved_type, valaddr);
3479 if (TYPE_DATA_LOCATION (resolved_type_no_typedef) != NULL
3480 && TYPE_DATA_LOCATION_KIND (resolved_type_no_typedef) == PROP_CONST)
3481 address = TYPE_DATA_LOCATION_ADDR (resolved_type_no_typedef);
3482 VALUE_LVAL (v) = lval_memory;
3483 set_value_address (v, address);
3484 return v;
3485 }
3486
3487 /* Create a value of type TYPE holding the contents CONTENTS.
3488 The new value is `not_lval'. */
3489
3490 struct value *
3491 value_from_contents (struct type *type, const gdb_byte *contents)
3492 {
3493 struct value *result;
3494
3495 result = allocate_value (type);
3496 memcpy (value_contents_raw (result), contents, TYPE_LENGTH (type));
3497 return result;
3498 }
3499
3500 /* Extract a value from the history file. Input will be of the form
3501 $digits or $$digits. See block comment above 'write_dollar_variable'
3502 for details. */
3503
3504 struct value *
3505 value_from_history_ref (const char *h, const char **endp)
3506 {
3507 int index, len;
3508
3509 if (h[0] == '$')
3510 len = 1;
3511 else
3512 return NULL;
3513
3514 if (h[1] == '$')
3515 len = 2;
3516
3517 /* Find length of numeral string. */
3518 for (; isdigit (h[len]); len++)
3519 ;
3520
3521 /* Make sure numeral string is not part of an identifier. */
3522 if (h[len] == '_' || isalpha (h[len]))
3523 return NULL;
3524
3525 /* Now collect the index value. */
3526 if (h[1] == '$')
3527 {
3528 if (len == 2)
3529 {
3530 /* For some bizarre reason, "$$" is equivalent to "$$1",
3531 rather than to "$$0" as it ought to be! */
3532 index = -1;
3533 *endp += len;
3534 }
3535 else
3536 {
3537 char *local_end;
3538
3539 index = -strtol (&h[2], &local_end, 10);
3540 *endp = local_end;
3541 }
3542 }
3543 else
3544 {
3545 if (len == 1)
3546 {
3547 /* "$" is equivalent to "$0". */
3548 index = 0;
3549 *endp += len;
3550 }
3551 else
3552 {
3553 char *local_end;
3554
3555 index = strtol (&h[1], &local_end, 10);
3556 *endp = local_end;
3557 }
3558 }
3559
3560 return access_value_history (index);
3561 }
3562
3563 /* Get the component value (offset by OFFSET bytes) of a struct or
3564 union WHOLE. Component's type is TYPE. */
3565
3566 struct value *
3567 value_from_component (struct value *whole, struct type *type, LONGEST offset)
3568 {
3569 struct value *v;
3570
3571 if (VALUE_LVAL (whole) == lval_memory && value_lazy (whole))
3572 v = allocate_value_lazy (type);
3573 else
3574 {
3575 v = allocate_value (type);
3576 value_contents_copy (v, value_embedded_offset (v),
3577 whole, value_embedded_offset (whole) + offset,
3578 type_length_units (type));
3579 }
3580 v->offset = value_offset (whole) + offset + value_embedded_offset (whole);
3581 set_value_component_location (v, whole);
3582
3583 return v;
3584 }
3585
3586 struct value *
3587 coerce_ref_if_computed (const struct value *arg)
3588 {
3589 const struct lval_funcs *funcs;
3590
3591 if (!TYPE_IS_REFERENCE (check_typedef (value_type (arg))))
3592 return NULL;
3593
3594 if (value_lval_const (arg) != lval_computed)
3595 return NULL;
3596
3597 funcs = value_computed_funcs (arg);
3598 if (funcs->coerce_ref == NULL)
3599 return NULL;
3600
3601 return funcs->coerce_ref (arg);
3602 }
3603
3604 /* Look at value.h for description. */
3605
3606 struct value *
3607 readjust_indirect_value_type (struct value *value, struct type *enc_type,
3608 const struct type *original_type,
3609 const struct value *original_value)
3610 {
3611 /* Re-adjust type. */
3612 deprecated_set_value_type (value, TYPE_TARGET_TYPE (original_type));
3613
3614 /* Add embedding info. */
3615 set_value_enclosing_type (value, enc_type);
3616 set_value_embedded_offset (value, value_pointed_to_offset (original_value));
3617
3618 /* We may be pointing to an object of some derived type. */
3619 return value_full_object (value, NULL, 0, 0, 0);
3620 }
3621
3622 struct value *
3623 coerce_ref (struct value *arg)
3624 {
3625 struct type *value_type_arg_tmp = check_typedef (value_type (arg));
3626 struct value *retval;
3627 struct type *enc_type;
3628
3629 retval = coerce_ref_if_computed (arg);
3630 if (retval)
3631 return retval;
3632
3633 if (!TYPE_IS_REFERENCE (value_type_arg_tmp))
3634 return arg;
3635
3636 enc_type = check_typedef (value_enclosing_type (arg));
3637 enc_type = TYPE_TARGET_TYPE (enc_type);
3638
3639 retval = value_at_lazy (enc_type,
3640 unpack_pointer (value_type (arg),
3641 value_contents (arg)));
3642 enc_type = value_type (retval);
3643 return readjust_indirect_value_type (retval, enc_type,
3644 value_type_arg_tmp, arg);
3645 }
3646
3647 struct value *
3648 coerce_array (struct value *arg)
3649 {
3650 struct type *type;
3651
3652 arg = coerce_ref (arg);
3653 type = check_typedef (value_type (arg));
3654
3655 switch (TYPE_CODE (type))
3656 {
3657 case TYPE_CODE_ARRAY:
3658 if (!TYPE_VECTOR (type) && current_language->c_style_arrays)
3659 arg = value_coerce_array (arg);
3660 break;
3661 case TYPE_CODE_FUNC:
3662 arg = value_coerce_function (arg);
3663 break;
3664 }
3665 return arg;
3666 }
3667 \f
3668
3669 /* Return the return value convention that will be used for the
3670 specified type. */
3671
3672 enum return_value_convention
3673 struct_return_convention (struct gdbarch *gdbarch,
3674 struct value *function, struct type *value_type)
3675 {
3676 enum type_code code = TYPE_CODE (value_type);
3677
3678 if (code == TYPE_CODE_ERROR)
3679 error (_("Function return type unknown."));
3680
3681 /* Probe the architecture for the return-value convention. */
3682 return gdbarch_return_value (gdbarch, function, value_type,
3683 NULL, NULL, NULL);
3684 }
3685
3686 /* Return true if the function returning the specified type is using
3687 the convention of returning structures in memory (passing in the
3688 address as a hidden first parameter). */
3689
3690 int
3691 using_struct_return (struct gdbarch *gdbarch,
3692 struct value *function, struct type *value_type)
3693 {
3694 if (TYPE_CODE (value_type) == TYPE_CODE_VOID)
3695 /* A void return value is never in memory. See also corresponding
3696 code in "print_return_value". */
3697 return 0;
3698
3699 return (struct_return_convention (gdbarch, function, value_type)
3700 != RETURN_VALUE_REGISTER_CONVENTION);
3701 }
3702
3703 /* Set the initialized field in a value struct. */
3704
3705 void
3706 set_value_initialized (struct value *val, int status)
3707 {
3708 val->initialized = status;
3709 }
3710
3711 /* Return the initialized field in a value struct. */
3712
3713 int
3714 value_initialized (const struct value *val)
3715 {
3716 return val->initialized;
3717 }
3718
3719 /* Load the actual content of a lazy value. Fetch the data from the
3720 user's process and clear the lazy flag to indicate that the data in
3721 the buffer is valid.
3722
3723 If the value is zero-length, we avoid calling read_memory, which
3724 would abort. We mark the value as fetched anyway -- all 0 bytes of
3725 it. */
3726
3727 void
3728 value_fetch_lazy (struct value *val)
3729 {
3730 gdb_assert (value_lazy (val));
3731 allocate_value_contents (val);
3732 /* A value is either lazy, or fully fetched. The
3733 availability/validity is only established as we try to fetch a
3734 value. */
3735 gdb_assert (val->optimized_out.empty ());
3736 gdb_assert (val->unavailable.empty ());
3737 if (value_bitsize (val))
3738 {
3739 /* To read a lazy bitfield, read the entire enclosing value. This
3740 prevents reading the same block of (possibly volatile) memory once
3741 per bitfield. It would be even better to read only the containing
3742 word, but we have no way to record that just specific bits of a
3743 value have been fetched. */
3744 struct type *type = check_typedef (value_type (val));
3745 struct value *parent = value_parent (val);
3746
3747 if (value_lazy (parent))
3748 value_fetch_lazy (parent);
3749
3750 unpack_value_bitfield (val,
3751 value_bitpos (val), value_bitsize (val),
3752 value_contents_for_printing (parent),
3753 value_offset (val), parent);
3754 }
3755 else if (VALUE_LVAL (val) == lval_memory)
3756 {
3757 CORE_ADDR addr = value_address (val);
3758 struct type *type = check_typedef (value_enclosing_type (val));
3759
3760 if (TYPE_LENGTH (type))
3761 read_value_memory (val, 0, value_stack (val),
3762 addr, value_contents_all_raw (val),
3763 type_length_units (type));
3764 }
3765 else if (VALUE_LVAL (val) == lval_register)
3766 {
3767 struct frame_info *next_frame;
3768 int regnum;
3769 struct type *type = check_typedef (value_type (val));
3770 struct value *new_val = val, *mark = value_mark ();
3771
3772 /* Offsets are not supported here; lazy register values must
3773 refer to the entire register. */
3774 gdb_assert (value_offset (val) == 0);
3775
3776 while (VALUE_LVAL (new_val) == lval_register && value_lazy (new_val))
3777 {
3778 struct frame_id next_frame_id = VALUE_NEXT_FRAME_ID (new_val);
3779
3780 next_frame = frame_find_by_id (next_frame_id);
3781 regnum = VALUE_REGNUM (new_val);
3782
3783 gdb_assert (next_frame != NULL);
3784
3785 /* Convertible register routines are used for multi-register
3786 values and for interpretation in different types
3787 (e.g. float or int from a double register). Lazy
3788 register values should have the register's natural type,
3789 so they do not apply. */
3790 gdb_assert (!gdbarch_convert_register_p (get_frame_arch (next_frame),
3791 regnum, type));
3792
3793 /* FRAME was obtained, above, via VALUE_NEXT_FRAME_ID.
3794 Since a "->next" operation was performed when setting
3795 this field, we do not need to perform a "next" operation
3796 again when unwinding the register. That's why
3797 frame_unwind_register_value() is called here instead of
3798 get_frame_register_value(). */
3799 new_val = frame_unwind_register_value (next_frame, regnum);
3800
3801 /* If we get another lazy lval_register value, it means the
3802 register is found by reading it from NEXT_FRAME's next frame.
3803 frame_unwind_register_value should never return a value with
3804 the frame id pointing to NEXT_FRAME. If it does, it means we
3805 either have two consecutive frames with the same frame id
3806 in the frame chain, or some code is trying to unwind
3807 behind get_prev_frame's back (e.g., a frame unwind
3808 sniffer trying to unwind), bypassing its validations. In
3809 any case, it should always be an internal error to end up
3810 in this situation. */
3811 if (VALUE_LVAL (new_val) == lval_register
3812 && value_lazy (new_val)
3813 && frame_id_eq (VALUE_NEXT_FRAME_ID (new_val), next_frame_id))
3814 internal_error (__FILE__, __LINE__,
3815 _("infinite loop while fetching a register"));
3816 }
3817
3818 /* If it's still lazy (for instance, a saved register on the
3819 stack), fetch it. */
3820 if (value_lazy (new_val))
3821 value_fetch_lazy (new_val);
3822
3823 /* Copy the contents and the unavailability/optimized-out
3824 meta-data from NEW_VAL to VAL. */
3825 set_value_lazy (val, 0);
3826 value_contents_copy (val, value_embedded_offset (val),
3827 new_val, value_embedded_offset (new_val),
3828 type_length_units (type));
3829
3830 if (frame_debug)
3831 {
3832 struct gdbarch *gdbarch;
3833 struct frame_info *frame;
3834 /* VALUE_FRAME_ID is used here, instead of VALUE_NEXT_FRAME_ID,
3835 so that the frame level will be shown correctly. */
3836 frame = frame_find_by_id (VALUE_FRAME_ID (val));
3837 regnum = VALUE_REGNUM (val);
3838 gdbarch = get_frame_arch (frame);
3839
3840 fprintf_unfiltered (gdb_stdlog,
3841 "{ value_fetch_lazy "
3842 "(frame=%d,regnum=%d(%s),...) ",
3843 frame_relative_level (frame), regnum,
3844 user_reg_map_regnum_to_name (gdbarch, regnum));
3845
3846 fprintf_unfiltered (gdb_stdlog, "->");
3847 if (value_optimized_out (new_val))
3848 {
3849 fprintf_unfiltered (gdb_stdlog, " ");
3850 val_print_optimized_out (new_val, gdb_stdlog);
3851 }
3852 else
3853 {
3854 int i;
3855 const gdb_byte *buf = value_contents (new_val);
3856
3857 if (VALUE_LVAL (new_val) == lval_register)
3858 fprintf_unfiltered (gdb_stdlog, " register=%d",
3859 VALUE_REGNUM (new_val));
3860 else if (VALUE_LVAL (new_val) == lval_memory)
3861 fprintf_unfiltered (gdb_stdlog, " address=%s",
3862 paddress (gdbarch,
3863 value_address (new_val)));
3864 else
3865 fprintf_unfiltered (gdb_stdlog, " computed");
3866
3867 fprintf_unfiltered (gdb_stdlog, " bytes=");
3868 fprintf_unfiltered (gdb_stdlog, "[");
3869 for (i = 0; i < register_size (gdbarch, regnum); i++)
3870 fprintf_unfiltered (gdb_stdlog, "%02x", buf[i]);
3871 fprintf_unfiltered (gdb_stdlog, "]");
3872 }
3873
3874 fprintf_unfiltered (gdb_stdlog, " }\n");
3875 }
3876
3877 /* Dispose of the intermediate values. This prevents
3878 watchpoints from trying to watch the saved frame pointer. */
3879 value_free_to_mark (mark);
3880 }
3881 else if (VALUE_LVAL (val) == lval_computed
3882 && value_computed_funcs (val)->read != NULL)
3883 value_computed_funcs (val)->read (val);
3884 else
3885 internal_error (__FILE__, __LINE__, _("Unexpected lazy value type."));
3886
3887 set_value_lazy (val, 0);
3888 }
3889
3890 /* Implementation of the convenience function $_isvoid. */
3891
3892 static struct value *
3893 isvoid_internal_fn (struct gdbarch *gdbarch,
3894 const struct language_defn *language,
3895 void *cookie, int argc, struct value **argv)
3896 {
3897 int ret;
3898
3899 if (argc != 1)
3900 error (_("You must provide one argument for $_isvoid."));
3901
3902 ret = TYPE_CODE (value_type (argv[0])) == TYPE_CODE_VOID;
3903
3904 return value_from_longest (builtin_type (gdbarch)->builtin_int, ret);
3905 }
3906
3907 void
3908 _initialize_values (void)
3909 {
3910 add_cmd ("convenience", no_class, show_convenience, _("\
3911 Debugger convenience (\"$foo\") variables and functions.\n\
3912 Convenience variables are created when you assign them values;\n\
3913 thus, \"set $foo=1\" gives \"$foo\" the value 1. Values may be any type.\n\
3914 \n\
3915 A few convenience variables are given values automatically:\n\
3916 \"$_\"holds the last address examined with \"x\" or \"info lines\",\n\
3917 \"$__\" holds the contents of the last address examined with \"x\"."
3918 #ifdef HAVE_PYTHON
3919 "\n\n\
3920 Convenience functions are defined via the Python API."
3921 #endif
3922 ), &showlist);
3923 add_alias_cmd ("conv", "convenience", no_class, 1, &showlist);
3924
3925 add_cmd ("values", no_set_class, show_values, _("\
3926 Elements of value history around item number IDX (or last ten)."),
3927 &showlist);
3928
3929 add_com ("init-if-undefined", class_vars, init_if_undefined_command, _("\
3930 Initialize a convenience variable if necessary.\n\
3931 init-if-undefined VARIABLE = EXPRESSION\n\
3932 Set an internal VARIABLE to the result of the EXPRESSION if it does not\n\
3933 exist or does not contain a value. The EXPRESSION is not evaluated if the\n\
3934 VARIABLE is already initialized."));
3935
3936 add_prefix_cmd ("function", no_class, function_command, _("\
3937 Placeholder command for showing help on convenience functions."),
3938 &functionlist, "function ", 0, &cmdlist);
3939
3940 add_internal_function ("_isvoid", _("\
3941 Check whether an expression is void.\n\
3942 Usage: $_isvoid (expression)\n\
3943 Return 1 if the expression is void, zero otherwise."),
3944 isvoid_internal_fn, NULL);
3945
3946 add_setshow_zuinteger_unlimited_cmd ("max-value-size",
3947 class_support, &max_value_size, _("\
3948 Set maximum sized value gdb will load from the inferior."), _("\
3949 Show maximum sized value gdb will load from the inferior."), _("\
3950 Use this to control the maximum size, in bytes, of a value that gdb\n\
3951 will load from the inferior. Setting this value to 'unlimited'\n\
3952 disables checking.\n\
3953 Setting this does not invalidate already allocated values, it only\n\
3954 prevents future values, larger than this size, from being allocated."),
3955 set_max_value_size,
3956 show_max_value_size,
3957 &setlist, &showlist);
3958 }
This page took 0.114541 seconds and 5 git commands to generate.