Implement RDRSEED, ADX and PRFCHW instructions
[deliverable/binutils-gdb.git] / gas / config / tc-i386.c
1 /* tc-i386.c -- Assemble code for the Intel 80386
2 Copyright 1989, 1991, 1992, 1993, 1994, 1995, 1996, 1997, 1998, 1999,
3 2000, 2001, 2002, 2003, 2004, 2005, 2006, 2007, 2008, 2009, 2010, 2011,
4 2012
5 Free Software Foundation, Inc.
6
7 This file is part of GAS, the GNU Assembler.
8
9 GAS is free software; you can redistribute it and/or modify
10 it under the terms of the GNU General Public License as published by
11 the Free Software Foundation; either version 3, or (at your option)
12 any later version.
13
14 GAS is distributed in the hope that it will be useful,
15 but WITHOUT ANY WARRANTY; without even the implied warranty of
16 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
17 GNU General Public License for more details.
18
19 You should have received a copy of the GNU General Public License
20 along with GAS; see the file COPYING. If not, write to the Free
21 Software Foundation, 51 Franklin Street - Fifth Floor, Boston, MA
22 02110-1301, USA. */
23
24 /* Intel 80386 machine specific gas.
25 Written by Eliot Dresselhaus (eliot@mgm.mit.edu).
26 x86_64 support by Jan Hubicka (jh@suse.cz)
27 VIA PadLock support by Michal Ludvig (mludvig@suse.cz)
28 Bugs & suggestions are completely welcome. This is free software.
29 Please help us make it better. */
30
31 #include "as.h"
32 #include "safe-ctype.h"
33 #include "subsegs.h"
34 #include "dwarf2dbg.h"
35 #include "dw2gencfi.h"
36 #include "elf/x86-64.h"
37 #include "opcodes/i386-init.h"
38
39 #ifndef REGISTER_WARNINGS
40 #define REGISTER_WARNINGS 1
41 #endif
42
43 #ifndef INFER_ADDR_PREFIX
44 #define INFER_ADDR_PREFIX 1
45 #endif
46
47 #ifndef DEFAULT_ARCH
48 #define DEFAULT_ARCH "i386"
49 #endif
50
51 #ifndef INLINE
52 #if __GNUC__ >= 2
53 #define INLINE __inline__
54 #else
55 #define INLINE
56 #endif
57 #endif
58
59 /* Prefixes will be emitted in the order defined below.
60 WAIT_PREFIX must be the first prefix since FWAIT is really is an
61 instruction, and so must come before any prefixes.
62 The preferred prefix order is SEG_PREFIX, ADDR_PREFIX, DATA_PREFIX,
63 REP_PREFIX/HLE_PREFIX, LOCK_PREFIX. */
64 #define WAIT_PREFIX 0
65 #define SEG_PREFIX 1
66 #define ADDR_PREFIX 2
67 #define DATA_PREFIX 3
68 #define REP_PREFIX 4
69 #define HLE_PREFIX REP_PREFIX
70 #define LOCK_PREFIX 5
71 #define REX_PREFIX 6 /* must come last. */
72 #define MAX_PREFIXES 7 /* max prefixes per opcode */
73
74 /* we define the syntax here (modulo base,index,scale syntax) */
75 #define REGISTER_PREFIX '%'
76 #define IMMEDIATE_PREFIX '$'
77 #define ABSOLUTE_PREFIX '*'
78
79 /* these are the instruction mnemonic suffixes in AT&T syntax or
80 memory operand size in Intel syntax. */
81 #define WORD_MNEM_SUFFIX 'w'
82 #define BYTE_MNEM_SUFFIX 'b'
83 #define SHORT_MNEM_SUFFIX 's'
84 #define LONG_MNEM_SUFFIX 'l'
85 #define QWORD_MNEM_SUFFIX 'q'
86 #define XMMWORD_MNEM_SUFFIX 'x'
87 #define YMMWORD_MNEM_SUFFIX 'y'
88 /* Intel Syntax. Use a non-ascii letter since since it never appears
89 in instructions. */
90 #define LONG_DOUBLE_MNEM_SUFFIX '\1'
91
92 #define END_OF_INSN '\0'
93
94 /*
95 'templates' is for grouping together 'template' structures for opcodes
96 of the same name. This is only used for storing the insns in the grand
97 ole hash table of insns.
98 The templates themselves start at START and range up to (but not including)
99 END.
100 */
101 typedef struct
102 {
103 const insn_template *start;
104 const insn_template *end;
105 }
106 templates;
107
108 /* 386 operand encoding bytes: see 386 book for details of this. */
109 typedef struct
110 {
111 unsigned int regmem; /* codes register or memory operand */
112 unsigned int reg; /* codes register operand (or extended opcode) */
113 unsigned int mode; /* how to interpret regmem & reg */
114 }
115 modrm_byte;
116
117 /* x86-64 extension prefix. */
118 typedef int rex_byte;
119
120 /* 386 opcode byte to code indirect addressing. */
121 typedef struct
122 {
123 unsigned base;
124 unsigned index;
125 unsigned scale;
126 }
127 sib_byte;
128
129 /* x86 arch names, types and features */
130 typedef struct
131 {
132 const char *name; /* arch name */
133 unsigned int len; /* arch string length */
134 enum processor_type type; /* arch type */
135 i386_cpu_flags flags; /* cpu feature flags */
136 unsigned int skip; /* show_arch should skip this. */
137 unsigned int negated; /* turn off indicated flags. */
138 }
139 arch_entry;
140
141 static void update_code_flag (int, int);
142 static void set_code_flag (int);
143 static void set_16bit_gcc_code_flag (int);
144 static void set_intel_syntax (int);
145 static void set_intel_mnemonic (int);
146 static void set_allow_index_reg (int);
147 static void set_sse_check (int);
148 static void set_cpu_arch (int);
149 #ifdef TE_PE
150 static void pe_directive_secrel (int);
151 #endif
152 static void signed_cons (int);
153 static char *output_invalid (int c);
154 static int i386_finalize_immediate (segT, expressionS *, i386_operand_type,
155 const char *);
156 static int i386_finalize_displacement (segT, expressionS *, i386_operand_type,
157 const char *);
158 static int i386_att_operand (char *);
159 static int i386_intel_operand (char *, int);
160 static int i386_intel_simplify (expressionS *);
161 static int i386_intel_parse_name (const char *, expressionS *);
162 static const reg_entry *parse_register (char *, char **);
163 static char *parse_insn (char *, char *);
164 static char *parse_operands (char *, const char *);
165 static void swap_operands (void);
166 static void swap_2_operands (int, int);
167 static void optimize_imm (void);
168 static void optimize_disp (void);
169 static const insn_template *match_template (void);
170 static int check_string (void);
171 static int process_suffix (void);
172 static int check_byte_reg (void);
173 static int check_long_reg (void);
174 static int check_qword_reg (void);
175 static int check_word_reg (void);
176 static int finalize_imm (void);
177 static int process_operands (void);
178 static const seg_entry *build_modrm_byte (void);
179 static void output_insn (void);
180 static void output_imm (fragS *, offsetT);
181 static void output_disp (fragS *, offsetT);
182 #ifndef I386COFF
183 static void s_bss (int);
184 #endif
185 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
186 static void handle_large_common (int small ATTRIBUTE_UNUSED);
187 #endif
188
189 static const char *default_arch = DEFAULT_ARCH;
190
191 /* VEX prefix. */
192 typedef struct
193 {
194 /* VEX prefix is either 2 byte or 3 byte. */
195 unsigned char bytes[3];
196 unsigned int length;
197 /* Destination or source register specifier. */
198 const reg_entry *register_specifier;
199 } vex_prefix;
200
201 /* 'md_assemble ()' gathers together information and puts it into a
202 i386_insn. */
203
204 union i386_op
205 {
206 expressionS *disps;
207 expressionS *imms;
208 const reg_entry *regs;
209 };
210
211 enum i386_error
212 {
213 operand_size_mismatch,
214 operand_type_mismatch,
215 register_type_mismatch,
216 number_of_operands_mismatch,
217 invalid_instruction_suffix,
218 bad_imm4,
219 old_gcc_only,
220 unsupported_with_intel_mnemonic,
221 unsupported_syntax,
222 unsupported,
223 invalid_vsib_address,
224 unsupported_vector_index_register
225 };
226
227 struct _i386_insn
228 {
229 /* TM holds the template for the insn were currently assembling. */
230 insn_template tm;
231
232 /* SUFFIX holds the instruction size suffix for byte, word, dword
233 or qword, if given. */
234 char suffix;
235
236 /* OPERANDS gives the number of given operands. */
237 unsigned int operands;
238
239 /* REG_OPERANDS, DISP_OPERANDS, MEM_OPERANDS, IMM_OPERANDS give the number
240 of given register, displacement, memory operands and immediate
241 operands. */
242 unsigned int reg_operands, disp_operands, mem_operands, imm_operands;
243
244 /* TYPES [i] is the type (see above #defines) which tells us how to
245 use OP[i] for the corresponding operand. */
246 i386_operand_type types[MAX_OPERANDS];
247
248 /* Displacement expression, immediate expression, or register for each
249 operand. */
250 union i386_op op[MAX_OPERANDS];
251
252 /* Flags for operands. */
253 unsigned int flags[MAX_OPERANDS];
254 #define Operand_PCrel 1
255
256 /* Relocation type for operand */
257 enum bfd_reloc_code_real reloc[MAX_OPERANDS];
258
259 /* BASE_REG, INDEX_REG, and LOG2_SCALE_FACTOR are used to encode
260 the base index byte below. */
261 const reg_entry *base_reg;
262 const reg_entry *index_reg;
263 unsigned int log2_scale_factor;
264
265 /* SEG gives the seg_entries of this insn. They are zero unless
266 explicit segment overrides are given. */
267 const seg_entry *seg[2];
268
269 /* PREFIX holds all the given prefix opcodes (usually null).
270 PREFIXES is the number of prefix opcodes. */
271 unsigned int prefixes;
272 unsigned char prefix[MAX_PREFIXES];
273
274 /* RM and SIB are the modrm byte and the sib byte where the
275 addressing modes of this insn are encoded. */
276 modrm_byte rm;
277 rex_byte rex;
278 sib_byte sib;
279 vex_prefix vex;
280
281 /* Swap operand in encoding. */
282 unsigned int swap_operand;
283
284 /* Prefer 8bit or 32bit displacement in encoding. */
285 enum
286 {
287 disp_encoding_default = 0,
288 disp_encoding_8bit,
289 disp_encoding_32bit
290 } disp_encoding;
291
292 /* Have HLE prefix. */
293 unsigned int have_hle;
294
295 /* Error message. */
296 enum i386_error error;
297 };
298
299 typedef struct _i386_insn i386_insn;
300
301 /* List of chars besides those in app.c:symbol_chars that can start an
302 operand. Used to prevent the scrubber eating vital white-space. */
303 const char extra_symbol_chars[] = "*%-(["
304 #ifdef LEX_AT
305 "@"
306 #endif
307 #ifdef LEX_QM
308 "?"
309 #endif
310 ;
311
312 #if (defined (TE_I386AIX) \
313 || ((defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)) \
314 && !defined (TE_GNU) \
315 && !defined (TE_LINUX) \
316 && !defined (TE_NACL) \
317 && !defined (TE_NETWARE) \
318 && !defined (TE_FreeBSD) \
319 && !defined (TE_DragonFly) \
320 && !defined (TE_NetBSD)))
321 /* This array holds the chars that always start a comment. If the
322 pre-processor is disabled, these aren't very useful. The option
323 --divide will remove '/' from this list. */
324 const char *i386_comment_chars = "#/";
325 #define SVR4_COMMENT_CHARS 1
326 #define PREFIX_SEPARATOR '\\'
327
328 #else
329 const char *i386_comment_chars = "#";
330 #define PREFIX_SEPARATOR '/'
331 #endif
332
333 /* This array holds the chars that only start a comment at the beginning of
334 a line. If the line seems to have the form '# 123 filename'
335 .line and .file directives will appear in the pre-processed output.
336 Note that input_file.c hand checks for '#' at the beginning of the
337 first line of the input file. This is because the compiler outputs
338 #NO_APP at the beginning of its output.
339 Also note that comments started like this one will always work if
340 '/' isn't otherwise defined. */
341 const char line_comment_chars[] = "#/";
342
343 const char line_separator_chars[] = ";";
344
345 /* Chars that can be used to separate mant from exp in floating point
346 nums. */
347 const char EXP_CHARS[] = "eE";
348
349 /* Chars that mean this number is a floating point constant
350 As in 0f12.456
351 or 0d1.2345e12. */
352 const char FLT_CHARS[] = "fFdDxX";
353
354 /* Tables for lexical analysis. */
355 static char mnemonic_chars[256];
356 static char register_chars[256];
357 static char operand_chars[256];
358 static char identifier_chars[256];
359 static char digit_chars[256];
360
361 /* Lexical macros. */
362 #define is_mnemonic_char(x) (mnemonic_chars[(unsigned char) x])
363 #define is_operand_char(x) (operand_chars[(unsigned char) x])
364 #define is_register_char(x) (register_chars[(unsigned char) x])
365 #define is_space_char(x) ((x) == ' ')
366 #define is_identifier_char(x) (identifier_chars[(unsigned char) x])
367 #define is_digit_char(x) (digit_chars[(unsigned char) x])
368
369 /* All non-digit non-letter characters that may occur in an operand. */
370 static char operand_special_chars[] = "%$-+(,)*._~/<>|&^!:[@]";
371
372 /* md_assemble() always leaves the strings it's passed unaltered. To
373 effect this we maintain a stack of saved characters that we've smashed
374 with '\0's (indicating end of strings for various sub-fields of the
375 assembler instruction). */
376 static char save_stack[32];
377 static char *save_stack_p;
378 #define END_STRING_AND_SAVE(s) \
379 do { *save_stack_p++ = *(s); *(s) = '\0'; } while (0)
380 #define RESTORE_END_STRING(s) \
381 do { *(s) = *--save_stack_p; } while (0)
382
383 /* The instruction we're assembling. */
384 static i386_insn i;
385
386 /* Possible templates for current insn. */
387 static const templates *current_templates;
388
389 /* Per instruction expressionS buffers: max displacements & immediates. */
390 static expressionS disp_expressions[MAX_MEMORY_OPERANDS];
391 static expressionS im_expressions[MAX_IMMEDIATE_OPERANDS];
392
393 /* Current operand we are working on. */
394 static int this_operand = -1;
395
396 /* We support four different modes. FLAG_CODE variable is used to distinguish
397 these. */
398
399 enum flag_code {
400 CODE_32BIT,
401 CODE_16BIT,
402 CODE_64BIT };
403
404 static enum flag_code flag_code;
405 static unsigned int object_64bit;
406 static unsigned int disallow_64bit_reloc;
407 static int use_rela_relocations = 0;
408
409 #if ((defined (OBJ_MAYBE_COFF) && defined (OBJ_MAYBE_AOUT)) \
410 || defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) \
411 || defined (TE_PE) || defined (TE_PEP) || defined (OBJ_MACH_O))
412
413 /* The ELF ABI to use. */
414 enum x86_elf_abi
415 {
416 I386_ABI,
417 X86_64_ABI,
418 X86_64_X32_ABI
419 };
420
421 static enum x86_elf_abi x86_elf_abi = I386_ABI;
422 #endif
423
424 /* The names used to print error messages. */
425 static const char *flag_code_names[] =
426 {
427 "32",
428 "16",
429 "64"
430 };
431
432 /* 1 for intel syntax,
433 0 if att syntax. */
434 static int intel_syntax = 0;
435
436 /* 1 for intel mnemonic,
437 0 if att mnemonic. */
438 static int intel_mnemonic = !SYSV386_COMPAT;
439
440 /* 1 if support old (<= 2.8.1) versions of gcc. */
441 static int old_gcc = OLDGCC_COMPAT;
442
443 /* 1 if pseudo registers are permitted. */
444 static int allow_pseudo_reg = 0;
445
446 /* 1 if register prefix % not required. */
447 static int allow_naked_reg = 0;
448
449 /* 1 if pseudo index register, eiz/riz, is allowed . */
450 static int allow_index_reg = 0;
451
452 static enum
453 {
454 sse_check_none = 0,
455 sse_check_warning,
456 sse_check_error
457 }
458 sse_check;
459
460 /* Register prefix used for error message. */
461 static const char *register_prefix = "%";
462
463 /* Used in 16 bit gcc mode to add an l suffix to call, ret, enter,
464 leave, push, and pop instructions so that gcc has the same stack
465 frame as in 32 bit mode. */
466 static char stackop_size = '\0';
467
468 /* Non-zero to optimize code alignment. */
469 int optimize_align_code = 1;
470
471 /* Non-zero to quieten some warnings. */
472 static int quiet_warnings = 0;
473
474 /* CPU name. */
475 static const char *cpu_arch_name = NULL;
476 static char *cpu_sub_arch_name = NULL;
477
478 /* CPU feature flags. */
479 static i386_cpu_flags cpu_arch_flags = CPU_UNKNOWN_FLAGS;
480
481 /* If we have selected a cpu we are generating instructions for. */
482 static int cpu_arch_tune_set = 0;
483
484 /* Cpu we are generating instructions for. */
485 enum processor_type cpu_arch_tune = PROCESSOR_UNKNOWN;
486
487 /* CPU feature flags of cpu we are generating instructions for. */
488 static i386_cpu_flags cpu_arch_tune_flags;
489
490 /* CPU instruction set architecture used. */
491 enum processor_type cpu_arch_isa = PROCESSOR_UNKNOWN;
492
493 /* CPU feature flags of instruction set architecture used. */
494 i386_cpu_flags cpu_arch_isa_flags;
495
496 /* If set, conditional jumps are not automatically promoted to handle
497 larger than a byte offset. */
498 static unsigned int no_cond_jump_promotion = 0;
499
500 /* Encode SSE instructions with VEX prefix. */
501 static unsigned int sse2avx;
502
503 /* Encode scalar AVX instructions with specific vector length. */
504 static enum
505 {
506 vex128 = 0,
507 vex256
508 } avxscalar;
509
510 /* Pre-defined "_GLOBAL_OFFSET_TABLE_". */
511 static symbolS *GOT_symbol;
512
513 /* The dwarf2 return column, adjusted for 32 or 64 bit. */
514 unsigned int x86_dwarf2_return_column;
515
516 /* The dwarf2 data alignment, adjusted for 32 or 64 bit. */
517 int x86_cie_data_alignment;
518
519 /* Interface to relax_segment.
520 There are 3 major relax states for 386 jump insns because the
521 different types of jumps add different sizes to frags when we're
522 figuring out what sort of jump to choose to reach a given label. */
523
524 /* Types. */
525 #define UNCOND_JUMP 0
526 #define COND_JUMP 1
527 #define COND_JUMP86 2
528
529 /* Sizes. */
530 #define CODE16 1
531 #define SMALL 0
532 #define SMALL16 (SMALL | CODE16)
533 #define BIG 2
534 #define BIG16 (BIG | CODE16)
535
536 #ifndef INLINE
537 #ifdef __GNUC__
538 #define INLINE __inline__
539 #else
540 #define INLINE
541 #endif
542 #endif
543
544 #define ENCODE_RELAX_STATE(type, size) \
545 ((relax_substateT) (((type) << 2) | (size)))
546 #define TYPE_FROM_RELAX_STATE(s) \
547 ((s) >> 2)
548 #define DISP_SIZE_FROM_RELAX_STATE(s) \
549 ((((s) & 3) == BIG ? 4 : (((s) & 3) == BIG16 ? 2 : 1)))
550
551 /* This table is used by relax_frag to promote short jumps to long
552 ones where necessary. SMALL (short) jumps may be promoted to BIG
553 (32 bit long) ones, and SMALL16 jumps to BIG16 (16 bit long). We
554 don't allow a short jump in a 32 bit code segment to be promoted to
555 a 16 bit offset jump because it's slower (requires data size
556 prefix), and doesn't work, unless the destination is in the bottom
557 64k of the code segment (The top 16 bits of eip are zeroed). */
558
559 const relax_typeS md_relax_table[] =
560 {
561 /* The fields are:
562 1) most positive reach of this state,
563 2) most negative reach of this state,
564 3) how many bytes this mode will have in the variable part of the frag
565 4) which index into the table to try if we can't fit into this one. */
566
567 /* UNCOND_JUMP states. */
568 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (UNCOND_JUMP, BIG)},
569 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (UNCOND_JUMP, BIG16)},
570 /* dword jmp adds 4 bytes to frag:
571 0 extra opcode bytes, 4 displacement bytes. */
572 {0, 0, 4, 0},
573 /* word jmp adds 2 byte2 to frag:
574 0 extra opcode bytes, 2 displacement bytes. */
575 {0, 0, 2, 0},
576
577 /* COND_JUMP states. */
578 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (COND_JUMP, BIG)},
579 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (COND_JUMP, BIG16)},
580 /* dword conditionals adds 5 bytes to frag:
581 1 extra opcode byte, 4 displacement bytes. */
582 {0, 0, 5, 0},
583 /* word conditionals add 3 bytes to frag:
584 1 extra opcode byte, 2 displacement bytes. */
585 {0, 0, 3, 0},
586
587 /* COND_JUMP86 states. */
588 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (COND_JUMP86, BIG)},
589 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (COND_JUMP86, BIG16)},
590 /* dword conditionals adds 5 bytes to frag:
591 1 extra opcode byte, 4 displacement bytes. */
592 {0, 0, 5, 0},
593 /* word conditionals add 4 bytes to frag:
594 1 displacement byte and a 3 byte long branch insn. */
595 {0, 0, 4, 0}
596 };
597
598 static const arch_entry cpu_arch[] =
599 {
600 /* Do not replace the first two entries - i386_target_format()
601 relies on them being there in this order. */
602 { STRING_COMMA_LEN ("generic32"), PROCESSOR_GENERIC32,
603 CPU_GENERIC32_FLAGS, 0, 0 },
604 { STRING_COMMA_LEN ("generic64"), PROCESSOR_GENERIC64,
605 CPU_GENERIC64_FLAGS, 0, 0 },
606 { STRING_COMMA_LEN ("i8086"), PROCESSOR_UNKNOWN,
607 CPU_NONE_FLAGS, 0, 0 },
608 { STRING_COMMA_LEN ("i186"), PROCESSOR_UNKNOWN,
609 CPU_I186_FLAGS, 0, 0 },
610 { STRING_COMMA_LEN ("i286"), PROCESSOR_UNKNOWN,
611 CPU_I286_FLAGS, 0, 0 },
612 { STRING_COMMA_LEN ("i386"), PROCESSOR_I386,
613 CPU_I386_FLAGS, 0, 0 },
614 { STRING_COMMA_LEN ("i486"), PROCESSOR_I486,
615 CPU_I486_FLAGS, 0, 0 },
616 { STRING_COMMA_LEN ("i586"), PROCESSOR_PENTIUM,
617 CPU_I586_FLAGS, 0, 0 },
618 { STRING_COMMA_LEN ("i686"), PROCESSOR_PENTIUMPRO,
619 CPU_I686_FLAGS, 0, 0 },
620 { STRING_COMMA_LEN ("pentium"), PROCESSOR_PENTIUM,
621 CPU_I586_FLAGS, 0, 0 },
622 { STRING_COMMA_LEN ("pentiumpro"), PROCESSOR_PENTIUMPRO,
623 CPU_PENTIUMPRO_FLAGS, 0, 0 },
624 { STRING_COMMA_LEN ("pentiumii"), PROCESSOR_PENTIUMPRO,
625 CPU_P2_FLAGS, 0, 0 },
626 { STRING_COMMA_LEN ("pentiumiii"),PROCESSOR_PENTIUMPRO,
627 CPU_P3_FLAGS, 0, 0 },
628 { STRING_COMMA_LEN ("pentium4"), PROCESSOR_PENTIUM4,
629 CPU_P4_FLAGS, 0, 0 },
630 { STRING_COMMA_LEN ("prescott"), PROCESSOR_NOCONA,
631 CPU_CORE_FLAGS, 0, 0 },
632 { STRING_COMMA_LEN ("nocona"), PROCESSOR_NOCONA,
633 CPU_NOCONA_FLAGS, 0, 0 },
634 { STRING_COMMA_LEN ("yonah"), PROCESSOR_CORE,
635 CPU_CORE_FLAGS, 1, 0 },
636 { STRING_COMMA_LEN ("core"), PROCESSOR_CORE,
637 CPU_CORE_FLAGS, 0, 0 },
638 { STRING_COMMA_LEN ("merom"), PROCESSOR_CORE2,
639 CPU_CORE2_FLAGS, 1, 0 },
640 { STRING_COMMA_LEN ("core2"), PROCESSOR_CORE2,
641 CPU_CORE2_FLAGS, 0, 0 },
642 { STRING_COMMA_LEN ("corei7"), PROCESSOR_COREI7,
643 CPU_COREI7_FLAGS, 0, 0 },
644 { STRING_COMMA_LEN ("l1om"), PROCESSOR_L1OM,
645 CPU_L1OM_FLAGS, 0, 0 },
646 { STRING_COMMA_LEN ("k1om"), PROCESSOR_K1OM,
647 CPU_K1OM_FLAGS, 0, 0 },
648 { STRING_COMMA_LEN ("k6"), PROCESSOR_K6,
649 CPU_K6_FLAGS, 0, 0 },
650 { STRING_COMMA_LEN ("k6_2"), PROCESSOR_K6,
651 CPU_K6_2_FLAGS, 0, 0 },
652 { STRING_COMMA_LEN ("athlon"), PROCESSOR_ATHLON,
653 CPU_ATHLON_FLAGS, 0, 0 },
654 { STRING_COMMA_LEN ("sledgehammer"), PROCESSOR_K8,
655 CPU_K8_FLAGS, 1, 0 },
656 { STRING_COMMA_LEN ("opteron"), PROCESSOR_K8,
657 CPU_K8_FLAGS, 0, 0 },
658 { STRING_COMMA_LEN ("k8"), PROCESSOR_K8,
659 CPU_K8_FLAGS, 0, 0 },
660 { STRING_COMMA_LEN ("amdfam10"), PROCESSOR_AMDFAM10,
661 CPU_AMDFAM10_FLAGS, 0, 0 },
662 { STRING_COMMA_LEN ("bdver1"), PROCESSOR_BD,
663 CPU_BDVER1_FLAGS, 0, 0 },
664 { STRING_COMMA_LEN ("bdver2"), PROCESSOR_BD,
665 CPU_BDVER2_FLAGS, 0, 0 },
666 { STRING_COMMA_LEN (".8087"), PROCESSOR_UNKNOWN,
667 CPU_8087_FLAGS, 0, 0 },
668 { STRING_COMMA_LEN (".287"), PROCESSOR_UNKNOWN,
669 CPU_287_FLAGS, 0, 0 },
670 { STRING_COMMA_LEN (".387"), PROCESSOR_UNKNOWN,
671 CPU_387_FLAGS, 0, 0 },
672 { STRING_COMMA_LEN (".no87"), PROCESSOR_UNKNOWN,
673 CPU_ANY87_FLAGS, 0, 1 },
674 { STRING_COMMA_LEN (".mmx"), PROCESSOR_UNKNOWN,
675 CPU_MMX_FLAGS, 0, 0 },
676 { STRING_COMMA_LEN (".nommx"), PROCESSOR_UNKNOWN,
677 CPU_3DNOWA_FLAGS, 0, 1 },
678 { STRING_COMMA_LEN (".sse"), PROCESSOR_UNKNOWN,
679 CPU_SSE_FLAGS, 0, 0 },
680 { STRING_COMMA_LEN (".sse2"), PROCESSOR_UNKNOWN,
681 CPU_SSE2_FLAGS, 0, 0 },
682 { STRING_COMMA_LEN (".sse3"), PROCESSOR_UNKNOWN,
683 CPU_SSE3_FLAGS, 0, 0 },
684 { STRING_COMMA_LEN (".ssse3"), PROCESSOR_UNKNOWN,
685 CPU_SSSE3_FLAGS, 0, 0 },
686 { STRING_COMMA_LEN (".sse4.1"), PROCESSOR_UNKNOWN,
687 CPU_SSE4_1_FLAGS, 0, 0 },
688 { STRING_COMMA_LEN (".sse4.2"), PROCESSOR_UNKNOWN,
689 CPU_SSE4_2_FLAGS, 0, 0 },
690 { STRING_COMMA_LEN (".sse4"), PROCESSOR_UNKNOWN,
691 CPU_SSE4_2_FLAGS, 0, 0 },
692 { STRING_COMMA_LEN (".nosse"), PROCESSOR_UNKNOWN,
693 CPU_ANY_SSE_FLAGS, 0, 1 },
694 { STRING_COMMA_LEN (".avx"), PROCESSOR_UNKNOWN,
695 CPU_AVX_FLAGS, 0, 0 },
696 { STRING_COMMA_LEN (".avx2"), PROCESSOR_UNKNOWN,
697 CPU_AVX2_FLAGS, 0, 0 },
698 { STRING_COMMA_LEN (".noavx"), PROCESSOR_UNKNOWN,
699 CPU_ANY_AVX_FLAGS, 0, 1 },
700 { STRING_COMMA_LEN (".vmx"), PROCESSOR_UNKNOWN,
701 CPU_VMX_FLAGS, 0, 0 },
702 { STRING_COMMA_LEN (".vmfunc"), PROCESSOR_UNKNOWN,
703 CPU_VMFUNC_FLAGS, 0, 0 },
704 { STRING_COMMA_LEN (".smx"), PROCESSOR_UNKNOWN,
705 CPU_SMX_FLAGS, 0, 0 },
706 { STRING_COMMA_LEN (".xsave"), PROCESSOR_UNKNOWN,
707 CPU_XSAVE_FLAGS, 0, 0 },
708 { STRING_COMMA_LEN (".xsaveopt"), PROCESSOR_UNKNOWN,
709 CPU_XSAVEOPT_FLAGS, 0, 0 },
710 { STRING_COMMA_LEN (".aes"), PROCESSOR_UNKNOWN,
711 CPU_AES_FLAGS, 0, 0 },
712 { STRING_COMMA_LEN (".pclmul"), PROCESSOR_UNKNOWN,
713 CPU_PCLMUL_FLAGS, 0, 0 },
714 { STRING_COMMA_LEN (".clmul"), PROCESSOR_UNKNOWN,
715 CPU_PCLMUL_FLAGS, 1, 0 },
716 { STRING_COMMA_LEN (".fsgsbase"), PROCESSOR_UNKNOWN,
717 CPU_FSGSBASE_FLAGS, 0, 0 },
718 { STRING_COMMA_LEN (".rdrnd"), PROCESSOR_UNKNOWN,
719 CPU_RDRND_FLAGS, 0, 0 },
720 { STRING_COMMA_LEN (".f16c"), PROCESSOR_UNKNOWN,
721 CPU_F16C_FLAGS, 0, 0 },
722 { STRING_COMMA_LEN (".bmi2"), PROCESSOR_UNKNOWN,
723 CPU_BMI2_FLAGS, 0, 0 },
724 { STRING_COMMA_LEN (".fma"), PROCESSOR_UNKNOWN,
725 CPU_FMA_FLAGS, 0, 0 },
726 { STRING_COMMA_LEN (".fma4"), PROCESSOR_UNKNOWN,
727 CPU_FMA4_FLAGS, 0, 0 },
728 { STRING_COMMA_LEN (".xop"), PROCESSOR_UNKNOWN,
729 CPU_XOP_FLAGS, 0, 0 },
730 { STRING_COMMA_LEN (".lwp"), PROCESSOR_UNKNOWN,
731 CPU_LWP_FLAGS, 0, 0 },
732 { STRING_COMMA_LEN (".movbe"), PROCESSOR_UNKNOWN,
733 CPU_MOVBE_FLAGS, 0, 0 },
734 { STRING_COMMA_LEN (".ept"), PROCESSOR_UNKNOWN,
735 CPU_EPT_FLAGS, 0, 0 },
736 { STRING_COMMA_LEN (".lzcnt"), PROCESSOR_UNKNOWN,
737 CPU_LZCNT_FLAGS, 0, 0 },
738 { STRING_COMMA_LEN (".hle"), PROCESSOR_UNKNOWN,
739 CPU_HLE_FLAGS, 0, 0 },
740 { STRING_COMMA_LEN (".rtm"), PROCESSOR_UNKNOWN,
741 CPU_RTM_FLAGS, 0, 0 },
742 { STRING_COMMA_LEN (".invpcid"), PROCESSOR_UNKNOWN,
743 CPU_INVPCID_FLAGS, 0, 0 },
744 { STRING_COMMA_LEN (".clflush"), PROCESSOR_UNKNOWN,
745 CPU_CLFLUSH_FLAGS, 0, 0 },
746 { STRING_COMMA_LEN (".nop"), PROCESSOR_UNKNOWN,
747 CPU_NOP_FLAGS, 0, 0 },
748 { STRING_COMMA_LEN (".syscall"), PROCESSOR_UNKNOWN,
749 CPU_SYSCALL_FLAGS, 0, 0 },
750 { STRING_COMMA_LEN (".rdtscp"), PROCESSOR_UNKNOWN,
751 CPU_RDTSCP_FLAGS, 0, 0 },
752 { STRING_COMMA_LEN (".3dnow"), PROCESSOR_UNKNOWN,
753 CPU_3DNOW_FLAGS, 0, 0 },
754 { STRING_COMMA_LEN (".3dnowa"), PROCESSOR_UNKNOWN,
755 CPU_3DNOWA_FLAGS, 0, 0 },
756 { STRING_COMMA_LEN (".padlock"), PROCESSOR_UNKNOWN,
757 CPU_PADLOCK_FLAGS, 0, 0 },
758 { STRING_COMMA_LEN (".pacifica"), PROCESSOR_UNKNOWN,
759 CPU_SVME_FLAGS, 1, 0 },
760 { STRING_COMMA_LEN (".svme"), PROCESSOR_UNKNOWN,
761 CPU_SVME_FLAGS, 0, 0 },
762 { STRING_COMMA_LEN (".sse4a"), PROCESSOR_UNKNOWN,
763 CPU_SSE4A_FLAGS, 0, 0 },
764 { STRING_COMMA_LEN (".abm"), PROCESSOR_UNKNOWN,
765 CPU_ABM_FLAGS, 0, 0 },
766 { STRING_COMMA_LEN (".bmi"), PROCESSOR_UNKNOWN,
767 CPU_BMI_FLAGS, 0, 0 },
768 { STRING_COMMA_LEN (".tbm"), PROCESSOR_UNKNOWN,
769 CPU_TBM_FLAGS, 0, 0 },
770 { STRING_COMMA_LEN (".adx"), PROCESSOR_UNKNOWN,
771 CPU_ADX_FLAGS, 0, 0 },
772 { STRING_COMMA_LEN (".rdseed"), PROCESSOR_UNKNOWN,
773 CPU_RDSEED_FLAGS, 0, 0 },
774 { STRING_COMMA_LEN (".prfchw"), PROCESSOR_UNKNOWN,
775 CPU_PRFCHW_FLAGS, 0, 0 },
776 };
777
778 #ifdef I386COFF
779 /* Like s_lcomm_internal in gas/read.c but the alignment string
780 is allowed to be optional. */
781
782 static symbolS *
783 pe_lcomm_internal (int needs_align, symbolS *symbolP, addressT size)
784 {
785 addressT align = 0;
786
787 SKIP_WHITESPACE ();
788
789 if (needs_align
790 && *input_line_pointer == ',')
791 {
792 align = parse_align (needs_align - 1);
793
794 if (align == (addressT) -1)
795 return NULL;
796 }
797 else
798 {
799 if (size >= 8)
800 align = 3;
801 else if (size >= 4)
802 align = 2;
803 else if (size >= 2)
804 align = 1;
805 else
806 align = 0;
807 }
808
809 bss_alloc (symbolP, size, align);
810 return symbolP;
811 }
812
813 static void
814 pe_lcomm (int needs_align)
815 {
816 s_comm_internal (needs_align * 2, pe_lcomm_internal);
817 }
818 #endif
819
820 const pseudo_typeS md_pseudo_table[] =
821 {
822 #if !defined(OBJ_AOUT) && !defined(USE_ALIGN_PTWO)
823 {"align", s_align_bytes, 0},
824 #else
825 {"align", s_align_ptwo, 0},
826 #endif
827 {"arch", set_cpu_arch, 0},
828 #ifndef I386COFF
829 {"bss", s_bss, 0},
830 #else
831 {"lcomm", pe_lcomm, 1},
832 #endif
833 {"ffloat", float_cons, 'f'},
834 {"dfloat", float_cons, 'd'},
835 {"tfloat", float_cons, 'x'},
836 {"value", cons, 2},
837 {"slong", signed_cons, 4},
838 {"noopt", s_ignore, 0},
839 {"optim", s_ignore, 0},
840 {"code16gcc", set_16bit_gcc_code_flag, CODE_16BIT},
841 {"code16", set_code_flag, CODE_16BIT},
842 {"code32", set_code_flag, CODE_32BIT},
843 {"code64", set_code_flag, CODE_64BIT},
844 {"intel_syntax", set_intel_syntax, 1},
845 {"att_syntax", set_intel_syntax, 0},
846 {"intel_mnemonic", set_intel_mnemonic, 1},
847 {"att_mnemonic", set_intel_mnemonic, 0},
848 {"allow_index_reg", set_allow_index_reg, 1},
849 {"disallow_index_reg", set_allow_index_reg, 0},
850 {"sse_check", set_sse_check, 0},
851 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
852 {"largecomm", handle_large_common, 0},
853 #else
854 {"file", (void (*) (int)) dwarf2_directive_file, 0},
855 {"loc", dwarf2_directive_loc, 0},
856 {"loc_mark_labels", dwarf2_directive_loc_mark_labels, 0},
857 #endif
858 #ifdef TE_PE
859 {"secrel32", pe_directive_secrel, 0},
860 #endif
861 {0, 0, 0}
862 };
863
864 /* For interface with expression (). */
865 extern char *input_line_pointer;
866
867 /* Hash table for instruction mnemonic lookup. */
868 static struct hash_control *op_hash;
869
870 /* Hash table for register lookup. */
871 static struct hash_control *reg_hash;
872 \f
873 void
874 i386_align_code (fragS *fragP, int count)
875 {
876 /* Various efficient no-op patterns for aligning code labels.
877 Note: Don't try to assemble the instructions in the comments.
878 0L and 0w are not legal. */
879 static const char f32_1[] =
880 {0x90}; /* nop */
881 static const char f32_2[] =
882 {0x66,0x90}; /* xchg %ax,%ax */
883 static const char f32_3[] =
884 {0x8d,0x76,0x00}; /* leal 0(%esi),%esi */
885 static const char f32_4[] =
886 {0x8d,0x74,0x26,0x00}; /* leal 0(%esi,1),%esi */
887 static const char f32_5[] =
888 {0x90, /* nop */
889 0x8d,0x74,0x26,0x00}; /* leal 0(%esi,1),%esi */
890 static const char f32_6[] =
891 {0x8d,0xb6,0x00,0x00,0x00,0x00}; /* leal 0L(%esi),%esi */
892 static const char f32_7[] =
893 {0x8d,0xb4,0x26,0x00,0x00,0x00,0x00}; /* leal 0L(%esi,1),%esi */
894 static const char f32_8[] =
895 {0x90, /* nop */
896 0x8d,0xb4,0x26,0x00,0x00,0x00,0x00}; /* leal 0L(%esi,1),%esi */
897 static const char f32_9[] =
898 {0x89,0xf6, /* movl %esi,%esi */
899 0x8d,0xbc,0x27,0x00,0x00,0x00,0x00}; /* leal 0L(%edi,1),%edi */
900 static const char f32_10[] =
901 {0x8d,0x76,0x00, /* leal 0(%esi),%esi */
902 0x8d,0xbc,0x27,0x00,0x00,0x00,0x00}; /* leal 0L(%edi,1),%edi */
903 static const char f32_11[] =
904 {0x8d,0x74,0x26,0x00, /* leal 0(%esi,1),%esi */
905 0x8d,0xbc,0x27,0x00,0x00,0x00,0x00}; /* leal 0L(%edi,1),%edi */
906 static const char f32_12[] =
907 {0x8d,0xb6,0x00,0x00,0x00,0x00, /* leal 0L(%esi),%esi */
908 0x8d,0xbf,0x00,0x00,0x00,0x00}; /* leal 0L(%edi),%edi */
909 static const char f32_13[] =
910 {0x8d,0xb6,0x00,0x00,0x00,0x00, /* leal 0L(%esi),%esi */
911 0x8d,0xbc,0x27,0x00,0x00,0x00,0x00}; /* leal 0L(%edi,1),%edi */
912 static const char f32_14[] =
913 {0x8d,0xb4,0x26,0x00,0x00,0x00,0x00, /* leal 0L(%esi,1),%esi */
914 0x8d,0xbc,0x27,0x00,0x00,0x00,0x00}; /* leal 0L(%edi,1),%edi */
915 static const char f16_3[] =
916 {0x8d,0x74,0x00}; /* lea 0(%esi),%esi */
917 static const char f16_4[] =
918 {0x8d,0xb4,0x00,0x00}; /* lea 0w(%si),%si */
919 static const char f16_5[] =
920 {0x90, /* nop */
921 0x8d,0xb4,0x00,0x00}; /* lea 0w(%si),%si */
922 static const char f16_6[] =
923 {0x89,0xf6, /* mov %si,%si */
924 0x8d,0xbd,0x00,0x00}; /* lea 0w(%di),%di */
925 static const char f16_7[] =
926 {0x8d,0x74,0x00, /* lea 0(%si),%si */
927 0x8d,0xbd,0x00,0x00}; /* lea 0w(%di),%di */
928 static const char f16_8[] =
929 {0x8d,0xb4,0x00,0x00, /* lea 0w(%si),%si */
930 0x8d,0xbd,0x00,0x00}; /* lea 0w(%di),%di */
931 static const char jump_31[] =
932 {0xeb,0x1d,0x90,0x90,0x90,0x90,0x90, /* jmp .+31; lotsa nops */
933 0x90,0x90,0x90,0x90,0x90,0x90,0x90,0x90,
934 0x90,0x90,0x90,0x90,0x90,0x90,0x90,0x90,
935 0x90,0x90,0x90,0x90,0x90,0x90,0x90,0x90};
936 static const char *const f32_patt[] = {
937 f32_1, f32_2, f32_3, f32_4, f32_5, f32_6, f32_7, f32_8,
938 f32_9, f32_10, f32_11, f32_12, f32_13, f32_14
939 };
940 static const char *const f16_patt[] = {
941 f32_1, f32_2, f16_3, f16_4, f16_5, f16_6, f16_7, f16_8
942 };
943 /* nopl (%[re]ax) */
944 static const char alt_3[] =
945 {0x0f,0x1f,0x00};
946 /* nopl 0(%[re]ax) */
947 static const char alt_4[] =
948 {0x0f,0x1f,0x40,0x00};
949 /* nopl 0(%[re]ax,%[re]ax,1) */
950 static const char alt_5[] =
951 {0x0f,0x1f,0x44,0x00,0x00};
952 /* nopw 0(%[re]ax,%[re]ax,1) */
953 static const char alt_6[] =
954 {0x66,0x0f,0x1f,0x44,0x00,0x00};
955 /* nopl 0L(%[re]ax) */
956 static const char alt_7[] =
957 {0x0f,0x1f,0x80,0x00,0x00,0x00,0x00};
958 /* nopl 0L(%[re]ax,%[re]ax,1) */
959 static const char alt_8[] =
960 {0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
961 /* nopw 0L(%[re]ax,%[re]ax,1) */
962 static const char alt_9[] =
963 {0x66,0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
964 /* nopw %cs:0L(%[re]ax,%[re]ax,1) */
965 static const char alt_10[] =
966 {0x66,0x2e,0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
967 /* data16
968 nopw %cs:0L(%[re]ax,%[re]ax,1) */
969 static const char alt_long_11[] =
970 {0x66,
971 0x66,0x2e,0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
972 /* data16
973 data16
974 nopw %cs:0L(%[re]ax,%[re]ax,1) */
975 static const char alt_long_12[] =
976 {0x66,
977 0x66,
978 0x66,0x2e,0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
979 /* data16
980 data16
981 data16
982 nopw %cs:0L(%[re]ax,%[re]ax,1) */
983 static const char alt_long_13[] =
984 {0x66,
985 0x66,
986 0x66,
987 0x66,0x2e,0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
988 /* data16
989 data16
990 data16
991 data16
992 nopw %cs:0L(%[re]ax,%[re]ax,1) */
993 static const char alt_long_14[] =
994 {0x66,
995 0x66,
996 0x66,
997 0x66,
998 0x66,0x2e,0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
999 /* data16
1000 data16
1001 data16
1002 data16
1003 data16
1004 nopw %cs:0L(%[re]ax,%[re]ax,1) */
1005 static const char alt_long_15[] =
1006 {0x66,
1007 0x66,
1008 0x66,
1009 0x66,
1010 0x66,
1011 0x66,0x2e,0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
1012 /* nopl 0(%[re]ax,%[re]ax,1)
1013 nopw 0(%[re]ax,%[re]ax,1) */
1014 static const char alt_short_11[] =
1015 {0x0f,0x1f,0x44,0x00,0x00,
1016 0x66,0x0f,0x1f,0x44,0x00,0x00};
1017 /* nopw 0(%[re]ax,%[re]ax,1)
1018 nopw 0(%[re]ax,%[re]ax,1) */
1019 static const char alt_short_12[] =
1020 {0x66,0x0f,0x1f,0x44,0x00,0x00,
1021 0x66,0x0f,0x1f,0x44,0x00,0x00};
1022 /* nopw 0(%[re]ax,%[re]ax,1)
1023 nopl 0L(%[re]ax) */
1024 static const char alt_short_13[] =
1025 {0x66,0x0f,0x1f,0x44,0x00,0x00,
1026 0x0f,0x1f,0x80,0x00,0x00,0x00,0x00};
1027 /* nopl 0L(%[re]ax)
1028 nopl 0L(%[re]ax) */
1029 static const char alt_short_14[] =
1030 {0x0f,0x1f,0x80,0x00,0x00,0x00,0x00,
1031 0x0f,0x1f,0x80,0x00,0x00,0x00,0x00};
1032 /* nopl 0L(%[re]ax)
1033 nopl 0L(%[re]ax,%[re]ax,1) */
1034 static const char alt_short_15[] =
1035 {0x0f,0x1f,0x80,0x00,0x00,0x00,0x00,
1036 0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
1037 static const char *const alt_short_patt[] = {
1038 f32_1, f32_2, alt_3, alt_4, alt_5, alt_6, alt_7, alt_8,
1039 alt_9, alt_10, alt_short_11, alt_short_12, alt_short_13,
1040 alt_short_14, alt_short_15
1041 };
1042 static const char *const alt_long_patt[] = {
1043 f32_1, f32_2, alt_3, alt_4, alt_5, alt_6, alt_7, alt_8,
1044 alt_9, alt_10, alt_long_11, alt_long_12, alt_long_13,
1045 alt_long_14, alt_long_15
1046 };
1047
1048 /* Only align for at least a positive non-zero boundary. */
1049 if (count <= 0 || count > MAX_MEM_FOR_RS_ALIGN_CODE)
1050 return;
1051
1052 /* We need to decide which NOP sequence to use for 32bit and
1053 64bit. When -mtune= is used:
1054
1055 1. For PROCESSOR_I386, PROCESSOR_I486, PROCESSOR_PENTIUM and
1056 PROCESSOR_GENERIC32, f32_patt will be used.
1057 2. For PROCESSOR_PENTIUMPRO, PROCESSOR_PENTIUM4, PROCESSOR_NOCONA,
1058 PROCESSOR_CORE, PROCESSOR_CORE2, PROCESSOR_COREI7, and
1059 PROCESSOR_GENERIC64, alt_long_patt will be used.
1060 3. For PROCESSOR_ATHLON, PROCESSOR_K6, PROCESSOR_K8 and
1061 PROCESSOR_AMDFAM10, and PROCESSOR_BD, alt_short_patt
1062 will be used.
1063
1064 When -mtune= isn't used, alt_long_patt will be used if
1065 cpu_arch_isa_flags has CpuNop. Otherwise, f32_patt will
1066 be used.
1067
1068 When -march= or .arch is used, we can't use anything beyond
1069 cpu_arch_isa_flags. */
1070
1071 if (flag_code == CODE_16BIT)
1072 {
1073 if (count > 8)
1074 {
1075 memcpy (fragP->fr_literal + fragP->fr_fix,
1076 jump_31, count);
1077 /* Adjust jump offset. */
1078 fragP->fr_literal[fragP->fr_fix + 1] = count - 2;
1079 }
1080 else
1081 memcpy (fragP->fr_literal + fragP->fr_fix,
1082 f16_patt[count - 1], count);
1083 }
1084 else
1085 {
1086 const char *const *patt = NULL;
1087
1088 if (fragP->tc_frag_data.isa == PROCESSOR_UNKNOWN)
1089 {
1090 /* PROCESSOR_UNKNOWN means that all ISAs may be used. */
1091 switch (cpu_arch_tune)
1092 {
1093 case PROCESSOR_UNKNOWN:
1094 /* We use cpu_arch_isa_flags to check if we SHOULD
1095 optimize with nops. */
1096 if (fragP->tc_frag_data.isa_flags.bitfield.cpunop)
1097 patt = alt_long_patt;
1098 else
1099 patt = f32_patt;
1100 break;
1101 case PROCESSOR_PENTIUM4:
1102 case PROCESSOR_NOCONA:
1103 case PROCESSOR_CORE:
1104 case PROCESSOR_CORE2:
1105 case PROCESSOR_COREI7:
1106 case PROCESSOR_L1OM:
1107 case PROCESSOR_K1OM:
1108 case PROCESSOR_GENERIC64:
1109 patt = alt_long_patt;
1110 break;
1111 case PROCESSOR_K6:
1112 case PROCESSOR_ATHLON:
1113 case PROCESSOR_K8:
1114 case PROCESSOR_AMDFAM10:
1115 case PROCESSOR_BD:
1116 patt = alt_short_patt;
1117 break;
1118 case PROCESSOR_I386:
1119 case PROCESSOR_I486:
1120 case PROCESSOR_PENTIUM:
1121 case PROCESSOR_PENTIUMPRO:
1122 case PROCESSOR_GENERIC32:
1123 patt = f32_patt;
1124 break;
1125 }
1126 }
1127 else
1128 {
1129 switch (fragP->tc_frag_data.tune)
1130 {
1131 case PROCESSOR_UNKNOWN:
1132 /* When cpu_arch_isa is set, cpu_arch_tune shouldn't be
1133 PROCESSOR_UNKNOWN. */
1134 abort ();
1135 break;
1136
1137 case PROCESSOR_I386:
1138 case PROCESSOR_I486:
1139 case PROCESSOR_PENTIUM:
1140 case PROCESSOR_K6:
1141 case PROCESSOR_ATHLON:
1142 case PROCESSOR_K8:
1143 case PROCESSOR_AMDFAM10:
1144 case PROCESSOR_BD:
1145 case PROCESSOR_GENERIC32:
1146 /* We use cpu_arch_isa_flags to check if we CAN optimize
1147 with nops. */
1148 if (fragP->tc_frag_data.isa_flags.bitfield.cpunop)
1149 patt = alt_short_patt;
1150 else
1151 patt = f32_patt;
1152 break;
1153 case PROCESSOR_PENTIUMPRO:
1154 case PROCESSOR_PENTIUM4:
1155 case PROCESSOR_NOCONA:
1156 case PROCESSOR_CORE:
1157 case PROCESSOR_CORE2:
1158 case PROCESSOR_COREI7:
1159 case PROCESSOR_L1OM:
1160 case PROCESSOR_K1OM:
1161 if (fragP->tc_frag_data.isa_flags.bitfield.cpunop)
1162 patt = alt_long_patt;
1163 else
1164 patt = f32_patt;
1165 break;
1166 case PROCESSOR_GENERIC64:
1167 patt = alt_long_patt;
1168 break;
1169 }
1170 }
1171
1172 if (patt == f32_patt)
1173 {
1174 /* If the padding is less than 15 bytes, we use the normal
1175 ones. Otherwise, we use a jump instruction and adjust
1176 its offset. */
1177 int limit;
1178
1179 /* For 64bit, the limit is 3 bytes. */
1180 if (flag_code == CODE_64BIT
1181 && fragP->tc_frag_data.isa_flags.bitfield.cpulm)
1182 limit = 3;
1183 else
1184 limit = 15;
1185 if (count < limit)
1186 memcpy (fragP->fr_literal + fragP->fr_fix,
1187 patt[count - 1], count);
1188 else
1189 {
1190 memcpy (fragP->fr_literal + fragP->fr_fix,
1191 jump_31, count);
1192 /* Adjust jump offset. */
1193 fragP->fr_literal[fragP->fr_fix + 1] = count - 2;
1194 }
1195 }
1196 else
1197 {
1198 /* Maximum length of an instruction is 15 byte. If the
1199 padding is greater than 15 bytes and we don't use jump,
1200 we have to break it into smaller pieces. */
1201 int padding = count;
1202 while (padding > 15)
1203 {
1204 padding -= 15;
1205 memcpy (fragP->fr_literal + fragP->fr_fix + padding,
1206 patt [14], 15);
1207 }
1208
1209 if (padding)
1210 memcpy (fragP->fr_literal + fragP->fr_fix,
1211 patt [padding - 1], padding);
1212 }
1213 }
1214 fragP->fr_var = count;
1215 }
1216
1217 static INLINE int
1218 operand_type_all_zero (const union i386_operand_type *x)
1219 {
1220 switch (ARRAY_SIZE(x->array))
1221 {
1222 case 3:
1223 if (x->array[2])
1224 return 0;
1225 case 2:
1226 if (x->array[1])
1227 return 0;
1228 case 1:
1229 return !x->array[0];
1230 default:
1231 abort ();
1232 }
1233 }
1234
1235 static INLINE void
1236 operand_type_set (union i386_operand_type *x, unsigned int v)
1237 {
1238 switch (ARRAY_SIZE(x->array))
1239 {
1240 case 3:
1241 x->array[2] = v;
1242 case 2:
1243 x->array[1] = v;
1244 case 1:
1245 x->array[0] = v;
1246 break;
1247 default:
1248 abort ();
1249 }
1250 }
1251
1252 static INLINE int
1253 operand_type_equal (const union i386_operand_type *x,
1254 const union i386_operand_type *y)
1255 {
1256 switch (ARRAY_SIZE(x->array))
1257 {
1258 case 3:
1259 if (x->array[2] != y->array[2])
1260 return 0;
1261 case 2:
1262 if (x->array[1] != y->array[1])
1263 return 0;
1264 case 1:
1265 return x->array[0] == y->array[0];
1266 break;
1267 default:
1268 abort ();
1269 }
1270 }
1271
1272 static INLINE int
1273 cpu_flags_all_zero (const union i386_cpu_flags *x)
1274 {
1275 switch (ARRAY_SIZE(x->array))
1276 {
1277 case 3:
1278 if (x->array[2])
1279 return 0;
1280 case 2:
1281 if (x->array[1])
1282 return 0;
1283 case 1:
1284 return !x->array[0];
1285 default:
1286 abort ();
1287 }
1288 }
1289
1290 static INLINE void
1291 cpu_flags_set (union i386_cpu_flags *x, unsigned int v)
1292 {
1293 switch (ARRAY_SIZE(x->array))
1294 {
1295 case 3:
1296 x->array[2] = v;
1297 case 2:
1298 x->array[1] = v;
1299 case 1:
1300 x->array[0] = v;
1301 break;
1302 default:
1303 abort ();
1304 }
1305 }
1306
1307 static INLINE int
1308 cpu_flags_equal (const union i386_cpu_flags *x,
1309 const union i386_cpu_flags *y)
1310 {
1311 switch (ARRAY_SIZE(x->array))
1312 {
1313 case 3:
1314 if (x->array[2] != y->array[2])
1315 return 0;
1316 case 2:
1317 if (x->array[1] != y->array[1])
1318 return 0;
1319 case 1:
1320 return x->array[0] == y->array[0];
1321 break;
1322 default:
1323 abort ();
1324 }
1325 }
1326
1327 static INLINE int
1328 cpu_flags_check_cpu64 (i386_cpu_flags f)
1329 {
1330 return !((flag_code == CODE_64BIT && f.bitfield.cpuno64)
1331 || (flag_code != CODE_64BIT && f.bitfield.cpu64));
1332 }
1333
1334 static INLINE i386_cpu_flags
1335 cpu_flags_and (i386_cpu_flags x, i386_cpu_flags y)
1336 {
1337 switch (ARRAY_SIZE (x.array))
1338 {
1339 case 3:
1340 x.array [2] &= y.array [2];
1341 case 2:
1342 x.array [1] &= y.array [1];
1343 case 1:
1344 x.array [0] &= y.array [0];
1345 break;
1346 default:
1347 abort ();
1348 }
1349 return x;
1350 }
1351
1352 static INLINE i386_cpu_flags
1353 cpu_flags_or (i386_cpu_flags x, i386_cpu_flags y)
1354 {
1355 switch (ARRAY_SIZE (x.array))
1356 {
1357 case 3:
1358 x.array [2] |= y.array [2];
1359 case 2:
1360 x.array [1] |= y.array [1];
1361 case 1:
1362 x.array [0] |= y.array [0];
1363 break;
1364 default:
1365 abort ();
1366 }
1367 return x;
1368 }
1369
1370 static INLINE i386_cpu_flags
1371 cpu_flags_and_not (i386_cpu_flags x, i386_cpu_flags y)
1372 {
1373 switch (ARRAY_SIZE (x.array))
1374 {
1375 case 3:
1376 x.array [2] &= ~y.array [2];
1377 case 2:
1378 x.array [1] &= ~y.array [1];
1379 case 1:
1380 x.array [0] &= ~y.array [0];
1381 break;
1382 default:
1383 abort ();
1384 }
1385 return x;
1386 }
1387
1388 #define CPU_FLAGS_ARCH_MATCH 0x1
1389 #define CPU_FLAGS_64BIT_MATCH 0x2
1390 #define CPU_FLAGS_AES_MATCH 0x4
1391 #define CPU_FLAGS_PCLMUL_MATCH 0x8
1392 #define CPU_FLAGS_AVX_MATCH 0x10
1393
1394 #define CPU_FLAGS_32BIT_MATCH \
1395 (CPU_FLAGS_ARCH_MATCH | CPU_FLAGS_AES_MATCH \
1396 | CPU_FLAGS_PCLMUL_MATCH | CPU_FLAGS_AVX_MATCH)
1397 #define CPU_FLAGS_PERFECT_MATCH \
1398 (CPU_FLAGS_32BIT_MATCH | CPU_FLAGS_64BIT_MATCH)
1399
1400 /* Return CPU flags match bits. */
1401
1402 static int
1403 cpu_flags_match (const insn_template *t)
1404 {
1405 i386_cpu_flags x = t->cpu_flags;
1406 int match = cpu_flags_check_cpu64 (x) ? CPU_FLAGS_64BIT_MATCH : 0;
1407
1408 x.bitfield.cpu64 = 0;
1409 x.bitfield.cpuno64 = 0;
1410
1411 if (cpu_flags_all_zero (&x))
1412 {
1413 /* This instruction is available on all archs. */
1414 match |= CPU_FLAGS_32BIT_MATCH;
1415 }
1416 else
1417 {
1418 /* This instruction is available only on some archs. */
1419 i386_cpu_flags cpu = cpu_arch_flags;
1420
1421 cpu.bitfield.cpu64 = 0;
1422 cpu.bitfield.cpuno64 = 0;
1423 cpu = cpu_flags_and (x, cpu);
1424 if (!cpu_flags_all_zero (&cpu))
1425 {
1426 if (x.bitfield.cpuavx)
1427 {
1428 /* We only need to check AES/PCLMUL/SSE2AVX with AVX. */
1429 if (cpu.bitfield.cpuavx)
1430 {
1431 /* Check SSE2AVX. */
1432 if (!t->opcode_modifier.sse2avx|| sse2avx)
1433 {
1434 match |= (CPU_FLAGS_ARCH_MATCH
1435 | CPU_FLAGS_AVX_MATCH);
1436 /* Check AES. */
1437 if (!x.bitfield.cpuaes || cpu.bitfield.cpuaes)
1438 match |= CPU_FLAGS_AES_MATCH;
1439 /* Check PCLMUL. */
1440 if (!x.bitfield.cpupclmul
1441 || cpu.bitfield.cpupclmul)
1442 match |= CPU_FLAGS_PCLMUL_MATCH;
1443 }
1444 }
1445 else
1446 match |= CPU_FLAGS_ARCH_MATCH;
1447 }
1448 else
1449 match |= CPU_FLAGS_32BIT_MATCH;
1450 }
1451 }
1452 return match;
1453 }
1454
1455 static INLINE i386_operand_type
1456 operand_type_and (i386_operand_type x, i386_operand_type y)
1457 {
1458 switch (ARRAY_SIZE (x.array))
1459 {
1460 case 3:
1461 x.array [2] &= y.array [2];
1462 case 2:
1463 x.array [1] &= y.array [1];
1464 case 1:
1465 x.array [0] &= y.array [0];
1466 break;
1467 default:
1468 abort ();
1469 }
1470 return x;
1471 }
1472
1473 static INLINE i386_operand_type
1474 operand_type_or (i386_operand_type x, i386_operand_type y)
1475 {
1476 switch (ARRAY_SIZE (x.array))
1477 {
1478 case 3:
1479 x.array [2] |= y.array [2];
1480 case 2:
1481 x.array [1] |= y.array [1];
1482 case 1:
1483 x.array [0] |= y.array [0];
1484 break;
1485 default:
1486 abort ();
1487 }
1488 return x;
1489 }
1490
1491 static INLINE i386_operand_type
1492 operand_type_xor (i386_operand_type x, i386_operand_type y)
1493 {
1494 switch (ARRAY_SIZE (x.array))
1495 {
1496 case 3:
1497 x.array [2] ^= y.array [2];
1498 case 2:
1499 x.array [1] ^= y.array [1];
1500 case 1:
1501 x.array [0] ^= y.array [0];
1502 break;
1503 default:
1504 abort ();
1505 }
1506 return x;
1507 }
1508
1509 static const i386_operand_type acc32 = OPERAND_TYPE_ACC32;
1510 static const i386_operand_type acc64 = OPERAND_TYPE_ACC64;
1511 static const i386_operand_type control = OPERAND_TYPE_CONTROL;
1512 static const i386_operand_type inoutportreg
1513 = OPERAND_TYPE_INOUTPORTREG;
1514 static const i386_operand_type reg16_inoutportreg
1515 = OPERAND_TYPE_REG16_INOUTPORTREG;
1516 static const i386_operand_type disp16 = OPERAND_TYPE_DISP16;
1517 static const i386_operand_type disp32 = OPERAND_TYPE_DISP32;
1518 static const i386_operand_type disp32s = OPERAND_TYPE_DISP32S;
1519 static const i386_operand_type disp16_32 = OPERAND_TYPE_DISP16_32;
1520 static const i386_operand_type anydisp
1521 = OPERAND_TYPE_ANYDISP;
1522 static const i386_operand_type regxmm = OPERAND_TYPE_REGXMM;
1523 static const i386_operand_type regymm = OPERAND_TYPE_REGYMM;
1524 static const i386_operand_type imm8 = OPERAND_TYPE_IMM8;
1525 static const i386_operand_type imm8s = OPERAND_TYPE_IMM8S;
1526 static const i386_operand_type imm16 = OPERAND_TYPE_IMM16;
1527 static const i386_operand_type imm32 = OPERAND_TYPE_IMM32;
1528 static const i386_operand_type imm32s = OPERAND_TYPE_IMM32S;
1529 static const i386_operand_type imm64 = OPERAND_TYPE_IMM64;
1530 static const i386_operand_type imm16_32 = OPERAND_TYPE_IMM16_32;
1531 static const i386_operand_type imm16_32s = OPERAND_TYPE_IMM16_32S;
1532 static const i386_operand_type imm16_32_32s = OPERAND_TYPE_IMM16_32_32S;
1533 static const i386_operand_type vec_imm4 = OPERAND_TYPE_VEC_IMM4;
1534
1535 enum operand_type
1536 {
1537 reg,
1538 imm,
1539 disp,
1540 anymem
1541 };
1542
1543 static INLINE int
1544 operand_type_check (i386_operand_type t, enum operand_type c)
1545 {
1546 switch (c)
1547 {
1548 case reg:
1549 return (t.bitfield.reg8
1550 || t.bitfield.reg16
1551 || t.bitfield.reg32
1552 || t.bitfield.reg64);
1553
1554 case imm:
1555 return (t.bitfield.imm8
1556 || t.bitfield.imm8s
1557 || t.bitfield.imm16
1558 || t.bitfield.imm32
1559 || t.bitfield.imm32s
1560 || t.bitfield.imm64);
1561
1562 case disp:
1563 return (t.bitfield.disp8
1564 || t.bitfield.disp16
1565 || t.bitfield.disp32
1566 || t.bitfield.disp32s
1567 || t.bitfield.disp64);
1568
1569 case anymem:
1570 return (t.bitfield.disp8
1571 || t.bitfield.disp16
1572 || t.bitfield.disp32
1573 || t.bitfield.disp32s
1574 || t.bitfield.disp64
1575 || t.bitfield.baseindex);
1576
1577 default:
1578 abort ();
1579 }
1580
1581 return 0;
1582 }
1583
1584 /* Return 1 if there is no conflict in 8bit/16bit/32bit/64bit on
1585 operand J for instruction template T. */
1586
1587 static INLINE int
1588 match_reg_size (const insn_template *t, unsigned int j)
1589 {
1590 return !((i.types[j].bitfield.byte
1591 && !t->operand_types[j].bitfield.byte)
1592 || (i.types[j].bitfield.word
1593 && !t->operand_types[j].bitfield.word)
1594 || (i.types[j].bitfield.dword
1595 && !t->operand_types[j].bitfield.dword)
1596 || (i.types[j].bitfield.qword
1597 && !t->operand_types[j].bitfield.qword));
1598 }
1599
1600 /* Return 1 if there is no conflict in any size on operand J for
1601 instruction template T. */
1602
1603 static INLINE int
1604 match_mem_size (const insn_template *t, unsigned int j)
1605 {
1606 return (match_reg_size (t, j)
1607 && !((i.types[j].bitfield.unspecified
1608 && !t->operand_types[j].bitfield.unspecified)
1609 || (i.types[j].bitfield.fword
1610 && !t->operand_types[j].bitfield.fword)
1611 || (i.types[j].bitfield.tbyte
1612 && !t->operand_types[j].bitfield.tbyte)
1613 || (i.types[j].bitfield.xmmword
1614 && !t->operand_types[j].bitfield.xmmword)
1615 || (i.types[j].bitfield.ymmword
1616 && !t->operand_types[j].bitfield.ymmword)));
1617 }
1618
1619 /* Return 1 if there is no size conflict on any operands for
1620 instruction template T. */
1621
1622 static INLINE int
1623 operand_size_match (const insn_template *t)
1624 {
1625 unsigned int j;
1626 int match = 1;
1627
1628 /* Don't check jump instructions. */
1629 if (t->opcode_modifier.jump
1630 || t->opcode_modifier.jumpbyte
1631 || t->opcode_modifier.jumpdword
1632 || t->opcode_modifier.jumpintersegment)
1633 return match;
1634
1635 /* Check memory and accumulator operand size. */
1636 for (j = 0; j < i.operands; j++)
1637 {
1638 if (t->operand_types[j].bitfield.anysize)
1639 continue;
1640
1641 if (t->operand_types[j].bitfield.acc && !match_reg_size (t, j))
1642 {
1643 match = 0;
1644 break;
1645 }
1646
1647 if (i.types[j].bitfield.mem && !match_mem_size (t, j))
1648 {
1649 match = 0;
1650 break;
1651 }
1652 }
1653
1654 if (match)
1655 return match;
1656 else if (!t->opcode_modifier.d && !t->opcode_modifier.floatd)
1657 {
1658 mismatch:
1659 i.error = operand_size_mismatch;
1660 return 0;
1661 }
1662
1663 /* Check reverse. */
1664 gas_assert (i.operands == 2);
1665
1666 match = 1;
1667 for (j = 0; j < 2; j++)
1668 {
1669 if (t->operand_types[j].bitfield.acc
1670 && !match_reg_size (t, j ? 0 : 1))
1671 goto mismatch;
1672
1673 if (i.types[j].bitfield.mem
1674 && !match_mem_size (t, j ? 0 : 1))
1675 goto mismatch;
1676 }
1677
1678 return match;
1679 }
1680
1681 static INLINE int
1682 operand_type_match (i386_operand_type overlap,
1683 i386_operand_type given)
1684 {
1685 i386_operand_type temp = overlap;
1686
1687 temp.bitfield.jumpabsolute = 0;
1688 temp.bitfield.unspecified = 0;
1689 temp.bitfield.byte = 0;
1690 temp.bitfield.word = 0;
1691 temp.bitfield.dword = 0;
1692 temp.bitfield.fword = 0;
1693 temp.bitfield.qword = 0;
1694 temp.bitfield.tbyte = 0;
1695 temp.bitfield.xmmword = 0;
1696 temp.bitfield.ymmword = 0;
1697 if (operand_type_all_zero (&temp))
1698 goto mismatch;
1699
1700 if (given.bitfield.baseindex == overlap.bitfield.baseindex
1701 && given.bitfield.jumpabsolute == overlap.bitfield.jumpabsolute)
1702 return 1;
1703
1704 mismatch:
1705 i.error = operand_type_mismatch;
1706 return 0;
1707 }
1708
1709 /* If given types g0 and g1 are registers they must be of the same type
1710 unless the expected operand type register overlap is null.
1711 Note that Acc in a template matches every size of reg. */
1712
1713 static INLINE int
1714 operand_type_register_match (i386_operand_type m0,
1715 i386_operand_type g0,
1716 i386_operand_type t0,
1717 i386_operand_type m1,
1718 i386_operand_type g1,
1719 i386_operand_type t1)
1720 {
1721 if (!operand_type_check (g0, reg))
1722 return 1;
1723
1724 if (!operand_type_check (g1, reg))
1725 return 1;
1726
1727 if (g0.bitfield.reg8 == g1.bitfield.reg8
1728 && g0.bitfield.reg16 == g1.bitfield.reg16
1729 && g0.bitfield.reg32 == g1.bitfield.reg32
1730 && g0.bitfield.reg64 == g1.bitfield.reg64)
1731 return 1;
1732
1733 if (m0.bitfield.acc)
1734 {
1735 t0.bitfield.reg8 = 1;
1736 t0.bitfield.reg16 = 1;
1737 t0.bitfield.reg32 = 1;
1738 t0.bitfield.reg64 = 1;
1739 }
1740
1741 if (m1.bitfield.acc)
1742 {
1743 t1.bitfield.reg8 = 1;
1744 t1.bitfield.reg16 = 1;
1745 t1.bitfield.reg32 = 1;
1746 t1.bitfield.reg64 = 1;
1747 }
1748
1749 if (!(t0.bitfield.reg8 & t1.bitfield.reg8)
1750 && !(t0.bitfield.reg16 & t1.bitfield.reg16)
1751 && !(t0.bitfield.reg32 & t1.bitfield.reg32)
1752 && !(t0.bitfield.reg64 & t1.bitfield.reg64))
1753 return 1;
1754
1755 i.error = register_type_mismatch;
1756
1757 return 0;
1758 }
1759
1760 static INLINE unsigned int
1761 mode_from_disp_size (i386_operand_type t)
1762 {
1763 if (t.bitfield.disp8)
1764 return 1;
1765 else if (t.bitfield.disp16
1766 || t.bitfield.disp32
1767 || t.bitfield.disp32s)
1768 return 2;
1769 else
1770 return 0;
1771 }
1772
1773 static INLINE int
1774 fits_in_signed_byte (offsetT num)
1775 {
1776 return (num >= -128) && (num <= 127);
1777 }
1778
1779 static INLINE int
1780 fits_in_unsigned_byte (offsetT num)
1781 {
1782 return (num & 0xff) == num;
1783 }
1784
1785 static INLINE int
1786 fits_in_unsigned_word (offsetT num)
1787 {
1788 return (num & 0xffff) == num;
1789 }
1790
1791 static INLINE int
1792 fits_in_signed_word (offsetT num)
1793 {
1794 return (-32768 <= num) && (num <= 32767);
1795 }
1796
1797 static INLINE int
1798 fits_in_signed_long (offsetT num ATTRIBUTE_UNUSED)
1799 {
1800 #ifndef BFD64
1801 return 1;
1802 #else
1803 return (!(((offsetT) -1 << 31) & num)
1804 || (((offsetT) -1 << 31) & num) == ((offsetT) -1 << 31));
1805 #endif
1806 } /* fits_in_signed_long() */
1807
1808 static INLINE int
1809 fits_in_unsigned_long (offsetT num ATTRIBUTE_UNUSED)
1810 {
1811 #ifndef BFD64
1812 return 1;
1813 #else
1814 return (num & (((offsetT) 2 << 31) - 1)) == num;
1815 #endif
1816 } /* fits_in_unsigned_long() */
1817
1818 static INLINE int
1819 fits_in_imm4 (offsetT num)
1820 {
1821 return (num & 0xf) == num;
1822 }
1823
1824 static i386_operand_type
1825 smallest_imm_type (offsetT num)
1826 {
1827 i386_operand_type t;
1828
1829 operand_type_set (&t, 0);
1830 t.bitfield.imm64 = 1;
1831
1832 if (cpu_arch_tune != PROCESSOR_I486 && num == 1)
1833 {
1834 /* This code is disabled on the 486 because all the Imm1 forms
1835 in the opcode table are slower on the i486. They're the
1836 versions with the implicitly specified single-position
1837 displacement, which has another syntax if you really want to
1838 use that form. */
1839 t.bitfield.imm1 = 1;
1840 t.bitfield.imm8 = 1;
1841 t.bitfield.imm8s = 1;
1842 t.bitfield.imm16 = 1;
1843 t.bitfield.imm32 = 1;
1844 t.bitfield.imm32s = 1;
1845 }
1846 else if (fits_in_signed_byte (num))
1847 {
1848 t.bitfield.imm8 = 1;
1849 t.bitfield.imm8s = 1;
1850 t.bitfield.imm16 = 1;
1851 t.bitfield.imm32 = 1;
1852 t.bitfield.imm32s = 1;
1853 }
1854 else if (fits_in_unsigned_byte (num))
1855 {
1856 t.bitfield.imm8 = 1;
1857 t.bitfield.imm16 = 1;
1858 t.bitfield.imm32 = 1;
1859 t.bitfield.imm32s = 1;
1860 }
1861 else if (fits_in_signed_word (num) || fits_in_unsigned_word (num))
1862 {
1863 t.bitfield.imm16 = 1;
1864 t.bitfield.imm32 = 1;
1865 t.bitfield.imm32s = 1;
1866 }
1867 else if (fits_in_signed_long (num))
1868 {
1869 t.bitfield.imm32 = 1;
1870 t.bitfield.imm32s = 1;
1871 }
1872 else if (fits_in_unsigned_long (num))
1873 t.bitfield.imm32 = 1;
1874
1875 return t;
1876 }
1877
1878 static offsetT
1879 offset_in_range (offsetT val, int size)
1880 {
1881 addressT mask;
1882
1883 switch (size)
1884 {
1885 case 1: mask = ((addressT) 1 << 8) - 1; break;
1886 case 2: mask = ((addressT) 1 << 16) - 1; break;
1887 case 4: mask = ((addressT) 2 << 31) - 1; break;
1888 #ifdef BFD64
1889 case 8: mask = ((addressT) 2 << 63) - 1; break;
1890 #endif
1891 default: abort ();
1892 }
1893
1894 #ifdef BFD64
1895 /* If BFD64, sign extend val for 32bit address mode. */
1896 if (flag_code != CODE_64BIT
1897 || i.prefix[ADDR_PREFIX])
1898 if ((val & ~(((addressT) 2 << 31) - 1)) == 0)
1899 val = (val ^ ((addressT) 1 << 31)) - ((addressT) 1 << 31);
1900 #endif
1901
1902 if ((val & ~mask) != 0 && (val & ~mask) != ~mask)
1903 {
1904 char buf1[40], buf2[40];
1905
1906 sprint_value (buf1, val);
1907 sprint_value (buf2, val & mask);
1908 as_warn (_("%s shortened to %s"), buf1, buf2);
1909 }
1910 return val & mask;
1911 }
1912
1913 enum PREFIX_GROUP
1914 {
1915 PREFIX_EXIST = 0,
1916 PREFIX_LOCK,
1917 PREFIX_REP,
1918 PREFIX_OTHER
1919 };
1920
1921 /* Returns
1922 a. PREFIX_EXIST if attempting to add a prefix where one from the
1923 same class already exists.
1924 b. PREFIX_LOCK if lock prefix is added.
1925 c. PREFIX_REP if rep/repne prefix is added.
1926 d. PREFIX_OTHER if other prefix is added.
1927 */
1928
1929 static enum PREFIX_GROUP
1930 add_prefix (unsigned int prefix)
1931 {
1932 enum PREFIX_GROUP ret = PREFIX_OTHER;
1933 unsigned int q;
1934
1935 if (prefix >= REX_OPCODE && prefix < REX_OPCODE + 16
1936 && flag_code == CODE_64BIT)
1937 {
1938 if ((i.prefix[REX_PREFIX] & prefix & REX_W)
1939 || ((i.prefix[REX_PREFIX] & (REX_R | REX_X | REX_B))
1940 && (prefix & (REX_R | REX_X | REX_B))))
1941 ret = PREFIX_EXIST;
1942 q = REX_PREFIX;
1943 }
1944 else
1945 {
1946 switch (prefix)
1947 {
1948 default:
1949 abort ();
1950
1951 case CS_PREFIX_OPCODE:
1952 case DS_PREFIX_OPCODE:
1953 case ES_PREFIX_OPCODE:
1954 case FS_PREFIX_OPCODE:
1955 case GS_PREFIX_OPCODE:
1956 case SS_PREFIX_OPCODE:
1957 q = SEG_PREFIX;
1958 break;
1959
1960 case REPNE_PREFIX_OPCODE:
1961 case REPE_PREFIX_OPCODE:
1962 q = REP_PREFIX;
1963 ret = PREFIX_REP;
1964 break;
1965
1966 case LOCK_PREFIX_OPCODE:
1967 q = LOCK_PREFIX;
1968 ret = PREFIX_LOCK;
1969 break;
1970
1971 case FWAIT_OPCODE:
1972 q = WAIT_PREFIX;
1973 break;
1974
1975 case ADDR_PREFIX_OPCODE:
1976 q = ADDR_PREFIX;
1977 break;
1978
1979 case DATA_PREFIX_OPCODE:
1980 q = DATA_PREFIX;
1981 break;
1982 }
1983 if (i.prefix[q] != 0)
1984 ret = PREFIX_EXIST;
1985 }
1986
1987 if (ret)
1988 {
1989 if (!i.prefix[q])
1990 ++i.prefixes;
1991 i.prefix[q] |= prefix;
1992 }
1993 else
1994 as_bad (_("same type of prefix used twice"));
1995
1996 return ret;
1997 }
1998
1999 static void
2000 update_code_flag (int value, int check)
2001 {
2002 PRINTF_LIKE ((*as_error));
2003
2004 flag_code = (enum flag_code) value;
2005 if (flag_code == CODE_64BIT)
2006 {
2007 cpu_arch_flags.bitfield.cpu64 = 1;
2008 cpu_arch_flags.bitfield.cpuno64 = 0;
2009 }
2010 else
2011 {
2012 cpu_arch_flags.bitfield.cpu64 = 0;
2013 cpu_arch_flags.bitfield.cpuno64 = 1;
2014 }
2015 if (value == CODE_64BIT && !cpu_arch_flags.bitfield.cpulm )
2016 {
2017 if (check)
2018 as_error = as_fatal;
2019 else
2020 as_error = as_bad;
2021 (*as_error) (_("64bit mode not supported on `%s'."),
2022 cpu_arch_name ? cpu_arch_name : default_arch);
2023 }
2024 if (value == CODE_32BIT && !cpu_arch_flags.bitfield.cpui386)
2025 {
2026 if (check)
2027 as_error = as_fatal;
2028 else
2029 as_error = as_bad;
2030 (*as_error) (_("32bit mode not supported on `%s'."),
2031 cpu_arch_name ? cpu_arch_name : default_arch);
2032 }
2033 stackop_size = '\0';
2034 }
2035
2036 static void
2037 set_code_flag (int value)
2038 {
2039 update_code_flag (value, 0);
2040 }
2041
2042 static void
2043 set_16bit_gcc_code_flag (int new_code_flag)
2044 {
2045 flag_code = (enum flag_code) new_code_flag;
2046 if (flag_code != CODE_16BIT)
2047 abort ();
2048 cpu_arch_flags.bitfield.cpu64 = 0;
2049 cpu_arch_flags.bitfield.cpuno64 = 1;
2050 stackop_size = LONG_MNEM_SUFFIX;
2051 }
2052
2053 static void
2054 set_intel_syntax (int syntax_flag)
2055 {
2056 /* Find out if register prefixing is specified. */
2057 int ask_naked_reg = 0;
2058
2059 SKIP_WHITESPACE ();
2060 if (!is_end_of_line[(unsigned char) *input_line_pointer])
2061 {
2062 char *string = input_line_pointer;
2063 int e = get_symbol_end ();
2064
2065 if (strcmp (string, "prefix") == 0)
2066 ask_naked_reg = 1;
2067 else if (strcmp (string, "noprefix") == 0)
2068 ask_naked_reg = -1;
2069 else
2070 as_bad (_("bad argument to syntax directive."));
2071 *input_line_pointer = e;
2072 }
2073 demand_empty_rest_of_line ();
2074
2075 intel_syntax = syntax_flag;
2076
2077 if (ask_naked_reg == 0)
2078 allow_naked_reg = (intel_syntax
2079 && (bfd_get_symbol_leading_char (stdoutput) != '\0'));
2080 else
2081 allow_naked_reg = (ask_naked_reg < 0);
2082
2083 expr_set_rank (O_full_ptr, syntax_flag ? 10 : 0);
2084
2085 identifier_chars['%'] = intel_syntax && allow_naked_reg ? '%' : 0;
2086 identifier_chars['$'] = intel_syntax ? '$' : 0;
2087 register_prefix = allow_naked_reg ? "" : "%";
2088 }
2089
2090 static void
2091 set_intel_mnemonic (int mnemonic_flag)
2092 {
2093 intel_mnemonic = mnemonic_flag;
2094 }
2095
2096 static void
2097 set_allow_index_reg (int flag)
2098 {
2099 allow_index_reg = flag;
2100 }
2101
2102 static void
2103 set_sse_check (int dummy ATTRIBUTE_UNUSED)
2104 {
2105 SKIP_WHITESPACE ();
2106
2107 if (!is_end_of_line[(unsigned char) *input_line_pointer])
2108 {
2109 char *string = input_line_pointer;
2110 int e = get_symbol_end ();
2111
2112 if (strcmp (string, "none") == 0)
2113 sse_check = sse_check_none;
2114 else if (strcmp (string, "warning") == 0)
2115 sse_check = sse_check_warning;
2116 else if (strcmp (string, "error") == 0)
2117 sse_check = sse_check_error;
2118 else
2119 as_bad (_("bad argument to sse_check directive."));
2120 *input_line_pointer = e;
2121 }
2122 else
2123 as_bad (_("missing argument for sse_check directive"));
2124
2125 demand_empty_rest_of_line ();
2126 }
2127
2128 static void
2129 check_cpu_arch_compatible (const char *name ATTRIBUTE_UNUSED,
2130 i386_cpu_flags new_flag ATTRIBUTE_UNUSED)
2131 {
2132 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
2133 static const char *arch;
2134
2135 /* Intel LIOM is only supported on ELF. */
2136 if (!IS_ELF)
2137 return;
2138
2139 if (!arch)
2140 {
2141 /* Use cpu_arch_name if it is set in md_parse_option. Otherwise
2142 use default_arch. */
2143 arch = cpu_arch_name;
2144 if (!arch)
2145 arch = default_arch;
2146 }
2147
2148 /* If we are targeting Intel L1OM, we must enable it. */
2149 if (get_elf_backend_data (stdoutput)->elf_machine_code != EM_L1OM
2150 || new_flag.bitfield.cpul1om)
2151 return;
2152
2153 /* If we are targeting Intel K1OM, we must enable it. */
2154 if (get_elf_backend_data (stdoutput)->elf_machine_code != EM_K1OM
2155 || new_flag.bitfield.cpuk1om)
2156 return;
2157
2158 as_bad (_("`%s' is not supported on `%s'"), name, arch);
2159 #endif
2160 }
2161
2162 static void
2163 set_cpu_arch (int dummy ATTRIBUTE_UNUSED)
2164 {
2165 SKIP_WHITESPACE ();
2166
2167 if (!is_end_of_line[(unsigned char) *input_line_pointer])
2168 {
2169 char *string = input_line_pointer;
2170 int e = get_symbol_end ();
2171 unsigned int j;
2172 i386_cpu_flags flags;
2173
2174 for (j = 0; j < ARRAY_SIZE (cpu_arch); j++)
2175 {
2176 if (strcmp (string, cpu_arch[j].name) == 0)
2177 {
2178 check_cpu_arch_compatible (string, cpu_arch[j].flags);
2179
2180 if (*string != '.')
2181 {
2182 cpu_arch_name = cpu_arch[j].name;
2183 cpu_sub_arch_name = NULL;
2184 cpu_arch_flags = cpu_arch[j].flags;
2185 if (flag_code == CODE_64BIT)
2186 {
2187 cpu_arch_flags.bitfield.cpu64 = 1;
2188 cpu_arch_flags.bitfield.cpuno64 = 0;
2189 }
2190 else
2191 {
2192 cpu_arch_flags.bitfield.cpu64 = 0;
2193 cpu_arch_flags.bitfield.cpuno64 = 1;
2194 }
2195 cpu_arch_isa = cpu_arch[j].type;
2196 cpu_arch_isa_flags = cpu_arch[j].flags;
2197 if (!cpu_arch_tune_set)
2198 {
2199 cpu_arch_tune = cpu_arch_isa;
2200 cpu_arch_tune_flags = cpu_arch_isa_flags;
2201 }
2202 break;
2203 }
2204
2205 if (!cpu_arch[j].negated)
2206 flags = cpu_flags_or (cpu_arch_flags,
2207 cpu_arch[j].flags);
2208 else
2209 flags = cpu_flags_and_not (cpu_arch_flags,
2210 cpu_arch[j].flags);
2211 if (!cpu_flags_equal (&flags, &cpu_arch_flags))
2212 {
2213 if (cpu_sub_arch_name)
2214 {
2215 char *name = cpu_sub_arch_name;
2216 cpu_sub_arch_name = concat (name,
2217 cpu_arch[j].name,
2218 (const char *) NULL);
2219 free (name);
2220 }
2221 else
2222 cpu_sub_arch_name = xstrdup (cpu_arch[j].name);
2223 cpu_arch_flags = flags;
2224 cpu_arch_isa_flags = flags;
2225 }
2226 *input_line_pointer = e;
2227 demand_empty_rest_of_line ();
2228 return;
2229 }
2230 }
2231 if (j >= ARRAY_SIZE (cpu_arch))
2232 as_bad (_("no such architecture: `%s'"), string);
2233
2234 *input_line_pointer = e;
2235 }
2236 else
2237 as_bad (_("missing cpu architecture"));
2238
2239 no_cond_jump_promotion = 0;
2240 if (*input_line_pointer == ','
2241 && !is_end_of_line[(unsigned char) input_line_pointer[1]])
2242 {
2243 char *string = ++input_line_pointer;
2244 int e = get_symbol_end ();
2245
2246 if (strcmp (string, "nojumps") == 0)
2247 no_cond_jump_promotion = 1;
2248 else if (strcmp (string, "jumps") == 0)
2249 ;
2250 else
2251 as_bad (_("no such architecture modifier: `%s'"), string);
2252
2253 *input_line_pointer = e;
2254 }
2255
2256 demand_empty_rest_of_line ();
2257 }
2258
2259 enum bfd_architecture
2260 i386_arch (void)
2261 {
2262 if (cpu_arch_isa == PROCESSOR_L1OM)
2263 {
2264 if (OUTPUT_FLAVOR != bfd_target_elf_flavour
2265 || flag_code != CODE_64BIT)
2266 as_fatal (_("Intel L1OM is 64bit ELF only"));
2267 return bfd_arch_l1om;
2268 }
2269 else if (cpu_arch_isa == PROCESSOR_K1OM)
2270 {
2271 if (OUTPUT_FLAVOR != bfd_target_elf_flavour
2272 || flag_code != CODE_64BIT)
2273 as_fatal (_("Intel K1OM is 64bit ELF only"));
2274 return bfd_arch_k1om;
2275 }
2276 else
2277 return bfd_arch_i386;
2278 }
2279
2280 unsigned long
2281 i386_mach (void)
2282 {
2283 if (!strncmp (default_arch, "x86_64", 6))
2284 {
2285 if (cpu_arch_isa == PROCESSOR_L1OM)
2286 {
2287 if (OUTPUT_FLAVOR != bfd_target_elf_flavour
2288 || default_arch[6] != '\0')
2289 as_fatal (_("Intel L1OM is 64bit ELF only"));
2290 return bfd_mach_l1om;
2291 }
2292 else if (cpu_arch_isa == PROCESSOR_K1OM)
2293 {
2294 if (OUTPUT_FLAVOR != bfd_target_elf_flavour
2295 || default_arch[6] != '\0')
2296 as_fatal (_("Intel K1OM is 64bit ELF only"));
2297 return bfd_mach_k1om;
2298 }
2299 else if (default_arch[6] == '\0')
2300 return bfd_mach_x86_64;
2301 else
2302 return bfd_mach_x64_32;
2303 }
2304 else if (!strcmp (default_arch, "i386"))
2305 return bfd_mach_i386_i386;
2306 else
2307 as_fatal (_("unknown architecture"));
2308 }
2309 \f
2310 void
2311 md_begin (void)
2312 {
2313 const char *hash_err;
2314
2315 /* Initialize op_hash hash table. */
2316 op_hash = hash_new ();
2317
2318 {
2319 const insn_template *optab;
2320 templates *core_optab;
2321
2322 /* Setup for loop. */
2323 optab = i386_optab;
2324 core_optab = (templates *) xmalloc (sizeof (templates));
2325 core_optab->start = optab;
2326
2327 while (1)
2328 {
2329 ++optab;
2330 if (optab->name == NULL
2331 || strcmp (optab->name, (optab - 1)->name) != 0)
2332 {
2333 /* different name --> ship out current template list;
2334 add to hash table; & begin anew. */
2335 core_optab->end = optab;
2336 hash_err = hash_insert (op_hash,
2337 (optab - 1)->name,
2338 (void *) core_optab);
2339 if (hash_err)
2340 {
2341 as_fatal (_("internal Error: Can't hash %s: %s"),
2342 (optab - 1)->name,
2343 hash_err);
2344 }
2345 if (optab->name == NULL)
2346 break;
2347 core_optab = (templates *) xmalloc (sizeof (templates));
2348 core_optab->start = optab;
2349 }
2350 }
2351 }
2352
2353 /* Initialize reg_hash hash table. */
2354 reg_hash = hash_new ();
2355 {
2356 const reg_entry *regtab;
2357 unsigned int regtab_size = i386_regtab_size;
2358
2359 for (regtab = i386_regtab; regtab_size--; regtab++)
2360 {
2361 hash_err = hash_insert (reg_hash, regtab->reg_name, (void *) regtab);
2362 if (hash_err)
2363 as_fatal (_("internal Error: Can't hash %s: %s"),
2364 regtab->reg_name,
2365 hash_err);
2366 }
2367 }
2368
2369 /* Fill in lexical tables: mnemonic_chars, operand_chars. */
2370 {
2371 int c;
2372 char *p;
2373
2374 for (c = 0; c < 256; c++)
2375 {
2376 if (ISDIGIT (c))
2377 {
2378 digit_chars[c] = c;
2379 mnemonic_chars[c] = c;
2380 register_chars[c] = c;
2381 operand_chars[c] = c;
2382 }
2383 else if (ISLOWER (c))
2384 {
2385 mnemonic_chars[c] = c;
2386 register_chars[c] = c;
2387 operand_chars[c] = c;
2388 }
2389 else if (ISUPPER (c))
2390 {
2391 mnemonic_chars[c] = TOLOWER (c);
2392 register_chars[c] = mnemonic_chars[c];
2393 operand_chars[c] = c;
2394 }
2395
2396 if (ISALPHA (c) || ISDIGIT (c))
2397 identifier_chars[c] = c;
2398 else if (c >= 128)
2399 {
2400 identifier_chars[c] = c;
2401 operand_chars[c] = c;
2402 }
2403 }
2404
2405 #ifdef LEX_AT
2406 identifier_chars['@'] = '@';
2407 #endif
2408 #ifdef LEX_QM
2409 identifier_chars['?'] = '?';
2410 operand_chars['?'] = '?';
2411 #endif
2412 digit_chars['-'] = '-';
2413 mnemonic_chars['_'] = '_';
2414 mnemonic_chars['-'] = '-';
2415 mnemonic_chars['.'] = '.';
2416 identifier_chars['_'] = '_';
2417 identifier_chars['.'] = '.';
2418
2419 for (p = operand_special_chars; *p != '\0'; p++)
2420 operand_chars[(unsigned char) *p] = *p;
2421 }
2422
2423 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
2424 if (IS_ELF)
2425 {
2426 record_alignment (text_section, 2);
2427 record_alignment (data_section, 2);
2428 record_alignment (bss_section, 2);
2429 }
2430 #endif
2431
2432 if (flag_code == CODE_64BIT)
2433 {
2434 #if defined (OBJ_COFF) && defined (TE_PE)
2435 x86_dwarf2_return_column = (OUTPUT_FLAVOR == bfd_target_coff_flavour
2436 ? 32 : 16);
2437 #else
2438 x86_dwarf2_return_column = 16;
2439 #endif
2440 x86_cie_data_alignment = -8;
2441 }
2442 else
2443 {
2444 x86_dwarf2_return_column = 8;
2445 x86_cie_data_alignment = -4;
2446 }
2447 }
2448
2449 void
2450 i386_print_statistics (FILE *file)
2451 {
2452 hash_print_statistics (file, "i386 opcode", op_hash);
2453 hash_print_statistics (file, "i386 register", reg_hash);
2454 }
2455 \f
2456 #ifdef DEBUG386
2457
2458 /* Debugging routines for md_assemble. */
2459 static void pte (insn_template *);
2460 static void pt (i386_operand_type);
2461 static void pe (expressionS *);
2462 static void ps (symbolS *);
2463
2464 static void
2465 pi (char *line, i386_insn *x)
2466 {
2467 unsigned int j;
2468
2469 fprintf (stdout, "%s: template ", line);
2470 pte (&x->tm);
2471 fprintf (stdout, " address: base %s index %s scale %x\n",
2472 x->base_reg ? x->base_reg->reg_name : "none",
2473 x->index_reg ? x->index_reg->reg_name : "none",
2474 x->log2_scale_factor);
2475 fprintf (stdout, " modrm: mode %x reg %x reg/mem %x\n",
2476 x->rm.mode, x->rm.reg, x->rm.regmem);
2477 fprintf (stdout, " sib: base %x index %x scale %x\n",
2478 x->sib.base, x->sib.index, x->sib.scale);
2479 fprintf (stdout, " rex: 64bit %x extX %x extY %x extZ %x\n",
2480 (x->rex & REX_W) != 0,
2481 (x->rex & REX_R) != 0,
2482 (x->rex & REX_X) != 0,
2483 (x->rex & REX_B) != 0);
2484 for (j = 0; j < x->operands; j++)
2485 {
2486 fprintf (stdout, " #%d: ", j + 1);
2487 pt (x->types[j]);
2488 fprintf (stdout, "\n");
2489 if (x->types[j].bitfield.reg8
2490 || x->types[j].bitfield.reg16
2491 || x->types[j].bitfield.reg32
2492 || x->types[j].bitfield.reg64
2493 || x->types[j].bitfield.regmmx
2494 || x->types[j].bitfield.regxmm
2495 || x->types[j].bitfield.regymm
2496 || x->types[j].bitfield.sreg2
2497 || x->types[j].bitfield.sreg3
2498 || x->types[j].bitfield.control
2499 || x->types[j].bitfield.debug
2500 || x->types[j].bitfield.test)
2501 fprintf (stdout, "%s\n", x->op[j].regs->reg_name);
2502 if (operand_type_check (x->types[j], imm))
2503 pe (x->op[j].imms);
2504 if (operand_type_check (x->types[j], disp))
2505 pe (x->op[j].disps);
2506 }
2507 }
2508
2509 static void
2510 pte (insn_template *t)
2511 {
2512 unsigned int j;
2513 fprintf (stdout, " %d operands ", t->operands);
2514 fprintf (stdout, "opcode %x ", t->base_opcode);
2515 if (t->extension_opcode != None)
2516 fprintf (stdout, "ext %x ", t->extension_opcode);
2517 if (t->opcode_modifier.d)
2518 fprintf (stdout, "D");
2519 if (t->opcode_modifier.w)
2520 fprintf (stdout, "W");
2521 fprintf (stdout, "\n");
2522 for (j = 0; j < t->operands; j++)
2523 {
2524 fprintf (stdout, " #%d type ", j + 1);
2525 pt (t->operand_types[j]);
2526 fprintf (stdout, "\n");
2527 }
2528 }
2529
2530 static void
2531 pe (expressionS *e)
2532 {
2533 fprintf (stdout, " operation %d\n", e->X_op);
2534 fprintf (stdout, " add_number %ld (%lx)\n",
2535 (long) e->X_add_number, (long) e->X_add_number);
2536 if (e->X_add_symbol)
2537 {
2538 fprintf (stdout, " add_symbol ");
2539 ps (e->X_add_symbol);
2540 fprintf (stdout, "\n");
2541 }
2542 if (e->X_op_symbol)
2543 {
2544 fprintf (stdout, " op_symbol ");
2545 ps (e->X_op_symbol);
2546 fprintf (stdout, "\n");
2547 }
2548 }
2549
2550 static void
2551 ps (symbolS *s)
2552 {
2553 fprintf (stdout, "%s type %s%s",
2554 S_GET_NAME (s),
2555 S_IS_EXTERNAL (s) ? "EXTERNAL " : "",
2556 segment_name (S_GET_SEGMENT (s)));
2557 }
2558
2559 static struct type_name
2560 {
2561 i386_operand_type mask;
2562 const char *name;
2563 }
2564 const type_names[] =
2565 {
2566 { OPERAND_TYPE_REG8, "r8" },
2567 { OPERAND_TYPE_REG16, "r16" },
2568 { OPERAND_TYPE_REG32, "r32" },
2569 { OPERAND_TYPE_REG64, "r64" },
2570 { OPERAND_TYPE_IMM8, "i8" },
2571 { OPERAND_TYPE_IMM8, "i8s" },
2572 { OPERAND_TYPE_IMM16, "i16" },
2573 { OPERAND_TYPE_IMM32, "i32" },
2574 { OPERAND_TYPE_IMM32S, "i32s" },
2575 { OPERAND_TYPE_IMM64, "i64" },
2576 { OPERAND_TYPE_IMM1, "i1" },
2577 { OPERAND_TYPE_BASEINDEX, "BaseIndex" },
2578 { OPERAND_TYPE_DISP8, "d8" },
2579 { OPERAND_TYPE_DISP16, "d16" },
2580 { OPERAND_TYPE_DISP32, "d32" },
2581 { OPERAND_TYPE_DISP32S, "d32s" },
2582 { OPERAND_TYPE_DISP64, "d64" },
2583 { OPERAND_TYPE_INOUTPORTREG, "InOutPortReg" },
2584 { OPERAND_TYPE_SHIFTCOUNT, "ShiftCount" },
2585 { OPERAND_TYPE_CONTROL, "control reg" },
2586 { OPERAND_TYPE_TEST, "test reg" },
2587 { OPERAND_TYPE_DEBUG, "debug reg" },
2588 { OPERAND_TYPE_FLOATREG, "FReg" },
2589 { OPERAND_TYPE_FLOATACC, "FAcc" },
2590 { OPERAND_TYPE_SREG2, "SReg2" },
2591 { OPERAND_TYPE_SREG3, "SReg3" },
2592 { OPERAND_TYPE_ACC, "Acc" },
2593 { OPERAND_TYPE_JUMPABSOLUTE, "Jump Absolute" },
2594 { OPERAND_TYPE_REGMMX, "rMMX" },
2595 { OPERAND_TYPE_REGXMM, "rXMM" },
2596 { OPERAND_TYPE_REGYMM, "rYMM" },
2597 { OPERAND_TYPE_ESSEG, "es" },
2598 };
2599
2600 static void
2601 pt (i386_operand_type t)
2602 {
2603 unsigned int j;
2604 i386_operand_type a;
2605
2606 for (j = 0; j < ARRAY_SIZE (type_names); j++)
2607 {
2608 a = operand_type_and (t, type_names[j].mask);
2609 if (!operand_type_all_zero (&a))
2610 fprintf (stdout, "%s, ", type_names[j].name);
2611 }
2612 fflush (stdout);
2613 }
2614
2615 #endif /* DEBUG386 */
2616 \f
2617 static bfd_reloc_code_real_type
2618 reloc (unsigned int size,
2619 int pcrel,
2620 int sign,
2621 bfd_reloc_code_real_type other)
2622 {
2623 if (other != NO_RELOC)
2624 {
2625 reloc_howto_type *rel;
2626
2627 if (size == 8)
2628 switch (other)
2629 {
2630 case BFD_RELOC_X86_64_GOT32:
2631 return BFD_RELOC_X86_64_GOT64;
2632 break;
2633 case BFD_RELOC_X86_64_PLTOFF64:
2634 return BFD_RELOC_X86_64_PLTOFF64;
2635 break;
2636 case BFD_RELOC_X86_64_GOTPC32:
2637 other = BFD_RELOC_X86_64_GOTPC64;
2638 break;
2639 case BFD_RELOC_X86_64_GOTPCREL:
2640 other = BFD_RELOC_X86_64_GOTPCREL64;
2641 break;
2642 case BFD_RELOC_X86_64_TPOFF32:
2643 other = BFD_RELOC_X86_64_TPOFF64;
2644 break;
2645 case BFD_RELOC_X86_64_DTPOFF32:
2646 other = BFD_RELOC_X86_64_DTPOFF64;
2647 break;
2648 default:
2649 break;
2650 }
2651
2652 /* Sign-checking 4-byte relocations in 16-/32-bit code is pointless. */
2653 if (size == 4 && (flag_code != CODE_64BIT || disallow_64bit_reloc))
2654 sign = -1;
2655
2656 rel = bfd_reloc_type_lookup (stdoutput, other);
2657 if (!rel)
2658 as_bad (_("unknown relocation (%u)"), other);
2659 else if (size != bfd_get_reloc_size (rel))
2660 as_bad (_("%u-byte relocation cannot be applied to %u-byte field"),
2661 bfd_get_reloc_size (rel),
2662 size);
2663 else if (pcrel && !rel->pc_relative)
2664 as_bad (_("non-pc-relative relocation for pc-relative field"));
2665 else if ((rel->complain_on_overflow == complain_overflow_signed
2666 && !sign)
2667 || (rel->complain_on_overflow == complain_overflow_unsigned
2668 && sign > 0))
2669 as_bad (_("relocated field and relocation type differ in signedness"));
2670 else
2671 return other;
2672 return NO_RELOC;
2673 }
2674
2675 if (pcrel)
2676 {
2677 if (!sign)
2678 as_bad (_("there are no unsigned pc-relative relocations"));
2679 switch (size)
2680 {
2681 case 1: return BFD_RELOC_8_PCREL;
2682 case 2: return BFD_RELOC_16_PCREL;
2683 case 4: return BFD_RELOC_32_PCREL;
2684 case 8: return BFD_RELOC_64_PCREL;
2685 }
2686 as_bad (_("cannot do %u byte pc-relative relocation"), size);
2687 }
2688 else
2689 {
2690 if (sign > 0)
2691 switch (size)
2692 {
2693 case 4: return BFD_RELOC_X86_64_32S;
2694 }
2695 else
2696 switch (size)
2697 {
2698 case 1: return BFD_RELOC_8;
2699 case 2: return BFD_RELOC_16;
2700 case 4: return BFD_RELOC_32;
2701 case 8: return BFD_RELOC_64;
2702 }
2703 as_bad (_("cannot do %s %u byte relocation"),
2704 sign > 0 ? "signed" : "unsigned", size);
2705 }
2706
2707 return NO_RELOC;
2708 }
2709
2710 /* Here we decide which fixups can be adjusted to make them relative to
2711 the beginning of the section instead of the symbol. Basically we need
2712 to make sure that the dynamic relocations are done correctly, so in
2713 some cases we force the original symbol to be used. */
2714
2715 int
2716 tc_i386_fix_adjustable (fixS *fixP ATTRIBUTE_UNUSED)
2717 {
2718 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
2719 if (!IS_ELF)
2720 return 1;
2721
2722 /* Don't adjust pc-relative references to merge sections in 64-bit
2723 mode. */
2724 if (use_rela_relocations
2725 && (S_GET_SEGMENT (fixP->fx_addsy)->flags & SEC_MERGE) != 0
2726 && fixP->fx_pcrel)
2727 return 0;
2728
2729 /* The x86_64 GOTPCREL are represented as 32bit PCrel relocations
2730 and changed later by validate_fix. */
2731 if (GOT_symbol && fixP->fx_subsy == GOT_symbol
2732 && fixP->fx_r_type == BFD_RELOC_32_PCREL)
2733 return 0;
2734
2735 /* adjust_reloc_syms doesn't know about the GOT. */
2736 if (fixP->fx_r_type == BFD_RELOC_386_GOTOFF
2737 || fixP->fx_r_type == BFD_RELOC_386_PLT32
2738 || fixP->fx_r_type == BFD_RELOC_386_GOT32
2739 || fixP->fx_r_type == BFD_RELOC_386_TLS_GD
2740 || fixP->fx_r_type == BFD_RELOC_386_TLS_LDM
2741 || fixP->fx_r_type == BFD_RELOC_386_TLS_LDO_32
2742 || fixP->fx_r_type == BFD_RELOC_386_TLS_IE_32
2743 || fixP->fx_r_type == BFD_RELOC_386_TLS_IE
2744 || fixP->fx_r_type == BFD_RELOC_386_TLS_GOTIE
2745 || fixP->fx_r_type == BFD_RELOC_386_TLS_LE_32
2746 || fixP->fx_r_type == BFD_RELOC_386_TLS_LE
2747 || fixP->fx_r_type == BFD_RELOC_386_TLS_GOTDESC
2748 || fixP->fx_r_type == BFD_RELOC_386_TLS_DESC_CALL
2749 || fixP->fx_r_type == BFD_RELOC_X86_64_PLT32
2750 || fixP->fx_r_type == BFD_RELOC_X86_64_GOT32
2751 || fixP->fx_r_type == BFD_RELOC_X86_64_GOTPCREL
2752 || fixP->fx_r_type == BFD_RELOC_X86_64_TLSGD
2753 || fixP->fx_r_type == BFD_RELOC_X86_64_TLSLD
2754 || fixP->fx_r_type == BFD_RELOC_X86_64_DTPOFF32
2755 || fixP->fx_r_type == BFD_RELOC_X86_64_DTPOFF64
2756 || fixP->fx_r_type == BFD_RELOC_X86_64_GOTTPOFF
2757 || fixP->fx_r_type == BFD_RELOC_X86_64_TPOFF32
2758 || fixP->fx_r_type == BFD_RELOC_X86_64_TPOFF64
2759 || fixP->fx_r_type == BFD_RELOC_X86_64_GOTOFF64
2760 || fixP->fx_r_type == BFD_RELOC_X86_64_GOTPC32_TLSDESC
2761 || fixP->fx_r_type == BFD_RELOC_X86_64_TLSDESC_CALL
2762 || fixP->fx_r_type == BFD_RELOC_VTABLE_INHERIT
2763 || fixP->fx_r_type == BFD_RELOC_VTABLE_ENTRY)
2764 return 0;
2765 #endif
2766 return 1;
2767 }
2768
2769 static int
2770 intel_float_operand (const char *mnemonic)
2771 {
2772 /* Note that the value returned is meaningful only for opcodes with (memory)
2773 operands, hence the code here is free to improperly handle opcodes that
2774 have no operands (for better performance and smaller code). */
2775
2776 if (mnemonic[0] != 'f')
2777 return 0; /* non-math */
2778
2779 switch (mnemonic[1])
2780 {
2781 /* fclex, fdecstp, fdisi, femms, feni, fincstp, finit, fsetpm, and
2782 the fs segment override prefix not currently handled because no
2783 call path can make opcodes without operands get here */
2784 case 'i':
2785 return 2 /* integer op */;
2786 case 'l':
2787 if (mnemonic[2] == 'd' && (mnemonic[3] == 'c' || mnemonic[3] == 'e'))
2788 return 3; /* fldcw/fldenv */
2789 break;
2790 case 'n':
2791 if (mnemonic[2] != 'o' /* fnop */)
2792 return 3; /* non-waiting control op */
2793 break;
2794 case 'r':
2795 if (mnemonic[2] == 's')
2796 return 3; /* frstor/frstpm */
2797 break;
2798 case 's':
2799 if (mnemonic[2] == 'a')
2800 return 3; /* fsave */
2801 if (mnemonic[2] == 't')
2802 {
2803 switch (mnemonic[3])
2804 {
2805 case 'c': /* fstcw */
2806 case 'd': /* fstdw */
2807 case 'e': /* fstenv */
2808 case 's': /* fsts[gw] */
2809 return 3;
2810 }
2811 }
2812 break;
2813 case 'x':
2814 if (mnemonic[2] == 'r' || mnemonic[2] == 's')
2815 return 0; /* fxsave/fxrstor are not really math ops */
2816 break;
2817 }
2818
2819 return 1;
2820 }
2821
2822 /* Build the VEX prefix. */
2823
2824 static void
2825 build_vex_prefix (const insn_template *t)
2826 {
2827 unsigned int register_specifier;
2828 unsigned int implied_prefix;
2829 unsigned int vector_length;
2830
2831 /* Check register specifier. */
2832 if (i.vex.register_specifier)
2833 {
2834 register_specifier = i.vex.register_specifier->reg_num;
2835 if ((i.vex.register_specifier->reg_flags & RegRex))
2836 register_specifier += 8;
2837 register_specifier = ~register_specifier & 0xf;
2838 }
2839 else
2840 register_specifier = 0xf;
2841
2842 /* Use 2-byte VEX prefix by swappping destination and source
2843 operand. */
2844 if (!i.swap_operand
2845 && i.operands == i.reg_operands
2846 && i.tm.opcode_modifier.vexopcode == VEX0F
2847 && i.tm.opcode_modifier.s
2848 && i.rex == REX_B)
2849 {
2850 unsigned int xchg = i.operands - 1;
2851 union i386_op temp_op;
2852 i386_operand_type temp_type;
2853
2854 temp_type = i.types[xchg];
2855 i.types[xchg] = i.types[0];
2856 i.types[0] = temp_type;
2857 temp_op = i.op[xchg];
2858 i.op[xchg] = i.op[0];
2859 i.op[0] = temp_op;
2860
2861 gas_assert (i.rm.mode == 3);
2862
2863 i.rex = REX_R;
2864 xchg = i.rm.regmem;
2865 i.rm.regmem = i.rm.reg;
2866 i.rm.reg = xchg;
2867
2868 /* Use the next insn. */
2869 i.tm = t[1];
2870 }
2871
2872 if (i.tm.opcode_modifier.vex == VEXScalar)
2873 vector_length = avxscalar;
2874 else
2875 vector_length = i.tm.opcode_modifier.vex == VEX256 ? 1 : 0;
2876
2877 switch ((i.tm.base_opcode >> 8) & 0xff)
2878 {
2879 case 0:
2880 implied_prefix = 0;
2881 break;
2882 case DATA_PREFIX_OPCODE:
2883 implied_prefix = 1;
2884 break;
2885 case REPE_PREFIX_OPCODE:
2886 implied_prefix = 2;
2887 break;
2888 case REPNE_PREFIX_OPCODE:
2889 implied_prefix = 3;
2890 break;
2891 default:
2892 abort ();
2893 }
2894
2895 /* Use 2-byte VEX prefix if possible. */
2896 if (i.tm.opcode_modifier.vexopcode == VEX0F
2897 && i.tm.opcode_modifier.vexw != VEXW1
2898 && (i.rex & (REX_W | REX_X | REX_B)) == 0)
2899 {
2900 /* 2-byte VEX prefix. */
2901 unsigned int r;
2902
2903 i.vex.length = 2;
2904 i.vex.bytes[0] = 0xc5;
2905
2906 /* Check the REX.R bit. */
2907 r = (i.rex & REX_R) ? 0 : 1;
2908 i.vex.bytes[1] = (r << 7
2909 | register_specifier << 3
2910 | vector_length << 2
2911 | implied_prefix);
2912 }
2913 else
2914 {
2915 /* 3-byte VEX prefix. */
2916 unsigned int m, w;
2917
2918 i.vex.length = 3;
2919
2920 switch (i.tm.opcode_modifier.vexopcode)
2921 {
2922 case VEX0F:
2923 m = 0x1;
2924 i.vex.bytes[0] = 0xc4;
2925 break;
2926 case VEX0F38:
2927 m = 0x2;
2928 i.vex.bytes[0] = 0xc4;
2929 break;
2930 case VEX0F3A:
2931 m = 0x3;
2932 i.vex.bytes[0] = 0xc4;
2933 break;
2934 case XOP08:
2935 m = 0x8;
2936 i.vex.bytes[0] = 0x8f;
2937 break;
2938 case XOP09:
2939 m = 0x9;
2940 i.vex.bytes[0] = 0x8f;
2941 break;
2942 case XOP0A:
2943 m = 0xa;
2944 i.vex.bytes[0] = 0x8f;
2945 break;
2946 default:
2947 abort ();
2948 }
2949
2950 /* The high 3 bits of the second VEX byte are 1's compliment
2951 of RXB bits from REX. */
2952 i.vex.bytes[1] = (~i.rex & 0x7) << 5 | m;
2953
2954 /* Check the REX.W bit. */
2955 w = (i.rex & REX_W) ? 1 : 0;
2956 if (i.tm.opcode_modifier.vexw)
2957 {
2958 if (w)
2959 abort ();
2960
2961 if (i.tm.opcode_modifier.vexw == VEXW1)
2962 w = 1;
2963 }
2964
2965 i.vex.bytes[2] = (w << 7
2966 | register_specifier << 3
2967 | vector_length << 2
2968 | implied_prefix);
2969 }
2970 }
2971
2972 static void
2973 process_immext (void)
2974 {
2975 expressionS *exp;
2976
2977 if (i.tm.cpu_flags.bitfield.cpusse3 && i.operands > 0)
2978 {
2979 /* SSE3 Instructions have the fixed operands with an opcode
2980 suffix which is coded in the same place as an 8-bit immediate
2981 field would be. Here we check those operands and remove them
2982 afterwards. */
2983 unsigned int x;
2984
2985 for (x = 0; x < i.operands; x++)
2986 if (i.op[x].regs->reg_num != x)
2987 as_bad (_("can't use register '%s%s' as operand %d in '%s'."),
2988 register_prefix, i.op[x].regs->reg_name, x + 1,
2989 i.tm.name);
2990
2991 i.operands = 0;
2992 }
2993
2994 /* These AMD 3DNow! and SSE2 instructions have an opcode suffix
2995 which is coded in the same place as an 8-bit immediate field
2996 would be. Here we fake an 8-bit immediate operand from the
2997 opcode suffix stored in tm.extension_opcode.
2998
2999 AVX instructions also use this encoding, for some of
3000 3 argument instructions. */
3001
3002 gas_assert (i.imm_operands == 0
3003 && (i.operands <= 2
3004 || (i.tm.opcode_modifier.vex
3005 && i.operands <= 4)));
3006
3007 exp = &im_expressions[i.imm_operands++];
3008 i.op[i.operands].imms = exp;
3009 i.types[i.operands] = imm8;
3010 i.operands++;
3011 exp->X_op = O_constant;
3012 exp->X_add_number = i.tm.extension_opcode;
3013 i.tm.extension_opcode = None;
3014 }
3015
3016
3017 static int
3018 check_hle (void)
3019 {
3020 switch (i.tm.opcode_modifier.hleprefixok)
3021 {
3022 default:
3023 abort ();
3024 case HLEPrefixNone:
3025 if (i.prefix[HLE_PREFIX] == XACQUIRE_PREFIX_OPCODE)
3026 as_bad (_("invalid instruction `%s' after `xacquire'"),
3027 i.tm.name);
3028 else
3029 as_bad (_("invalid instruction `%s' after `xrelease'"),
3030 i.tm.name);
3031 return 0;
3032 case HLEPrefixLock:
3033 if (i.prefix[LOCK_PREFIX])
3034 return 1;
3035 if (i.prefix[HLE_PREFIX] == XACQUIRE_PREFIX_OPCODE)
3036 as_bad (_("missing `lock' with `xacquire'"));
3037 else
3038 as_bad (_("missing `lock' with `xrelease'"));
3039 return 0;
3040 case HLEPrefixAny:
3041 return 1;
3042 case HLEPrefixRelease:
3043 if (i.prefix[HLE_PREFIX] != XRELEASE_PREFIX_OPCODE)
3044 {
3045 as_bad (_("instruction `%s' after `xacquire' not allowed"),
3046 i.tm.name);
3047 return 0;
3048 }
3049 if (i.mem_operands == 0
3050 || !operand_type_check (i.types[i.operands - 1], anymem))
3051 {
3052 as_bad (_("memory destination needed for instruction `%s'"
3053 " after `xrelease'"), i.tm.name);
3054 return 0;
3055 }
3056 return 1;
3057 }
3058 }
3059
3060 /* This is the guts of the machine-dependent assembler. LINE points to a
3061 machine dependent instruction. This function is supposed to emit
3062 the frags/bytes it assembles to. */
3063
3064 void
3065 md_assemble (char *line)
3066 {
3067 unsigned int j;
3068 char mnemonic[MAX_MNEM_SIZE];
3069 const insn_template *t;
3070
3071 /* Initialize globals. */
3072 memset (&i, '\0', sizeof (i));
3073 for (j = 0; j < MAX_OPERANDS; j++)
3074 i.reloc[j] = NO_RELOC;
3075 memset (disp_expressions, '\0', sizeof (disp_expressions));
3076 memset (im_expressions, '\0', sizeof (im_expressions));
3077 save_stack_p = save_stack;
3078
3079 /* First parse an instruction mnemonic & call i386_operand for the operands.
3080 We assume that the scrubber has arranged it so that line[0] is the valid
3081 start of a (possibly prefixed) mnemonic. */
3082
3083 line = parse_insn (line, mnemonic);
3084 if (line == NULL)
3085 return;
3086
3087 line = parse_operands (line, mnemonic);
3088 this_operand = -1;
3089 if (line == NULL)
3090 return;
3091
3092 /* Now we've parsed the mnemonic into a set of templates, and have the
3093 operands at hand. */
3094
3095 /* All intel opcodes have reversed operands except for "bound" and
3096 "enter". We also don't reverse intersegment "jmp" and "call"
3097 instructions with 2 immediate operands so that the immediate segment
3098 precedes the offset, as it does when in AT&T mode. */
3099 if (intel_syntax
3100 && i.operands > 1
3101 && (strcmp (mnemonic, "bound") != 0)
3102 && (strcmp (mnemonic, "invlpga") != 0)
3103 && !(operand_type_check (i.types[0], imm)
3104 && operand_type_check (i.types[1], imm)))
3105 swap_operands ();
3106
3107 /* The order of the immediates should be reversed
3108 for 2 immediates extrq and insertq instructions */
3109 if (i.imm_operands == 2
3110 && (strcmp (mnemonic, "extrq") == 0
3111 || strcmp (mnemonic, "insertq") == 0))
3112 swap_2_operands (0, 1);
3113
3114 if (i.imm_operands)
3115 optimize_imm ();
3116
3117 /* Don't optimize displacement for movabs since it only takes 64bit
3118 displacement. */
3119 if (i.disp_operands
3120 && i.disp_encoding != disp_encoding_32bit
3121 && (flag_code != CODE_64BIT
3122 || strcmp (mnemonic, "movabs") != 0))
3123 optimize_disp ();
3124
3125 /* Next, we find a template that matches the given insn,
3126 making sure the overlap of the given operands types is consistent
3127 with the template operand types. */
3128
3129 if (!(t = match_template ()))
3130 return;
3131
3132 if (sse_check != sse_check_none
3133 && !i.tm.opcode_modifier.noavx
3134 && (i.tm.cpu_flags.bitfield.cpusse
3135 || i.tm.cpu_flags.bitfield.cpusse2
3136 || i.tm.cpu_flags.bitfield.cpusse3
3137 || i.tm.cpu_flags.bitfield.cpussse3
3138 || i.tm.cpu_flags.bitfield.cpusse4_1
3139 || i.tm.cpu_flags.bitfield.cpusse4_2))
3140 {
3141 (sse_check == sse_check_warning
3142 ? as_warn
3143 : as_bad) (_("SSE instruction `%s' is used"), i.tm.name);
3144 }
3145
3146 /* Zap movzx and movsx suffix. The suffix has been set from
3147 "word ptr" or "byte ptr" on the source operand in Intel syntax
3148 or extracted from mnemonic in AT&T syntax. But we'll use
3149 the destination register to choose the suffix for encoding. */
3150 if ((i.tm.base_opcode & ~9) == 0x0fb6)
3151 {
3152 /* In Intel syntax, there must be a suffix. In AT&T syntax, if
3153 there is no suffix, the default will be byte extension. */
3154 if (i.reg_operands != 2
3155 && !i.suffix
3156 && intel_syntax)
3157 as_bad (_("ambiguous operand size for `%s'"), i.tm.name);
3158
3159 i.suffix = 0;
3160 }
3161
3162 if (i.tm.opcode_modifier.fwait)
3163 if (!add_prefix (FWAIT_OPCODE))
3164 return;
3165
3166 /* Check for lock without a lockable instruction. Destination operand
3167 must be memory unless it is xchg (0x86). */
3168 if (i.prefix[LOCK_PREFIX]
3169 && (!i.tm.opcode_modifier.islockable
3170 || i.mem_operands == 0
3171 || (i.tm.base_opcode != 0x86
3172 && !operand_type_check (i.types[i.operands - 1], anymem))))
3173 {
3174 as_bad (_("expecting lockable instruction after `lock'"));
3175 return;
3176 }
3177
3178 /* Check if HLE prefix is OK. */
3179 if (i.have_hle && !check_hle ())
3180 return;
3181
3182 /* Check string instruction segment overrides. */
3183 if (i.tm.opcode_modifier.isstring && i.mem_operands != 0)
3184 {
3185 if (!check_string ())
3186 return;
3187 i.disp_operands = 0;
3188 }
3189
3190 if (!process_suffix ())
3191 return;
3192
3193 /* Update operand types. */
3194 for (j = 0; j < i.operands; j++)
3195 i.types[j] = operand_type_and (i.types[j], i.tm.operand_types[j]);
3196
3197 /* Make still unresolved immediate matches conform to size of immediate
3198 given in i.suffix. */
3199 if (!finalize_imm ())
3200 return;
3201
3202 if (i.types[0].bitfield.imm1)
3203 i.imm_operands = 0; /* kludge for shift insns. */
3204
3205 /* We only need to check those implicit registers for instructions
3206 with 3 operands or less. */
3207 if (i.operands <= 3)
3208 for (j = 0; j < i.operands; j++)
3209 if (i.types[j].bitfield.inoutportreg
3210 || i.types[j].bitfield.shiftcount
3211 || i.types[j].bitfield.acc
3212 || i.types[j].bitfield.floatacc)
3213 i.reg_operands--;
3214
3215 /* ImmExt should be processed after SSE2AVX. */
3216 if (!i.tm.opcode_modifier.sse2avx
3217 && i.tm.opcode_modifier.immext)
3218 process_immext ();
3219
3220 /* For insns with operands there are more diddles to do to the opcode. */
3221 if (i.operands)
3222 {
3223 if (!process_operands ())
3224 return;
3225 }
3226 else if (!quiet_warnings && i.tm.opcode_modifier.ugh)
3227 {
3228 /* UnixWare fsub no args is alias for fsubp, fadd -> faddp, etc. */
3229 as_warn (_("translating to `%sp'"), i.tm.name);
3230 }
3231
3232 if (i.tm.opcode_modifier.vex)
3233 build_vex_prefix (t);
3234
3235 /* Handle conversion of 'int $3' --> special int3 insn. XOP or FMA4
3236 instructions may define INT_OPCODE as well, so avoid this corner
3237 case for those instructions that use MODRM. */
3238 if (i.tm.base_opcode == INT_OPCODE
3239 && !i.tm.opcode_modifier.modrm
3240 && i.op[0].imms->X_add_number == 3)
3241 {
3242 i.tm.base_opcode = INT3_OPCODE;
3243 i.imm_operands = 0;
3244 }
3245
3246 if ((i.tm.opcode_modifier.jump
3247 || i.tm.opcode_modifier.jumpbyte
3248 || i.tm.opcode_modifier.jumpdword)
3249 && i.op[0].disps->X_op == O_constant)
3250 {
3251 /* Convert "jmp constant" (and "call constant") to a jump (call) to
3252 the absolute address given by the constant. Since ix86 jumps and
3253 calls are pc relative, we need to generate a reloc. */
3254 i.op[0].disps->X_add_symbol = &abs_symbol;
3255 i.op[0].disps->X_op = O_symbol;
3256 }
3257
3258 if (i.tm.opcode_modifier.rex64)
3259 i.rex |= REX_W;
3260
3261 /* For 8 bit registers we need an empty rex prefix. Also if the
3262 instruction already has a prefix, we need to convert old
3263 registers to new ones. */
3264
3265 if ((i.types[0].bitfield.reg8
3266 && (i.op[0].regs->reg_flags & RegRex64) != 0)
3267 || (i.types[1].bitfield.reg8
3268 && (i.op[1].regs->reg_flags & RegRex64) != 0)
3269 || ((i.types[0].bitfield.reg8
3270 || i.types[1].bitfield.reg8)
3271 && i.rex != 0))
3272 {
3273 int x;
3274
3275 i.rex |= REX_OPCODE;
3276 for (x = 0; x < 2; x++)
3277 {
3278 /* Look for 8 bit operand that uses old registers. */
3279 if (i.types[x].bitfield.reg8
3280 && (i.op[x].regs->reg_flags & RegRex64) == 0)
3281 {
3282 /* In case it is "hi" register, give up. */
3283 if (i.op[x].regs->reg_num > 3)
3284 as_bad (_("can't encode register '%s%s' in an "
3285 "instruction requiring REX prefix."),
3286 register_prefix, i.op[x].regs->reg_name);
3287
3288 /* Otherwise it is equivalent to the extended register.
3289 Since the encoding doesn't change this is merely
3290 cosmetic cleanup for debug output. */
3291
3292 i.op[x].regs = i.op[x].regs + 8;
3293 }
3294 }
3295 }
3296
3297 if (i.rex != 0)
3298 add_prefix (REX_OPCODE | i.rex);
3299
3300 /* We are ready to output the insn. */
3301 output_insn ();
3302 }
3303
3304 static char *
3305 parse_insn (char *line, char *mnemonic)
3306 {
3307 char *l = line;
3308 char *token_start = l;
3309 char *mnem_p;
3310 int supported;
3311 const insn_template *t;
3312 char *dot_p = NULL;
3313
3314 /* Non-zero if we found a prefix only acceptable with string insns. */
3315 const char *expecting_string_instruction = NULL;
3316
3317 while (1)
3318 {
3319 mnem_p = mnemonic;
3320 while ((*mnem_p = mnemonic_chars[(unsigned char) *l]) != 0)
3321 {
3322 if (*mnem_p == '.')
3323 dot_p = mnem_p;
3324 mnem_p++;
3325 if (mnem_p >= mnemonic + MAX_MNEM_SIZE)
3326 {
3327 as_bad (_("no such instruction: `%s'"), token_start);
3328 return NULL;
3329 }
3330 l++;
3331 }
3332 if (!is_space_char (*l)
3333 && *l != END_OF_INSN
3334 && (intel_syntax
3335 || (*l != PREFIX_SEPARATOR
3336 && *l != ',')))
3337 {
3338 as_bad (_("invalid character %s in mnemonic"),
3339 output_invalid (*l));
3340 return NULL;
3341 }
3342 if (token_start == l)
3343 {
3344 if (!intel_syntax && *l == PREFIX_SEPARATOR)
3345 as_bad (_("expecting prefix; got nothing"));
3346 else
3347 as_bad (_("expecting mnemonic; got nothing"));
3348 return NULL;
3349 }
3350
3351 /* Look up instruction (or prefix) via hash table. */
3352 current_templates = (const templates *) hash_find (op_hash, mnemonic);
3353
3354 if (*l != END_OF_INSN
3355 && (!is_space_char (*l) || l[1] != END_OF_INSN)
3356 && current_templates
3357 && current_templates->start->opcode_modifier.isprefix)
3358 {
3359 if (!cpu_flags_check_cpu64 (current_templates->start->cpu_flags))
3360 {
3361 as_bad ((flag_code != CODE_64BIT
3362 ? _("`%s' is only supported in 64-bit mode")
3363 : _("`%s' is not supported in 64-bit mode")),
3364 current_templates->start->name);
3365 return NULL;
3366 }
3367 /* If we are in 16-bit mode, do not allow addr16 or data16.
3368 Similarly, in 32-bit mode, do not allow addr32 or data32. */
3369 if ((current_templates->start->opcode_modifier.size16
3370 || current_templates->start->opcode_modifier.size32)
3371 && flag_code != CODE_64BIT
3372 && (current_templates->start->opcode_modifier.size32
3373 ^ (flag_code == CODE_16BIT)))
3374 {
3375 as_bad (_("redundant %s prefix"),
3376 current_templates->start->name);
3377 return NULL;
3378 }
3379 /* Add prefix, checking for repeated prefixes. */
3380 switch (add_prefix (current_templates->start->base_opcode))
3381 {
3382 case PREFIX_EXIST:
3383 return NULL;
3384 case PREFIX_REP:
3385 if (current_templates->start->cpu_flags.bitfield.cpuhle)
3386 i.have_hle = 1;
3387 else
3388 expecting_string_instruction = current_templates->start->name;
3389 break;
3390 default:
3391 break;
3392 }
3393 /* Skip past PREFIX_SEPARATOR and reset token_start. */
3394 token_start = ++l;
3395 }
3396 else
3397 break;
3398 }
3399
3400 if (!current_templates)
3401 {
3402 /* Check if we should swap operand or force 32bit displacement in
3403 encoding. */
3404 if (mnem_p - 2 == dot_p && dot_p[1] == 's')
3405 i.swap_operand = 1;
3406 else if (mnem_p - 3 == dot_p
3407 && dot_p[1] == 'd'
3408 && dot_p[2] == '8')
3409 i.disp_encoding = disp_encoding_8bit;
3410 else if (mnem_p - 4 == dot_p
3411 && dot_p[1] == 'd'
3412 && dot_p[2] == '3'
3413 && dot_p[3] == '2')
3414 i.disp_encoding = disp_encoding_32bit;
3415 else
3416 goto check_suffix;
3417 mnem_p = dot_p;
3418 *dot_p = '\0';
3419 current_templates = (const templates *) hash_find (op_hash, mnemonic);
3420 }
3421
3422 if (!current_templates)
3423 {
3424 check_suffix:
3425 /* See if we can get a match by trimming off a suffix. */
3426 switch (mnem_p[-1])
3427 {
3428 case WORD_MNEM_SUFFIX:
3429 if (intel_syntax && (intel_float_operand (mnemonic) & 2))
3430 i.suffix = SHORT_MNEM_SUFFIX;
3431 else
3432 case BYTE_MNEM_SUFFIX:
3433 case QWORD_MNEM_SUFFIX:
3434 i.suffix = mnem_p[-1];
3435 mnem_p[-1] = '\0';
3436 current_templates = (const templates *) hash_find (op_hash,
3437 mnemonic);
3438 break;
3439 case SHORT_MNEM_SUFFIX:
3440 case LONG_MNEM_SUFFIX:
3441 if (!intel_syntax)
3442 {
3443 i.suffix = mnem_p[-1];
3444 mnem_p[-1] = '\0';
3445 current_templates = (const templates *) hash_find (op_hash,
3446 mnemonic);
3447 }
3448 break;
3449
3450 /* Intel Syntax. */
3451 case 'd':
3452 if (intel_syntax)
3453 {
3454 if (intel_float_operand (mnemonic) == 1)
3455 i.suffix = SHORT_MNEM_SUFFIX;
3456 else
3457 i.suffix = LONG_MNEM_SUFFIX;
3458 mnem_p[-1] = '\0';
3459 current_templates = (const templates *) hash_find (op_hash,
3460 mnemonic);
3461 }
3462 break;
3463 }
3464 if (!current_templates)
3465 {
3466 as_bad (_("no such instruction: `%s'"), token_start);
3467 return NULL;
3468 }
3469 }
3470
3471 if (current_templates->start->opcode_modifier.jump
3472 || current_templates->start->opcode_modifier.jumpbyte)
3473 {
3474 /* Check for a branch hint. We allow ",pt" and ",pn" for
3475 predict taken and predict not taken respectively.
3476 I'm not sure that branch hints actually do anything on loop
3477 and jcxz insns (JumpByte) for current Pentium4 chips. They
3478 may work in the future and it doesn't hurt to accept them
3479 now. */
3480 if (l[0] == ',' && l[1] == 'p')
3481 {
3482 if (l[2] == 't')
3483 {
3484 if (!add_prefix (DS_PREFIX_OPCODE))
3485 return NULL;
3486 l += 3;
3487 }
3488 else if (l[2] == 'n')
3489 {
3490 if (!add_prefix (CS_PREFIX_OPCODE))
3491 return NULL;
3492 l += 3;
3493 }
3494 }
3495 }
3496 /* Any other comma loses. */
3497 if (*l == ',')
3498 {
3499 as_bad (_("invalid character %s in mnemonic"),
3500 output_invalid (*l));
3501 return NULL;
3502 }
3503
3504 /* Check if instruction is supported on specified architecture. */
3505 supported = 0;
3506 for (t = current_templates->start; t < current_templates->end; ++t)
3507 {
3508 supported |= cpu_flags_match (t);
3509 if (supported == CPU_FLAGS_PERFECT_MATCH)
3510 goto skip;
3511 }
3512
3513 if (!(supported & CPU_FLAGS_64BIT_MATCH))
3514 {
3515 as_bad (flag_code == CODE_64BIT
3516 ? _("`%s' is not supported in 64-bit mode")
3517 : _("`%s' is only supported in 64-bit mode"),
3518 current_templates->start->name);
3519 return NULL;
3520 }
3521 if (supported != CPU_FLAGS_PERFECT_MATCH)
3522 {
3523 as_bad (_("`%s' is not supported on `%s%s'"),
3524 current_templates->start->name,
3525 cpu_arch_name ? cpu_arch_name : default_arch,
3526 cpu_sub_arch_name ? cpu_sub_arch_name : "");
3527 return NULL;
3528 }
3529
3530 skip:
3531 if (!cpu_arch_flags.bitfield.cpui386
3532 && (flag_code != CODE_16BIT))
3533 {
3534 as_warn (_("use .code16 to ensure correct addressing mode"));
3535 }
3536
3537 /* Check for rep/repne without a string (or other allowed) instruction. */
3538 if (expecting_string_instruction)
3539 {
3540 static templates override;
3541
3542 for (t = current_templates->start; t < current_templates->end; ++t)
3543 if (t->opcode_modifier.repprefixok)
3544 break;
3545 if (t >= current_templates->end)
3546 {
3547 as_bad (_("expecting string instruction after `%s'"),
3548 expecting_string_instruction);
3549 return NULL;
3550 }
3551 for (override.start = t; t < current_templates->end; ++t)
3552 if (!t->opcode_modifier.repprefixok)
3553 break;
3554 override.end = t;
3555 current_templates = &override;
3556 }
3557
3558 return l;
3559 }
3560
3561 static char *
3562 parse_operands (char *l, const char *mnemonic)
3563 {
3564 char *token_start;
3565
3566 /* 1 if operand is pending after ','. */
3567 unsigned int expecting_operand = 0;
3568
3569 /* Non-zero if operand parens not balanced. */
3570 unsigned int paren_not_balanced;
3571
3572 while (*l != END_OF_INSN)
3573 {
3574 /* Skip optional white space before operand. */
3575 if (is_space_char (*l))
3576 ++l;
3577 if (!is_operand_char (*l) && *l != END_OF_INSN)
3578 {
3579 as_bad (_("invalid character %s before operand %d"),
3580 output_invalid (*l),
3581 i.operands + 1);
3582 return NULL;
3583 }
3584 token_start = l; /* after white space */
3585 paren_not_balanced = 0;
3586 while (paren_not_balanced || *l != ',')
3587 {
3588 if (*l == END_OF_INSN)
3589 {
3590 if (paren_not_balanced)
3591 {
3592 if (!intel_syntax)
3593 as_bad (_("unbalanced parenthesis in operand %d."),
3594 i.operands + 1);
3595 else
3596 as_bad (_("unbalanced brackets in operand %d."),
3597 i.operands + 1);
3598 return NULL;
3599 }
3600 else
3601 break; /* we are done */
3602 }
3603 else if (!is_operand_char (*l) && !is_space_char (*l))
3604 {
3605 as_bad (_("invalid character %s in operand %d"),
3606 output_invalid (*l),
3607 i.operands + 1);
3608 return NULL;
3609 }
3610 if (!intel_syntax)
3611 {
3612 if (*l == '(')
3613 ++paren_not_balanced;
3614 if (*l == ')')
3615 --paren_not_balanced;
3616 }
3617 else
3618 {
3619 if (*l == '[')
3620 ++paren_not_balanced;
3621 if (*l == ']')
3622 --paren_not_balanced;
3623 }
3624 l++;
3625 }
3626 if (l != token_start)
3627 { /* Yes, we've read in another operand. */
3628 unsigned int operand_ok;
3629 this_operand = i.operands++;
3630 i.types[this_operand].bitfield.unspecified = 1;
3631 if (i.operands > MAX_OPERANDS)
3632 {
3633 as_bad (_("spurious operands; (%d operands/instruction max)"),
3634 MAX_OPERANDS);
3635 return NULL;
3636 }
3637 /* Now parse operand adding info to 'i' as we go along. */
3638 END_STRING_AND_SAVE (l);
3639
3640 if (intel_syntax)
3641 operand_ok =
3642 i386_intel_operand (token_start,
3643 intel_float_operand (mnemonic));
3644 else
3645 operand_ok = i386_att_operand (token_start);
3646
3647 RESTORE_END_STRING (l);
3648 if (!operand_ok)
3649 return NULL;
3650 }
3651 else
3652 {
3653 if (expecting_operand)
3654 {
3655 expecting_operand_after_comma:
3656 as_bad (_("expecting operand after ','; got nothing"));
3657 return NULL;
3658 }
3659 if (*l == ',')
3660 {
3661 as_bad (_("expecting operand before ','; got nothing"));
3662 return NULL;
3663 }
3664 }
3665
3666 /* Now *l must be either ',' or END_OF_INSN. */
3667 if (*l == ',')
3668 {
3669 if (*++l == END_OF_INSN)
3670 {
3671 /* Just skip it, if it's \n complain. */
3672 goto expecting_operand_after_comma;
3673 }
3674 expecting_operand = 1;
3675 }
3676 }
3677 return l;
3678 }
3679
3680 static void
3681 swap_2_operands (int xchg1, int xchg2)
3682 {
3683 union i386_op temp_op;
3684 i386_operand_type temp_type;
3685 enum bfd_reloc_code_real temp_reloc;
3686
3687 temp_type = i.types[xchg2];
3688 i.types[xchg2] = i.types[xchg1];
3689 i.types[xchg1] = temp_type;
3690 temp_op = i.op[xchg2];
3691 i.op[xchg2] = i.op[xchg1];
3692 i.op[xchg1] = temp_op;
3693 temp_reloc = i.reloc[xchg2];
3694 i.reloc[xchg2] = i.reloc[xchg1];
3695 i.reloc[xchg1] = temp_reloc;
3696 }
3697
3698 static void
3699 swap_operands (void)
3700 {
3701 switch (i.operands)
3702 {
3703 case 5:
3704 case 4:
3705 swap_2_operands (1, i.operands - 2);
3706 case 3:
3707 case 2:
3708 swap_2_operands (0, i.operands - 1);
3709 break;
3710 default:
3711 abort ();
3712 }
3713
3714 if (i.mem_operands == 2)
3715 {
3716 const seg_entry *temp_seg;
3717 temp_seg = i.seg[0];
3718 i.seg[0] = i.seg[1];
3719 i.seg[1] = temp_seg;
3720 }
3721 }
3722
3723 /* Try to ensure constant immediates are represented in the smallest
3724 opcode possible. */
3725 static void
3726 optimize_imm (void)
3727 {
3728 char guess_suffix = 0;
3729 int op;
3730
3731 if (i.suffix)
3732 guess_suffix = i.suffix;
3733 else if (i.reg_operands)
3734 {
3735 /* Figure out a suffix from the last register operand specified.
3736 We can't do this properly yet, ie. excluding InOutPortReg,
3737 but the following works for instructions with immediates.
3738 In any case, we can't set i.suffix yet. */
3739 for (op = i.operands; --op >= 0;)
3740 if (i.types[op].bitfield.reg8)
3741 {
3742 guess_suffix = BYTE_MNEM_SUFFIX;
3743 break;
3744 }
3745 else if (i.types[op].bitfield.reg16)
3746 {
3747 guess_suffix = WORD_MNEM_SUFFIX;
3748 break;
3749 }
3750 else if (i.types[op].bitfield.reg32)
3751 {
3752 guess_suffix = LONG_MNEM_SUFFIX;
3753 break;
3754 }
3755 else if (i.types[op].bitfield.reg64)
3756 {
3757 guess_suffix = QWORD_MNEM_SUFFIX;
3758 break;
3759 }
3760 }
3761 else if ((flag_code == CODE_16BIT) ^ (i.prefix[DATA_PREFIX] != 0))
3762 guess_suffix = WORD_MNEM_SUFFIX;
3763
3764 for (op = i.operands; --op >= 0;)
3765 if (operand_type_check (i.types[op], imm))
3766 {
3767 switch (i.op[op].imms->X_op)
3768 {
3769 case O_constant:
3770 /* If a suffix is given, this operand may be shortened. */
3771 switch (guess_suffix)
3772 {
3773 case LONG_MNEM_SUFFIX:
3774 i.types[op].bitfield.imm32 = 1;
3775 i.types[op].bitfield.imm64 = 1;
3776 break;
3777 case WORD_MNEM_SUFFIX:
3778 i.types[op].bitfield.imm16 = 1;
3779 i.types[op].bitfield.imm32 = 1;
3780 i.types[op].bitfield.imm32s = 1;
3781 i.types[op].bitfield.imm64 = 1;
3782 break;
3783 case BYTE_MNEM_SUFFIX:
3784 i.types[op].bitfield.imm8 = 1;
3785 i.types[op].bitfield.imm8s = 1;
3786 i.types[op].bitfield.imm16 = 1;
3787 i.types[op].bitfield.imm32 = 1;
3788 i.types[op].bitfield.imm32s = 1;
3789 i.types[op].bitfield.imm64 = 1;
3790 break;
3791 }
3792
3793 /* If this operand is at most 16 bits, convert it
3794 to a signed 16 bit number before trying to see
3795 whether it will fit in an even smaller size.
3796 This allows a 16-bit operand such as $0xffe0 to
3797 be recognised as within Imm8S range. */
3798 if ((i.types[op].bitfield.imm16)
3799 && (i.op[op].imms->X_add_number & ~(offsetT) 0xffff) == 0)
3800 {
3801 i.op[op].imms->X_add_number =
3802 (((i.op[op].imms->X_add_number & 0xffff) ^ 0x8000) - 0x8000);
3803 }
3804 if ((i.types[op].bitfield.imm32)
3805 && ((i.op[op].imms->X_add_number & ~(((offsetT) 2 << 31) - 1))
3806 == 0))
3807 {
3808 i.op[op].imms->X_add_number = ((i.op[op].imms->X_add_number
3809 ^ ((offsetT) 1 << 31))
3810 - ((offsetT) 1 << 31));
3811 }
3812 i.types[op]
3813 = operand_type_or (i.types[op],
3814 smallest_imm_type (i.op[op].imms->X_add_number));
3815
3816 /* We must avoid matching of Imm32 templates when 64bit
3817 only immediate is available. */
3818 if (guess_suffix == QWORD_MNEM_SUFFIX)
3819 i.types[op].bitfield.imm32 = 0;
3820 break;
3821
3822 case O_absent:
3823 case O_register:
3824 abort ();
3825
3826 /* Symbols and expressions. */
3827 default:
3828 /* Convert symbolic operand to proper sizes for matching, but don't
3829 prevent matching a set of insns that only supports sizes other
3830 than those matching the insn suffix. */
3831 {
3832 i386_operand_type mask, allowed;
3833 const insn_template *t;
3834
3835 operand_type_set (&mask, 0);
3836 operand_type_set (&allowed, 0);
3837
3838 for (t = current_templates->start;
3839 t < current_templates->end;
3840 ++t)
3841 allowed = operand_type_or (allowed,
3842 t->operand_types[op]);
3843 switch (guess_suffix)
3844 {
3845 case QWORD_MNEM_SUFFIX:
3846 mask.bitfield.imm64 = 1;
3847 mask.bitfield.imm32s = 1;
3848 break;
3849 case LONG_MNEM_SUFFIX:
3850 mask.bitfield.imm32 = 1;
3851 break;
3852 case WORD_MNEM_SUFFIX:
3853 mask.bitfield.imm16 = 1;
3854 break;
3855 case BYTE_MNEM_SUFFIX:
3856 mask.bitfield.imm8 = 1;
3857 break;
3858 default:
3859 break;
3860 }
3861 allowed = operand_type_and (mask, allowed);
3862 if (!operand_type_all_zero (&allowed))
3863 i.types[op] = operand_type_and (i.types[op], mask);
3864 }
3865 break;
3866 }
3867 }
3868 }
3869
3870 /* Try to use the smallest displacement type too. */
3871 static void
3872 optimize_disp (void)
3873 {
3874 int op;
3875
3876 for (op = i.operands; --op >= 0;)
3877 if (operand_type_check (i.types[op], disp))
3878 {
3879 if (i.op[op].disps->X_op == O_constant)
3880 {
3881 offsetT op_disp = i.op[op].disps->X_add_number;
3882
3883 if (i.types[op].bitfield.disp16
3884 && (op_disp & ~(offsetT) 0xffff) == 0)
3885 {
3886 /* If this operand is at most 16 bits, convert
3887 to a signed 16 bit number and don't use 64bit
3888 displacement. */
3889 op_disp = (((op_disp & 0xffff) ^ 0x8000) - 0x8000);
3890 i.types[op].bitfield.disp64 = 0;
3891 }
3892 if (i.types[op].bitfield.disp32
3893 && (op_disp & ~(((offsetT) 2 << 31) - 1)) == 0)
3894 {
3895 /* If this operand is at most 32 bits, convert
3896 to a signed 32 bit number and don't use 64bit
3897 displacement. */
3898 op_disp &= (((offsetT) 2 << 31) - 1);
3899 op_disp = (op_disp ^ ((offsetT) 1 << 31)) - ((addressT) 1 << 31);
3900 i.types[op].bitfield.disp64 = 0;
3901 }
3902 if (!op_disp && i.types[op].bitfield.baseindex)
3903 {
3904 i.types[op].bitfield.disp8 = 0;
3905 i.types[op].bitfield.disp16 = 0;
3906 i.types[op].bitfield.disp32 = 0;
3907 i.types[op].bitfield.disp32s = 0;
3908 i.types[op].bitfield.disp64 = 0;
3909 i.op[op].disps = 0;
3910 i.disp_operands--;
3911 }
3912 else if (flag_code == CODE_64BIT)
3913 {
3914 if (fits_in_signed_long (op_disp))
3915 {
3916 i.types[op].bitfield.disp64 = 0;
3917 i.types[op].bitfield.disp32s = 1;
3918 }
3919 if (i.prefix[ADDR_PREFIX]
3920 && fits_in_unsigned_long (op_disp))
3921 i.types[op].bitfield.disp32 = 1;
3922 }
3923 if ((i.types[op].bitfield.disp32
3924 || i.types[op].bitfield.disp32s
3925 || i.types[op].bitfield.disp16)
3926 && fits_in_signed_byte (op_disp))
3927 i.types[op].bitfield.disp8 = 1;
3928 }
3929 else if (i.reloc[op] == BFD_RELOC_386_TLS_DESC_CALL
3930 || i.reloc[op] == BFD_RELOC_X86_64_TLSDESC_CALL)
3931 {
3932 fix_new_exp (frag_now, frag_more (0) - frag_now->fr_literal, 0,
3933 i.op[op].disps, 0, i.reloc[op]);
3934 i.types[op].bitfield.disp8 = 0;
3935 i.types[op].bitfield.disp16 = 0;
3936 i.types[op].bitfield.disp32 = 0;
3937 i.types[op].bitfield.disp32s = 0;
3938 i.types[op].bitfield.disp64 = 0;
3939 }
3940 else
3941 /* We only support 64bit displacement on constants. */
3942 i.types[op].bitfield.disp64 = 0;
3943 }
3944 }
3945
3946 /* Check if operands are valid for the instruction. */
3947
3948 static int
3949 check_VecOperands (const insn_template *t)
3950 {
3951 /* Without VSIB byte, we can't have a vector register for index. */
3952 if (!t->opcode_modifier.vecsib
3953 && i.index_reg
3954 && (i.index_reg->reg_type.bitfield.regxmm
3955 || i.index_reg->reg_type.bitfield.regymm))
3956 {
3957 i.error = unsupported_vector_index_register;
3958 return 1;
3959 }
3960
3961 /* For VSIB byte, we need a vector register for index and no PC
3962 relative addressing is allowed. */
3963 if (t->opcode_modifier.vecsib
3964 && (!i.index_reg
3965 || !((t->opcode_modifier.vecsib == VecSIB128
3966 && i.index_reg->reg_type.bitfield.regxmm)
3967 || (t->opcode_modifier.vecsib == VecSIB256
3968 && i.index_reg->reg_type.bitfield.regymm))
3969 || (i.base_reg && i.base_reg->reg_num == RegRip)))
3970 {
3971 i.error = invalid_vsib_address;
3972 return 1;
3973 }
3974
3975 return 0;
3976 }
3977
3978 /* Check if operands are valid for the instruction. Update VEX
3979 operand types. */
3980
3981 static int
3982 VEX_check_operands (const insn_template *t)
3983 {
3984 if (!t->opcode_modifier.vex)
3985 return 0;
3986
3987 /* Only check VEX_Imm4, which must be the first operand. */
3988 if (t->operand_types[0].bitfield.vec_imm4)
3989 {
3990 if (i.op[0].imms->X_op != O_constant
3991 || !fits_in_imm4 (i.op[0].imms->X_add_number))
3992 {
3993 i.error = bad_imm4;
3994 return 1;
3995 }
3996
3997 /* Turn off Imm8 so that update_imm won't complain. */
3998 i.types[0] = vec_imm4;
3999 }
4000
4001 return 0;
4002 }
4003
4004 static const insn_template *
4005 match_template (void)
4006 {
4007 /* Points to template once we've found it. */
4008 const insn_template *t;
4009 i386_operand_type overlap0, overlap1, overlap2, overlap3;
4010 i386_operand_type overlap4;
4011 unsigned int found_reverse_match;
4012 i386_opcode_modifier suffix_check;
4013 i386_operand_type operand_types [MAX_OPERANDS];
4014 int addr_prefix_disp;
4015 unsigned int j;
4016 unsigned int found_cpu_match;
4017 unsigned int check_register;
4018
4019 #if MAX_OPERANDS != 5
4020 # error "MAX_OPERANDS must be 5."
4021 #endif
4022
4023 found_reverse_match = 0;
4024 addr_prefix_disp = -1;
4025
4026 memset (&suffix_check, 0, sizeof (suffix_check));
4027 if (i.suffix == BYTE_MNEM_SUFFIX)
4028 suffix_check.no_bsuf = 1;
4029 else if (i.suffix == WORD_MNEM_SUFFIX)
4030 suffix_check.no_wsuf = 1;
4031 else if (i.suffix == SHORT_MNEM_SUFFIX)
4032 suffix_check.no_ssuf = 1;
4033 else if (i.suffix == LONG_MNEM_SUFFIX)
4034 suffix_check.no_lsuf = 1;
4035 else if (i.suffix == QWORD_MNEM_SUFFIX)
4036 suffix_check.no_qsuf = 1;
4037 else if (i.suffix == LONG_DOUBLE_MNEM_SUFFIX)
4038 suffix_check.no_ldsuf = 1;
4039
4040 /* Must have right number of operands. */
4041 i.error = number_of_operands_mismatch;
4042
4043 for (t = current_templates->start; t < current_templates->end; t++)
4044 {
4045 addr_prefix_disp = -1;
4046
4047 if (i.operands != t->operands)
4048 continue;
4049
4050 /* Check processor support. */
4051 i.error = unsupported;
4052 found_cpu_match = (cpu_flags_match (t)
4053 == CPU_FLAGS_PERFECT_MATCH);
4054 if (!found_cpu_match)
4055 continue;
4056
4057 /* Check old gcc support. */
4058 i.error = old_gcc_only;
4059 if (!old_gcc && t->opcode_modifier.oldgcc)
4060 continue;
4061
4062 /* Check AT&T mnemonic. */
4063 i.error = unsupported_with_intel_mnemonic;
4064 if (intel_mnemonic && t->opcode_modifier.attmnemonic)
4065 continue;
4066
4067 /* Check AT&T/Intel syntax. */
4068 i.error = unsupported_syntax;
4069 if ((intel_syntax && t->opcode_modifier.attsyntax)
4070 || (!intel_syntax && t->opcode_modifier.intelsyntax))
4071 continue;
4072
4073 /* Check the suffix, except for some instructions in intel mode. */
4074 i.error = invalid_instruction_suffix;
4075 if ((!intel_syntax || !t->opcode_modifier.ignoresize)
4076 && ((t->opcode_modifier.no_bsuf && suffix_check.no_bsuf)
4077 || (t->opcode_modifier.no_wsuf && suffix_check.no_wsuf)
4078 || (t->opcode_modifier.no_lsuf && suffix_check.no_lsuf)
4079 || (t->opcode_modifier.no_ssuf && suffix_check.no_ssuf)
4080 || (t->opcode_modifier.no_qsuf && suffix_check.no_qsuf)
4081 || (t->opcode_modifier.no_ldsuf && suffix_check.no_ldsuf)))
4082 continue;
4083
4084 if (!operand_size_match (t))
4085 continue;
4086
4087 for (j = 0; j < MAX_OPERANDS; j++)
4088 operand_types[j] = t->operand_types[j];
4089
4090 /* In general, don't allow 64-bit operands in 32-bit mode. */
4091 if (i.suffix == QWORD_MNEM_SUFFIX
4092 && flag_code != CODE_64BIT
4093 && (intel_syntax
4094 ? (!t->opcode_modifier.ignoresize
4095 && !intel_float_operand (t->name))
4096 : intel_float_operand (t->name) != 2)
4097 && ((!operand_types[0].bitfield.regmmx
4098 && !operand_types[0].bitfield.regxmm
4099 && !operand_types[0].bitfield.regymm)
4100 || (!operand_types[t->operands > 1].bitfield.regmmx
4101 && !!operand_types[t->operands > 1].bitfield.regxmm
4102 && !!operand_types[t->operands > 1].bitfield.regymm))
4103 && (t->base_opcode != 0x0fc7
4104 || t->extension_opcode != 1 /* cmpxchg8b */))
4105 continue;
4106
4107 /* In general, don't allow 32-bit operands on pre-386. */
4108 else if (i.suffix == LONG_MNEM_SUFFIX
4109 && !cpu_arch_flags.bitfield.cpui386
4110 && (intel_syntax
4111 ? (!t->opcode_modifier.ignoresize
4112 && !intel_float_operand (t->name))
4113 : intel_float_operand (t->name) != 2)
4114 && ((!operand_types[0].bitfield.regmmx
4115 && !operand_types[0].bitfield.regxmm)
4116 || (!operand_types[t->operands > 1].bitfield.regmmx
4117 && !!operand_types[t->operands > 1].bitfield.regxmm)))
4118 continue;
4119
4120 /* Do not verify operands when there are none. */
4121 else
4122 {
4123 if (!t->operands)
4124 /* We've found a match; break out of loop. */
4125 break;
4126 }
4127
4128 /* Address size prefix will turn Disp64/Disp32/Disp16 operand
4129 into Disp32/Disp16/Disp32 operand. */
4130 if (i.prefix[ADDR_PREFIX] != 0)
4131 {
4132 /* There should be only one Disp operand. */
4133 switch (flag_code)
4134 {
4135 case CODE_16BIT:
4136 for (j = 0; j < MAX_OPERANDS; j++)
4137 {
4138 if (operand_types[j].bitfield.disp16)
4139 {
4140 addr_prefix_disp = j;
4141 operand_types[j].bitfield.disp32 = 1;
4142 operand_types[j].bitfield.disp16 = 0;
4143 break;
4144 }
4145 }
4146 break;
4147 case CODE_32BIT:
4148 for (j = 0; j < MAX_OPERANDS; j++)
4149 {
4150 if (operand_types[j].bitfield.disp32)
4151 {
4152 addr_prefix_disp = j;
4153 operand_types[j].bitfield.disp32 = 0;
4154 operand_types[j].bitfield.disp16 = 1;
4155 break;
4156 }
4157 }
4158 break;
4159 case CODE_64BIT:
4160 for (j = 0; j < MAX_OPERANDS; j++)
4161 {
4162 if (operand_types[j].bitfield.disp64)
4163 {
4164 addr_prefix_disp = j;
4165 operand_types[j].bitfield.disp64 = 0;
4166 operand_types[j].bitfield.disp32 = 1;
4167 break;
4168 }
4169 }
4170 break;
4171 }
4172 }
4173
4174 /* We check register size if needed. */
4175 check_register = t->opcode_modifier.checkregsize;
4176 overlap0 = operand_type_and (i.types[0], operand_types[0]);
4177 switch (t->operands)
4178 {
4179 case 1:
4180 if (!operand_type_match (overlap0, i.types[0]))
4181 continue;
4182 break;
4183 case 2:
4184 /* xchg %eax, %eax is a special case. It is an aliase for nop
4185 only in 32bit mode and we can use opcode 0x90. In 64bit
4186 mode, we can't use 0x90 for xchg %eax, %eax since it should
4187 zero-extend %eax to %rax. */
4188 if (flag_code == CODE_64BIT
4189 && t->base_opcode == 0x90
4190 && operand_type_equal (&i.types [0], &acc32)
4191 && operand_type_equal (&i.types [1], &acc32))
4192 continue;
4193 if (i.swap_operand)
4194 {
4195 /* If we swap operand in encoding, we either match
4196 the next one or reverse direction of operands. */
4197 if (t->opcode_modifier.s)
4198 continue;
4199 else if (t->opcode_modifier.d)
4200 goto check_reverse;
4201 }
4202
4203 case 3:
4204 /* If we swap operand in encoding, we match the next one. */
4205 if (i.swap_operand && t->opcode_modifier.s)
4206 continue;
4207 case 4:
4208 case 5:
4209 overlap1 = operand_type_and (i.types[1], operand_types[1]);
4210 if (!operand_type_match (overlap0, i.types[0])
4211 || !operand_type_match (overlap1, i.types[1])
4212 || (check_register
4213 && !operand_type_register_match (overlap0, i.types[0],
4214 operand_types[0],
4215 overlap1, i.types[1],
4216 operand_types[1])))
4217 {
4218 /* Check if other direction is valid ... */
4219 if (!t->opcode_modifier.d && !t->opcode_modifier.floatd)
4220 continue;
4221
4222 check_reverse:
4223 /* Try reversing direction of operands. */
4224 overlap0 = operand_type_and (i.types[0], operand_types[1]);
4225 overlap1 = operand_type_and (i.types[1], operand_types[0]);
4226 if (!operand_type_match (overlap0, i.types[0])
4227 || !operand_type_match (overlap1, i.types[1])
4228 || (check_register
4229 && !operand_type_register_match (overlap0,
4230 i.types[0],
4231 operand_types[1],
4232 overlap1,
4233 i.types[1],
4234 operand_types[0])))
4235 {
4236 /* Does not match either direction. */
4237 continue;
4238 }
4239 /* found_reverse_match holds which of D or FloatDR
4240 we've found. */
4241 if (t->opcode_modifier.d)
4242 found_reverse_match = Opcode_D;
4243 else if (t->opcode_modifier.floatd)
4244 found_reverse_match = Opcode_FloatD;
4245 else
4246 found_reverse_match = 0;
4247 if (t->opcode_modifier.floatr)
4248 found_reverse_match |= Opcode_FloatR;
4249 }
4250 else
4251 {
4252 /* Found a forward 2 operand match here. */
4253 switch (t->operands)
4254 {
4255 case 5:
4256 overlap4 = operand_type_and (i.types[4],
4257 operand_types[4]);
4258 case 4:
4259 overlap3 = operand_type_and (i.types[3],
4260 operand_types[3]);
4261 case 3:
4262 overlap2 = operand_type_and (i.types[2],
4263 operand_types[2]);
4264 break;
4265 }
4266
4267 switch (t->operands)
4268 {
4269 case 5:
4270 if (!operand_type_match (overlap4, i.types[4])
4271 || !operand_type_register_match (overlap3,
4272 i.types[3],
4273 operand_types[3],
4274 overlap4,
4275 i.types[4],
4276 operand_types[4]))
4277 continue;
4278 case 4:
4279 if (!operand_type_match (overlap3, i.types[3])
4280 || (check_register
4281 && !operand_type_register_match (overlap2,
4282 i.types[2],
4283 operand_types[2],
4284 overlap3,
4285 i.types[3],
4286 operand_types[3])))
4287 continue;
4288 case 3:
4289 /* Here we make use of the fact that there are no
4290 reverse match 3 operand instructions, and all 3
4291 operand instructions only need to be checked for
4292 register consistency between operands 2 and 3. */
4293 if (!operand_type_match (overlap2, i.types[2])
4294 || (check_register
4295 && !operand_type_register_match (overlap1,
4296 i.types[1],
4297 operand_types[1],
4298 overlap2,
4299 i.types[2],
4300 operand_types[2])))
4301 continue;
4302 break;
4303 }
4304 }
4305 /* Found either forward/reverse 2, 3 or 4 operand match here:
4306 slip through to break. */
4307 }
4308 if (!found_cpu_match)
4309 {
4310 found_reverse_match = 0;
4311 continue;
4312 }
4313
4314 /* Check if vector operands are valid. */
4315 if (check_VecOperands (t))
4316 continue;
4317
4318 /* Check if VEX operands are valid. */
4319 if (VEX_check_operands (t))
4320 continue;
4321
4322 /* We've found a match; break out of loop. */
4323 break;
4324 }
4325
4326 if (t == current_templates->end)
4327 {
4328 /* We found no match. */
4329 const char *err_msg;
4330 switch (i.error)
4331 {
4332 default:
4333 abort ();
4334 case operand_size_mismatch:
4335 err_msg = _("operand size mismatch");
4336 break;
4337 case operand_type_mismatch:
4338 err_msg = _("operand type mismatch");
4339 break;
4340 case register_type_mismatch:
4341 err_msg = _("register type mismatch");
4342 break;
4343 case number_of_operands_mismatch:
4344 err_msg = _("number of operands mismatch");
4345 break;
4346 case invalid_instruction_suffix:
4347 err_msg = _("invalid instruction suffix");
4348 break;
4349 case bad_imm4:
4350 err_msg = _("Imm4 isn't the first operand");
4351 break;
4352 case old_gcc_only:
4353 err_msg = _("only supported with old gcc");
4354 break;
4355 case unsupported_with_intel_mnemonic:
4356 err_msg = _("unsupported with Intel mnemonic");
4357 break;
4358 case unsupported_syntax:
4359 err_msg = _("unsupported syntax");
4360 break;
4361 case unsupported:
4362 as_bad (_("unsupported instruction `%s'"),
4363 current_templates->start->name);
4364 return NULL;
4365 case invalid_vsib_address:
4366 err_msg = _("invalid VSIB address");
4367 break;
4368 case unsupported_vector_index_register:
4369 err_msg = _("unsupported vector index register");
4370 break;
4371 }
4372 as_bad (_("%s for `%s'"), err_msg,
4373 current_templates->start->name);
4374 return NULL;
4375 }
4376
4377 if (!quiet_warnings)
4378 {
4379 if (!intel_syntax
4380 && (i.types[0].bitfield.jumpabsolute
4381 != operand_types[0].bitfield.jumpabsolute))
4382 {
4383 as_warn (_("indirect %s without `*'"), t->name);
4384 }
4385
4386 if (t->opcode_modifier.isprefix
4387 && t->opcode_modifier.ignoresize)
4388 {
4389 /* Warn them that a data or address size prefix doesn't
4390 affect assembly of the next line of code. */
4391 as_warn (_("stand-alone `%s' prefix"), t->name);
4392 }
4393 }
4394
4395 /* Copy the template we found. */
4396 i.tm = *t;
4397
4398 if (addr_prefix_disp != -1)
4399 i.tm.operand_types[addr_prefix_disp]
4400 = operand_types[addr_prefix_disp];
4401
4402 if (found_reverse_match)
4403 {
4404 /* If we found a reverse match we must alter the opcode
4405 direction bit. found_reverse_match holds bits to change
4406 (different for int & float insns). */
4407
4408 i.tm.base_opcode ^= found_reverse_match;
4409
4410 i.tm.operand_types[0] = operand_types[1];
4411 i.tm.operand_types[1] = operand_types[0];
4412 }
4413
4414 return t;
4415 }
4416
4417 static int
4418 check_string (void)
4419 {
4420 int mem_op = operand_type_check (i.types[0], anymem) ? 0 : 1;
4421 if (i.tm.operand_types[mem_op].bitfield.esseg)
4422 {
4423 if (i.seg[0] != NULL && i.seg[0] != &es)
4424 {
4425 as_bad (_("`%s' operand %d must use `%ses' segment"),
4426 i.tm.name,
4427 mem_op + 1,
4428 register_prefix);
4429 return 0;
4430 }
4431 /* There's only ever one segment override allowed per instruction.
4432 This instruction possibly has a legal segment override on the
4433 second operand, so copy the segment to where non-string
4434 instructions store it, allowing common code. */
4435 i.seg[0] = i.seg[1];
4436 }
4437 else if (i.tm.operand_types[mem_op + 1].bitfield.esseg)
4438 {
4439 if (i.seg[1] != NULL && i.seg[1] != &es)
4440 {
4441 as_bad (_("`%s' operand %d must use `%ses' segment"),
4442 i.tm.name,
4443 mem_op + 2,
4444 register_prefix);
4445 return 0;
4446 }
4447 }
4448 return 1;
4449 }
4450
4451 static int
4452 process_suffix (void)
4453 {
4454 /* If matched instruction specifies an explicit instruction mnemonic
4455 suffix, use it. */
4456 if (i.tm.opcode_modifier.size16)
4457 i.suffix = WORD_MNEM_SUFFIX;
4458 else if (i.tm.opcode_modifier.size32)
4459 i.suffix = LONG_MNEM_SUFFIX;
4460 else if (i.tm.opcode_modifier.size64)
4461 i.suffix = QWORD_MNEM_SUFFIX;
4462 else if (i.reg_operands)
4463 {
4464 /* If there's no instruction mnemonic suffix we try to invent one
4465 based on register operands. */
4466 if (!i.suffix)
4467 {
4468 /* We take i.suffix from the last register operand specified,
4469 Destination register type is more significant than source
4470 register type. crc32 in SSE4.2 prefers source register
4471 type. */
4472 if (i.tm.base_opcode == 0xf20f38f1)
4473 {
4474 if (i.types[0].bitfield.reg16)
4475 i.suffix = WORD_MNEM_SUFFIX;
4476 else if (i.types[0].bitfield.reg32)
4477 i.suffix = LONG_MNEM_SUFFIX;
4478 else if (i.types[0].bitfield.reg64)
4479 i.suffix = QWORD_MNEM_SUFFIX;
4480 }
4481 else if (i.tm.base_opcode == 0xf20f38f0)
4482 {
4483 if (i.types[0].bitfield.reg8)
4484 i.suffix = BYTE_MNEM_SUFFIX;
4485 }
4486
4487 if (!i.suffix)
4488 {
4489 int op;
4490
4491 if (i.tm.base_opcode == 0xf20f38f1
4492 || i.tm.base_opcode == 0xf20f38f0)
4493 {
4494 /* We have to know the operand size for crc32. */
4495 as_bad (_("ambiguous memory operand size for `%s`"),
4496 i.tm.name);
4497 return 0;
4498 }
4499
4500 for (op = i.operands; --op >= 0;)
4501 if (!i.tm.operand_types[op].bitfield.inoutportreg)
4502 {
4503 if (i.types[op].bitfield.reg8)
4504 {
4505 i.suffix = BYTE_MNEM_SUFFIX;
4506 break;
4507 }
4508 else if (i.types[op].bitfield.reg16)
4509 {
4510 i.suffix = WORD_MNEM_SUFFIX;
4511 break;
4512 }
4513 else if (i.types[op].bitfield.reg32)
4514 {
4515 i.suffix = LONG_MNEM_SUFFIX;
4516 break;
4517 }
4518 else if (i.types[op].bitfield.reg64)
4519 {
4520 i.suffix = QWORD_MNEM_SUFFIX;
4521 break;
4522 }
4523 }
4524 }
4525 }
4526 else if (i.suffix == BYTE_MNEM_SUFFIX)
4527 {
4528 if (intel_syntax
4529 && i.tm.opcode_modifier.ignoresize
4530 && i.tm.opcode_modifier.no_bsuf)
4531 i.suffix = 0;
4532 else if (!check_byte_reg ())
4533 return 0;
4534 }
4535 else if (i.suffix == LONG_MNEM_SUFFIX)
4536 {
4537 if (intel_syntax
4538 && i.tm.opcode_modifier.ignoresize
4539 && i.tm.opcode_modifier.no_lsuf)
4540 i.suffix = 0;
4541 else if (!check_long_reg ())
4542 return 0;
4543 }
4544 else if (i.suffix == QWORD_MNEM_SUFFIX)
4545 {
4546 if (intel_syntax
4547 && i.tm.opcode_modifier.ignoresize
4548 && i.tm.opcode_modifier.no_qsuf)
4549 i.suffix = 0;
4550 else if (!check_qword_reg ())
4551 return 0;
4552 }
4553 else if (i.suffix == WORD_MNEM_SUFFIX)
4554 {
4555 if (intel_syntax
4556 && i.tm.opcode_modifier.ignoresize
4557 && i.tm.opcode_modifier.no_wsuf)
4558 i.suffix = 0;
4559 else if (!check_word_reg ())
4560 return 0;
4561 }
4562 else if (i.suffix == XMMWORD_MNEM_SUFFIX
4563 || i.suffix == YMMWORD_MNEM_SUFFIX)
4564 {
4565 /* Skip if the instruction has x/y suffix. match_template
4566 should check if it is a valid suffix. */
4567 }
4568 else if (intel_syntax && i.tm.opcode_modifier.ignoresize)
4569 /* Do nothing if the instruction is going to ignore the prefix. */
4570 ;
4571 else
4572 abort ();
4573 }
4574 else if (i.tm.opcode_modifier.defaultsize
4575 && !i.suffix
4576 /* exclude fldenv/frstor/fsave/fstenv */
4577 && i.tm.opcode_modifier.no_ssuf)
4578 {
4579 i.suffix = stackop_size;
4580 }
4581 else if (intel_syntax
4582 && !i.suffix
4583 && (i.tm.operand_types[0].bitfield.jumpabsolute
4584 || i.tm.opcode_modifier.jumpbyte
4585 || i.tm.opcode_modifier.jumpintersegment
4586 || (i.tm.base_opcode == 0x0f01 /* [ls][gi]dt */
4587 && i.tm.extension_opcode <= 3)))
4588 {
4589 switch (flag_code)
4590 {
4591 case CODE_64BIT:
4592 if (!i.tm.opcode_modifier.no_qsuf)
4593 {
4594 i.suffix = QWORD_MNEM_SUFFIX;
4595 break;
4596 }
4597 case CODE_32BIT:
4598 if (!i.tm.opcode_modifier.no_lsuf)
4599 i.suffix = LONG_MNEM_SUFFIX;
4600 break;
4601 case CODE_16BIT:
4602 if (!i.tm.opcode_modifier.no_wsuf)
4603 i.suffix = WORD_MNEM_SUFFIX;
4604 break;
4605 }
4606 }
4607
4608 if (!i.suffix)
4609 {
4610 if (!intel_syntax)
4611 {
4612 if (i.tm.opcode_modifier.w)
4613 {
4614 as_bad (_("no instruction mnemonic suffix given and "
4615 "no register operands; can't size instruction"));
4616 return 0;
4617 }
4618 }
4619 else
4620 {
4621 unsigned int suffixes;
4622
4623 suffixes = !i.tm.opcode_modifier.no_bsuf;
4624 if (!i.tm.opcode_modifier.no_wsuf)
4625 suffixes |= 1 << 1;
4626 if (!i.tm.opcode_modifier.no_lsuf)
4627 suffixes |= 1 << 2;
4628 if (!i.tm.opcode_modifier.no_ldsuf)
4629 suffixes |= 1 << 3;
4630 if (!i.tm.opcode_modifier.no_ssuf)
4631 suffixes |= 1 << 4;
4632 if (!i.tm.opcode_modifier.no_qsuf)
4633 suffixes |= 1 << 5;
4634
4635 /* There are more than suffix matches. */
4636 if (i.tm.opcode_modifier.w
4637 || ((suffixes & (suffixes - 1))
4638 && !i.tm.opcode_modifier.defaultsize
4639 && !i.tm.opcode_modifier.ignoresize))
4640 {
4641 as_bad (_("ambiguous operand size for `%s'"), i.tm.name);
4642 return 0;
4643 }
4644 }
4645 }
4646
4647 /* Change the opcode based on the operand size given by i.suffix;
4648 We don't need to change things for byte insns. */
4649
4650 if (i.suffix
4651 && i.suffix != BYTE_MNEM_SUFFIX
4652 && i.suffix != XMMWORD_MNEM_SUFFIX
4653 && i.suffix != YMMWORD_MNEM_SUFFIX)
4654 {
4655 /* It's not a byte, select word/dword operation. */
4656 if (i.tm.opcode_modifier.w)
4657 {
4658 if (i.tm.opcode_modifier.shortform)
4659 i.tm.base_opcode |= 8;
4660 else
4661 i.tm.base_opcode |= 1;
4662 }
4663
4664 /* Now select between word & dword operations via the operand
4665 size prefix, except for instructions that will ignore this
4666 prefix anyway. */
4667 if (i.tm.opcode_modifier.addrprefixop0)
4668 {
4669 /* The address size override prefix changes the size of the
4670 first operand. */
4671 if ((flag_code == CODE_32BIT
4672 && i.op->regs[0].reg_type.bitfield.reg16)
4673 || (flag_code != CODE_32BIT
4674 && i.op->regs[0].reg_type.bitfield.reg32))
4675 if (!add_prefix (ADDR_PREFIX_OPCODE))
4676 return 0;
4677 }
4678 else if (i.suffix != QWORD_MNEM_SUFFIX
4679 && i.suffix != LONG_DOUBLE_MNEM_SUFFIX
4680 && !i.tm.opcode_modifier.ignoresize
4681 && !i.tm.opcode_modifier.floatmf
4682 && ((i.suffix == LONG_MNEM_SUFFIX) == (flag_code == CODE_16BIT)
4683 || (flag_code == CODE_64BIT
4684 && i.tm.opcode_modifier.jumpbyte)))
4685 {
4686 unsigned int prefix = DATA_PREFIX_OPCODE;
4687
4688 if (i.tm.opcode_modifier.jumpbyte) /* jcxz, loop */
4689 prefix = ADDR_PREFIX_OPCODE;
4690
4691 if (!add_prefix (prefix))
4692 return 0;
4693 }
4694
4695 /* Set mode64 for an operand. */
4696 if (i.suffix == QWORD_MNEM_SUFFIX
4697 && flag_code == CODE_64BIT
4698 && !i.tm.opcode_modifier.norex64)
4699 {
4700 /* Special case for xchg %rax,%rax. It is NOP and doesn't
4701 need rex64. cmpxchg8b is also a special case. */
4702 if (! (i.operands == 2
4703 && i.tm.base_opcode == 0x90
4704 && i.tm.extension_opcode == None
4705 && operand_type_equal (&i.types [0], &acc64)
4706 && operand_type_equal (&i.types [1], &acc64))
4707 && ! (i.operands == 1
4708 && i.tm.base_opcode == 0xfc7
4709 && i.tm.extension_opcode == 1
4710 && !operand_type_check (i.types [0], reg)
4711 && operand_type_check (i.types [0], anymem)))
4712 i.rex |= REX_W;
4713 }
4714
4715 /* Size floating point instruction. */
4716 if (i.suffix == LONG_MNEM_SUFFIX)
4717 if (i.tm.opcode_modifier.floatmf)
4718 i.tm.base_opcode ^= 4;
4719 }
4720
4721 return 1;
4722 }
4723
4724 static int
4725 check_byte_reg (void)
4726 {
4727 int op;
4728
4729 for (op = i.operands; --op >= 0;)
4730 {
4731 /* If this is an eight bit register, it's OK. If it's the 16 or
4732 32 bit version of an eight bit register, we will just use the
4733 low portion, and that's OK too. */
4734 if (i.types[op].bitfield.reg8)
4735 continue;
4736
4737 /* crc32 doesn't generate this warning. */
4738 if (i.tm.base_opcode == 0xf20f38f0)
4739 continue;
4740
4741 if ((i.types[op].bitfield.reg16
4742 || i.types[op].bitfield.reg32
4743 || i.types[op].bitfield.reg64)
4744 && i.op[op].regs->reg_num < 4)
4745 {
4746 /* Prohibit these changes in the 64bit mode, since the
4747 lowering is more complicated. */
4748 if (flag_code == CODE_64BIT
4749 && !i.tm.operand_types[op].bitfield.inoutportreg)
4750 {
4751 as_bad (_("incorrect register `%s%s' used with `%c' suffix"),
4752 register_prefix, i.op[op].regs->reg_name,
4753 i.suffix);
4754 return 0;
4755 }
4756 #if REGISTER_WARNINGS
4757 if (!quiet_warnings
4758 && !i.tm.operand_types[op].bitfield.inoutportreg)
4759 as_warn (_("using `%s%s' instead of `%s%s' due to `%c' suffix"),
4760 register_prefix,
4761 (i.op[op].regs + (i.types[op].bitfield.reg16
4762 ? REGNAM_AL - REGNAM_AX
4763 : REGNAM_AL - REGNAM_EAX))->reg_name,
4764 register_prefix,
4765 i.op[op].regs->reg_name,
4766 i.suffix);
4767 #endif
4768 continue;
4769 }
4770 /* Any other register is bad. */
4771 if (i.types[op].bitfield.reg16
4772 || i.types[op].bitfield.reg32
4773 || i.types[op].bitfield.reg64
4774 || i.types[op].bitfield.regmmx
4775 || i.types[op].bitfield.regxmm
4776 || i.types[op].bitfield.regymm
4777 || i.types[op].bitfield.sreg2
4778 || i.types[op].bitfield.sreg3
4779 || i.types[op].bitfield.control
4780 || i.types[op].bitfield.debug
4781 || i.types[op].bitfield.test
4782 || i.types[op].bitfield.floatreg
4783 || i.types[op].bitfield.floatacc)
4784 {
4785 as_bad (_("`%s%s' not allowed with `%s%c'"),
4786 register_prefix,
4787 i.op[op].regs->reg_name,
4788 i.tm.name,
4789 i.suffix);
4790 return 0;
4791 }
4792 }
4793 return 1;
4794 }
4795
4796 static int
4797 check_long_reg (void)
4798 {
4799 int op;
4800
4801 for (op = i.operands; --op >= 0;)
4802 /* Reject eight bit registers, except where the template requires
4803 them. (eg. movzb) */
4804 if (i.types[op].bitfield.reg8
4805 && (i.tm.operand_types[op].bitfield.reg16
4806 || i.tm.operand_types[op].bitfield.reg32
4807 || i.tm.operand_types[op].bitfield.acc))
4808 {
4809 as_bad (_("`%s%s' not allowed with `%s%c'"),
4810 register_prefix,
4811 i.op[op].regs->reg_name,
4812 i.tm.name,
4813 i.suffix);
4814 return 0;
4815 }
4816 /* Warn if the e prefix on a general reg is missing. */
4817 else if ((!quiet_warnings || flag_code == CODE_64BIT)
4818 && i.types[op].bitfield.reg16
4819 && (i.tm.operand_types[op].bitfield.reg32
4820 || i.tm.operand_types[op].bitfield.acc))
4821 {
4822 /* Prohibit these changes in the 64bit mode, since the
4823 lowering is more complicated. */
4824 if (flag_code == CODE_64BIT)
4825 {
4826 as_bad (_("incorrect register `%s%s' used with `%c' suffix"),
4827 register_prefix, i.op[op].regs->reg_name,
4828 i.suffix);
4829 return 0;
4830 }
4831 #if REGISTER_WARNINGS
4832 else
4833 as_warn (_("using `%s%s' instead of `%s%s' due to `%c' suffix"),
4834 register_prefix,
4835 (i.op[op].regs + REGNAM_EAX - REGNAM_AX)->reg_name,
4836 register_prefix,
4837 i.op[op].regs->reg_name,
4838 i.suffix);
4839 #endif
4840 }
4841 /* Warn if the r prefix on a general reg is missing. */
4842 else if (i.types[op].bitfield.reg64
4843 && (i.tm.operand_types[op].bitfield.reg32
4844 || i.tm.operand_types[op].bitfield.acc))
4845 {
4846 if (intel_syntax
4847 && i.tm.opcode_modifier.toqword
4848 && !i.types[0].bitfield.regxmm)
4849 {
4850 /* Convert to QWORD. We want REX byte. */
4851 i.suffix = QWORD_MNEM_SUFFIX;
4852 }
4853 else
4854 {
4855 as_bad (_("incorrect register `%s%s' used with `%c' suffix"),
4856 register_prefix, i.op[op].regs->reg_name,
4857 i.suffix);
4858 return 0;
4859 }
4860 }
4861 return 1;
4862 }
4863
4864 static int
4865 check_qword_reg (void)
4866 {
4867 int op;
4868
4869 for (op = i.operands; --op >= 0; )
4870 /* Reject eight bit registers, except where the template requires
4871 them. (eg. movzb) */
4872 if (i.types[op].bitfield.reg8
4873 && (i.tm.operand_types[op].bitfield.reg16
4874 || i.tm.operand_types[op].bitfield.reg32
4875 || i.tm.operand_types[op].bitfield.acc))
4876 {
4877 as_bad (_("`%s%s' not allowed with `%s%c'"),
4878 register_prefix,
4879 i.op[op].regs->reg_name,
4880 i.tm.name,
4881 i.suffix);
4882 return 0;
4883 }
4884 /* Warn if the e prefix on a general reg is missing. */
4885 else if ((i.types[op].bitfield.reg16
4886 || i.types[op].bitfield.reg32)
4887 && (i.tm.operand_types[op].bitfield.reg32
4888 || i.tm.operand_types[op].bitfield.acc))
4889 {
4890 /* Prohibit these changes in the 64bit mode, since the
4891 lowering is more complicated. */
4892 if (intel_syntax
4893 && i.tm.opcode_modifier.todword
4894 && !i.types[0].bitfield.regxmm)
4895 {
4896 /* Convert to DWORD. We don't want REX byte. */
4897 i.suffix = LONG_MNEM_SUFFIX;
4898 }
4899 else
4900 {
4901 as_bad (_("incorrect register `%s%s' used with `%c' suffix"),
4902 register_prefix, i.op[op].regs->reg_name,
4903 i.suffix);
4904 return 0;
4905 }
4906 }
4907 return 1;
4908 }
4909
4910 static int
4911 check_word_reg (void)
4912 {
4913 int op;
4914 for (op = i.operands; --op >= 0;)
4915 /* Reject eight bit registers, except where the template requires
4916 them. (eg. movzb) */
4917 if (i.types[op].bitfield.reg8
4918 && (i.tm.operand_types[op].bitfield.reg16
4919 || i.tm.operand_types[op].bitfield.reg32
4920 || i.tm.operand_types[op].bitfield.acc))
4921 {
4922 as_bad (_("`%s%s' not allowed with `%s%c'"),
4923 register_prefix,
4924 i.op[op].regs->reg_name,
4925 i.tm.name,
4926 i.suffix);
4927 return 0;
4928 }
4929 /* Warn if the e prefix on a general reg is present. */
4930 else if ((!quiet_warnings || flag_code == CODE_64BIT)
4931 && i.types[op].bitfield.reg32
4932 && (i.tm.operand_types[op].bitfield.reg16
4933 || i.tm.operand_types[op].bitfield.acc))
4934 {
4935 /* Prohibit these changes in the 64bit mode, since the
4936 lowering is more complicated. */
4937 if (flag_code == CODE_64BIT)
4938 {
4939 as_bad (_("incorrect register `%s%s' used with `%c' suffix"),
4940 register_prefix, i.op[op].regs->reg_name,
4941 i.suffix);
4942 return 0;
4943 }
4944 else
4945 #if REGISTER_WARNINGS
4946 as_warn (_("using `%s%s' instead of `%s%s' due to `%c' suffix"),
4947 register_prefix,
4948 (i.op[op].regs + REGNAM_AX - REGNAM_EAX)->reg_name,
4949 register_prefix,
4950 i.op[op].regs->reg_name,
4951 i.suffix);
4952 #endif
4953 }
4954 return 1;
4955 }
4956
4957 static int
4958 update_imm (unsigned int j)
4959 {
4960 i386_operand_type overlap = i.types[j];
4961 if ((overlap.bitfield.imm8
4962 || overlap.bitfield.imm8s
4963 || overlap.bitfield.imm16
4964 || overlap.bitfield.imm32
4965 || overlap.bitfield.imm32s
4966 || overlap.bitfield.imm64)
4967 && !operand_type_equal (&overlap, &imm8)
4968 && !operand_type_equal (&overlap, &imm8s)
4969 && !operand_type_equal (&overlap, &imm16)
4970 && !operand_type_equal (&overlap, &imm32)
4971 && !operand_type_equal (&overlap, &imm32s)
4972 && !operand_type_equal (&overlap, &imm64))
4973 {
4974 if (i.suffix)
4975 {
4976 i386_operand_type temp;
4977
4978 operand_type_set (&temp, 0);
4979 if (i.suffix == BYTE_MNEM_SUFFIX)
4980 {
4981 temp.bitfield.imm8 = overlap.bitfield.imm8;
4982 temp.bitfield.imm8s = overlap.bitfield.imm8s;
4983 }
4984 else if (i.suffix == WORD_MNEM_SUFFIX)
4985 temp.bitfield.imm16 = overlap.bitfield.imm16;
4986 else if (i.suffix == QWORD_MNEM_SUFFIX)
4987 {
4988 temp.bitfield.imm64 = overlap.bitfield.imm64;
4989 temp.bitfield.imm32s = overlap.bitfield.imm32s;
4990 }
4991 else
4992 temp.bitfield.imm32 = overlap.bitfield.imm32;
4993 overlap = temp;
4994 }
4995 else if (operand_type_equal (&overlap, &imm16_32_32s)
4996 || operand_type_equal (&overlap, &imm16_32)
4997 || operand_type_equal (&overlap, &imm16_32s))
4998 {
4999 if ((flag_code == CODE_16BIT) ^ (i.prefix[DATA_PREFIX] != 0))
5000 overlap = imm16;
5001 else
5002 overlap = imm32s;
5003 }
5004 if (!operand_type_equal (&overlap, &imm8)
5005 && !operand_type_equal (&overlap, &imm8s)
5006 && !operand_type_equal (&overlap, &imm16)
5007 && !operand_type_equal (&overlap, &imm32)
5008 && !operand_type_equal (&overlap, &imm32s)
5009 && !operand_type_equal (&overlap, &imm64))
5010 {
5011 as_bad (_("no instruction mnemonic suffix given; "
5012 "can't determine immediate size"));
5013 return 0;
5014 }
5015 }
5016 i.types[j] = overlap;
5017
5018 return 1;
5019 }
5020
5021 static int
5022 finalize_imm (void)
5023 {
5024 unsigned int j, n;
5025
5026 /* Update the first 2 immediate operands. */
5027 n = i.operands > 2 ? 2 : i.operands;
5028 if (n)
5029 {
5030 for (j = 0; j < n; j++)
5031 if (update_imm (j) == 0)
5032 return 0;
5033
5034 /* The 3rd operand can't be immediate operand. */
5035 gas_assert (operand_type_check (i.types[2], imm) == 0);
5036 }
5037
5038 return 1;
5039 }
5040
5041 static int
5042 bad_implicit_operand (int xmm)
5043 {
5044 const char *ireg = xmm ? "xmm0" : "ymm0";
5045
5046 if (intel_syntax)
5047 as_bad (_("the last operand of `%s' must be `%s%s'"),
5048 i.tm.name, register_prefix, ireg);
5049 else
5050 as_bad (_("the first operand of `%s' must be `%s%s'"),
5051 i.tm.name, register_prefix, ireg);
5052 return 0;
5053 }
5054
5055 static int
5056 process_operands (void)
5057 {
5058 /* Default segment register this instruction will use for memory
5059 accesses. 0 means unknown. This is only for optimizing out
5060 unnecessary segment overrides. */
5061 const seg_entry *default_seg = 0;
5062
5063 if (i.tm.opcode_modifier.sse2avx && i.tm.opcode_modifier.vexvvvv)
5064 {
5065 unsigned int dupl = i.operands;
5066 unsigned int dest = dupl - 1;
5067 unsigned int j;
5068
5069 /* The destination must be an xmm register. */
5070 gas_assert (i.reg_operands
5071 && MAX_OPERANDS > dupl
5072 && operand_type_equal (&i.types[dest], &regxmm));
5073
5074 if (i.tm.opcode_modifier.firstxmm0)
5075 {
5076 /* The first operand is implicit and must be xmm0. */
5077 gas_assert (operand_type_equal (&i.types[0], &regxmm));
5078 if (i.op[0].regs->reg_num != 0)
5079 return bad_implicit_operand (1);
5080
5081 if (i.tm.opcode_modifier.vexsources == VEX3SOURCES)
5082 {
5083 /* Keep xmm0 for instructions with VEX prefix and 3
5084 sources. */
5085 goto duplicate;
5086 }
5087 else
5088 {
5089 /* We remove the first xmm0 and keep the number of
5090 operands unchanged, which in fact duplicates the
5091 destination. */
5092 for (j = 1; j < i.operands; j++)
5093 {
5094 i.op[j - 1] = i.op[j];
5095 i.types[j - 1] = i.types[j];
5096 i.tm.operand_types[j - 1] = i.tm.operand_types[j];
5097 }
5098 }
5099 }
5100 else if (i.tm.opcode_modifier.implicit1stxmm0)
5101 {
5102 gas_assert ((MAX_OPERANDS - 1) > dupl
5103 && (i.tm.opcode_modifier.vexsources
5104 == VEX3SOURCES));
5105
5106 /* Add the implicit xmm0 for instructions with VEX prefix
5107 and 3 sources. */
5108 for (j = i.operands; j > 0; j--)
5109 {
5110 i.op[j] = i.op[j - 1];
5111 i.types[j] = i.types[j - 1];
5112 i.tm.operand_types[j] = i.tm.operand_types[j - 1];
5113 }
5114 i.op[0].regs
5115 = (const reg_entry *) hash_find (reg_hash, "xmm0");
5116 i.types[0] = regxmm;
5117 i.tm.operand_types[0] = regxmm;
5118
5119 i.operands += 2;
5120 i.reg_operands += 2;
5121 i.tm.operands += 2;
5122
5123 dupl++;
5124 dest++;
5125 i.op[dupl] = i.op[dest];
5126 i.types[dupl] = i.types[dest];
5127 i.tm.operand_types[dupl] = i.tm.operand_types[dest];
5128 }
5129 else
5130 {
5131 duplicate:
5132 i.operands++;
5133 i.reg_operands++;
5134 i.tm.operands++;
5135
5136 i.op[dupl] = i.op[dest];
5137 i.types[dupl] = i.types[dest];
5138 i.tm.operand_types[dupl] = i.tm.operand_types[dest];
5139 }
5140
5141 if (i.tm.opcode_modifier.immext)
5142 process_immext ();
5143 }
5144 else if (i.tm.opcode_modifier.firstxmm0)
5145 {
5146 unsigned int j;
5147
5148 /* The first operand is implicit and must be xmm0/ymm0. */
5149 gas_assert (i.reg_operands
5150 && (operand_type_equal (&i.types[0], &regxmm)
5151 || operand_type_equal (&i.types[0], &regymm)));
5152 if (i.op[0].regs->reg_num != 0)
5153 return bad_implicit_operand (i.types[0].bitfield.regxmm);
5154
5155 for (j = 1; j < i.operands; j++)
5156 {
5157 i.op[j - 1] = i.op[j];
5158 i.types[j - 1] = i.types[j];
5159
5160 /* We need to adjust fields in i.tm since they are used by
5161 build_modrm_byte. */
5162 i.tm.operand_types [j - 1] = i.tm.operand_types [j];
5163 }
5164
5165 i.operands--;
5166 i.reg_operands--;
5167 i.tm.operands--;
5168 }
5169 else if (i.tm.opcode_modifier.regkludge)
5170 {
5171 /* The imul $imm, %reg instruction is converted into
5172 imul $imm, %reg, %reg, and the clr %reg instruction
5173 is converted into xor %reg, %reg. */
5174
5175 unsigned int first_reg_op;
5176
5177 if (operand_type_check (i.types[0], reg))
5178 first_reg_op = 0;
5179 else
5180 first_reg_op = 1;
5181 /* Pretend we saw the extra register operand. */
5182 gas_assert (i.reg_operands == 1
5183 && i.op[first_reg_op + 1].regs == 0);
5184 i.op[first_reg_op + 1].regs = i.op[first_reg_op].regs;
5185 i.types[first_reg_op + 1] = i.types[first_reg_op];
5186 i.operands++;
5187 i.reg_operands++;
5188 }
5189
5190 if (i.tm.opcode_modifier.shortform)
5191 {
5192 if (i.types[0].bitfield.sreg2
5193 || i.types[0].bitfield.sreg3)
5194 {
5195 if (i.tm.base_opcode == POP_SEG_SHORT
5196 && i.op[0].regs->reg_num == 1)
5197 {
5198 as_bad (_("you can't `pop %scs'"), register_prefix);
5199 return 0;
5200 }
5201 i.tm.base_opcode |= (i.op[0].regs->reg_num << 3);
5202 if ((i.op[0].regs->reg_flags & RegRex) != 0)
5203 i.rex |= REX_B;
5204 }
5205 else
5206 {
5207 /* The register or float register operand is in operand
5208 0 or 1. */
5209 unsigned int op;
5210
5211 if (i.types[0].bitfield.floatreg
5212 || operand_type_check (i.types[0], reg))
5213 op = 0;
5214 else
5215 op = 1;
5216 /* Register goes in low 3 bits of opcode. */
5217 i.tm.base_opcode |= i.op[op].regs->reg_num;
5218 if ((i.op[op].regs->reg_flags & RegRex) != 0)
5219 i.rex |= REX_B;
5220 if (!quiet_warnings && i.tm.opcode_modifier.ugh)
5221 {
5222 /* Warn about some common errors, but press on regardless.
5223 The first case can be generated by gcc (<= 2.8.1). */
5224 if (i.operands == 2)
5225 {
5226 /* Reversed arguments on faddp, fsubp, etc. */
5227 as_warn (_("translating to `%s %s%s,%s%s'"), i.tm.name,
5228 register_prefix, i.op[!intel_syntax].regs->reg_name,
5229 register_prefix, i.op[intel_syntax].regs->reg_name);
5230 }
5231 else
5232 {
5233 /* Extraneous `l' suffix on fp insn. */
5234 as_warn (_("translating to `%s %s%s'"), i.tm.name,
5235 register_prefix, i.op[0].regs->reg_name);
5236 }
5237 }
5238 }
5239 }
5240 else if (i.tm.opcode_modifier.modrm)
5241 {
5242 /* The opcode is completed (modulo i.tm.extension_opcode which
5243 must be put into the modrm byte). Now, we make the modrm and
5244 index base bytes based on all the info we've collected. */
5245
5246 default_seg = build_modrm_byte ();
5247 }
5248 else if ((i.tm.base_opcode & ~0x3) == MOV_AX_DISP32)
5249 {
5250 default_seg = &ds;
5251 }
5252 else if (i.tm.opcode_modifier.isstring)
5253 {
5254 /* For the string instructions that allow a segment override
5255 on one of their operands, the default segment is ds. */
5256 default_seg = &ds;
5257 }
5258
5259 if (i.tm.base_opcode == 0x8d /* lea */
5260 && i.seg[0]
5261 && !quiet_warnings)
5262 as_warn (_("segment override on `%s' is ineffectual"), i.tm.name);
5263
5264 /* If a segment was explicitly specified, and the specified segment
5265 is not the default, use an opcode prefix to select it. If we
5266 never figured out what the default segment is, then default_seg
5267 will be zero at this point, and the specified segment prefix will
5268 always be used. */
5269 if ((i.seg[0]) && (i.seg[0] != default_seg))
5270 {
5271 if (!add_prefix (i.seg[0]->seg_prefix))
5272 return 0;
5273 }
5274 return 1;
5275 }
5276
5277 static const seg_entry *
5278 build_modrm_byte (void)
5279 {
5280 const seg_entry *default_seg = 0;
5281 unsigned int source, dest;
5282 int vex_3_sources;
5283
5284 /* The first operand of instructions with VEX prefix and 3 sources
5285 must be VEX_Imm4. */
5286 vex_3_sources = i.tm.opcode_modifier.vexsources == VEX3SOURCES;
5287 if (vex_3_sources)
5288 {
5289 unsigned int nds, reg_slot;
5290 expressionS *exp;
5291
5292 if (i.tm.opcode_modifier.veximmext
5293 && i.tm.opcode_modifier.immext)
5294 {
5295 dest = i.operands - 2;
5296 gas_assert (dest == 3);
5297 }
5298 else
5299 dest = i.operands - 1;
5300 nds = dest - 1;
5301
5302 /* There are 2 kinds of instructions:
5303 1. 5 operands: 4 register operands or 3 register operands
5304 plus 1 memory operand plus one Vec_Imm4 operand, VexXDS, and
5305 VexW0 or VexW1. The destination must be either XMM or YMM
5306 register.
5307 2. 4 operands: 4 register operands or 3 register operands
5308 plus 1 memory operand, VexXDS, and VexImmExt */
5309 gas_assert ((i.reg_operands == 4
5310 || (i.reg_operands == 3 && i.mem_operands == 1))
5311 && i.tm.opcode_modifier.vexvvvv == VEXXDS
5312 && (i.tm.opcode_modifier.veximmext
5313 || (i.imm_operands == 1
5314 && i.types[0].bitfield.vec_imm4
5315 && (i.tm.opcode_modifier.vexw == VEXW0
5316 || i.tm.opcode_modifier.vexw == VEXW1)
5317 && (operand_type_equal (&i.tm.operand_types[dest], &regxmm)
5318 || operand_type_equal (&i.tm.operand_types[dest], &regymm)))));
5319
5320 if (i.imm_operands == 0)
5321 {
5322 /* When there is no immediate operand, generate an 8bit
5323 immediate operand to encode the first operand. */
5324 exp = &im_expressions[i.imm_operands++];
5325 i.op[i.operands].imms = exp;
5326 i.types[i.operands] = imm8;
5327 i.operands++;
5328 /* If VexW1 is set, the first operand is the source and
5329 the second operand is encoded in the immediate operand. */
5330 if (i.tm.opcode_modifier.vexw == VEXW1)
5331 {
5332 source = 0;
5333 reg_slot = 1;
5334 }
5335 else
5336 {
5337 source = 1;
5338 reg_slot = 0;
5339 }
5340
5341 /* FMA swaps REG and NDS. */
5342 if (i.tm.cpu_flags.bitfield.cpufma)
5343 {
5344 unsigned int tmp;
5345 tmp = reg_slot;
5346 reg_slot = nds;
5347 nds = tmp;
5348 }
5349
5350 gas_assert (operand_type_equal (&i.tm.operand_types[reg_slot],
5351 &regxmm)
5352 || operand_type_equal (&i.tm.operand_types[reg_slot],
5353 &regymm));
5354 exp->X_op = O_constant;
5355 exp->X_add_number
5356 = ((i.op[reg_slot].regs->reg_num
5357 + ((i.op[reg_slot].regs->reg_flags & RegRex) ? 8 : 0))
5358 << 4);
5359 }
5360 else
5361 {
5362 unsigned int imm_slot;
5363
5364 if (i.tm.opcode_modifier.vexw == VEXW0)
5365 {
5366 /* If VexW0 is set, the third operand is the source and
5367 the second operand is encoded in the immediate
5368 operand. */
5369 source = 2;
5370 reg_slot = 1;
5371 }
5372 else
5373 {
5374 /* VexW1 is set, the second operand is the source and
5375 the third operand is encoded in the immediate
5376 operand. */
5377 source = 1;
5378 reg_slot = 2;
5379 }
5380
5381 if (i.tm.opcode_modifier.immext)
5382 {
5383 /* When ImmExt is set, the immdiate byte is the last
5384 operand. */
5385 imm_slot = i.operands - 1;
5386 source--;
5387 reg_slot--;
5388 }
5389 else
5390 {
5391 imm_slot = 0;
5392
5393 /* Turn on Imm8 so that output_imm will generate it. */
5394 i.types[imm_slot].bitfield.imm8 = 1;
5395 }
5396
5397 gas_assert (operand_type_equal (&i.tm.operand_types[reg_slot],
5398 &regxmm)
5399 || operand_type_equal (&i.tm.operand_types[reg_slot],
5400 &regymm));
5401 i.op[imm_slot].imms->X_add_number
5402 |= ((i.op[reg_slot].regs->reg_num
5403 + ((i.op[reg_slot].regs->reg_flags & RegRex) ? 8 : 0))
5404 << 4);
5405 }
5406
5407 gas_assert (operand_type_equal (&i.tm.operand_types[nds], &regxmm)
5408 || operand_type_equal (&i.tm.operand_types[nds],
5409 &regymm));
5410 i.vex.register_specifier = i.op[nds].regs;
5411 }
5412 else
5413 source = dest = 0;
5414
5415 /* i.reg_operands MUST be the number of real register operands;
5416 implicit registers do not count. If there are 3 register
5417 operands, it must be a instruction with VexNDS. For a
5418 instruction with VexNDD, the destination register is encoded
5419 in VEX prefix. If there are 4 register operands, it must be
5420 a instruction with VEX prefix and 3 sources. */
5421 if (i.mem_operands == 0
5422 && ((i.reg_operands == 2
5423 && i.tm.opcode_modifier.vexvvvv <= VEXXDS)
5424 || (i.reg_operands == 3
5425 && i.tm.opcode_modifier.vexvvvv == VEXXDS)
5426 || (i.reg_operands == 4 && vex_3_sources)))
5427 {
5428 switch (i.operands)
5429 {
5430 case 2:
5431 source = 0;
5432 break;
5433 case 3:
5434 /* When there are 3 operands, one of them may be immediate,
5435 which may be the first or the last operand. Otherwise,
5436 the first operand must be shift count register (cl) or it
5437 is an instruction with VexNDS. */
5438 gas_assert (i.imm_operands == 1
5439 || (i.imm_operands == 0
5440 && (i.tm.opcode_modifier.vexvvvv == VEXXDS
5441 || i.types[0].bitfield.shiftcount)));
5442 if (operand_type_check (i.types[0], imm)
5443 || i.types[0].bitfield.shiftcount)
5444 source = 1;
5445 else
5446 source = 0;
5447 break;
5448 case 4:
5449 /* When there are 4 operands, the first two must be 8bit
5450 immediate operands. The source operand will be the 3rd
5451 one.
5452
5453 For instructions with VexNDS, if the first operand
5454 an imm8, the source operand is the 2nd one. If the last
5455 operand is imm8, the source operand is the first one. */
5456 gas_assert ((i.imm_operands == 2
5457 && i.types[0].bitfield.imm8
5458 && i.types[1].bitfield.imm8)
5459 || (i.tm.opcode_modifier.vexvvvv == VEXXDS
5460 && i.imm_operands == 1
5461 && (i.types[0].bitfield.imm8
5462 || i.types[i.operands - 1].bitfield.imm8)));
5463 if (i.imm_operands == 2)
5464 source = 2;
5465 else
5466 {
5467 if (i.types[0].bitfield.imm8)
5468 source = 1;
5469 else
5470 source = 0;
5471 }
5472 break;
5473 case 5:
5474 break;
5475 default:
5476 abort ();
5477 }
5478
5479 if (!vex_3_sources)
5480 {
5481 dest = source + 1;
5482
5483 if (i.tm.opcode_modifier.vexvvvv == VEXXDS)
5484 {
5485 /* For instructions with VexNDS, the register-only
5486 source operand must be 32/64bit integer, XMM or
5487 YMM register. It is encoded in VEX prefix. We
5488 need to clear RegMem bit before calling
5489 operand_type_equal. */
5490
5491 i386_operand_type op;
5492 unsigned int vvvv;
5493
5494 /* Check register-only source operand when two source
5495 operands are swapped. */
5496 if (!i.tm.operand_types[source].bitfield.baseindex
5497 && i.tm.operand_types[dest].bitfield.baseindex)
5498 {
5499 vvvv = source;
5500 source = dest;
5501 }
5502 else
5503 vvvv = dest;
5504
5505 op = i.tm.operand_types[vvvv];
5506 op.bitfield.regmem = 0;
5507 if ((dest + 1) >= i.operands
5508 || (op.bitfield.reg32 != 1
5509 && !op.bitfield.reg64 != 1
5510 && !operand_type_equal (&op, &regxmm)
5511 && !operand_type_equal (&op, &regymm)))
5512 abort ();
5513 i.vex.register_specifier = i.op[vvvv].regs;
5514 dest++;
5515 }
5516 }
5517
5518 i.rm.mode = 3;
5519 /* One of the register operands will be encoded in the i.tm.reg
5520 field, the other in the combined i.tm.mode and i.tm.regmem
5521 fields. If no form of this instruction supports a memory
5522 destination operand, then we assume the source operand may
5523 sometimes be a memory operand and so we need to store the
5524 destination in the i.rm.reg field. */
5525 if (!i.tm.operand_types[dest].bitfield.regmem
5526 && operand_type_check (i.tm.operand_types[dest], anymem) == 0)
5527 {
5528 i.rm.reg = i.op[dest].regs->reg_num;
5529 i.rm.regmem = i.op[source].regs->reg_num;
5530 if ((i.op[dest].regs->reg_flags & RegRex) != 0)
5531 i.rex |= REX_R;
5532 if ((i.op[source].regs->reg_flags & RegRex) != 0)
5533 i.rex |= REX_B;
5534 }
5535 else
5536 {
5537 i.rm.reg = i.op[source].regs->reg_num;
5538 i.rm.regmem = i.op[dest].regs->reg_num;
5539 if ((i.op[dest].regs->reg_flags & RegRex) != 0)
5540 i.rex |= REX_B;
5541 if ((i.op[source].regs->reg_flags & RegRex) != 0)
5542 i.rex |= REX_R;
5543 }
5544 if (flag_code != CODE_64BIT && (i.rex & (REX_R | REX_B)))
5545 {
5546 if (!i.types[0].bitfield.control
5547 && !i.types[1].bitfield.control)
5548 abort ();
5549 i.rex &= ~(REX_R | REX_B);
5550 add_prefix (LOCK_PREFIX_OPCODE);
5551 }
5552 }
5553 else
5554 { /* If it's not 2 reg operands... */
5555 unsigned int mem;
5556
5557 if (i.mem_operands)
5558 {
5559 unsigned int fake_zero_displacement = 0;
5560 unsigned int op;
5561
5562 for (op = 0; op < i.operands; op++)
5563 if (operand_type_check (i.types[op], anymem))
5564 break;
5565 gas_assert (op < i.operands);
5566
5567 if (i.tm.opcode_modifier.vecsib)
5568 {
5569 if (i.index_reg->reg_num == RegEiz
5570 || i.index_reg->reg_num == RegRiz)
5571 abort ();
5572
5573 i.rm.regmem = ESCAPE_TO_TWO_BYTE_ADDRESSING;
5574 if (!i.base_reg)
5575 {
5576 i.sib.base = NO_BASE_REGISTER;
5577 i.sib.scale = i.log2_scale_factor;
5578 i.types[op].bitfield.disp8 = 0;
5579 i.types[op].bitfield.disp16 = 0;
5580 i.types[op].bitfield.disp64 = 0;
5581 if (flag_code != CODE_64BIT)
5582 {
5583 /* Must be 32 bit */
5584 i.types[op].bitfield.disp32 = 1;
5585 i.types[op].bitfield.disp32s = 0;
5586 }
5587 else
5588 {
5589 i.types[op].bitfield.disp32 = 0;
5590 i.types[op].bitfield.disp32s = 1;
5591 }
5592 }
5593 i.sib.index = i.index_reg->reg_num;
5594 if ((i.index_reg->reg_flags & RegRex) != 0)
5595 i.rex |= REX_X;
5596 }
5597
5598 default_seg = &ds;
5599
5600 if (i.base_reg == 0)
5601 {
5602 i.rm.mode = 0;
5603 if (!i.disp_operands)
5604 {
5605 fake_zero_displacement = 1;
5606 /* Instructions with VSIB byte need 32bit displacement
5607 if there is no base register. */
5608 if (i.tm.opcode_modifier.vecsib)
5609 i.types[op].bitfield.disp32 = 1;
5610 }
5611 if (i.index_reg == 0)
5612 {
5613 gas_assert (!i.tm.opcode_modifier.vecsib);
5614 /* Operand is just <disp> */
5615 if (flag_code == CODE_64BIT)
5616 {
5617 /* 64bit mode overwrites the 32bit absolute
5618 addressing by RIP relative addressing and
5619 absolute addressing is encoded by one of the
5620 redundant SIB forms. */
5621 i.rm.regmem = ESCAPE_TO_TWO_BYTE_ADDRESSING;
5622 i.sib.base = NO_BASE_REGISTER;
5623 i.sib.index = NO_INDEX_REGISTER;
5624 i.types[op] = ((i.prefix[ADDR_PREFIX] == 0)
5625 ? disp32s : disp32);
5626 }
5627 else if ((flag_code == CODE_16BIT)
5628 ^ (i.prefix[ADDR_PREFIX] != 0))
5629 {
5630 i.rm.regmem = NO_BASE_REGISTER_16;
5631 i.types[op] = disp16;
5632 }
5633 else
5634 {
5635 i.rm.regmem = NO_BASE_REGISTER;
5636 i.types[op] = disp32;
5637 }
5638 }
5639 else if (!i.tm.opcode_modifier.vecsib)
5640 {
5641 /* !i.base_reg && i.index_reg */
5642 if (i.index_reg->reg_num == RegEiz
5643 || i.index_reg->reg_num == RegRiz)
5644 i.sib.index = NO_INDEX_REGISTER;
5645 else
5646 i.sib.index = i.index_reg->reg_num;
5647 i.sib.base = NO_BASE_REGISTER;
5648 i.sib.scale = i.log2_scale_factor;
5649 i.rm.regmem = ESCAPE_TO_TWO_BYTE_ADDRESSING;
5650 i.types[op].bitfield.disp8 = 0;
5651 i.types[op].bitfield.disp16 = 0;
5652 i.types[op].bitfield.disp64 = 0;
5653 if (flag_code != CODE_64BIT)
5654 {
5655 /* Must be 32 bit */
5656 i.types[op].bitfield.disp32 = 1;
5657 i.types[op].bitfield.disp32s = 0;
5658 }
5659 else
5660 {
5661 i.types[op].bitfield.disp32 = 0;
5662 i.types[op].bitfield.disp32s = 1;
5663 }
5664 if ((i.index_reg->reg_flags & RegRex) != 0)
5665 i.rex |= REX_X;
5666 }
5667 }
5668 /* RIP addressing for 64bit mode. */
5669 else if (i.base_reg->reg_num == RegRip ||
5670 i.base_reg->reg_num == RegEip)
5671 {
5672 gas_assert (!i.tm.opcode_modifier.vecsib);
5673 i.rm.regmem = NO_BASE_REGISTER;
5674 i.types[op].bitfield.disp8 = 0;
5675 i.types[op].bitfield.disp16 = 0;
5676 i.types[op].bitfield.disp32 = 0;
5677 i.types[op].bitfield.disp32s = 1;
5678 i.types[op].bitfield.disp64 = 0;
5679 i.flags[op] |= Operand_PCrel;
5680 if (! i.disp_operands)
5681 fake_zero_displacement = 1;
5682 }
5683 else if (i.base_reg->reg_type.bitfield.reg16)
5684 {
5685 gas_assert (!i.tm.opcode_modifier.vecsib);
5686 switch (i.base_reg->reg_num)
5687 {
5688 case 3: /* (%bx) */
5689 if (i.index_reg == 0)
5690 i.rm.regmem = 7;
5691 else /* (%bx,%si) -> 0, or (%bx,%di) -> 1 */
5692 i.rm.regmem = i.index_reg->reg_num - 6;
5693 break;
5694 case 5: /* (%bp) */
5695 default_seg = &ss;
5696 if (i.index_reg == 0)
5697 {
5698 i.rm.regmem = 6;
5699 if (operand_type_check (i.types[op], disp) == 0)
5700 {
5701 /* fake (%bp) into 0(%bp) */
5702 i.types[op].bitfield.disp8 = 1;
5703 fake_zero_displacement = 1;
5704 }
5705 }
5706 else /* (%bp,%si) -> 2, or (%bp,%di) -> 3 */
5707 i.rm.regmem = i.index_reg->reg_num - 6 + 2;
5708 break;
5709 default: /* (%si) -> 4 or (%di) -> 5 */
5710 i.rm.regmem = i.base_reg->reg_num - 6 + 4;
5711 }
5712 i.rm.mode = mode_from_disp_size (i.types[op]);
5713 }
5714 else /* i.base_reg and 32/64 bit mode */
5715 {
5716 if (flag_code == CODE_64BIT
5717 && operand_type_check (i.types[op], disp))
5718 {
5719 i386_operand_type temp;
5720 operand_type_set (&temp, 0);
5721 temp.bitfield.disp8 = i.types[op].bitfield.disp8;
5722 i.types[op] = temp;
5723 if (i.prefix[ADDR_PREFIX] == 0)
5724 i.types[op].bitfield.disp32s = 1;
5725 else
5726 i.types[op].bitfield.disp32 = 1;
5727 }
5728
5729 if (!i.tm.opcode_modifier.vecsib)
5730 i.rm.regmem = i.base_reg->reg_num;
5731 if ((i.base_reg->reg_flags & RegRex) != 0)
5732 i.rex |= REX_B;
5733 i.sib.base = i.base_reg->reg_num;
5734 /* x86-64 ignores REX prefix bit here to avoid decoder
5735 complications. */
5736 if ((i.base_reg->reg_num & 7) == EBP_REG_NUM)
5737 {
5738 default_seg = &ss;
5739 if (i.disp_operands == 0)
5740 {
5741 fake_zero_displacement = 1;
5742 i.types[op].bitfield.disp8 = 1;
5743 }
5744 }
5745 else if (i.base_reg->reg_num == ESP_REG_NUM)
5746 {
5747 default_seg = &ss;
5748 }
5749 i.sib.scale = i.log2_scale_factor;
5750 if (i.index_reg == 0)
5751 {
5752 gas_assert (!i.tm.opcode_modifier.vecsib);
5753 /* <disp>(%esp) becomes two byte modrm with no index
5754 register. We've already stored the code for esp
5755 in i.rm.regmem ie. ESCAPE_TO_TWO_BYTE_ADDRESSING.
5756 Any base register besides %esp will not use the
5757 extra modrm byte. */
5758 i.sib.index = NO_INDEX_REGISTER;
5759 }
5760 else if (!i.tm.opcode_modifier.vecsib)
5761 {
5762 if (i.index_reg->reg_num == RegEiz
5763 || i.index_reg->reg_num == RegRiz)
5764 i.sib.index = NO_INDEX_REGISTER;
5765 else
5766 i.sib.index = i.index_reg->reg_num;
5767 i.rm.regmem = ESCAPE_TO_TWO_BYTE_ADDRESSING;
5768 if ((i.index_reg->reg_flags & RegRex) != 0)
5769 i.rex |= REX_X;
5770 }
5771
5772 if (i.disp_operands
5773 && (i.reloc[op] == BFD_RELOC_386_TLS_DESC_CALL
5774 || i.reloc[op] == BFD_RELOC_X86_64_TLSDESC_CALL))
5775 i.rm.mode = 0;
5776 else
5777 {
5778 if (!fake_zero_displacement
5779 && !i.disp_operands
5780 && i.disp_encoding)
5781 {
5782 fake_zero_displacement = 1;
5783 if (i.disp_encoding == disp_encoding_8bit)
5784 i.types[op].bitfield.disp8 = 1;
5785 else
5786 i.types[op].bitfield.disp32 = 1;
5787 }
5788 i.rm.mode = mode_from_disp_size (i.types[op]);
5789 }
5790 }
5791
5792 if (fake_zero_displacement)
5793 {
5794 /* Fakes a zero displacement assuming that i.types[op]
5795 holds the correct displacement size. */
5796 expressionS *exp;
5797
5798 gas_assert (i.op[op].disps == 0);
5799 exp = &disp_expressions[i.disp_operands++];
5800 i.op[op].disps = exp;
5801 exp->X_op = O_constant;
5802 exp->X_add_number = 0;
5803 exp->X_add_symbol = (symbolS *) 0;
5804 exp->X_op_symbol = (symbolS *) 0;
5805 }
5806
5807 mem = op;
5808 }
5809 else
5810 mem = ~0;
5811
5812 if (i.tm.opcode_modifier.vexsources == XOP2SOURCES)
5813 {
5814 if (operand_type_check (i.types[0], imm))
5815 i.vex.register_specifier = NULL;
5816 else
5817 {
5818 /* VEX.vvvv encodes one of the sources when the first
5819 operand is not an immediate. */
5820 if (i.tm.opcode_modifier.vexw == VEXW0)
5821 i.vex.register_specifier = i.op[0].regs;
5822 else
5823 i.vex.register_specifier = i.op[1].regs;
5824 }
5825
5826 /* Destination is a XMM register encoded in the ModRM.reg
5827 and VEX.R bit. */
5828 i.rm.reg = i.op[2].regs->reg_num;
5829 if ((i.op[2].regs->reg_flags & RegRex) != 0)
5830 i.rex |= REX_R;
5831
5832 /* ModRM.rm and VEX.B encodes the other source. */
5833 if (!i.mem_operands)
5834 {
5835 i.rm.mode = 3;
5836
5837 if (i.tm.opcode_modifier.vexw == VEXW0)
5838 i.rm.regmem = i.op[1].regs->reg_num;
5839 else
5840 i.rm.regmem = i.op[0].regs->reg_num;
5841
5842 if ((i.op[1].regs->reg_flags & RegRex) != 0)
5843 i.rex |= REX_B;
5844 }
5845 }
5846 else if (i.tm.opcode_modifier.vexvvvv == VEXLWP)
5847 {
5848 i.vex.register_specifier = i.op[2].regs;
5849 if (!i.mem_operands)
5850 {
5851 i.rm.mode = 3;
5852 i.rm.regmem = i.op[1].regs->reg_num;
5853 if ((i.op[1].regs->reg_flags & RegRex) != 0)
5854 i.rex |= REX_B;
5855 }
5856 }
5857 /* Fill in i.rm.reg or i.rm.regmem field with register operand
5858 (if any) based on i.tm.extension_opcode. Again, we must be
5859 careful to make sure that segment/control/debug/test/MMX
5860 registers are coded into the i.rm.reg field. */
5861 else if (i.reg_operands)
5862 {
5863 unsigned int op;
5864 unsigned int vex_reg = ~0;
5865
5866 for (op = 0; op < i.operands; op++)
5867 if (i.types[op].bitfield.reg8
5868 || i.types[op].bitfield.reg16
5869 || i.types[op].bitfield.reg32
5870 || i.types[op].bitfield.reg64
5871 || i.types[op].bitfield.regmmx
5872 || i.types[op].bitfield.regxmm
5873 || i.types[op].bitfield.regymm
5874 || i.types[op].bitfield.sreg2
5875 || i.types[op].bitfield.sreg3
5876 || i.types[op].bitfield.control
5877 || i.types[op].bitfield.debug
5878 || i.types[op].bitfield.test)
5879 break;
5880
5881 if (vex_3_sources)
5882 op = dest;
5883 else if (i.tm.opcode_modifier.vexvvvv == VEXXDS)
5884 {
5885 /* For instructions with VexNDS, the register-only
5886 source operand is encoded in VEX prefix. */
5887 gas_assert (mem != (unsigned int) ~0);
5888
5889 if (op > mem)
5890 {
5891 vex_reg = op++;
5892 gas_assert (op < i.operands);
5893 }
5894 else
5895 {
5896 /* Check register-only source operand when two source
5897 operands are swapped. */
5898 if (!i.tm.operand_types[op].bitfield.baseindex
5899 && i.tm.operand_types[op + 1].bitfield.baseindex)
5900 {
5901 vex_reg = op;
5902 op += 2;
5903 gas_assert (mem == (vex_reg + 1)
5904 && op < i.operands);
5905 }
5906 else
5907 {
5908 vex_reg = op + 1;
5909 gas_assert (vex_reg < i.operands);
5910 }
5911 }
5912 }
5913 else if (i.tm.opcode_modifier.vexvvvv == VEXNDD)
5914 {
5915 /* For instructions with VexNDD, the register destination
5916 is encoded in VEX prefix. */
5917 if (i.mem_operands == 0)
5918 {
5919 /* There is no memory operand. */
5920 gas_assert ((op + 2) == i.operands);
5921 vex_reg = op + 1;
5922 }
5923 else
5924 {
5925 /* There are only 2 operands. */
5926 gas_assert (op < 2 && i.operands == 2);
5927 vex_reg = 1;
5928 }
5929 }
5930 else
5931 gas_assert (op < i.operands);
5932
5933 if (vex_reg != (unsigned int) ~0)
5934 {
5935 i386_operand_type *type = &i.tm.operand_types[vex_reg];
5936
5937 if (type->bitfield.reg32 != 1
5938 && type->bitfield.reg64 != 1
5939 && !operand_type_equal (type, &regxmm)
5940 && !operand_type_equal (type, &regymm))
5941 abort ();
5942
5943 i.vex.register_specifier = i.op[vex_reg].regs;
5944 }
5945
5946 /* Don't set OP operand twice. */
5947 if (vex_reg != op)
5948 {
5949 /* If there is an extension opcode to put here, the
5950 register number must be put into the regmem field. */
5951 if (i.tm.extension_opcode != None)
5952 {
5953 i.rm.regmem = i.op[op].regs->reg_num;
5954 if ((i.op[op].regs->reg_flags & RegRex) != 0)
5955 i.rex |= REX_B;
5956 }
5957 else
5958 {
5959 i.rm.reg = i.op[op].regs->reg_num;
5960 if ((i.op[op].regs->reg_flags & RegRex) != 0)
5961 i.rex |= REX_R;
5962 }
5963 }
5964
5965 /* Now, if no memory operand has set i.rm.mode = 0, 1, 2 we
5966 must set it to 3 to indicate this is a register operand
5967 in the regmem field. */
5968 if (!i.mem_operands)
5969 i.rm.mode = 3;
5970 }
5971
5972 /* Fill in i.rm.reg field with extension opcode (if any). */
5973 if (i.tm.extension_opcode != None)
5974 i.rm.reg = i.tm.extension_opcode;
5975 }
5976 return default_seg;
5977 }
5978
5979 static void
5980 output_branch (void)
5981 {
5982 char *p;
5983 int size;
5984 int code16;
5985 int prefix;
5986 relax_substateT subtype;
5987 symbolS *sym;
5988 offsetT off;
5989
5990 code16 = flag_code == CODE_16BIT ? CODE16 : 0;
5991 size = i.disp_encoding == disp_encoding_32bit ? BIG : SMALL;
5992
5993 prefix = 0;
5994 if (i.prefix[DATA_PREFIX] != 0)
5995 {
5996 prefix = 1;
5997 i.prefixes -= 1;
5998 code16 ^= CODE16;
5999 }
6000 /* Pentium4 branch hints. */
6001 if (i.prefix[SEG_PREFIX] == CS_PREFIX_OPCODE /* not taken */
6002 || i.prefix[SEG_PREFIX] == DS_PREFIX_OPCODE /* taken */)
6003 {
6004 prefix++;
6005 i.prefixes--;
6006 }
6007 if (i.prefix[REX_PREFIX] != 0)
6008 {
6009 prefix++;
6010 i.prefixes--;
6011 }
6012
6013 if (i.prefixes != 0 && !intel_syntax)
6014 as_warn (_("skipping prefixes on this instruction"));
6015
6016 /* It's always a symbol; End frag & setup for relax.
6017 Make sure there is enough room in this frag for the largest
6018 instruction we may generate in md_convert_frag. This is 2
6019 bytes for the opcode and room for the prefix and largest
6020 displacement. */
6021 frag_grow (prefix + 2 + 4);
6022 /* Prefix and 1 opcode byte go in fr_fix. */
6023 p = frag_more (prefix + 1);
6024 if (i.prefix[DATA_PREFIX] != 0)
6025 *p++ = DATA_PREFIX_OPCODE;
6026 if (i.prefix[SEG_PREFIX] == CS_PREFIX_OPCODE
6027 || i.prefix[SEG_PREFIX] == DS_PREFIX_OPCODE)
6028 *p++ = i.prefix[SEG_PREFIX];
6029 if (i.prefix[REX_PREFIX] != 0)
6030 *p++ = i.prefix[REX_PREFIX];
6031 *p = i.tm.base_opcode;
6032
6033 if ((unsigned char) *p == JUMP_PC_RELATIVE)
6034 subtype = ENCODE_RELAX_STATE (UNCOND_JUMP, size);
6035 else if (cpu_arch_flags.bitfield.cpui386)
6036 subtype = ENCODE_RELAX_STATE (COND_JUMP, size);
6037 else
6038 subtype = ENCODE_RELAX_STATE (COND_JUMP86, size);
6039 subtype |= code16;
6040
6041 sym = i.op[0].disps->X_add_symbol;
6042 off = i.op[0].disps->X_add_number;
6043
6044 if (i.op[0].disps->X_op != O_constant
6045 && i.op[0].disps->X_op != O_symbol)
6046 {
6047 /* Handle complex expressions. */
6048 sym = make_expr_symbol (i.op[0].disps);
6049 off = 0;
6050 }
6051
6052 /* 1 possible extra opcode + 4 byte displacement go in var part.
6053 Pass reloc in fr_var. */
6054 frag_var (rs_machine_dependent, 5, i.reloc[0], subtype, sym, off, p);
6055 }
6056
6057 static void
6058 output_jump (void)
6059 {
6060 char *p;
6061 int size;
6062 fixS *fixP;
6063
6064 if (i.tm.opcode_modifier.jumpbyte)
6065 {
6066 /* This is a loop or jecxz type instruction. */
6067 size = 1;
6068 if (i.prefix[ADDR_PREFIX] != 0)
6069 {
6070 FRAG_APPEND_1_CHAR (ADDR_PREFIX_OPCODE);
6071 i.prefixes -= 1;
6072 }
6073 /* Pentium4 branch hints. */
6074 if (i.prefix[SEG_PREFIX] == CS_PREFIX_OPCODE /* not taken */
6075 || i.prefix[SEG_PREFIX] == DS_PREFIX_OPCODE /* taken */)
6076 {
6077 FRAG_APPEND_1_CHAR (i.prefix[SEG_PREFIX]);
6078 i.prefixes--;
6079 }
6080 }
6081 else
6082 {
6083 int code16;
6084
6085 code16 = 0;
6086 if (flag_code == CODE_16BIT)
6087 code16 = CODE16;
6088
6089 if (i.prefix[DATA_PREFIX] != 0)
6090 {
6091 FRAG_APPEND_1_CHAR (DATA_PREFIX_OPCODE);
6092 i.prefixes -= 1;
6093 code16 ^= CODE16;
6094 }
6095
6096 size = 4;
6097 if (code16)
6098 size = 2;
6099 }
6100
6101 if (i.prefix[REX_PREFIX] != 0)
6102 {
6103 FRAG_APPEND_1_CHAR (i.prefix[REX_PREFIX]);
6104 i.prefixes -= 1;
6105 }
6106
6107 if (i.prefixes != 0 && !intel_syntax)
6108 as_warn (_("skipping prefixes on this instruction"));
6109
6110 p = frag_more (i.tm.opcode_length + size);
6111 switch (i.tm.opcode_length)
6112 {
6113 case 2:
6114 *p++ = i.tm.base_opcode >> 8;
6115 case 1:
6116 *p++ = i.tm.base_opcode;
6117 break;
6118 default:
6119 abort ();
6120 }
6121
6122 fixP = fix_new_exp (frag_now, p - frag_now->fr_literal, size,
6123 i.op[0].disps, 1, reloc (size, 1, 1, i.reloc[0]));
6124
6125 /* All jumps handled here are signed, but don't use a signed limit
6126 check for 32 and 16 bit jumps as we want to allow wrap around at
6127 4G and 64k respectively. */
6128 if (size == 1)
6129 fixP->fx_signed = 1;
6130 }
6131
6132 static void
6133 output_interseg_jump (void)
6134 {
6135 char *p;
6136 int size;
6137 int prefix;
6138 int code16;
6139
6140 code16 = 0;
6141 if (flag_code == CODE_16BIT)
6142 code16 = CODE16;
6143
6144 prefix = 0;
6145 if (i.prefix[DATA_PREFIX] != 0)
6146 {
6147 prefix = 1;
6148 i.prefixes -= 1;
6149 code16 ^= CODE16;
6150 }
6151 if (i.prefix[REX_PREFIX] != 0)
6152 {
6153 prefix++;
6154 i.prefixes -= 1;
6155 }
6156
6157 size = 4;
6158 if (code16)
6159 size = 2;
6160
6161 if (i.prefixes != 0 && !intel_syntax)
6162 as_warn (_("skipping prefixes on this instruction"));
6163
6164 /* 1 opcode; 2 segment; offset */
6165 p = frag_more (prefix + 1 + 2 + size);
6166
6167 if (i.prefix[DATA_PREFIX] != 0)
6168 *p++ = DATA_PREFIX_OPCODE;
6169
6170 if (i.prefix[REX_PREFIX] != 0)
6171 *p++ = i.prefix[REX_PREFIX];
6172
6173 *p++ = i.tm.base_opcode;
6174 if (i.op[1].imms->X_op == O_constant)
6175 {
6176 offsetT n = i.op[1].imms->X_add_number;
6177
6178 if (size == 2
6179 && !fits_in_unsigned_word (n)
6180 && !fits_in_signed_word (n))
6181 {
6182 as_bad (_("16-bit jump out of range"));
6183 return;
6184 }
6185 md_number_to_chars (p, n, size);
6186 }
6187 else
6188 fix_new_exp (frag_now, p - frag_now->fr_literal, size,
6189 i.op[1].imms, 0, reloc (size, 0, 0, i.reloc[1]));
6190 if (i.op[0].imms->X_op != O_constant)
6191 as_bad (_("can't handle non absolute segment in `%s'"),
6192 i.tm.name);
6193 md_number_to_chars (p + size, (valueT) i.op[0].imms->X_add_number, 2);
6194 }
6195
6196 static void
6197 output_insn (void)
6198 {
6199 fragS *insn_start_frag;
6200 offsetT insn_start_off;
6201
6202 /* Tie dwarf2 debug info to the address at the start of the insn.
6203 We can't do this after the insn has been output as the current
6204 frag may have been closed off. eg. by frag_var. */
6205 dwarf2_emit_insn (0);
6206
6207 insn_start_frag = frag_now;
6208 insn_start_off = frag_now_fix ();
6209
6210 /* Output jumps. */
6211 if (i.tm.opcode_modifier.jump)
6212 output_branch ();
6213 else if (i.tm.opcode_modifier.jumpbyte
6214 || i.tm.opcode_modifier.jumpdword)
6215 output_jump ();
6216 else if (i.tm.opcode_modifier.jumpintersegment)
6217 output_interseg_jump ();
6218 else
6219 {
6220 /* Output normal instructions here. */
6221 char *p;
6222 unsigned char *q;
6223 unsigned int j;
6224 unsigned int prefix;
6225
6226 /* Since the VEX prefix contains the implicit prefix, we don't
6227 need the explicit prefix. */
6228 if (!i.tm.opcode_modifier.vex)
6229 {
6230 switch (i.tm.opcode_length)
6231 {
6232 case 3:
6233 if (i.tm.base_opcode & 0xff000000)
6234 {
6235 prefix = (i.tm.base_opcode >> 24) & 0xff;
6236 goto check_prefix;
6237 }
6238 break;
6239 case 2:
6240 if ((i.tm.base_opcode & 0xff0000) != 0)
6241 {
6242 prefix = (i.tm.base_opcode >> 16) & 0xff;
6243 if (i.tm.cpu_flags.bitfield.cpupadlock)
6244 {
6245 check_prefix:
6246 if (prefix != REPE_PREFIX_OPCODE
6247 || (i.prefix[REP_PREFIX]
6248 != REPE_PREFIX_OPCODE))
6249 add_prefix (prefix);
6250 }
6251 else
6252 add_prefix (prefix);
6253 }
6254 break;
6255 case 1:
6256 break;
6257 default:
6258 abort ();
6259 }
6260
6261 /* The prefix bytes. */
6262 for (j = ARRAY_SIZE (i.prefix), q = i.prefix; j > 0; j--, q++)
6263 if (*q)
6264 FRAG_APPEND_1_CHAR (*q);
6265 }
6266 else
6267 {
6268 for (j = 0, q = i.prefix; j < ARRAY_SIZE (i.prefix); j++, q++)
6269 if (*q)
6270 switch (j)
6271 {
6272 case REX_PREFIX:
6273 /* REX byte is encoded in VEX prefix. */
6274 break;
6275 case SEG_PREFIX:
6276 case ADDR_PREFIX:
6277 FRAG_APPEND_1_CHAR (*q);
6278 break;
6279 default:
6280 /* There should be no other prefixes for instructions
6281 with VEX prefix. */
6282 abort ();
6283 }
6284
6285 /* Now the VEX prefix. */
6286 p = frag_more (i.vex.length);
6287 for (j = 0; j < i.vex.length; j++)
6288 p[j] = i.vex.bytes[j];
6289 }
6290
6291 /* Now the opcode; be careful about word order here! */
6292 if (i.tm.opcode_length == 1)
6293 {
6294 FRAG_APPEND_1_CHAR (i.tm.base_opcode);
6295 }
6296 else
6297 {
6298 switch (i.tm.opcode_length)
6299 {
6300 case 3:
6301 p = frag_more (3);
6302 *p++ = (i.tm.base_opcode >> 16) & 0xff;
6303 break;
6304 case 2:
6305 p = frag_more (2);
6306 break;
6307 default:
6308 abort ();
6309 break;
6310 }
6311
6312 /* Put out high byte first: can't use md_number_to_chars! */
6313 *p++ = (i.tm.base_opcode >> 8) & 0xff;
6314 *p = i.tm.base_opcode & 0xff;
6315 }
6316
6317 /* Now the modrm byte and sib byte (if present). */
6318 if (i.tm.opcode_modifier.modrm)
6319 {
6320 FRAG_APPEND_1_CHAR ((i.rm.regmem << 0
6321 | i.rm.reg << 3
6322 | i.rm.mode << 6));
6323 /* If i.rm.regmem == ESP (4)
6324 && i.rm.mode != (Register mode)
6325 && not 16 bit
6326 ==> need second modrm byte. */
6327 if (i.rm.regmem == ESCAPE_TO_TWO_BYTE_ADDRESSING
6328 && i.rm.mode != 3
6329 && !(i.base_reg && i.base_reg->reg_type.bitfield.reg16))
6330 FRAG_APPEND_1_CHAR ((i.sib.base << 0
6331 | i.sib.index << 3
6332 | i.sib.scale << 6));
6333 }
6334
6335 if (i.disp_operands)
6336 output_disp (insn_start_frag, insn_start_off);
6337
6338 if (i.imm_operands)
6339 output_imm (insn_start_frag, insn_start_off);
6340 }
6341
6342 #ifdef DEBUG386
6343 if (flag_debug)
6344 {
6345 pi ("" /*line*/, &i);
6346 }
6347 #endif /* DEBUG386 */
6348 }
6349
6350 /* Return the size of the displacement operand N. */
6351
6352 static int
6353 disp_size (unsigned int n)
6354 {
6355 int size = 4;
6356 if (i.types[n].bitfield.disp64)
6357 size = 8;
6358 else if (i.types[n].bitfield.disp8)
6359 size = 1;
6360 else if (i.types[n].bitfield.disp16)
6361 size = 2;
6362 return size;
6363 }
6364
6365 /* Return the size of the immediate operand N. */
6366
6367 static int
6368 imm_size (unsigned int n)
6369 {
6370 int size = 4;
6371 if (i.types[n].bitfield.imm64)
6372 size = 8;
6373 else if (i.types[n].bitfield.imm8 || i.types[n].bitfield.imm8s)
6374 size = 1;
6375 else if (i.types[n].bitfield.imm16)
6376 size = 2;
6377 return size;
6378 }
6379
6380 static void
6381 output_disp (fragS *insn_start_frag, offsetT insn_start_off)
6382 {
6383 char *p;
6384 unsigned int n;
6385
6386 for (n = 0; n < i.operands; n++)
6387 {
6388 if (operand_type_check (i.types[n], disp))
6389 {
6390 if (i.op[n].disps->X_op == O_constant)
6391 {
6392 int size = disp_size (n);
6393 offsetT val;
6394
6395 val = offset_in_range (i.op[n].disps->X_add_number,
6396 size);
6397 p = frag_more (size);
6398 md_number_to_chars (p, val, size);
6399 }
6400 else
6401 {
6402 enum bfd_reloc_code_real reloc_type;
6403 int size = disp_size (n);
6404 int sign = i.types[n].bitfield.disp32s;
6405 int pcrel = (i.flags[n] & Operand_PCrel) != 0;
6406
6407 /* We can't have 8 bit displacement here. */
6408 gas_assert (!i.types[n].bitfield.disp8);
6409
6410 /* The PC relative address is computed relative
6411 to the instruction boundary, so in case immediate
6412 fields follows, we need to adjust the value. */
6413 if (pcrel && i.imm_operands)
6414 {
6415 unsigned int n1;
6416 int sz = 0;
6417
6418 for (n1 = 0; n1 < i.operands; n1++)
6419 if (operand_type_check (i.types[n1], imm))
6420 {
6421 /* Only one immediate is allowed for PC
6422 relative address. */
6423 gas_assert (sz == 0);
6424 sz = imm_size (n1);
6425 i.op[n].disps->X_add_number -= sz;
6426 }
6427 /* We should find the immediate. */
6428 gas_assert (sz != 0);
6429 }
6430
6431 p = frag_more (size);
6432 reloc_type = reloc (size, pcrel, sign, i.reloc[n]);
6433 if (GOT_symbol
6434 && GOT_symbol == i.op[n].disps->X_add_symbol
6435 && (((reloc_type == BFD_RELOC_32
6436 || reloc_type == BFD_RELOC_X86_64_32S
6437 || (reloc_type == BFD_RELOC_64
6438 && object_64bit))
6439 && (i.op[n].disps->X_op == O_symbol
6440 || (i.op[n].disps->X_op == O_add
6441 && ((symbol_get_value_expression
6442 (i.op[n].disps->X_op_symbol)->X_op)
6443 == O_subtract))))
6444 || reloc_type == BFD_RELOC_32_PCREL))
6445 {
6446 offsetT add;
6447
6448 if (insn_start_frag == frag_now)
6449 add = (p - frag_now->fr_literal) - insn_start_off;
6450 else
6451 {
6452 fragS *fr;
6453
6454 add = insn_start_frag->fr_fix - insn_start_off;
6455 for (fr = insn_start_frag->fr_next;
6456 fr && fr != frag_now; fr = fr->fr_next)
6457 add += fr->fr_fix;
6458 add += p - frag_now->fr_literal;
6459 }
6460
6461 if (!object_64bit)
6462 {
6463 reloc_type = BFD_RELOC_386_GOTPC;
6464 i.op[n].imms->X_add_number += add;
6465 }
6466 else if (reloc_type == BFD_RELOC_64)
6467 reloc_type = BFD_RELOC_X86_64_GOTPC64;
6468 else
6469 /* Don't do the adjustment for x86-64, as there
6470 the pcrel addressing is relative to the _next_
6471 insn, and that is taken care of in other code. */
6472 reloc_type = BFD_RELOC_X86_64_GOTPC32;
6473 }
6474 fix_new_exp (frag_now, p - frag_now->fr_literal, size,
6475 i.op[n].disps, pcrel, reloc_type);
6476 }
6477 }
6478 }
6479 }
6480
6481 static void
6482 output_imm (fragS *insn_start_frag, offsetT insn_start_off)
6483 {
6484 char *p;
6485 unsigned int n;
6486
6487 for (n = 0; n < i.operands; n++)
6488 {
6489 if (operand_type_check (i.types[n], imm))
6490 {
6491 if (i.op[n].imms->X_op == O_constant)
6492 {
6493 int size = imm_size (n);
6494 offsetT val;
6495
6496 val = offset_in_range (i.op[n].imms->X_add_number,
6497 size);
6498 p = frag_more (size);
6499 md_number_to_chars (p, val, size);
6500 }
6501 else
6502 {
6503 /* Not absolute_section.
6504 Need a 32-bit fixup (don't support 8bit
6505 non-absolute imms). Try to support other
6506 sizes ... */
6507 enum bfd_reloc_code_real reloc_type;
6508 int size = imm_size (n);
6509 int sign;
6510
6511 if (i.types[n].bitfield.imm32s
6512 && (i.suffix == QWORD_MNEM_SUFFIX
6513 || (!i.suffix && i.tm.opcode_modifier.no_lsuf)))
6514 sign = 1;
6515 else
6516 sign = 0;
6517
6518 p = frag_more (size);
6519 reloc_type = reloc (size, 0, sign, i.reloc[n]);
6520
6521 /* This is tough to explain. We end up with this one if we
6522 * have operands that look like
6523 * "_GLOBAL_OFFSET_TABLE_+[.-.L284]". The goal here is to
6524 * obtain the absolute address of the GOT, and it is strongly
6525 * preferable from a performance point of view to avoid using
6526 * a runtime relocation for this. The actual sequence of
6527 * instructions often look something like:
6528 *
6529 * call .L66
6530 * .L66:
6531 * popl %ebx
6532 * addl $_GLOBAL_OFFSET_TABLE_+[.-.L66],%ebx
6533 *
6534 * The call and pop essentially return the absolute address
6535 * of the label .L66 and store it in %ebx. The linker itself
6536 * will ultimately change the first operand of the addl so
6537 * that %ebx points to the GOT, but to keep things simple, the
6538 * .o file must have this operand set so that it generates not
6539 * the absolute address of .L66, but the absolute address of
6540 * itself. This allows the linker itself simply treat a GOTPC
6541 * relocation as asking for a pcrel offset to the GOT to be
6542 * added in, and the addend of the relocation is stored in the
6543 * operand field for the instruction itself.
6544 *
6545 * Our job here is to fix the operand so that it would add
6546 * the correct offset so that %ebx would point to itself. The
6547 * thing that is tricky is that .-.L66 will point to the
6548 * beginning of the instruction, so we need to further modify
6549 * the operand so that it will point to itself. There are
6550 * other cases where you have something like:
6551 *
6552 * .long $_GLOBAL_OFFSET_TABLE_+[.-.L66]
6553 *
6554 * and here no correction would be required. Internally in
6555 * the assembler we treat operands of this form as not being
6556 * pcrel since the '.' is explicitly mentioned, and I wonder
6557 * whether it would simplify matters to do it this way. Who
6558 * knows. In earlier versions of the PIC patches, the
6559 * pcrel_adjust field was used to store the correction, but
6560 * since the expression is not pcrel, I felt it would be
6561 * confusing to do it this way. */
6562
6563 if ((reloc_type == BFD_RELOC_32
6564 || reloc_type == BFD_RELOC_X86_64_32S
6565 || reloc_type == BFD_RELOC_64)
6566 && GOT_symbol
6567 && GOT_symbol == i.op[n].imms->X_add_symbol
6568 && (i.op[n].imms->X_op == O_symbol
6569 || (i.op[n].imms->X_op == O_add
6570 && ((symbol_get_value_expression
6571 (i.op[n].imms->X_op_symbol)->X_op)
6572 == O_subtract))))
6573 {
6574 offsetT add;
6575
6576 if (insn_start_frag == frag_now)
6577 add = (p - frag_now->fr_literal) - insn_start_off;
6578 else
6579 {
6580 fragS *fr;
6581
6582 add = insn_start_frag->fr_fix - insn_start_off;
6583 for (fr = insn_start_frag->fr_next;
6584 fr && fr != frag_now; fr = fr->fr_next)
6585 add += fr->fr_fix;
6586 add += p - frag_now->fr_literal;
6587 }
6588
6589 if (!object_64bit)
6590 reloc_type = BFD_RELOC_386_GOTPC;
6591 else if (size == 4)
6592 reloc_type = BFD_RELOC_X86_64_GOTPC32;
6593 else if (size == 8)
6594 reloc_type = BFD_RELOC_X86_64_GOTPC64;
6595 i.op[n].imms->X_add_number += add;
6596 }
6597 fix_new_exp (frag_now, p - frag_now->fr_literal, size,
6598 i.op[n].imms, 0, reloc_type);
6599 }
6600 }
6601 }
6602 }
6603 \f
6604 /* x86_cons_fix_new is called via the expression parsing code when a
6605 reloc is needed. We use this hook to get the correct .got reloc. */
6606 static enum bfd_reloc_code_real got_reloc = NO_RELOC;
6607 static int cons_sign = -1;
6608
6609 void
6610 x86_cons_fix_new (fragS *frag, unsigned int off, unsigned int len,
6611 expressionS *exp)
6612 {
6613 enum bfd_reloc_code_real r = reloc (len, 0, cons_sign, got_reloc);
6614
6615 got_reloc = NO_RELOC;
6616
6617 #ifdef TE_PE
6618 if (exp->X_op == O_secrel)
6619 {
6620 exp->X_op = O_symbol;
6621 r = BFD_RELOC_32_SECREL;
6622 }
6623 #endif
6624
6625 fix_new_exp (frag, off, len, exp, 0, r);
6626 }
6627
6628 /* Export the ABI address size for use by TC_ADDRESS_BYTES for the
6629 purpose of the `.dc.a' internal pseudo-op. */
6630
6631 int
6632 x86_address_bytes (void)
6633 {
6634 if ((stdoutput->arch_info->mach & bfd_mach_x64_32))
6635 return 4;
6636 return stdoutput->arch_info->bits_per_address / 8;
6637 }
6638
6639 #if !(defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) || defined (OBJ_MACH_O)) \
6640 || defined (LEX_AT)
6641 # define lex_got(reloc, adjust, types) NULL
6642 #else
6643 /* Parse operands of the form
6644 <symbol>@GOTOFF+<nnn>
6645 and similar .plt or .got references.
6646
6647 If we find one, set up the correct relocation in RELOC and copy the
6648 input string, minus the `@GOTOFF' into a malloc'd buffer for
6649 parsing by the calling routine. Return this buffer, and if ADJUST
6650 is non-null set it to the length of the string we removed from the
6651 input line. Otherwise return NULL. */
6652 static char *
6653 lex_got (enum bfd_reloc_code_real *rel,
6654 int *adjust,
6655 i386_operand_type *types)
6656 {
6657 /* Some of the relocations depend on the size of what field is to
6658 be relocated. But in our callers i386_immediate and i386_displacement
6659 we don't yet know the operand size (this will be set by insn
6660 matching). Hence we record the word32 relocation here,
6661 and adjust the reloc according to the real size in reloc(). */
6662 static const struct {
6663 const char *str;
6664 int len;
6665 const enum bfd_reloc_code_real rel[2];
6666 const i386_operand_type types64;
6667 } gotrel[] = {
6668 { STRING_COMMA_LEN ("PLTOFF"), { _dummy_first_bfd_reloc_code_real,
6669 BFD_RELOC_X86_64_PLTOFF64 },
6670 OPERAND_TYPE_IMM64 },
6671 { STRING_COMMA_LEN ("PLT"), { BFD_RELOC_386_PLT32,
6672 BFD_RELOC_X86_64_PLT32 },
6673 OPERAND_TYPE_IMM32_32S_DISP32 },
6674 { STRING_COMMA_LEN ("GOTPLT"), { _dummy_first_bfd_reloc_code_real,
6675 BFD_RELOC_X86_64_GOTPLT64 },
6676 OPERAND_TYPE_IMM64_DISP64 },
6677 { STRING_COMMA_LEN ("GOTOFF"), { BFD_RELOC_386_GOTOFF,
6678 BFD_RELOC_X86_64_GOTOFF64 },
6679 OPERAND_TYPE_IMM64_DISP64 },
6680 { STRING_COMMA_LEN ("GOTPCREL"), { _dummy_first_bfd_reloc_code_real,
6681 BFD_RELOC_X86_64_GOTPCREL },
6682 OPERAND_TYPE_IMM32_32S_DISP32 },
6683 { STRING_COMMA_LEN ("TLSGD"), { BFD_RELOC_386_TLS_GD,
6684 BFD_RELOC_X86_64_TLSGD },
6685 OPERAND_TYPE_IMM32_32S_DISP32 },
6686 { STRING_COMMA_LEN ("TLSLDM"), { BFD_RELOC_386_TLS_LDM,
6687 _dummy_first_bfd_reloc_code_real },
6688 OPERAND_TYPE_NONE },
6689 { STRING_COMMA_LEN ("TLSLD"), { _dummy_first_bfd_reloc_code_real,
6690 BFD_RELOC_X86_64_TLSLD },
6691 OPERAND_TYPE_IMM32_32S_DISP32 },
6692 { STRING_COMMA_LEN ("GOTTPOFF"), { BFD_RELOC_386_TLS_IE_32,
6693 BFD_RELOC_X86_64_GOTTPOFF },
6694 OPERAND_TYPE_IMM32_32S_DISP32 },
6695 { STRING_COMMA_LEN ("TPOFF"), { BFD_RELOC_386_TLS_LE_32,
6696 BFD_RELOC_X86_64_TPOFF32 },
6697 OPERAND_TYPE_IMM32_32S_64_DISP32_64 },
6698 { STRING_COMMA_LEN ("NTPOFF"), { BFD_RELOC_386_TLS_LE,
6699 _dummy_first_bfd_reloc_code_real },
6700 OPERAND_TYPE_NONE },
6701 { STRING_COMMA_LEN ("DTPOFF"), { BFD_RELOC_386_TLS_LDO_32,
6702 BFD_RELOC_X86_64_DTPOFF32 },
6703 OPERAND_TYPE_IMM32_32S_64_DISP32_64 },
6704 { STRING_COMMA_LEN ("GOTNTPOFF"),{ BFD_RELOC_386_TLS_GOTIE,
6705 _dummy_first_bfd_reloc_code_real },
6706 OPERAND_TYPE_NONE },
6707 { STRING_COMMA_LEN ("INDNTPOFF"),{ BFD_RELOC_386_TLS_IE,
6708 _dummy_first_bfd_reloc_code_real },
6709 OPERAND_TYPE_NONE },
6710 { STRING_COMMA_LEN ("GOT"), { BFD_RELOC_386_GOT32,
6711 BFD_RELOC_X86_64_GOT32 },
6712 OPERAND_TYPE_IMM32_32S_64_DISP32 },
6713 { STRING_COMMA_LEN ("TLSDESC"), { BFD_RELOC_386_TLS_GOTDESC,
6714 BFD_RELOC_X86_64_GOTPC32_TLSDESC },
6715 OPERAND_TYPE_IMM32_32S_DISP32 },
6716 { STRING_COMMA_LEN ("TLSCALL"), { BFD_RELOC_386_TLS_DESC_CALL,
6717 BFD_RELOC_X86_64_TLSDESC_CALL },
6718 OPERAND_TYPE_IMM32_32S_DISP32 },
6719 };
6720 char *cp;
6721 unsigned int j;
6722
6723 #if defined (OBJ_MAYBE_ELF)
6724 if (!IS_ELF)
6725 return NULL;
6726 #endif
6727
6728 for (cp = input_line_pointer; *cp != '@'; cp++)
6729 if (is_end_of_line[(unsigned char) *cp] || *cp == ',')
6730 return NULL;
6731
6732 for (j = 0; j < ARRAY_SIZE (gotrel); j++)
6733 {
6734 int len = gotrel[j].len;
6735 if (strncasecmp (cp + 1, gotrel[j].str, len) == 0)
6736 {
6737 if (gotrel[j].rel[object_64bit] != 0)
6738 {
6739 int first, second;
6740 char *tmpbuf, *past_reloc;
6741
6742 *rel = gotrel[j].rel[object_64bit];
6743 if (adjust)
6744 *adjust = len;
6745
6746 if (types)
6747 {
6748 if (flag_code != CODE_64BIT)
6749 {
6750 types->bitfield.imm32 = 1;
6751 types->bitfield.disp32 = 1;
6752 }
6753 else
6754 *types = gotrel[j].types64;
6755 }
6756
6757 if (GOT_symbol == NULL)
6758 GOT_symbol = symbol_find_or_make (GLOBAL_OFFSET_TABLE_NAME);
6759
6760 /* The length of the first part of our input line. */
6761 first = cp - input_line_pointer;
6762
6763 /* The second part goes from after the reloc token until
6764 (and including) an end_of_line char or comma. */
6765 past_reloc = cp + 1 + len;
6766 cp = past_reloc;
6767 while (!is_end_of_line[(unsigned char) *cp] && *cp != ',')
6768 ++cp;
6769 second = cp + 1 - past_reloc;
6770
6771 /* Allocate and copy string. The trailing NUL shouldn't
6772 be necessary, but be safe. */
6773 tmpbuf = (char *) xmalloc (first + second + 2);
6774 memcpy (tmpbuf, input_line_pointer, first);
6775 if (second != 0 && *past_reloc != ' ')
6776 /* Replace the relocation token with ' ', so that
6777 errors like foo@GOTOFF1 will be detected. */
6778 tmpbuf[first++] = ' ';
6779 memcpy (tmpbuf + first, past_reloc, second);
6780 tmpbuf[first + second] = '\0';
6781 return tmpbuf;
6782 }
6783
6784 as_bad (_("@%s reloc is not supported with %d-bit output format"),
6785 gotrel[j].str, 1 << (5 + object_64bit));
6786 return NULL;
6787 }
6788 }
6789
6790 /* Might be a symbol version string. Don't as_bad here. */
6791 return NULL;
6792 }
6793 #endif
6794
6795 void
6796 x86_cons (expressionS *exp, int size)
6797 {
6798 intel_syntax = -intel_syntax;
6799
6800 exp->X_md = 0;
6801 if (size == 4 || (object_64bit && size == 8))
6802 {
6803 /* Handle @GOTOFF and the like in an expression. */
6804 char *save;
6805 char *gotfree_input_line;
6806 int adjust = 0;
6807
6808 save = input_line_pointer;
6809 gotfree_input_line = lex_got (&got_reloc, &adjust, NULL);
6810 if (gotfree_input_line)
6811 input_line_pointer = gotfree_input_line;
6812
6813 expression (exp);
6814
6815 if (gotfree_input_line)
6816 {
6817 /* expression () has merrily parsed up to the end of line,
6818 or a comma - in the wrong buffer. Transfer how far
6819 input_line_pointer has moved to the right buffer. */
6820 input_line_pointer = (save
6821 + (input_line_pointer - gotfree_input_line)
6822 + adjust);
6823 free (gotfree_input_line);
6824 if (exp->X_op == O_constant
6825 || exp->X_op == O_absent
6826 || exp->X_op == O_illegal
6827 || exp->X_op == O_register
6828 || exp->X_op == O_big)
6829 {
6830 char c = *input_line_pointer;
6831 *input_line_pointer = 0;
6832 as_bad (_("missing or invalid expression `%s'"), save);
6833 *input_line_pointer = c;
6834 }
6835 }
6836 }
6837 else
6838 expression (exp);
6839
6840 intel_syntax = -intel_syntax;
6841
6842 if (intel_syntax)
6843 i386_intel_simplify (exp);
6844 }
6845
6846 static void
6847 signed_cons (int size)
6848 {
6849 if (flag_code == CODE_64BIT)
6850 cons_sign = 1;
6851 cons (size);
6852 cons_sign = -1;
6853 }
6854
6855 #ifdef TE_PE
6856 static void
6857 pe_directive_secrel (int dummy ATTRIBUTE_UNUSED)
6858 {
6859 expressionS exp;
6860
6861 do
6862 {
6863 expression (&exp);
6864 if (exp.X_op == O_symbol)
6865 exp.X_op = O_secrel;
6866
6867 emit_expr (&exp, 4);
6868 }
6869 while (*input_line_pointer++ == ',');
6870
6871 input_line_pointer--;
6872 demand_empty_rest_of_line ();
6873 }
6874 #endif
6875
6876 static int
6877 i386_immediate (char *imm_start)
6878 {
6879 char *save_input_line_pointer;
6880 char *gotfree_input_line;
6881 segT exp_seg = 0;
6882 expressionS *exp;
6883 i386_operand_type types;
6884
6885 operand_type_set (&types, ~0);
6886
6887 if (i.imm_operands == MAX_IMMEDIATE_OPERANDS)
6888 {
6889 as_bad (_("at most %d immediate operands are allowed"),
6890 MAX_IMMEDIATE_OPERANDS);
6891 return 0;
6892 }
6893
6894 exp = &im_expressions[i.imm_operands++];
6895 i.op[this_operand].imms = exp;
6896
6897 if (is_space_char (*imm_start))
6898 ++imm_start;
6899
6900 save_input_line_pointer = input_line_pointer;
6901 input_line_pointer = imm_start;
6902
6903 gotfree_input_line = lex_got (&i.reloc[this_operand], NULL, &types);
6904 if (gotfree_input_line)
6905 input_line_pointer = gotfree_input_line;
6906
6907 exp_seg = expression (exp);
6908
6909 SKIP_WHITESPACE ();
6910 if (*input_line_pointer)
6911 as_bad (_("junk `%s' after expression"), input_line_pointer);
6912
6913 input_line_pointer = save_input_line_pointer;
6914 if (gotfree_input_line)
6915 {
6916 free (gotfree_input_line);
6917
6918 if (exp->X_op == O_constant || exp->X_op == O_register)
6919 exp->X_op = O_illegal;
6920 }
6921
6922 return i386_finalize_immediate (exp_seg, exp, types, imm_start);
6923 }
6924
6925 static int
6926 i386_finalize_immediate (segT exp_seg ATTRIBUTE_UNUSED, expressionS *exp,
6927 i386_operand_type types, const char *imm_start)
6928 {
6929 if (exp->X_op == O_absent || exp->X_op == O_illegal || exp->X_op == O_big)
6930 {
6931 if (imm_start)
6932 as_bad (_("missing or invalid immediate expression `%s'"),
6933 imm_start);
6934 return 0;
6935 }
6936 else if (exp->X_op == O_constant)
6937 {
6938 /* Size it properly later. */
6939 i.types[this_operand].bitfield.imm64 = 1;
6940 /* If not 64bit, sign extend val. */
6941 if (flag_code != CODE_64BIT
6942 && (exp->X_add_number & ~(((addressT) 2 << 31) - 1)) == 0)
6943 exp->X_add_number
6944 = (exp->X_add_number ^ ((addressT) 1 << 31)) - ((addressT) 1 << 31);
6945 }
6946 #if (defined (OBJ_AOUT) || defined (OBJ_MAYBE_AOUT))
6947 else if (OUTPUT_FLAVOR == bfd_target_aout_flavour
6948 && exp_seg != absolute_section
6949 && exp_seg != text_section
6950 && exp_seg != data_section
6951 && exp_seg != bss_section
6952 && exp_seg != undefined_section
6953 && !bfd_is_com_section (exp_seg))
6954 {
6955 as_bad (_("unimplemented segment %s in operand"), exp_seg->name);
6956 return 0;
6957 }
6958 #endif
6959 else if (!intel_syntax && exp->X_op == O_register)
6960 {
6961 if (imm_start)
6962 as_bad (_("illegal immediate register operand %s"), imm_start);
6963 return 0;
6964 }
6965 else
6966 {
6967 /* This is an address. The size of the address will be
6968 determined later, depending on destination register,
6969 suffix, or the default for the section. */
6970 i.types[this_operand].bitfield.imm8 = 1;
6971 i.types[this_operand].bitfield.imm16 = 1;
6972 i.types[this_operand].bitfield.imm32 = 1;
6973 i.types[this_operand].bitfield.imm32s = 1;
6974 i.types[this_operand].bitfield.imm64 = 1;
6975 i.types[this_operand] = operand_type_and (i.types[this_operand],
6976 types);
6977 }
6978
6979 return 1;
6980 }
6981
6982 static char *
6983 i386_scale (char *scale)
6984 {
6985 offsetT val;
6986 char *save = input_line_pointer;
6987
6988 input_line_pointer = scale;
6989 val = get_absolute_expression ();
6990
6991 switch (val)
6992 {
6993 case 1:
6994 i.log2_scale_factor = 0;
6995 break;
6996 case 2:
6997 i.log2_scale_factor = 1;
6998 break;
6999 case 4:
7000 i.log2_scale_factor = 2;
7001 break;
7002 case 8:
7003 i.log2_scale_factor = 3;
7004 break;
7005 default:
7006 {
7007 char sep = *input_line_pointer;
7008
7009 *input_line_pointer = '\0';
7010 as_bad (_("expecting scale factor of 1, 2, 4, or 8: got `%s'"),
7011 scale);
7012 *input_line_pointer = sep;
7013 input_line_pointer = save;
7014 return NULL;
7015 }
7016 }
7017 if (i.log2_scale_factor != 0 && i.index_reg == 0)
7018 {
7019 as_warn (_("scale factor of %d without an index register"),
7020 1 << i.log2_scale_factor);
7021 i.log2_scale_factor = 0;
7022 }
7023 scale = input_line_pointer;
7024 input_line_pointer = save;
7025 return scale;
7026 }
7027
7028 static int
7029 i386_displacement (char *disp_start, char *disp_end)
7030 {
7031 expressionS *exp;
7032 segT exp_seg = 0;
7033 char *save_input_line_pointer;
7034 char *gotfree_input_line;
7035 int override;
7036 i386_operand_type bigdisp, types = anydisp;
7037 int ret;
7038
7039 if (i.disp_operands == MAX_MEMORY_OPERANDS)
7040 {
7041 as_bad (_("at most %d displacement operands are allowed"),
7042 MAX_MEMORY_OPERANDS);
7043 return 0;
7044 }
7045
7046 operand_type_set (&bigdisp, 0);
7047 if ((i.types[this_operand].bitfield.jumpabsolute)
7048 || (!current_templates->start->opcode_modifier.jump
7049 && !current_templates->start->opcode_modifier.jumpdword))
7050 {
7051 bigdisp.bitfield.disp32 = 1;
7052 override = (i.prefix[ADDR_PREFIX] != 0);
7053 if (flag_code == CODE_64BIT)
7054 {
7055 if (!override)
7056 {
7057 bigdisp.bitfield.disp32s = 1;
7058 bigdisp.bitfield.disp64 = 1;
7059 }
7060 }
7061 else if ((flag_code == CODE_16BIT) ^ override)
7062 {
7063 bigdisp.bitfield.disp32 = 0;
7064 bigdisp.bitfield.disp16 = 1;
7065 }
7066 }
7067 else
7068 {
7069 /* For PC-relative branches, the width of the displacement
7070 is dependent upon data size, not address size. */
7071 override = (i.prefix[DATA_PREFIX] != 0);
7072 if (flag_code == CODE_64BIT)
7073 {
7074 if (override || i.suffix == WORD_MNEM_SUFFIX)
7075 bigdisp.bitfield.disp16 = 1;
7076 else
7077 {
7078 bigdisp.bitfield.disp32 = 1;
7079 bigdisp.bitfield.disp32s = 1;
7080 }
7081 }
7082 else
7083 {
7084 if (!override)
7085 override = (i.suffix == (flag_code != CODE_16BIT
7086 ? WORD_MNEM_SUFFIX
7087 : LONG_MNEM_SUFFIX));
7088 bigdisp.bitfield.disp32 = 1;
7089 if ((flag_code == CODE_16BIT) ^ override)
7090 {
7091 bigdisp.bitfield.disp32 = 0;
7092 bigdisp.bitfield.disp16 = 1;
7093 }
7094 }
7095 }
7096 i.types[this_operand] = operand_type_or (i.types[this_operand],
7097 bigdisp);
7098
7099 exp = &disp_expressions[i.disp_operands];
7100 i.op[this_operand].disps = exp;
7101 i.disp_operands++;
7102 save_input_line_pointer = input_line_pointer;
7103 input_line_pointer = disp_start;
7104 END_STRING_AND_SAVE (disp_end);
7105
7106 #ifndef GCC_ASM_O_HACK
7107 #define GCC_ASM_O_HACK 0
7108 #endif
7109 #if GCC_ASM_O_HACK
7110 END_STRING_AND_SAVE (disp_end + 1);
7111 if (i.types[this_operand].bitfield.baseIndex
7112 && displacement_string_end[-1] == '+')
7113 {
7114 /* This hack is to avoid a warning when using the "o"
7115 constraint within gcc asm statements.
7116 For instance:
7117
7118 #define _set_tssldt_desc(n,addr,limit,type) \
7119 __asm__ __volatile__ ( \
7120 "movw %w2,%0\n\t" \
7121 "movw %w1,2+%0\n\t" \
7122 "rorl $16,%1\n\t" \
7123 "movb %b1,4+%0\n\t" \
7124 "movb %4,5+%0\n\t" \
7125 "movb $0,6+%0\n\t" \
7126 "movb %h1,7+%0\n\t" \
7127 "rorl $16,%1" \
7128 : "=o"(*(n)) : "q" (addr), "ri"(limit), "i"(type))
7129
7130 This works great except that the output assembler ends
7131 up looking a bit weird if it turns out that there is
7132 no offset. You end up producing code that looks like:
7133
7134 #APP
7135 movw $235,(%eax)
7136 movw %dx,2+(%eax)
7137 rorl $16,%edx
7138 movb %dl,4+(%eax)
7139 movb $137,5+(%eax)
7140 movb $0,6+(%eax)
7141 movb %dh,7+(%eax)
7142 rorl $16,%edx
7143 #NO_APP
7144
7145 So here we provide the missing zero. */
7146
7147 *displacement_string_end = '0';
7148 }
7149 #endif
7150 gotfree_input_line = lex_got (&i.reloc[this_operand], NULL, &types);
7151 if (gotfree_input_line)
7152 input_line_pointer = gotfree_input_line;
7153
7154 exp_seg = expression (exp);
7155
7156 SKIP_WHITESPACE ();
7157 if (*input_line_pointer)
7158 as_bad (_("junk `%s' after expression"), input_line_pointer);
7159 #if GCC_ASM_O_HACK
7160 RESTORE_END_STRING (disp_end + 1);
7161 #endif
7162 input_line_pointer = save_input_line_pointer;
7163 if (gotfree_input_line)
7164 {
7165 free (gotfree_input_line);
7166
7167 if (exp->X_op == O_constant || exp->X_op == O_register)
7168 exp->X_op = O_illegal;
7169 }
7170
7171 ret = i386_finalize_displacement (exp_seg, exp, types, disp_start);
7172
7173 RESTORE_END_STRING (disp_end);
7174
7175 return ret;
7176 }
7177
7178 static int
7179 i386_finalize_displacement (segT exp_seg ATTRIBUTE_UNUSED, expressionS *exp,
7180 i386_operand_type types, const char *disp_start)
7181 {
7182 i386_operand_type bigdisp;
7183 int ret = 1;
7184
7185 /* We do this to make sure that the section symbol is in
7186 the symbol table. We will ultimately change the relocation
7187 to be relative to the beginning of the section. */
7188 if (i.reloc[this_operand] == BFD_RELOC_386_GOTOFF
7189 || i.reloc[this_operand] == BFD_RELOC_X86_64_GOTPCREL
7190 || i.reloc[this_operand] == BFD_RELOC_X86_64_GOTOFF64)
7191 {
7192 if (exp->X_op != O_symbol)
7193 goto inv_disp;
7194
7195 if (S_IS_LOCAL (exp->X_add_symbol)
7196 && S_GET_SEGMENT (exp->X_add_symbol) != undefined_section
7197 && S_GET_SEGMENT (exp->X_add_symbol) != expr_section)
7198 section_symbol (S_GET_SEGMENT (exp->X_add_symbol));
7199 exp->X_op = O_subtract;
7200 exp->X_op_symbol = GOT_symbol;
7201 if (i.reloc[this_operand] == BFD_RELOC_X86_64_GOTPCREL)
7202 i.reloc[this_operand] = BFD_RELOC_32_PCREL;
7203 else if (i.reloc[this_operand] == BFD_RELOC_X86_64_GOTOFF64)
7204 i.reloc[this_operand] = BFD_RELOC_64;
7205 else
7206 i.reloc[this_operand] = BFD_RELOC_32;
7207 }
7208
7209 else if (exp->X_op == O_absent
7210 || exp->X_op == O_illegal
7211 || exp->X_op == O_big)
7212 {
7213 inv_disp:
7214 as_bad (_("missing or invalid displacement expression `%s'"),
7215 disp_start);
7216 ret = 0;
7217 }
7218
7219 else if (flag_code == CODE_64BIT
7220 && !i.prefix[ADDR_PREFIX]
7221 && exp->X_op == O_constant)
7222 {
7223 /* Since displacement is signed extended to 64bit, don't allow
7224 disp32 and turn off disp32s if they are out of range. */
7225 i.types[this_operand].bitfield.disp32 = 0;
7226 if (!fits_in_signed_long (exp->X_add_number))
7227 {
7228 i.types[this_operand].bitfield.disp32s = 0;
7229 if (i.types[this_operand].bitfield.baseindex)
7230 {
7231 as_bad (_("0x%lx out range of signed 32bit displacement"),
7232 (long) exp->X_add_number);
7233 ret = 0;
7234 }
7235 }
7236 }
7237
7238 #if (defined (OBJ_AOUT) || defined (OBJ_MAYBE_AOUT))
7239 else if (exp->X_op != O_constant
7240 && OUTPUT_FLAVOR == bfd_target_aout_flavour
7241 && exp_seg != absolute_section
7242 && exp_seg != text_section
7243 && exp_seg != data_section
7244 && exp_seg != bss_section
7245 && exp_seg != undefined_section
7246 && !bfd_is_com_section (exp_seg))
7247 {
7248 as_bad (_("unimplemented segment %s in operand"), exp_seg->name);
7249 ret = 0;
7250 }
7251 #endif
7252
7253 /* Check if this is a displacement only operand. */
7254 bigdisp = i.types[this_operand];
7255 bigdisp.bitfield.disp8 = 0;
7256 bigdisp.bitfield.disp16 = 0;
7257 bigdisp.bitfield.disp32 = 0;
7258 bigdisp.bitfield.disp32s = 0;
7259 bigdisp.bitfield.disp64 = 0;
7260 if (operand_type_all_zero (&bigdisp))
7261 i.types[this_operand] = operand_type_and (i.types[this_operand],
7262 types);
7263
7264 return ret;
7265 }
7266
7267 /* Make sure the memory operand we've been dealt is valid.
7268 Return 1 on success, 0 on a failure. */
7269
7270 static int
7271 i386_index_check (const char *operand_string)
7272 {
7273 int ok;
7274 const char *kind = "base/index";
7275 #if INFER_ADDR_PREFIX
7276 int fudged = 0;
7277
7278 tryprefix:
7279 #endif
7280 ok = 1;
7281 if (current_templates->start->opcode_modifier.isstring
7282 && !current_templates->start->opcode_modifier.immext
7283 && (current_templates->end[-1].opcode_modifier.isstring
7284 || i.mem_operands))
7285 {
7286 /* Memory operands of string insns are special in that they only allow
7287 a single register (rDI, rSI, or rBX) as their memory address. */
7288 unsigned int expected;
7289
7290 kind = "string address";
7291
7292 if (current_templates->start->opcode_modifier.w)
7293 {
7294 i386_operand_type type = current_templates->end[-1].operand_types[0];
7295
7296 if (!type.bitfield.baseindex
7297 || ((!i.mem_operands != !intel_syntax)
7298 && current_templates->end[-1].operand_types[1]
7299 .bitfield.baseindex))
7300 type = current_templates->end[-1].operand_types[1];
7301 expected = type.bitfield.esseg ? 7 /* rDI */ : 6 /* rSI */;
7302 }
7303 else
7304 expected = 3 /* rBX */;
7305
7306 if (!i.base_reg || i.index_reg
7307 || operand_type_check (i.types[this_operand], disp))
7308 ok = -1;
7309 else if (!(flag_code == CODE_64BIT
7310 ? i.prefix[ADDR_PREFIX]
7311 ? i.base_reg->reg_type.bitfield.reg32
7312 : i.base_reg->reg_type.bitfield.reg64
7313 : (flag_code == CODE_16BIT) ^ !i.prefix[ADDR_PREFIX]
7314 ? i.base_reg->reg_type.bitfield.reg32
7315 : i.base_reg->reg_type.bitfield.reg16))
7316 ok = 0;
7317 else if (i.base_reg->reg_num != expected)
7318 ok = -1;
7319
7320 if (ok < 0)
7321 {
7322 unsigned int j;
7323
7324 for (j = 0; j < i386_regtab_size; ++j)
7325 if ((flag_code == CODE_64BIT
7326 ? i.prefix[ADDR_PREFIX]
7327 ? i386_regtab[j].reg_type.bitfield.reg32
7328 : i386_regtab[j].reg_type.bitfield.reg64
7329 : (flag_code == CODE_16BIT) ^ !i.prefix[ADDR_PREFIX]
7330 ? i386_regtab[j].reg_type.bitfield.reg32
7331 : i386_regtab[j].reg_type.bitfield.reg16)
7332 && i386_regtab[j].reg_num == expected)
7333 break;
7334 gas_assert (j < i386_regtab_size);
7335 as_warn (_("`%s' is not valid here (expected `%c%s%s%c')"),
7336 operand_string,
7337 intel_syntax ? '[' : '(',
7338 register_prefix,
7339 i386_regtab[j].reg_name,
7340 intel_syntax ? ']' : ')');
7341 ok = 1;
7342 }
7343 }
7344 else if (flag_code == CODE_64BIT)
7345 {
7346 if ((i.base_reg
7347 && ((i.prefix[ADDR_PREFIX] == 0
7348 && !i.base_reg->reg_type.bitfield.reg64)
7349 || (i.prefix[ADDR_PREFIX]
7350 && !i.base_reg->reg_type.bitfield.reg32))
7351 && (i.index_reg
7352 || i.base_reg->reg_num !=
7353 (i.prefix[ADDR_PREFIX] == 0 ? RegRip : RegEip)))
7354 || (i.index_reg
7355 && !(i.index_reg->reg_type.bitfield.regxmm
7356 || i.index_reg->reg_type.bitfield.regymm)
7357 && (!i.index_reg->reg_type.bitfield.baseindex
7358 || (i.prefix[ADDR_PREFIX] == 0
7359 && i.index_reg->reg_num != RegRiz
7360 && !i.index_reg->reg_type.bitfield.reg64
7361 )
7362 || (i.prefix[ADDR_PREFIX]
7363 && i.index_reg->reg_num != RegEiz
7364 && !i.index_reg->reg_type.bitfield.reg32))))
7365 ok = 0;
7366 }
7367 else
7368 {
7369 if ((flag_code == CODE_16BIT) ^ (i.prefix[ADDR_PREFIX] != 0))
7370 {
7371 /* 16bit checks. */
7372 if ((i.base_reg
7373 && (!i.base_reg->reg_type.bitfield.reg16
7374 || !i.base_reg->reg_type.bitfield.baseindex))
7375 || (i.index_reg
7376 && (!i.index_reg->reg_type.bitfield.reg16
7377 || !i.index_reg->reg_type.bitfield.baseindex
7378 || !(i.base_reg
7379 && i.base_reg->reg_num < 6
7380 && i.index_reg->reg_num >= 6
7381 && i.log2_scale_factor == 0))))
7382 ok = 0;
7383 }
7384 else
7385 {
7386 /* 32bit checks. */
7387 if ((i.base_reg
7388 && !i.base_reg->reg_type.bitfield.reg32)
7389 || (i.index_reg
7390 && !i.index_reg->reg_type.bitfield.regxmm
7391 && !i.index_reg->reg_type.bitfield.regymm
7392 && ((!i.index_reg->reg_type.bitfield.reg32
7393 && i.index_reg->reg_num != RegEiz)
7394 || !i.index_reg->reg_type.bitfield.baseindex)))
7395 ok = 0;
7396 }
7397 }
7398 if (!ok)
7399 {
7400 #if INFER_ADDR_PREFIX
7401 if (!i.mem_operands && !i.prefix[ADDR_PREFIX])
7402 {
7403 i.prefix[ADDR_PREFIX] = ADDR_PREFIX_OPCODE;
7404 i.prefixes += 1;
7405 /* Change the size of any displacement too. At most one of
7406 Disp16 or Disp32 is set.
7407 FIXME. There doesn't seem to be any real need for separate
7408 Disp16 and Disp32 flags. The same goes for Imm16 and Imm32.
7409 Removing them would probably clean up the code quite a lot. */
7410 if (flag_code != CODE_64BIT
7411 && (i.types[this_operand].bitfield.disp16
7412 || i.types[this_operand].bitfield.disp32))
7413 i.types[this_operand]
7414 = operand_type_xor (i.types[this_operand], disp16_32);
7415 fudged = 1;
7416 goto tryprefix;
7417 }
7418 if (fudged)
7419 as_bad (_("`%s' is not a valid %s expression"),
7420 operand_string,
7421 kind);
7422 else
7423 #endif
7424 as_bad (_("`%s' is not a valid %s-bit %s expression"),
7425 operand_string,
7426 flag_code_names[i.prefix[ADDR_PREFIX]
7427 ? flag_code == CODE_32BIT
7428 ? CODE_16BIT
7429 : CODE_32BIT
7430 : flag_code],
7431 kind);
7432 }
7433 return ok;
7434 }
7435
7436 /* Parse OPERAND_STRING into the i386_insn structure I. Returns zero
7437 on error. */
7438
7439 static int
7440 i386_att_operand (char *operand_string)
7441 {
7442 const reg_entry *r;
7443 char *end_op;
7444 char *op_string = operand_string;
7445
7446 if (is_space_char (*op_string))
7447 ++op_string;
7448
7449 /* We check for an absolute prefix (differentiating,
7450 for example, 'jmp pc_relative_label' from 'jmp *absolute_label'. */
7451 if (*op_string == ABSOLUTE_PREFIX)
7452 {
7453 ++op_string;
7454 if (is_space_char (*op_string))
7455 ++op_string;
7456 i.types[this_operand].bitfield.jumpabsolute = 1;
7457 }
7458
7459 /* Check if operand is a register. */
7460 if ((r = parse_register (op_string, &end_op)) != NULL)
7461 {
7462 i386_operand_type temp;
7463
7464 /* Check for a segment override by searching for ':' after a
7465 segment register. */
7466 op_string = end_op;
7467 if (is_space_char (*op_string))
7468 ++op_string;
7469 if (*op_string == ':'
7470 && (r->reg_type.bitfield.sreg2
7471 || r->reg_type.bitfield.sreg3))
7472 {
7473 switch (r->reg_num)
7474 {
7475 case 0:
7476 i.seg[i.mem_operands] = &es;
7477 break;
7478 case 1:
7479 i.seg[i.mem_operands] = &cs;
7480 break;
7481 case 2:
7482 i.seg[i.mem_operands] = &ss;
7483 break;
7484 case 3:
7485 i.seg[i.mem_operands] = &ds;
7486 break;
7487 case 4:
7488 i.seg[i.mem_operands] = &fs;
7489 break;
7490 case 5:
7491 i.seg[i.mem_operands] = &gs;
7492 break;
7493 }
7494
7495 /* Skip the ':' and whitespace. */
7496 ++op_string;
7497 if (is_space_char (*op_string))
7498 ++op_string;
7499
7500 if (!is_digit_char (*op_string)
7501 && !is_identifier_char (*op_string)
7502 && *op_string != '('
7503 && *op_string != ABSOLUTE_PREFIX)
7504 {
7505 as_bad (_("bad memory operand `%s'"), op_string);
7506 return 0;
7507 }
7508 /* Handle case of %es:*foo. */
7509 if (*op_string == ABSOLUTE_PREFIX)
7510 {
7511 ++op_string;
7512 if (is_space_char (*op_string))
7513 ++op_string;
7514 i.types[this_operand].bitfield.jumpabsolute = 1;
7515 }
7516 goto do_memory_reference;
7517 }
7518 if (*op_string)
7519 {
7520 as_bad (_("junk `%s' after register"), op_string);
7521 return 0;
7522 }
7523 temp = r->reg_type;
7524 temp.bitfield.baseindex = 0;
7525 i.types[this_operand] = operand_type_or (i.types[this_operand],
7526 temp);
7527 i.types[this_operand].bitfield.unspecified = 0;
7528 i.op[this_operand].regs = r;
7529 i.reg_operands++;
7530 }
7531 else if (*op_string == REGISTER_PREFIX)
7532 {
7533 as_bad (_("bad register name `%s'"), op_string);
7534 return 0;
7535 }
7536 else if (*op_string == IMMEDIATE_PREFIX)
7537 {
7538 ++op_string;
7539 if (i.types[this_operand].bitfield.jumpabsolute)
7540 {
7541 as_bad (_("immediate operand illegal with absolute jump"));
7542 return 0;
7543 }
7544 if (!i386_immediate (op_string))
7545 return 0;
7546 }
7547 else if (is_digit_char (*op_string)
7548 || is_identifier_char (*op_string)
7549 || *op_string == '(')
7550 {
7551 /* This is a memory reference of some sort. */
7552 char *base_string;
7553
7554 /* Start and end of displacement string expression (if found). */
7555 char *displacement_string_start;
7556 char *displacement_string_end;
7557
7558 do_memory_reference:
7559 if ((i.mem_operands == 1
7560 && !current_templates->start->opcode_modifier.isstring)
7561 || i.mem_operands == 2)
7562 {
7563 as_bad (_("too many memory references for `%s'"),
7564 current_templates->start->name);
7565 return 0;
7566 }
7567
7568 /* Check for base index form. We detect the base index form by
7569 looking for an ')' at the end of the operand, searching
7570 for the '(' matching it, and finding a REGISTER_PREFIX or ','
7571 after the '('. */
7572 base_string = op_string + strlen (op_string);
7573
7574 --base_string;
7575 if (is_space_char (*base_string))
7576 --base_string;
7577
7578 /* If we only have a displacement, set-up for it to be parsed later. */
7579 displacement_string_start = op_string;
7580 displacement_string_end = base_string + 1;
7581
7582 if (*base_string == ')')
7583 {
7584 char *temp_string;
7585 unsigned int parens_balanced = 1;
7586 /* We've already checked that the number of left & right ()'s are
7587 equal, so this loop will not be infinite. */
7588 do
7589 {
7590 base_string--;
7591 if (*base_string == ')')
7592 parens_balanced++;
7593 if (*base_string == '(')
7594 parens_balanced--;
7595 }
7596 while (parens_balanced);
7597
7598 temp_string = base_string;
7599
7600 /* Skip past '(' and whitespace. */
7601 ++base_string;
7602 if (is_space_char (*base_string))
7603 ++base_string;
7604
7605 if (*base_string == ','
7606 || ((i.base_reg = parse_register (base_string, &end_op))
7607 != NULL))
7608 {
7609 displacement_string_end = temp_string;
7610
7611 i.types[this_operand].bitfield.baseindex = 1;
7612
7613 if (i.base_reg)
7614 {
7615 base_string = end_op;
7616 if (is_space_char (*base_string))
7617 ++base_string;
7618 }
7619
7620 /* There may be an index reg or scale factor here. */
7621 if (*base_string == ',')
7622 {
7623 ++base_string;
7624 if (is_space_char (*base_string))
7625 ++base_string;
7626
7627 if ((i.index_reg = parse_register (base_string, &end_op))
7628 != NULL)
7629 {
7630 base_string = end_op;
7631 if (is_space_char (*base_string))
7632 ++base_string;
7633 if (*base_string == ',')
7634 {
7635 ++base_string;
7636 if (is_space_char (*base_string))
7637 ++base_string;
7638 }
7639 else if (*base_string != ')')
7640 {
7641 as_bad (_("expecting `,' or `)' "
7642 "after index register in `%s'"),
7643 operand_string);
7644 return 0;
7645 }
7646 }
7647 else if (*base_string == REGISTER_PREFIX)
7648 {
7649 as_bad (_("bad register name `%s'"), base_string);
7650 return 0;
7651 }
7652
7653 /* Check for scale factor. */
7654 if (*base_string != ')')
7655 {
7656 char *end_scale = i386_scale (base_string);
7657
7658 if (!end_scale)
7659 return 0;
7660
7661 base_string = end_scale;
7662 if (is_space_char (*base_string))
7663 ++base_string;
7664 if (*base_string != ')')
7665 {
7666 as_bad (_("expecting `)' "
7667 "after scale factor in `%s'"),
7668 operand_string);
7669 return 0;
7670 }
7671 }
7672 else if (!i.index_reg)
7673 {
7674 as_bad (_("expecting index register or scale factor "
7675 "after `,'; got '%c'"),
7676 *base_string);
7677 return 0;
7678 }
7679 }
7680 else if (*base_string != ')')
7681 {
7682 as_bad (_("expecting `,' or `)' "
7683 "after base register in `%s'"),
7684 operand_string);
7685 return 0;
7686 }
7687 }
7688 else if (*base_string == REGISTER_PREFIX)
7689 {
7690 as_bad (_("bad register name `%s'"), base_string);
7691 return 0;
7692 }
7693 }
7694
7695 /* If there's an expression beginning the operand, parse it,
7696 assuming displacement_string_start and
7697 displacement_string_end are meaningful. */
7698 if (displacement_string_start != displacement_string_end)
7699 {
7700 if (!i386_displacement (displacement_string_start,
7701 displacement_string_end))
7702 return 0;
7703 }
7704
7705 /* Special case for (%dx) while doing input/output op. */
7706 if (i.base_reg
7707 && operand_type_equal (&i.base_reg->reg_type,
7708 &reg16_inoutportreg)
7709 && i.index_reg == 0
7710 && i.log2_scale_factor == 0
7711 && i.seg[i.mem_operands] == 0
7712 && !operand_type_check (i.types[this_operand], disp))
7713 {
7714 i.types[this_operand] = inoutportreg;
7715 return 1;
7716 }
7717
7718 if (i386_index_check (operand_string) == 0)
7719 return 0;
7720 i.types[this_operand].bitfield.mem = 1;
7721 i.mem_operands++;
7722 }
7723 else
7724 {
7725 /* It's not a memory operand; argh! */
7726 as_bad (_("invalid char %s beginning operand %d `%s'"),
7727 output_invalid (*op_string),
7728 this_operand + 1,
7729 op_string);
7730 return 0;
7731 }
7732 return 1; /* Normal return. */
7733 }
7734 \f
7735 /* Calculate the maximum variable size (i.e., excluding fr_fix)
7736 that an rs_machine_dependent frag may reach. */
7737
7738 unsigned int
7739 i386_frag_max_var (fragS *frag)
7740 {
7741 /* The only relaxable frags are for jumps.
7742 Unconditional jumps can grow by 4 bytes and others by 5 bytes. */
7743 gas_assert (frag->fr_type == rs_machine_dependent);
7744 return TYPE_FROM_RELAX_STATE (frag->fr_subtype) == UNCOND_JUMP ? 4 : 5;
7745 }
7746
7747 /* md_estimate_size_before_relax()
7748
7749 Called just before relax() for rs_machine_dependent frags. The x86
7750 assembler uses these frags to handle variable size jump
7751 instructions.
7752
7753 Any symbol that is now undefined will not become defined.
7754 Return the correct fr_subtype in the frag.
7755 Return the initial "guess for variable size of frag" to caller.
7756 The guess is actually the growth beyond the fixed part. Whatever
7757 we do to grow the fixed or variable part contributes to our
7758 returned value. */
7759
7760 int
7761 md_estimate_size_before_relax (fragS *fragP, segT segment)
7762 {
7763 /* We've already got fragP->fr_subtype right; all we have to do is
7764 check for un-relaxable symbols. On an ELF system, we can't relax
7765 an externally visible symbol, because it may be overridden by a
7766 shared library. */
7767 if (S_GET_SEGMENT (fragP->fr_symbol) != segment
7768 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
7769 || (IS_ELF
7770 && (S_IS_EXTERNAL (fragP->fr_symbol)
7771 || S_IS_WEAK (fragP->fr_symbol)
7772 || ((symbol_get_bfdsym (fragP->fr_symbol)->flags
7773 & BSF_GNU_INDIRECT_FUNCTION))))
7774 #endif
7775 #if defined (OBJ_COFF) && defined (TE_PE)
7776 || (OUTPUT_FLAVOR == bfd_target_coff_flavour
7777 && S_IS_WEAK (fragP->fr_symbol))
7778 #endif
7779 )
7780 {
7781 /* Symbol is undefined in this segment, or we need to keep a
7782 reloc so that weak symbols can be overridden. */
7783 int size = (fragP->fr_subtype & CODE16) ? 2 : 4;
7784 enum bfd_reloc_code_real reloc_type;
7785 unsigned char *opcode;
7786 int old_fr_fix;
7787
7788 if (fragP->fr_var != NO_RELOC)
7789 reloc_type = (enum bfd_reloc_code_real) fragP->fr_var;
7790 else if (size == 2)
7791 reloc_type = BFD_RELOC_16_PCREL;
7792 else
7793 reloc_type = BFD_RELOC_32_PCREL;
7794
7795 old_fr_fix = fragP->fr_fix;
7796 opcode = (unsigned char *) fragP->fr_opcode;
7797
7798 switch (TYPE_FROM_RELAX_STATE (fragP->fr_subtype))
7799 {
7800 case UNCOND_JUMP:
7801 /* Make jmp (0xeb) a (d)word displacement jump. */
7802 opcode[0] = 0xe9;
7803 fragP->fr_fix += size;
7804 fix_new (fragP, old_fr_fix, size,
7805 fragP->fr_symbol,
7806 fragP->fr_offset, 1,
7807 reloc_type);
7808 break;
7809
7810 case COND_JUMP86:
7811 if (size == 2
7812 && (!no_cond_jump_promotion || fragP->fr_var != NO_RELOC))
7813 {
7814 /* Negate the condition, and branch past an
7815 unconditional jump. */
7816 opcode[0] ^= 1;
7817 opcode[1] = 3;
7818 /* Insert an unconditional jump. */
7819 opcode[2] = 0xe9;
7820 /* We added two extra opcode bytes, and have a two byte
7821 offset. */
7822 fragP->fr_fix += 2 + 2;
7823 fix_new (fragP, old_fr_fix + 2, 2,
7824 fragP->fr_symbol,
7825 fragP->fr_offset, 1,
7826 reloc_type);
7827 break;
7828 }
7829 /* Fall through. */
7830
7831 case COND_JUMP:
7832 if (no_cond_jump_promotion && fragP->fr_var == NO_RELOC)
7833 {
7834 fixS *fixP;
7835
7836 fragP->fr_fix += 1;
7837 fixP = fix_new (fragP, old_fr_fix, 1,
7838 fragP->fr_symbol,
7839 fragP->fr_offset, 1,
7840 BFD_RELOC_8_PCREL);
7841 fixP->fx_signed = 1;
7842 break;
7843 }
7844
7845 /* This changes the byte-displacement jump 0x7N
7846 to the (d)word-displacement jump 0x0f,0x8N. */
7847 opcode[1] = opcode[0] + 0x10;
7848 opcode[0] = TWO_BYTE_OPCODE_ESCAPE;
7849 /* We've added an opcode byte. */
7850 fragP->fr_fix += 1 + size;
7851 fix_new (fragP, old_fr_fix + 1, size,
7852 fragP->fr_symbol,
7853 fragP->fr_offset, 1,
7854 reloc_type);
7855 break;
7856
7857 default:
7858 BAD_CASE (fragP->fr_subtype);
7859 break;
7860 }
7861 frag_wane (fragP);
7862 return fragP->fr_fix - old_fr_fix;
7863 }
7864
7865 /* Guess size depending on current relax state. Initially the relax
7866 state will correspond to a short jump and we return 1, because
7867 the variable part of the frag (the branch offset) is one byte
7868 long. However, we can relax a section more than once and in that
7869 case we must either set fr_subtype back to the unrelaxed state,
7870 or return the value for the appropriate branch. */
7871 return md_relax_table[fragP->fr_subtype].rlx_length;
7872 }
7873
7874 /* Called after relax() is finished.
7875
7876 In: Address of frag.
7877 fr_type == rs_machine_dependent.
7878 fr_subtype is what the address relaxed to.
7879
7880 Out: Any fixSs and constants are set up.
7881 Caller will turn frag into a ".space 0". */
7882
7883 void
7884 md_convert_frag (bfd *abfd ATTRIBUTE_UNUSED, segT sec ATTRIBUTE_UNUSED,
7885 fragS *fragP)
7886 {
7887 unsigned char *opcode;
7888 unsigned char *where_to_put_displacement = NULL;
7889 offsetT target_address;
7890 offsetT opcode_address;
7891 unsigned int extension = 0;
7892 offsetT displacement_from_opcode_start;
7893
7894 opcode = (unsigned char *) fragP->fr_opcode;
7895
7896 /* Address we want to reach in file space. */
7897 target_address = S_GET_VALUE (fragP->fr_symbol) + fragP->fr_offset;
7898
7899 /* Address opcode resides at in file space. */
7900 opcode_address = fragP->fr_address + fragP->fr_fix;
7901
7902 /* Displacement from opcode start to fill into instruction. */
7903 displacement_from_opcode_start = target_address - opcode_address;
7904
7905 if ((fragP->fr_subtype & BIG) == 0)
7906 {
7907 /* Don't have to change opcode. */
7908 extension = 1; /* 1 opcode + 1 displacement */
7909 where_to_put_displacement = &opcode[1];
7910 }
7911 else
7912 {
7913 if (no_cond_jump_promotion
7914 && TYPE_FROM_RELAX_STATE (fragP->fr_subtype) != UNCOND_JUMP)
7915 as_warn_where (fragP->fr_file, fragP->fr_line,
7916 _("long jump required"));
7917
7918 switch (fragP->fr_subtype)
7919 {
7920 case ENCODE_RELAX_STATE (UNCOND_JUMP, BIG):
7921 extension = 4; /* 1 opcode + 4 displacement */
7922 opcode[0] = 0xe9;
7923 where_to_put_displacement = &opcode[1];
7924 break;
7925
7926 case ENCODE_RELAX_STATE (UNCOND_JUMP, BIG16):
7927 extension = 2; /* 1 opcode + 2 displacement */
7928 opcode[0] = 0xe9;
7929 where_to_put_displacement = &opcode[1];
7930 break;
7931
7932 case ENCODE_RELAX_STATE (COND_JUMP, BIG):
7933 case ENCODE_RELAX_STATE (COND_JUMP86, BIG):
7934 extension = 5; /* 2 opcode + 4 displacement */
7935 opcode[1] = opcode[0] + 0x10;
7936 opcode[0] = TWO_BYTE_OPCODE_ESCAPE;
7937 where_to_put_displacement = &opcode[2];
7938 break;
7939
7940 case ENCODE_RELAX_STATE (COND_JUMP, BIG16):
7941 extension = 3; /* 2 opcode + 2 displacement */
7942 opcode[1] = opcode[0] + 0x10;
7943 opcode[0] = TWO_BYTE_OPCODE_ESCAPE;
7944 where_to_put_displacement = &opcode[2];
7945 break;
7946
7947 case ENCODE_RELAX_STATE (COND_JUMP86, BIG16):
7948 extension = 4;
7949 opcode[0] ^= 1;
7950 opcode[1] = 3;
7951 opcode[2] = 0xe9;
7952 where_to_put_displacement = &opcode[3];
7953 break;
7954
7955 default:
7956 BAD_CASE (fragP->fr_subtype);
7957 break;
7958 }
7959 }
7960
7961 /* If size if less then four we are sure that the operand fits,
7962 but if it's 4, then it could be that the displacement is larger
7963 then -/+ 2GB. */
7964 if (DISP_SIZE_FROM_RELAX_STATE (fragP->fr_subtype) == 4
7965 && object_64bit
7966 && ((addressT) (displacement_from_opcode_start - extension
7967 + ((addressT) 1 << 31))
7968 > (((addressT) 2 << 31) - 1)))
7969 {
7970 as_bad_where (fragP->fr_file, fragP->fr_line,
7971 _("jump target out of range"));
7972 /* Make us emit 0. */
7973 displacement_from_opcode_start = extension;
7974 }
7975 /* Now put displacement after opcode. */
7976 md_number_to_chars ((char *) where_to_put_displacement,
7977 (valueT) (displacement_from_opcode_start - extension),
7978 DISP_SIZE_FROM_RELAX_STATE (fragP->fr_subtype));
7979 fragP->fr_fix += extension;
7980 }
7981 \f
7982 /* Apply a fixup (fixP) to segment data, once it has been determined
7983 by our caller that we have all the info we need to fix it up.
7984
7985 Parameter valP is the pointer to the value of the bits.
7986
7987 On the 386, immediates, displacements, and data pointers are all in
7988 the same (little-endian) format, so we don't need to care about which
7989 we are handling. */
7990
7991 void
7992 md_apply_fix (fixS *fixP, valueT *valP, segT seg ATTRIBUTE_UNUSED)
7993 {
7994 char *p = fixP->fx_where + fixP->fx_frag->fr_literal;
7995 valueT value = *valP;
7996
7997 #if !defined (TE_Mach)
7998 if (fixP->fx_pcrel)
7999 {
8000 switch (fixP->fx_r_type)
8001 {
8002 default:
8003 break;
8004
8005 case BFD_RELOC_64:
8006 fixP->fx_r_type = BFD_RELOC_64_PCREL;
8007 break;
8008 case BFD_RELOC_32:
8009 case BFD_RELOC_X86_64_32S:
8010 fixP->fx_r_type = BFD_RELOC_32_PCREL;
8011 break;
8012 case BFD_RELOC_16:
8013 fixP->fx_r_type = BFD_RELOC_16_PCREL;
8014 break;
8015 case BFD_RELOC_8:
8016 fixP->fx_r_type = BFD_RELOC_8_PCREL;
8017 break;
8018 }
8019 }
8020
8021 if (fixP->fx_addsy != NULL
8022 && (fixP->fx_r_type == BFD_RELOC_32_PCREL
8023 || fixP->fx_r_type == BFD_RELOC_64_PCREL
8024 || fixP->fx_r_type == BFD_RELOC_16_PCREL
8025 || fixP->fx_r_type == BFD_RELOC_8_PCREL)
8026 && !use_rela_relocations)
8027 {
8028 /* This is a hack. There should be a better way to handle this.
8029 This covers for the fact that bfd_install_relocation will
8030 subtract the current location (for partial_inplace, PC relative
8031 relocations); see more below. */
8032 #ifndef OBJ_AOUT
8033 if (IS_ELF
8034 #ifdef TE_PE
8035 || OUTPUT_FLAVOR == bfd_target_coff_flavour
8036 #endif
8037 )
8038 value += fixP->fx_where + fixP->fx_frag->fr_address;
8039 #endif
8040 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
8041 if (IS_ELF)
8042 {
8043 segT sym_seg = S_GET_SEGMENT (fixP->fx_addsy);
8044
8045 if ((sym_seg == seg
8046 || (symbol_section_p (fixP->fx_addsy)
8047 && sym_seg != absolute_section))
8048 && !generic_force_reloc (fixP))
8049 {
8050 /* Yes, we add the values in twice. This is because
8051 bfd_install_relocation subtracts them out again. I think
8052 bfd_install_relocation is broken, but I don't dare change
8053 it. FIXME. */
8054 value += fixP->fx_where + fixP->fx_frag->fr_address;
8055 }
8056 }
8057 #endif
8058 #if defined (OBJ_COFF) && defined (TE_PE)
8059 /* For some reason, the PE format does not store a
8060 section address offset for a PC relative symbol. */
8061 if (S_GET_SEGMENT (fixP->fx_addsy) != seg
8062 || S_IS_WEAK (fixP->fx_addsy))
8063 value += md_pcrel_from (fixP);
8064 #endif
8065 }
8066 #if defined (OBJ_COFF) && defined (TE_PE)
8067 if (fixP->fx_addsy != NULL && S_IS_WEAK (fixP->fx_addsy))
8068 {
8069 value -= S_GET_VALUE (fixP->fx_addsy);
8070 }
8071 #endif
8072
8073 /* Fix a few things - the dynamic linker expects certain values here,
8074 and we must not disappoint it. */
8075 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
8076 if (IS_ELF && fixP->fx_addsy)
8077 switch (fixP->fx_r_type)
8078 {
8079 case BFD_RELOC_386_PLT32:
8080 case BFD_RELOC_X86_64_PLT32:
8081 /* Make the jump instruction point to the address of the operand. At
8082 runtime we merely add the offset to the actual PLT entry. */
8083 value = -4;
8084 break;
8085
8086 case BFD_RELOC_386_TLS_GD:
8087 case BFD_RELOC_386_TLS_LDM:
8088 case BFD_RELOC_386_TLS_IE_32:
8089 case BFD_RELOC_386_TLS_IE:
8090 case BFD_RELOC_386_TLS_GOTIE:
8091 case BFD_RELOC_386_TLS_GOTDESC:
8092 case BFD_RELOC_X86_64_TLSGD:
8093 case BFD_RELOC_X86_64_TLSLD:
8094 case BFD_RELOC_X86_64_GOTTPOFF:
8095 case BFD_RELOC_X86_64_GOTPC32_TLSDESC:
8096 value = 0; /* Fully resolved at runtime. No addend. */
8097 /* Fallthrough */
8098 case BFD_RELOC_386_TLS_LE:
8099 case BFD_RELOC_386_TLS_LDO_32:
8100 case BFD_RELOC_386_TLS_LE_32:
8101 case BFD_RELOC_X86_64_DTPOFF32:
8102 case BFD_RELOC_X86_64_DTPOFF64:
8103 case BFD_RELOC_X86_64_TPOFF32:
8104 case BFD_RELOC_X86_64_TPOFF64:
8105 S_SET_THREAD_LOCAL (fixP->fx_addsy);
8106 break;
8107
8108 case BFD_RELOC_386_TLS_DESC_CALL:
8109 case BFD_RELOC_X86_64_TLSDESC_CALL:
8110 value = 0; /* Fully resolved at runtime. No addend. */
8111 S_SET_THREAD_LOCAL (fixP->fx_addsy);
8112 fixP->fx_done = 0;
8113 return;
8114
8115 case BFD_RELOC_386_GOT32:
8116 case BFD_RELOC_X86_64_GOT32:
8117 value = 0; /* Fully resolved at runtime. No addend. */
8118 break;
8119
8120 case BFD_RELOC_VTABLE_INHERIT:
8121 case BFD_RELOC_VTABLE_ENTRY:
8122 fixP->fx_done = 0;
8123 return;
8124
8125 default:
8126 break;
8127 }
8128 #endif /* defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) */
8129 *valP = value;
8130 #endif /* !defined (TE_Mach) */
8131
8132 /* Are we finished with this relocation now? */
8133 if (fixP->fx_addsy == NULL)
8134 fixP->fx_done = 1;
8135 #if defined (OBJ_COFF) && defined (TE_PE)
8136 else if (fixP->fx_addsy != NULL && S_IS_WEAK (fixP->fx_addsy))
8137 {
8138 fixP->fx_done = 0;
8139 /* Remember value for tc_gen_reloc. */
8140 fixP->fx_addnumber = value;
8141 /* Clear out the frag for now. */
8142 value = 0;
8143 }
8144 #endif
8145 else if (use_rela_relocations)
8146 {
8147 fixP->fx_no_overflow = 1;
8148 /* Remember value for tc_gen_reloc. */
8149 fixP->fx_addnumber = value;
8150 value = 0;
8151 }
8152
8153 md_number_to_chars (p, value, fixP->fx_size);
8154 }
8155 \f
8156 char *
8157 md_atof (int type, char *litP, int *sizeP)
8158 {
8159 /* This outputs the LITTLENUMs in REVERSE order;
8160 in accord with the bigendian 386. */
8161 return ieee_md_atof (type, litP, sizeP, FALSE);
8162 }
8163 \f
8164 static char output_invalid_buf[sizeof (unsigned char) * 2 + 6];
8165
8166 static char *
8167 output_invalid (int c)
8168 {
8169 if (ISPRINT (c))
8170 snprintf (output_invalid_buf, sizeof (output_invalid_buf),
8171 "'%c'", c);
8172 else
8173 snprintf (output_invalid_buf, sizeof (output_invalid_buf),
8174 "(0x%x)", (unsigned char) c);
8175 return output_invalid_buf;
8176 }
8177
8178 /* REG_STRING starts *before* REGISTER_PREFIX. */
8179
8180 static const reg_entry *
8181 parse_real_register (char *reg_string, char **end_op)
8182 {
8183 char *s = reg_string;
8184 char *p;
8185 char reg_name_given[MAX_REG_NAME_SIZE + 1];
8186 const reg_entry *r;
8187
8188 /* Skip possible REGISTER_PREFIX and possible whitespace. */
8189 if (*s == REGISTER_PREFIX)
8190 ++s;
8191
8192 if (is_space_char (*s))
8193 ++s;
8194
8195 p = reg_name_given;
8196 while ((*p++ = register_chars[(unsigned char) *s]) != '\0')
8197 {
8198 if (p >= reg_name_given + MAX_REG_NAME_SIZE)
8199 return (const reg_entry *) NULL;
8200 s++;
8201 }
8202
8203 /* For naked regs, make sure that we are not dealing with an identifier.
8204 This prevents confusing an identifier like `eax_var' with register
8205 `eax'. */
8206 if (allow_naked_reg && identifier_chars[(unsigned char) *s])
8207 return (const reg_entry *) NULL;
8208
8209 *end_op = s;
8210
8211 r = (const reg_entry *) hash_find (reg_hash, reg_name_given);
8212
8213 /* Handle floating point regs, allowing spaces in the (i) part. */
8214 if (r == i386_regtab /* %st is first entry of table */)
8215 {
8216 if (is_space_char (*s))
8217 ++s;
8218 if (*s == '(')
8219 {
8220 ++s;
8221 if (is_space_char (*s))
8222 ++s;
8223 if (*s >= '0' && *s <= '7')
8224 {
8225 int fpr = *s - '0';
8226 ++s;
8227 if (is_space_char (*s))
8228 ++s;
8229 if (*s == ')')
8230 {
8231 *end_op = s + 1;
8232 r = (const reg_entry *) hash_find (reg_hash, "st(0)");
8233 know (r);
8234 return r + fpr;
8235 }
8236 }
8237 /* We have "%st(" then garbage. */
8238 return (const reg_entry *) NULL;
8239 }
8240 }
8241
8242 if (r == NULL || allow_pseudo_reg)
8243 return r;
8244
8245 if (operand_type_all_zero (&r->reg_type))
8246 return (const reg_entry *) NULL;
8247
8248 if ((r->reg_type.bitfield.reg32
8249 || r->reg_type.bitfield.sreg3
8250 || r->reg_type.bitfield.control
8251 || r->reg_type.bitfield.debug
8252 || r->reg_type.bitfield.test)
8253 && !cpu_arch_flags.bitfield.cpui386)
8254 return (const reg_entry *) NULL;
8255
8256 if (r->reg_type.bitfield.floatreg
8257 && !cpu_arch_flags.bitfield.cpu8087
8258 && !cpu_arch_flags.bitfield.cpu287
8259 && !cpu_arch_flags.bitfield.cpu387)
8260 return (const reg_entry *) NULL;
8261
8262 if (r->reg_type.bitfield.regmmx && !cpu_arch_flags.bitfield.cpummx)
8263 return (const reg_entry *) NULL;
8264
8265 if (r->reg_type.bitfield.regxmm && !cpu_arch_flags.bitfield.cpusse)
8266 return (const reg_entry *) NULL;
8267
8268 if (r->reg_type.bitfield.regymm && !cpu_arch_flags.bitfield.cpuavx)
8269 return (const reg_entry *) NULL;
8270
8271 /* Don't allow fake index register unless allow_index_reg isn't 0. */
8272 if (!allow_index_reg
8273 && (r->reg_num == RegEiz || r->reg_num == RegRiz))
8274 return (const reg_entry *) NULL;
8275
8276 if (((r->reg_flags & (RegRex64 | RegRex))
8277 || r->reg_type.bitfield.reg64)
8278 && (!cpu_arch_flags.bitfield.cpulm
8279 || !operand_type_equal (&r->reg_type, &control))
8280 && flag_code != CODE_64BIT)
8281 return (const reg_entry *) NULL;
8282
8283 if (r->reg_type.bitfield.sreg3 && r->reg_num == RegFlat && !intel_syntax)
8284 return (const reg_entry *) NULL;
8285
8286 return r;
8287 }
8288
8289 /* REG_STRING starts *before* REGISTER_PREFIX. */
8290
8291 static const reg_entry *
8292 parse_register (char *reg_string, char **end_op)
8293 {
8294 const reg_entry *r;
8295
8296 if (*reg_string == REGISTER_PREFIX || allow_naked_reg)
8297 r = parse_real_register (reg_string, end_op);
8298 else
8299 r = NULL;
8300 if (!r)
8301 {
8302 char *save = input_line_pointer;
8303 char c;
8304 symbolS *symbolP;
8305
8306 input_line_pointer = reg_string;
8307 c = get_symbol_end ();
8308 symbolP = symbol_find (reg_string);
8309 if (symbolP && S_GET_SEGMENT (symbolP) == reg_section)
8310 {
8311 const expressionS *e = symbol_get_value_expression (symbolP);
8312
8313 know (e->X_op == O_register);
8314 know (e->X_add_number >= 0
8315 && (valueT) e->X_add_number < i386_regtab_size);
8316 r = i386_regtab + e->X_add_number;
8317 *end_op = input_line_pointer;
8318 }
8319 *input_line_pointer = c;
8320 input_line_pointer = save;
8321 }
8322 return r;
8323 }
8324
8325 int
8326 i386_parse_name (char *name, expressionS *e, char *nextcharP)
8327 {
8328 const reg_entry *r;
8329 char *end = input_line_pointer;
8330
8331 *end = *nextcharP;
8332 r = parse_register (name, &input_line_pointer);
8333 if (r && end <= input_line_pointer)
8334 {
8335 *nextcharP = *input_line_pointer;
8336 *input_line_pointer = 0;
8337 e->X_op = O_register;
8338 e->X_add_number = r - i386_regtab;
8339 return 1;
8340 }
8341 input_line_pointer = end;
8342 *end = 0;
8343 return intel_syntax ? i386_intel_parse_name (name, e) : 0;
8344 }
8345
8346 void
8347 md_operand (expressionS *e)
8348 {
8349 char *end;
8350 const reg_entry *r;
8351
8352 switch (*input_line_pointer)
8353 {
8354 case REGISTER_PREFIX:
8355 r = parse_real_register (input_line_pointer, &end);
8356 if (r)
8357 {
8358 e->X_op = O_register;
8359 e->X_add_number = r - i386_regtab;
8360 input_line_pointer = end;
8361 }
8362 break;
8363
8364 case '[':
8365 gas_assert (intel_syntax);
8366 end = input_line_pointer++;
8367 expression (e);
8368 if (*input_line_pointer == ']')
8369 {
8370 ++input_line_pointer;
8371 e->X_op_symbol = make_expr_symbol (e);
8372 e->X_add_symbol = NULL;
8373 e->X_add_number = 0;
8374 e->X_op = O_index;
8375 }
8376 else
8377 {
8378 e->X_op = O_absent;
8379 input_line_pointer = end;
8380 }
8381 break;
8382 }
8383 }
8384
8385 \f
8386 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
8387 const char *md_shortopts = "kVQ:sqn";
8388 #else
8389 const char *md_shortopts = "qn";
8390 #endif
8391
8392 #define OPTION_32 (OPTION_MD_BASE + 0)
8393 #define OPTION_64 (OPTION_MD_BASE + 1)
8394 #define OPTION_DIVIDE (OPTION_MD_BASE + 2)
8395 #define OPTION_MARCH (OPTION_MD_BASE + 3)
8396 #define OPTION_MTUNE (OPTION_MD_BASE + 4)
8397 #define OPTION_MMNEMONIC (OPTION_MD_BASE + 5)
8398 #define OPTION_MSYNTAX (OPTION_MD_BASE + 6)
8399 #define OPTION_MINDEX_REG (OPTION_MD_BASE + 7)
8400 #define OPTION_MNAKED_REG (OPTION_MD_BASE + 8)
8401 #define OPTION_MOLD_GCC (OPTION_MD_BASE + 9)
8402 #define OPTION_MSSE2AVX (OPTION_MD_BASE + 10)
8403 #define OPTION_MSSE_CHECK (OPTION_MD_BASE + 11)
8404 #define OPTION_MAVXSCALAR (OPTION_MD_BASE + 12)
8405 #define OPTION_X32 (OPTION_MD_BASE + 13)
8406
8407 struct option md_longopts[] =
8408 {
8409 {"32", no_argument, NULL, OPTION_32},
8410 #if (defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) \
8411 || defined (TE_PE) || defined (TE_PEP) || defined (OBJ_MACH_O))
8412 {"64", no_argument, NULL, OPTION_64},
8413 #endif
8414 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
8415 {"x32", no_argument, NULL, OPTION_X32},
8416 #endif
8417 {"divide", no_argument, NULL, OPTION_DIVIDE},
8418 {"march", required_argument, NULL, OPTION_MARCH},
8419 {"mtune", required_argument, NULL, OPTION_MTUNE},
8420 {"mmnemonic", required_argument, NULL, OPTION_MMNEMONIC},
8421 {"msyntax", required_argument, NULL, OPTION_MSYNTAX},
8422 {"mindex-reg", no_argument, NULL, OPTION_MINDEX_REG},
8423 {"mnaked-reg", no_argument, NULL, OPTION_MNAKED_REG},
8424 {"mold-gcc", no_argument, NULL, OPTION_MOLD_GCC},
8425 {"msse2avx", no_argument, NULL, OPTION_MSSE2AVX},
8426 {"msse-check", required_argument, NULL, OPTION_MSSE_CHECK},
8427 {"mavxscalar", required_argument, NULL, OPTION_MAVXSCALAR},
8428 {NULL, no_argument, NULL, 0}
8429 };
8430 size_t md_longopts_size = sizeof (md_longopts);
8431
8432 int
8433 md_parse_option (int c, char *arg)
8434 {
8435 unsigned int j;
8436 char *arch, *next;
8437
8438 switch (c)
8439 {
8440 case 'n':
8441 optimize_align_code = 0;
8442 break;
8443
8444 case 'q':
8445 quiet_warnings = 1;
8446 break;
8447
8448 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
8449 /* -Qy, -Qn: SVR4 arguments controlling whether a .comment section
8450 should be emitted or not. FIXME: Not implemented. */
8451 case 'Q':
8452 break;
8453
8454 /* -V: SVR4 argument to print version ID. */
8455 case 'V':
8456 print_version_id ();
8457 break;
8458
8459 /* -k: Ignore for FreeBSD compatibility. */
8460 case 'k':
8461 break;
8462
8463 case 's':
8464 /* -s: On i386 Solaris, this tells the native assembler to use
8465 .stab instead of .stab.excl. We always use .stab anyhow. */
8466 break;
8467 #endif
8468 #if (defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) \
8469 || defined (TE_PE) || defined (TE_PEP) || defined (OBJ_MACH_O))
8470 case OPTION_64:
8471 {
8472 const char **list, **l;
8473
8474 list = bfd_target_list ();
8475 for (l = list; *l != NULL; l++)
8476 if (CONST_STRNEQ (*l, "elf64-x86-64")
8477 || strcmp (*l, "coff-x86-64") == 0
8478 || strcmp (*l, "pe-x86-64") == 0
8479 || strcmp (*l, "pei-x86-64") == 0
8480 || strcmp (*l, "mach-o-x86-64") == 0)
8481 {
8482 default_arch = "x86_64";
8483 break;
8484 }
8485 if (*l == NULL)
8486 as_fatal (_("no compiled in support for x86_64"));
8487 free (list);
8488 }
8489 break;
8490 #endif
8491
8492 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
8493 case OPTION_X32:
8494 if (IS_ELF)
8495 {
8496 const char **list, **l;
8497
8498 list = bfd_target_list ();
8499 for (l = list; *l != NULL; l++)
8500 if (CONST_STRNEQ (*l, "elf32-x86-64"))
8501 {
8502 default_arch = "x86_64:32";
8503 break;
8504 }
8505 if (*l == NULL)
8506 as_fatal (_("no compiled in support for 32bit x86_64"));
8507 free (list);
8508 }
8509 else
8510 as_fatal (_("32bit x86_64 is only supported for ELF"));
8511 break;
8512 #endif
8513
8514 case OPTION_32:
8515 default_arch = "i386";
8516 break;
8517
8518 case OPTION_DIVIDE:
8519 #ifdef SVR4_COMMENT_CHARS
8520 {
8521 char *n, *t;
8522 const char *s;
8523
8524 n = (char *) xmalloc (strlen (i386_comment_chars) + 1);
8525 t = n;
8526 for (s = i386_comment_chars; *s != '\0'; s++)
8527 if (*s != '/')
8528 *t++ = *s;
8529 *t = '\0';
8530 i386_comment_chars = n;
8531 }
8532 #endif
8533 break;
8534
8535 case OPTION_MARCH:
8536 arch = xstrdup (arg);
8537 do
8538 {
8539 if (*arch == '.')
8540 as_fatal (_("invalid -march= option: `%s'"), arg);
8541 next = strchr (arch, '+');
8542 if (next)
8543 *next++ = '\0';
8544 for (j = 0; j < ARRAY_SIZE (cpu_arch); j++)
8545 {
8546 if (strcmp (arch, cpu_arch [j].name) == 0)
8547 {
8548 /* Processor. */
8549 if (! cpu_arch[j].flags.bitfield.cpui386)
8550 continue;
8551
8552 cpu_arch_name = cpu_arch[j].name;
8553 cpu_sub_arch_name = NULL;
8554 cpu_arch_flags = cpu_arch[j].flags;
8555 cpu_arch_isa = cpu_arch[j].type;
8556 cpu_arch_isa_flags = cpu_arch[j].flags;
8557 if (!cpu_arch_tune_set)
8558 {
8559 cpu_arch_tune = cpu_arch_isa;
8560 cpu_arch_tune_flags = cpu_arch_isa_flags;
8561 }
8562 break;
8563 }
8564 else if (*cpu_arch [j].name == '.'
8565 && strcmp (arch, cpu_arch [j].name + 1) == 0)
8566 {
8567 /* ISA entension. */
8568 i386_cpu_flags flags;
8569
8570 if (!cpu_arch[j].negated)
8571 flags = cpu_flags_or (cpu_arch_flags,
8572 cpu_arch[j].flags);
8573 else
8574 flags = cpu_flags_and_not (cpu_arch_flags,
8575 cpu_arch[j].flags);
8576 if (!cpu_flags_equal (&flags, &cpu_arch_flags))
8577 {
8578 if (cpu_sub_arch_name)
8579 {
8580 char *name = cpu_sub_arch_name;
8581 cpu_sub_arch_name = concat (name,
8582 cpu_arch[j].name,
8583 (const char *) NULL);
8584 free (name);
8585 }
8586 else
8587 cpu_sub_arch_name = xstrdup (cpu_arch[j].name);
8588 cpu_arch_flags = flags;
8589 cpu_arch_isa_flags = flags;
8590 }
8591 break;
8592 }
8593 }
8594
8595 if (j >= ARRAY_SIZE (cpu_arch))
8596 as_fatal (_("invalid -march= option: `%s'"), arg);
8597
8598 arch = next;
8599 }
8600 while (next != NULL );
8601 break;
8602
8603 case OPTION_MTUNE:
8604 if (*arg == '.')
8605 as_fatal (_("invalid -mtune= option: `%s'"), arg);
8606 for (j = 0; j < ARRAY_SIZE (cpu_arch); j++)
8607 {
8608 if (strcmp (arg, cpu_arch [j].name) == 0)
8609 {
8610 cpu_arch_tune_set = 1;
8611 cpu_arch_tune = cpu_arch [j].type;
8612 cpu_arch_tune_flags = cpu_arch[j].flags;
8613 break;
8614 }
8615 }
8616 if (j >= ARRAY_SIZE (cpu_arch))
8617 as_fatal (_("invalid -mtune= option: `%s'"), arg);
8618 break;
8619
8620 case OPTION_MMNEMONIC:
8621 if (strcasecmp (arg, "att") == 0)
8622 intel_mnemonic = 0;
8623 else if (strcasecmp (arg, "intel") == 0)
8624 intel_mnemonic = 1;
8625 else
8626 as_fatal (_("invalid -mmnemonic= option: `%s'"), arg);
8627 break;
8628
8629 case OPTION_MSYNTAX:
8630 if (strcasecmp (arg, "att") == 0)
8631 intel_syntax = 0;
8632 else if (strcasecmp (arg, "intel") == 0)
8633 intel_syntax = 1;
8634 else
8635 as_fatal (_("invalid -msyntax= option: `%s'"), arg);
8636 break;
8637
8638 case OPTION_MINDEX_REG:
8639 allow_index_reg = 1;
8640 break;
8641
8642 case OPTION_MNAKED_REG:
8643 allow_naked_reg = 1;
8644 break;
8645
8646 case OPTION_MOLD_GCC:
8647 old_gcc = 1;
8648 break;
8649
8650 case OPTION_MSSE2AVX:
8651 sse2avx = 1;
8652 break;
8653
8654 case OPTION_MSSE_CHECK:
8655 if (strcasecmp (arg, "error") == 0)
8656 sse_check = sse_check_error;
8657 else if (strcasecmp (arg, "warning") == 0)
8658 sse_check = sse_check_warning;
8659 else if (strcasecmp (arg, "none") == 0)
8660 sse_check = sse_check_none;
8661 else
8662 as_fatal (_("invalid -msse-check= option: `%s'"), arg);
8663 break;
8664
8665 case OPTION_MAVXSCALAR:
8666 if (strcasecmp (arg, "128") == 0)
8667 avxscalar = vex128;
8668 else if (strcasecmp (arg, "256") == 0)
8669 avxscalar = vex256;
8670 else
8671 as_fatal (_("invalid -mavxscalar= option: `%s'"), arg);
8672 break;
8673
8674 default:
8675 return 0;
8676 }
8677 return 1;
8678 }
8679
8680 #define MESSAGE_TEMPLATE \
8681 " "
8682
8683 static void
8684 show_arch (FILE *stream, int ext, int check)
8685 {
8686 static char message[] = MESSAGE_TEMPLATE;
8687 char *start = message + 27;
8688 char *p;
8689 int size = sizeof (MESSAGE_TEMPLATE);
8690 int left;
8691 const char *name;
8692 int len;
8693 unsigned int j;
8694
8695 p = start;
8696 left = size - (start - message);
8697 for (j = 0; j < ARRAY_SIZE (cpu_arch); j++)
8698 {
8699 /* Should it be skipped? */
8700 if (cpu_arch [j].skip)
8701 continue;
8702
8703 name = cpu_arch [j].name;
8704 len = cpu_arch [j].len;
8705 if (*name == '.')
8706 {
8707 /* It is an extension. Skip if we aren't asked to show it. */
8708 if (ext)
8709 {
8710 name++;
8711 len--;
8712 }
8713 else
8714 continue;
8715 }
8716 else if (ext)
8717 {
8718 /* It is an processor. Skip if we show only extension. */
8719 continue;
8720 }
8721 else if (check && ! cpu_arch[j].flags.bitfield.cpui386)
8722 {
8723 /* It is an impossible processor - skip. */
8724 continue;
8725 }
8726
8727 /* Reserve 2 spaces for ", " or ",\0" */
8728 left -= len + 2;
8729
8730 /* Check if there is any room. */
8731 if (left >= 0)
8732 {
8733 if (p != start)
8734 {
8735 *p++ = ',';
8736 *p++ = ' ';
8737 }
8738 p = mempcpy (p, name, len);
8739 }
8740 else
8741 {
8742 /* Output the current message now and start a new one. */
8743 *p++ = ',';
8744 *p = '\0';
8745 fprintf (stream, "%s\n", message);
8746 p = start;
8747 left = size - (start - message) - len - 2;
8748
8749 gas_assert (left >= 0);
8750
8751 p = mempcpy (p, name, len);
8752 }
8753 }
8754
8755 *p = '\0';
8756 fprintf (stream, "%s\n", message);
8757 }
8758
8759 void
8760 md_show_usage (FILE *stream)
8761 {
8762 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
8763 fprintf (stream, _("\
8764 -Q ignored\n\
8765 -V print assembler version number\n\
8766 -k ignored\n"));
8767 #endif
8768 fprintf (stream, _("\
8769 -n Do not optimize code alignment\n\
8770 -q quieten some warnings\n"));
8771 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
8772 fprintf (stream, _("\
8773 -s ignored\n"));
8774 #endif
8775 #if (defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) \
8776 || defined (TE_PE) || defined (TE_PEP))
8777 fprintf (stream, _("\
8778 --32/--64/--x32 generate 32bit/64bit/x32 code\n"));
8779 #endif
8780 #ifdef SVR4_COMMENT_CHARS
8781 fprintf (stream, _("\
8782 --divide do not treat `/' as a comment character\n"));
8783 #else
8784 fprintf (stream, _("\
8785 --divide ignored\n"));
8786 #endif
8787 fprintf (stream, _("\
8788 -march=CPU[,+EXTENSION...]\n\
8789 generate code for CPU and EXTENSION, CPU is one of:\n"));
8790 show_arch (stream, 0, 1);
8791 fprintf (stream, _("\
8792 EXTENSION is combination of:\n"));
8793 show_arch (stream, 1, 0);
8794 fprintf (stream, _("\
8795 -mtune=CPU optimize for CPU, CPU is one of:\n"));
8796 show_arch (stream, 0, 0);
8797 fprintf (stream, _("\
8798 -msse2avx encode SSE instructions with VEX prefix\n"));
8799 fprintf (stream, _("\
8800 -msse-check=[none|error|warning]\n\
8801 check SSE instructions\n"));
8802 fprintf (stream, _("\
8803 -mavxscalar=[128|256] encode scalar AVX instructions with specific vector\n\
8804 length\n"));
8805 fprintf (stream, _("\
8806 -mmnemonic=[att|intel] use AT&T/Intel mnemonic\n"));
8807 fprintf (stream, _("\
8808 -msyntax=[att|intel] use AT&T/Intel syntax\n"));
8809 fprintf (stream, _("\
8810 -mindex-reg support pseudo index registers\n"));
8811 fprintf (stream, _("\
8812 -mnaked-reg don't require `%%' prefix for registers\n"));
8813 fprintf (stream, _("\
8814 -mold-gcc support old (<= 2.8.1) versions of gcc\n"));
8815 }
8816
8817 #if ((defined (OBJ_MAYBE_COFF) && defined (OBJ_MAYBE_AOUT)) \
8818 || defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) \
8819 || defined (TE_PE) || defined (TE_PEP) || defined (OBJ_MACH_O))
8820
8821 /* Pick the target format to use. */
8822
8823 const char *
8824 i386_target_format (void)
8825 {
8826 if (!strncmp (default_arch, "x86_64", 6))
8827 {
8828 update_code_flag (CODE_64BIT, 1);
8829 if (default_arch[6] == '\0')
8830 x86_elf_abi = X86_64_ABI;
8831 else
8832 x86_elf_abi = X86_64_X32_ABI;
8833 }
8834 else if (!strcmp (default_arch, "i386"))
8835 update_code_flag (CODE_32BIT, 1);
8836 else
8837 as_fatal (_("unknown architecture"));
8838
8839 if (cpu_flags_all_zero (&cpu_arch_isa_flags))
8840 cpu_arch_isa_flags = cpu_arch[flag_code == CODE_64BIT].flags;
8841 if (cpu_flags_all_zero (&cpu_arch_tune_flags))
8842 cpu_arch_tune_flags = cpu_arch[flag_code == CODE_64BIT].flags;
8843
8844 switch (OUTPUT_FLAVOR)
8845 {
8846 #if defined (OBJ_MAYBE_AOUT) || defined (OBJ_AOUT)
8847 case bfd_target_aout_flavour:
8848 return AOUT_TARGET_FORMAT;
8849 #endif
8850 #if defined (OBJ_MAYBE_COFF) || defined (OBJ_COFF)
8851 # if defined (TE_PE) || defined (TE_PEP)
8852 case bfd_target_coff_flavour:
8853 return flag_code == CODE_64BIT ? "pe-x86-64" : "pe-i386";
8854 # elif defined (TE_GO32)
8855 case bfd_target_coff_flavour:
8856 return "coff-go32";
8857 # else
8858 case bfd_target_coff_flavour:
8859 return "coff-i386";
8860 # endif
8861 #endif
8862 #if defined (OBJ_MAYBE_ELF) || defined (OBJ_ELF)
8863 case bfd_target_elf_flavour:
8864 {
8865 const char *format;
8866
8867 switch (x86_elf_abi)
8868 {
8869 default:
8870 format = ELF_TARGET_FORMAT;
8871 break;
8872 case X86_64_ABI:
8873 use_rela_relocations = 1;
8874 object_64bit = 1;
8875 format = ELF_TARGET_FORMAT64;
8876 break;
8877 case X86_64_X32_ABI:
8878 use_rela_relocations = 1;
8879 object_64bit = 1;
8880 disallow_64bit_reloc = 1;
8881 format = ELF_TARGET_FORMAT32;
8882 break;
8883 }
8884 if (cpu_arch_isa == PROCESSOR_L1OM)
8885 {
8886 if (x86_elf_abi != X86_64_ABI)
8887 as_fatal (_("Intel L1OM is 64bit only"));
8888 return ELF_TARGET_L1OM_FORMAT;
8889 }
8890 if (cpu_arch_isa == PROCESSOR_K1OM)
8891 {
8892 if (x86_elf_abi != X86_64_ABI)
8893 as_fatal (_("Intel K1OM is 64bit only"));
8894 return ELF_TARGET_K1OM_FORMAT;
8895 }
8896 else
8897 return format;
8898 }
8899 #endif
8900 #if defined (OBJ_MACH_O)
8901 case bfd_target_mach_o_flavour:
8902 if (flag_code == CODE_64BIT)
8903 {
8904 use_rela_relocations = 1;
8905 object_64bit = 1;
8906 return "mach-o-x86-64";
8907 }
8908 else
8909 return "mach-o-i386";
8910 #endif
8911 default:
8912 abort ();
8913 return NULL;
8914 }
8915 }
8916
8917 #endif /* OBJ_MAYBE_ more than one */
8918
8919 #if (defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF))
8920 void
8921 i386_elf_emit_arch_note (void)
8922 {
8923 if (IS_ELF && cpu_arch_name != NULL)
8924 {
8925 char *p;
8926 asection *seg = now_seg;
8927 subsegT subseg = now_subseg;
8928 Elf_Internal_Note i_note;
8929 Elf_External_Note e_note;
8930 asection *note_secp;
8931 int len;
8932
8933 /* Create the .note section. */
8934 note_secp = subseg_new (".note", 0);
8935 bfd_set_section_flags (stdoutput,
8936 note_secp,
8937 SEC_HAS_CONTENTS | SEC_READONLY);
8938
8939 /* Process the arch string. */
8940 len = strlen (cpu_arch_name);
8941
8942 i_note.namesz = len + 1;
8943 i_note.descsz = 0;
8944 i_note.type = NT_ARCH;
8945 p = frag_more (sizeof (e_note.namesz));
8946 md_number_to_chars (p, (valueT) i_note.namesz, sizeof (e_note.namesz));
8947 p = frag_more (sizeof (e_note.descsz));
8948 md_number_to_chars (p, (valueT) i_note.descsz, sizeof (e_note.descsz));
8949 p = frag_more (sizeof (e_note.type));
8950 md_number_to_chars (p, (valueT) i_note.type, sizeof (e_note.type));
8951 p = frag_more (len + 1);
8952 strcpy (p, cpu_arch_name);
8953
8954 frag_align (2, 0, 0);
8955
8956 subseg_set (seg, subseg);
8957 }
8958 }
8959 #endif
8960 \f
8961 symbolS *
8962 md_undefined_symbol (char *name)
8963 {
8964 if (name[0] == GLOBAL_OFFSET_TABLE_NAME[0]
8965 && name[1] == GLOBAL_OFFSET_TABLE_NAME[1]
8966 && name[2] == GLOBAL_OFFSET_TABLE_NAME[2]
8967 && strcmp (name, GLOBAL_OFFSET_TABLE_NAME) == 0)
8968 {
8969 if (!GOT_symbol)
8970 {
8971 if (symbol_find (name))
8972 as_bad (_("GOT already in symbol table"));
8973 GOT_symbol = symbol_new (name, undefined_section,
8974 (valueT) 0, &zero_address_frag);
8975 };
8976 return GOT_symbol;
8977 }
8978 return 0;
8979 }
8980
8981 /* Round up a section size to the appropriate boundary. */
8982
8983 valueT
8984 md_section_align (segT segment ATTRIBUTE_UNUSED, valueT size)
8985 {
8986 #if (defined (OBJ_AOUT) || defined (OBJ_MAYBE_AOUT))
8987 if (OUTPUT_FLAVOR == bfd_target_aout_flavour)
8988 {
8989 /* For a.out, force the section size to be aligned. If we don't do
8990 this, BFD will align it for us, but it will not write out the
8991 final bytes of the section. This may be a bug in BFD, but it is
8992 easier to fix it here since that is how the other a.out targets
8993 work. */
8994 int align;
8995
8996 align = bfd_get_section_alignment (stdoutput, segment);
8997 size = ((size + (1 << align) - 1) & ((valueT) -1 << align));
8998 }
8999 #endif
9000
9001 return size;
9002 }
9003
9004 /* On the i386, PC-relative offsets are relative to the start of the
9005 next instruction. That is, the address of the offset, plus its
9006 size, since the offset is always the last part of the insn. */
9007
9008 long
9009 md_pcrel_from (fixS *fixP)
9010 {
9011 return fixP->fx_size + fixP->fx_where + fixP->fx_frag->fr_address;
9012 }
9013
9014 #ifndef I386COFF
9015
9016 static void
9017 s_bss (int ignore ATTRIBUTE_UNUSED)
9018 {
9019 int temp;
9020
9021 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
9022 if (IS_ELF)
9023 obj_elf_section_change_hook ();
9024 #endif
9025 temp = get_absolute_expression ();
9026 subseg_set (bss_section, (subsegT) temp);
9027 demand_empty_rest_of_line ();
9028 }
9029
9030 #endif
9031
9032 void
9033 i386_validate_fix (fixS *fixp)
9034 {
9035 if (fixp->fx_subsy && fixp->fx_subsy == GOT_symbol)
9036 {
9037 if (fixp->fx_r_type == BFD_RELOC_32_PCREL)
9038 {
9039 if (!object_64bit)
9040 abort ();
9041 fixp->fx_r_type = BFD_RELOC_X86_64_GOTPCREL;
9042 }
9043 else
9044 {
9045 if (!object_64bit)
9046 fixp->fx_r_type = BFD_RELOC_386_GOTOFF;
9047 else
9048 fixp->fx_r_type = BFD_RELOC_X86_64_GOTOFF64;
9049 }
9050 fixp->fx_subsy = 0;
9051 }
9052 }
9053
9054 arelent *
9055 tc_gen_reloc (asection *section ATTRIBUTE_UNUSED, fixS *fixp)
9056 {
9057 arelent *rel;
9058 bfd_reloc_code_real_type code;
9059
9060 switch (fixp->fx_r_type)
9061 {
9062 case BFD_RELOC_X86_64_PLT32:
9063 case BFD_RELOC_X86_64_GOT32:
9064 case BFD_RELOC_X86_64_GOTPCREL:
9065 case BFD_RELOC_386_PLT32:
9066 case BFD_RELOC_386_GOT32:
9067 case BFD_RELOC_386_GOTOFF:
9068 case BFD_RELOC_386_GOTPC:
9069 case BFD_RELOC_386_TLS_GD:
9070 case BFD_RELOC_386_TLS_LDM:
9071 case BFD_RELOC_386_TLS_LDO_32:
9072 case BFD_RELOC_386_TLS_IE_32:
9073 case BFD_RELOC_386_TLS_IE:
9074 case BFD_RELOC_386_TLS_GOTIE:
9075 case BFD_RELOC_386_TLS_LE_32:
9076 case BFD_RELOC_386_TLS_LE:
9077 case BFD_RELOC_386_TLS_GOTDESC:
9078 case BFD_RELOC_386_TLS_DESC_CALL:
9079 case BFD_RELOC_X86_64_TLSGD:
9080 case BFD_RELOC_X86_64_TLSLD:
9081 case BFD_RELOC_X86_64_DTPOFF32:
9082 case BFD_RELOC_X86_64_DTPOFF64:
9083 case BFD_RELOC_X86_64_GOTTPOFF:
9084 case BFD_RELOC_X86_64_TPOFF32:
9085 case BFD_RELOC_X86_64_TPOFF64:
9086 case BFD_RELOC_X86_64_GOTOFF64:
9087 case BFD_RELOC_X86_64_GOTPC32:
9088 case BFD_RELOC_X86_64_GOT64:
9089 case BFD_RELOC_X86_64_GOTPCREL64:
9090 case BFD_RELOC_X86_64_GOTPC64:
9091 case BFD_RELOC_X86_64_GOTPLT64:
9092 case BFD_RELOC_X86_64_PLTOFF64:
9093 case BFD_RELOC_X86_64_GOTPC32_TLSDESC:
9094 case BFD_RELOC_X86_64_TLSDESC_CALL:
9095 case BFD_RELOC_RVA:
9096 case BFD_RELOC_VTABLE_ENTRY:
9097 case BFD_RELOC_VTABLE_INHERIT:
9098 #ifdef TE_PE
9099 case BFD_RELOC_32_SECREL:
9100 #endif
9101 code = fixp->fx_r_type;
9102 break;
9103 case BFD_RELOC_X86_64_32S:
9104 if (!fixp->fx_pcrel)
9105 {
9106 /* Don't turn BFD_RELOC_X86_64_32S into BFD_RELOC_32. */
9107 code = fixp->fx_r_type;
9108 break;
9109 }
9110 default:
9111 if (fixp->fx_pcrel)
9112 {
9113 switch (fixp->fx_size)
9114 {
9115 default:
9116 as_bad_where (fixp->fx_file, fixp->fx_line,
9117 _("can not do %d byte pc-relative relocation"),
9118 fixp->fx_size);
9119 code = BFD_RELOC_32_PCREL;
9120 break;
9121 case 1: code = BFD_RELOC_8_PCREL; break;
9122 case 2: code = BFD_RELOC_16_PCREL; break;
9123 case 4: code = BFD_RELOC_32_PCREL; break;
9124 #ifdef BFD64
9125 case 8: code = BFD_RELOC_64_PCREL; break;
9126 #endif
9127 }
9128 }
9129 else
9130 {
9131 switch (fixp->fx_size)
9132 {
9133 default:
9134 as_bad_where (fixp->fx_file, fixp->fx_line,
9135 _("can not do %d byte relocation"),
9136 fixp->fx_size);
9137 code = BFD_RELOC_32;
9138 break;
9139 case 1: code = BFD_RELOC_8; break;
9140 case 2: code = BFD_RELOC_16; break;
9141 case 4: code = BFD_RELOC_32; break;
9142 #ifdef BFD64
9143 case 8: code = BFD_RELOC_64; break;
9144 #endif
9145 }
9146 }
9147 break;
9148 }
9149
9150 if ((code == BFD_RELOC_32
9151 || code == BFD_RELOC_32_PCREL
9152 || code == BFD_RELOC_X86_64_32S)
9153 && GOT_symbol
9154 && fixp->fx_addsy == GOT_symbol)
9155 {
9156 if (!object_64bit)
9157 code = BFD_RELOC_386_GOTPC;
9158 else
9159 code = BFD_RELOC_X86_64_GOTPC32;
9160 }
9161 if ((code == BFD_RELOC_64 || code == BFD_RELOC_64_PCREL)
9162 && GOT_symbol
9163 && fixp->fx_addsy == GOT_symbol)
9164 {
9165 code = BFD_RELOC_X86_64_GOTPC64;
9166 }
9167
9168 rel = (arelent *) xmalloc (sizeof (arelent));
9169 rel->sym_ptr_ptr = (asymbol **) xmalloc (sizeof (asymbol *));
9170 *rel->sym_ptr_ptr = symbol_get_bfdsym (fixp->fx_addsy);
9171
9172 rel->address = fixp->fx_frag->fr_address + fixp->fx_where;
9173
9174 if (!use_rela_relocations)
9175 {
9176 /* HACK: Since i386 ELF uses Rel instead of Rela, encode the
9177 vtable entry to be used in the relocation's section offset. */
9178 if (fixp->fx_r_type == BFD_RELOC_VTABLE_ENTRY)
9179 rel->address = fixp->fx_offset;
9180 #if defined (OBJ_COFF) && defined (TE_PE)
9181 else if (fixp->fx_addsy && S_IS_WEAK (fixp->fx_addsy))
9182 rel->addend = fixp->fx_addnumber - (S_GET_VALUE (fixp->fx_addsy) * 2);
9183 else
9184 #endif
9185 rel->addend = 0;
9186 }
9187 /* Use the rela in 64bit mode. */
9188 else
9189 {
9190 if (disallow_64bit_reloc)
9191 switch (code)
9192 {
9193 case BFD_RELOC_X86_64_DTPOFF64:
9194 case BFD_RELOC_X86_64_TPOFF64:
9195 case BFD_RELOC_64_PCREL:
9196 case BFD_RELOC_X86_64_GOTOFF64:
9197 case BFD_RELOC_X86_64_GOT64:
9198 case BFD_RELOC_X86_64_GOTPCREL64:
9199 case BFD_RELOC_X86_64_GOTPC64:
9200 case BFD_RELOC_X86_64_GOTPLT64:
9201 case BFD_RELOC_X86_64_PLTOFF64:
9202 as_bad_where (fixp->fx_file, fixp->fx_line,
9203 _("cannot represent relocation type %s in x32 mode"),
9204 bfd_get_reloc_code_name (code));
9205 break;
9206 default:
9207 break;
9208 }
9209
9210 if (!fixp->fx_pcrel)
9211 rel->addend = fixp->fx_offset;
9212 else
9213 switch (code)
9214 {
9215 case BFD_RELOC_X86_64_PLT32:
9216 case BFD_RELOC_X86_64_GOT32:
9217 case BFD_RELOC_X86_64_GOTPCREL:
9218 case BFD_RELOC_X86_64_TLSGD:
9219 case BFD_RELOC_X86_64_TLSLD:
9220 case BFD_RELOC_X86_64_GOTTPOFF:
9221 case BFD_RELOC_X86_64_GOTPC32_TLSDESC:
9222 case BFD_RELOC_X86_64_TLSDESC_CALL:
9223 rel->addend = fixp->fx_offset - fixp->fx_size;
9224 break;
9225 default:
9226 rel->addend = (section->vma
9227 - fixp->fx_size
9228 + fixp->fx_addnumber
9229 + md_pcrel_from (fixp));
9230 break;
9231 }
9232 }
9233
9234 rel->howto = bfd_reloc_type_lookup (stdoutput, code);
9235 if (rel->howto == NULL)
9236 {
9237 as_bad_where (fixp->fx_file, fixp->fx_line,
9238 _("cannot represent relocation type %s"),
9239 bfd_get_reloc_code_name (code));
9240 /* Set howto to a garbage value so that we can keep going. */
9241 rel->howto = bfd_reloc_type_lookup (stdoutput, BFD_RELOC_32);
9242 gas_assert (rel->howto != NULL);
9243 }
9244
9245 return rel;
9246 }
9247
9248 #include "tc-i386-intel.c"
9249
9250 void
9251 tc_x86_parse_to_dw2regnum (expressionS *exp)
9252 {
9253 int saved_naked_reg;
9254 char saved_register_dot;
9255
9256 saved_naked_reg = allow_naked_reg;
9257 allow_naked_reg = 1;
9258 saved_register_dot = register_chars['.'];
9259 register_chars['.'] = '.';
9260 allow_pseudo_reg = 1;
9261 expression_and_evaluate (exp);
9262 allow_pseudo_reg = 0;
9263 register_chars['.'] = saved_register_dot;
9264 allow_naked_reg = saved_naked_reg;
9265
9266 if (exp->X_op == O_register && exp->X_add_number >= 0)
9267 {
9268 if ((addressT) exp->X_add_number < i386_regtab_size)
9269 {
9270 exp->X_op = O_constant;
9271 exp->X_add_number = i386_regtab[exp->X_add_number]
9272 .dw2_regnum[flag_code >> 1];
9273 }
9274 else
9275 exp->X_op = O_illegal;
9276 }
9277 }
9278
9279 void
9280 tc_x86_frame_initial_instructions (void)
9281 {
9282 static unsigned int sp_regno[2];
9283
9284 if (!sp_regno[flag_code >> 1])
9285 {
9286 char *saved_input = input_line_pointer;
9287 char sp[][4] = {"esp", "rsp"};
9288 expressionS exp;
9289
9290 input_line_pointer = sp[flag_code >> 1];
9291 tc_x86_parse_to_dw2regnum (&exp);
9292 gas_assert (exp.X_op == O_constant);
9293 sp_regno[flag_code >> 1] = exp.X_add_number;
9294 input_line_pointer = saved_input;
9295 }
9296
9297 cfi_add_CFA_def_cfa (sp_regno[flag_code >> 1], -x86_cie_data_alignment);
9298 cfi_add_CFA_offset (x86_dwarf2_return_column, x86_cie_data_alignment);
9299 }
9300
9301 int
9302 x86_dwarf2_addr_size (void)
9303 {
9304 #if defined (OBJ_MAYBE_ELF) || defined (OBJ_ELF)
9305 if (x86_elf_abi == X86_64_X32_ABI)
9306 return 4;
9307 #endif
9308 return bfd_arch_bits_per_address (stdoutput) / 8;
9309 }
9310
9311 int
9312 i386_elf_section_type (const char *str, size_t len)
9313 {
9314 if (flag_code == CODE_64BIT
9315 && len == sizeof ("unwind") - 1
9316 && strncmp (str, "unwind", 6) == 0)
9317 return SHT_X86_64_UNWIND;
9318
9319 return -1;
9320 }
9321
9322 #ifdef TE_SOLARIS
9323 void
9324 i386_solaris_fix_up_eh_frame (segT sec)
9325 {
9326 if (flag_code == CODE_64BIT)
9327 elf_section_type (sec) = SHT_X86_64_UNWIND;
9328 }
9329 #endif
9330
9331 #ifdef TE_PE
9332 void
9333 tc_pe_dwarf2_emit_offset (symbolS *symbol, unsigned int size)
9334 {
9335 expressionS exp;
9336
9337 exp.X_op = O_secrel;
9338 exp.X_add_symbol = symbol;
9339 exp.X_add_number = 0;
9340 emit_expr (&exp, size);
9341 }
9342 #endif
9343
9344 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
9345 /* For ELF on x86-64, add support for SHF_X86_64_LARGE. */
9346
9347 bfd_vma
9348 x86_64_section_letter (int letter, char **ptr_msg)
9349 {
9350 if (flag_code == CODE_64BIT)
9351 {
9352 if (letter == 'l')
9353 return SHF_X86_64_LARGE;
9354
9355 *ptr_msg = _("bad .section directive: want a,l,w,x,M,S,G,T in string");
9356 }
9357 else
9358 *ptr_msg = _("bad .section directive: want a,w,x,M,S,G,T in string");
9359 return -1;
9360 }
9361
9362 bfd_vma
9363 x86_64_section_word (char *str, size_t len)
9364 {
9365 if (len == 5 && flag_code == CODE_64BIT && CONST_STRNEQ (str, "large"))
9366 return SHF_X86_64_LARGE;
9367
9368 return -1;
9369 }
9370
9371 static void
9372 handle_large_common (int small ATTRIBUTE_UNUSED)
9373 {
9374 if (flag_code != CODE_64BIT)
9375 {
9376 s_comm_internal (0, elf_common_parse);
9377 as_warn (_(".largecomm supported only in 64bit mode, producing .comm"));
9378 }
9379 else
9380 {
9381 static segT lbss_section;
9382 asection *saved_com_section_ptr = elf_com_section_ptr;
9383 asection *saved_bss_section = bss_section;
9384
9385 if (lbss_section == NULL)
9386 {
9387 flagword applicable;
9388 segT seg = now_seg;
9389 subsegT subseg = now_subseg;
9390
9391 /* The .lbss section is for local .largecomm symbols. */
9392 lbss_section = subseg_new (".lbss", 0);
9393 applicable = bfd_applicable_section_flags (stdoutput);
9394 bfd_set_section_flags (stdoutput, lbss_section,
9395 applicable & SEC_ALLOC);
9396 seg_info (lbss_section)->bss = 1;
9397
9398 subseg_set (seg, subseg);
9399 }
9400
9401 elf_com_section_ptr = &_bfd_elf_large_com_section;
9402 bss_section = lbss_section;
9403
9404 s_comm_internal (0, elf_common_parse);
9405
9406 elf_com_section_ptr = saved_com_section_ptr;
9407 bss_section = saved_bss_section;
9408 }
9409 }
9410 #endif /* OBJ_ELF || OBJ_MAYBE_ELF */
This page took 0.22938 seconds and 4 git commands to generate.