* config/tc-i386.c (lex_got): Provide implementation for PE
[deliverable/binutils-gdb.git] / gas / config / tc-i386.c
1 /* tc-i386.c -- Assemble code for the Intel 80386
2 Copyright 1989, 1991, 1992, 1993, 1994, 1995, 1996, 1997, 1998, 1999,
3 2000, 2001, 2002, 2003, 2004, 2005, 2006, 2007, 2008, 2009, 2010, 2011,
4 2012
5 Free Software Foundation, Inc.
6
7 This file is part of GAS, the GNU Assembler.
8
9 GAS is free software; you can redistribute it and/or modify
10 it under the terms of the GNU General Public License as published by
11 the Free Software Foundation; either version 3, or (at your option)
12 any later version.
13
14 GAS is distributed in the hope that it will be useful,
15 but WITHOUT ANY WARRANTY; without even the implied warranty of
16 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
17 GNU General Public License for more details.
18
19 You should have received a copy of the GNU General Public License
20 along with GAS; see the file COPYING. If not, write to the Free
21 Software Foundation, 51 Franklin Street - Fifth Floor, Boston, MA
22 02110-1301, USA. */
23
24 /* Intel 80386 machine specific gas.
25 Written by Eliot Dresselhaus (eliot@mgm.mit.edu).
26 x86_64 support by Jan Hubicka (jh@suse.cz)
27 VIA PadLock support by Michal Ludvig (mludvig@suse.cz)
28 Bugs & suggestions are completely welcome. This is free software.
29 Please help us make it better. */
30
31 #include "as.h"
32 #include "safe-ctype.h"
33 #include "subsegs.h"
34 #include "dwarf2dbg.h"
35 #include "dw2gencfi.h"
36 #include "elf/x86-64.h"
37 #include "opcodes/i386-init.h"
38
39 #ifndef REGISTER_WARNINGS
40 #define REGISTER_WARNINGS 1
41 #endif
42
43 #ifndef INFER_ADDR_PREFIX
44 #define INFER_ADDR_PREFIX 1
45 #endif
46
47 #ifndef DEFAULT_ARCH
48 #define DEFAULT_ARCH "i386"
49 #endif
50
51 #ifndef INLINE
52 #if __GNUC__ >= 2
53 #define INLINE __inline__
54 #else
55 #define INLINE
56 #endif
57 #endif
58
59 /* Prefixes will be emitted in the order defined below.
60 WAIT_PREFIX must be the first prefix since FWAIT is really is an
61 instruction, and so must come before any prefixes.
62 The preferred prefix order is SEG_PREFIX, ADDR_PREFIX, DATA_PREFIX,
63 REP_PREFIX/HLE_PREFIX, LOCK_PREFIX. */
64 #define WAIT_PREFIX 0
65 #define SEG_PREFIX 1
66 #define ADDR_PREFIX 2
67 #define DATA_PREFIX 3
68 #define REP_PREFIX 4
69 #define HLE_PREFIX REP_PREFIX
70 #define LOCK_PREFIX 5
71 #define REX_PREFIX 6 /* must come last. */
72 #define MAX_PREFIXES 7 /* max prefixes per opcode */
73
74 /* we define the syntax here (modulo base,index,scale syntax) */
75 #define REGISTER_PREFIX '%'
76 #define IMMEDIATE_PREFIX '$'
77 #define ABSOLUTE_PREFIX '*'
78
79 /* these are the instruction mnemonic suffixes in AT&T syntax or
80 memory operand size in Intel syntax. */
81 #define WORD_MNEM_SUFFIX 'w'
82 #define BYTE_MNEM_SUFFIX 'b'
83 #define SHORT_MNEM_SUFFIX 's'
84 #define LONG_MNEM_SUFFIX 'l'
85 #define QWORD_MNEM_SUFFIX 'q'
86 #define XMMWORD_MNEM_SUFFIX 'x'
87 #define YMMWORD_MNEM_SUFFIX 'y'
88 /* Intel Syntax. Use a non-ascii letter since since it never appears
89 in instructions. */
90 #define LONG_DOUBLE_MNEM_SUFFIX '\1'
91
92 #define END_OF_INSN '\0'
93
94 /*
95 'templates' is for grouping together 'template' structures for opcodes
96 of the same name. This is only used for storing the insns in the grand
97 ole hash table of insns.
98 The templates themselves start at START and range up to (but not including)
99 END.
100 */
101 typedef struct
102 {
103 const insn_template *start;
104 const insn_template *end;
105 }
106 templates;
107
108 /* 386 operand encoding bytes: see 386 book for details of this. */
109 typedef struct
110 {
111 unsigned int regmem; /* codes register or memory operand */
112 unsigned int reg; /* codes register operand (or extended opcode) */
113 unsigned int mode; /* how to interpret regmem & reg */
114 }
115 modrm_byte;
116
117 /* x86-64 extension prefix. */
118 typedef int rex_byte;
119
120 /* 386 opcode byte to code indirect addressing. */
121 typedef struct
122 {
123 unsigned base;
124 unsigned index;
125 unsigned scale;
126 }
127 sib_byte;
128
129 /* x86 arch names, types and features */
130 typedef struct
131 {
132 const char *name; /* arch name */
133 unsigned int len; /* arch string length */
134 enum processor_type type; /* arch type */
135 i386_cpu_flags flags; /* cpu feature flags */
136 unsigned int skip; /* show_arch should skip this. */
137 unsigned int negated; /* turn off indicated flags. */
138 }
139 arch_entry;
140
141 static void update_code_flag (int, int);
142 static void set_code_flag (int);
143 static void set_16bit_gcc_code_flag (int);
144 static void set_intel_syntax (int);
145 static void set_intel_mnemonic (int);
146 static void set_allow_index_reg (int);
147 static void set_sse_check (int);
148 static void set_cpu_arch (int);
149 #ifdef TE_PE
150 static void pe_directive_secrel (int);
151 #endif
152 static void signed_cons (int);
153 static char *output_invalid (int c);
154 static int i386_finalize_immediate (segT, expressionS *, i386_operand_type,
155 const char *);
156 static int i386_finalize_displacement (segT, expressionS *, i386_operand_type,
157 const char *);
158 static int i386_att_operand (char *);
159 static int i386_intel_operand (char *, int);
160 static int i386_intel_simplify (expressionS *);
161 static int i386_intel_parse_name (const char *, expressionS *);
162 static const reg_entry *parse_register (char *, char **);
163 static char *parse_insn (char *, char *);
164 static char *parse_operands (char *, const char *);
165 static void swap_operands (void);
166 static void swap_2_operands (int, int);
167 static void optimize_imm (void);
168 static void optimize_disp (void);
169 static const insn_template *match_template (void);
170 static int check_string (void);
171 static int process_suffix (void);
172 static int check_byte_reg (void);
173 static int check_long_reg (void);
174 static int check_qword_reg (void);
175 static int check_word_reg (void);
176 static int finalize_imm (void);
177 static int process_operands (void);
178 static const seg_entry *build_modrm_byte (void);
179 static void output_insn (void);
180 static void output_imm (fragS *, offsetT);
181 static void output_disp (fragS *, offsetT);
182 #ifndef I386COFF
183 static void s_bss (int);
184 #endif
185 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
186 static void handle_large_common (int small ATTRIBUTE_UNUSED);
187 #endif
188
189 static const char *default_arch = DEFAULT_ARCH;
190
191 /* VEX prefix. */
192 typedef struct
193 {
194 /* VEX prefix is either 2 byte or 3 byte. */
195 unsigned char bytes[3];
196 unsigned int length;
197 /* Destination or source register specifier. */
198 const reg_entry *register_specifier;
199 } vex_prefix;
200
201 /* 'md_assemble ()' gathers together information and puts it into a
202 i386_insn. */
203
204 union i386_op
205 {
206 expressionS *disps;
207 expressionS *imms;
208 const reg_entry *regs;
209 };
210
211 enum i386_error
212 {
213 operand_size_mismatch,
214 operand_type_mismatch,
215 register_type_mismatch,
216 number_of_operands_mismatch,
217 invalid_instruction_suffix,
218 bad_imm4,
219 old_gcc_only,
220 unsupported_with_intel_mnemonic,
221 unsupported_syntax,
222 unsupported,
223 invalid_vsib_address,
224 unsupported_vector_index_register
225 };
226
227 struct _i386_insn
228 {
229 /* TM holds the template for the insn were currently assembling. */
230 insn_template tm;
231
232 /* SUFFIX holds the instruction size suffix for byte, word, dword
233 or qword, if given. */
234 char suffix;
235
236 /* OPERANDS gives the number of given operands. */
237 unsigned int operands;
238
239 /* REG_OPERANDS, DISP_OPERANDS, MEM_OPERANDS, IMM_OPERANDS give the number
240 of given register, displacement, memory operands and immediate
241 operands. */
242 unsigned int reg_operands, disp_operands, mem_operands, imm_operands;
243
244 /* TYPES [i] is the type (see above #defines) which tells us how to
245 use OP[i] for the corresponding operand. */
246 i386_operand_type types[MAX_OPERANDS];
247
248 /* Displacement expression, immediate expression, or register for each
249 operand. */
250 union i386_op op[MAX_OPERANDS];
251
252 /* Flags for operands. */
253 unsigned int flags[MAX_OPERANDS];
254 #define Operand_PCrel 1
255
256 /* Relocation type for operand */
257 enum bfd_reloc_code_real reloc[MAX_OPERANDS];
258
259 /* BASE_REG, INDEX_REG, and LOG2_SCALE_FACTOR are used to encode
260 the base index byte below. */
261 const reg_entry *base_reg;
262 const reg_entry *index_reg;
263 unsigned int log2_scale_factor;
264
265 /* SEG gives the seg_entries of this insn. They are zero unless
266 explicit segment overrides are given. */
267 const seg_entry *seg[2];
268
269 /* PREFIX holds all the given prefix opcodes (usually null).
270 PREFIXES is the number of prefix opcodes. */
271 unsigned int prefixes;
272 unsigned char prefix[MAX_PREFIXES];
273
274 /* RM and SIB are the modrm byte and the sib byte where the
275 addressing modes of this insn are encoded. */
276 modrm_byte rm;
277 rex_byte rex;
278 sib_byte sib;
279 vex_prefix vex;
280
281 /* Swap operand in encoding. */
282 unsigned int swap_operand;
283
284 /* Prefer 8bit or 32bit displacement in encoding. */
285 enum
286 {
287 disp_encoding_default = 0,
288 disp_encoding_8bit,
289 disp_encoding_32bit
290 } disp_encoding;
291
292 /* Have HLE prefix. */
293 unsigned int have_hle;
294
295 /* Error message. */
296 enum i386_error error;
297 };
298
299 typedef struct _i386_insn i386_insn;
300
301 /* List of chars besides those in app.c:symbol_chars that can start an
302 operand. Used to prevent the scrubber eating vital white-space. */
303 const char extra_symbol_chars[] = "*%-(["
304 #ifdef LEX_AT
305 "@"
306 #endif
307 #ifdef LEX_QM
308 "?"
309 #endif
310 ;
311
312 #if (defined (TE_I386AIX) \
313 || ((defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)) \
314 && !defined (TE_GNU) \
315 && !defined (TE_LINUX) \
316 && !defined (TE_NACL) \
317 && !defined (TE_NETWARE) \
318 && !defined (TE_FreeBSD) \
319 && !defined (TE_DragonFly) \
320 && !defined (TE_NetBSD)))
321 /* This array holds the chars that always start a comment. If the
322 pre-processor is disabled, these aren't very useful. The option
323 --divide will remove '/' from this list. */
324 const char *i386_comment_chars = "#/";
325 #define SVR4_COMMENT_CHARS 1
326 #define PREFIX_SEPARATOR '\\'
327
328 #else
329 const char *i386_comment_chars = "#";
330 #define PREFIX_SEPARATOR '/'
331 #endif
332
333 /* This array holds the chars that only start a comment at the beginning of
334 a line. If the line seems to have the form '# 123 filename'
335 .line and .file directives will appear in the pre-processed output.
336 Note that input_file.c hand checks for '#' at the beginning of the
337 first line of the input file. This is because the compiler outputs
338 #NO_APP at the beginning of its output.
339 Also note that comments started like this one will always work if
340 '/' isn't otherwise defined. */
341 const char line_comment_chars[] = "#/";
342
343 const char line_separator_chars[] = ";";
344
345 /* Chars that can be used to separate mant from exp in floating point
346 nums. */
347 const char EXP_CHARS[] = "eE";
348
349 /* Chars that mean this number is a floating point constant
350 As in 0f12.456
351 or 0d1.2345e12. */
352 const char FLT_CHARS[] = "fFdDxX";
353
354 /* Tables for lexical analysis. */
355 static char mnemonic_chars[256];
356 static char register_chars[256];
357 static char operand_chars[256];
358 static char identifier_chars[256];
359 static char digit_chars[256];
360
361 /* Lexical macros. */
362 #define is_mnemonic_char(x) (mnemonic_chars[(unsigned char) x])
363 #define is_operand_char(x) (operand_chars[(unsigned char) x])
364 #define is_register_char(x) (register_chars[(unsigned char) x])
365 #define is_space_char(x) ((x) == ' ')
366 #define is_identifier_char(x) (identifier_chars[(unsigned char) x])
367 #define is_digit_char(x) (digit_chars[(unsigned char) x])
368
369 /* All non-digit non-letter characters that may occur in an operand. */
370 static char operand_special_chars[] = "%$-+(,)*._~/<>|&^!:[@]";
371
372 /* md_assemble() always leaves the strings it's passed unaltered. To
373 effect this we maintain a stack of saved characters that we've smashed
374 with '\0's (indicating end of strings for various sub-fields of the
375 assembler instruction). */
376 static char save_stack[32];
377 static char *save_stack_p;
378 #define END_STRING_AND_SAVE(s) \
379 do { *save_stack_p++ = *(s); *(s) = '\0'; } while (0)
380 #define RESTORE_END_STRING(s) \
381 do { *(s) = *--save_stack_p; } while (0)
382
383 /* The instruction we're assembling. */
384 static i386_insn i;
385
386 /* Possible templates for current insn. */
387 static const templates *current_templates;
388
389 /* Per instruction expressionS buffers: max displacements & immediates. */
390 static expressionS disp_expressions[MAX_MEMORY_OPERANDS];
391 static expressionS im_expressions[MAX_IMMEDIATE_OPERANDS];
392
393 /* Current operand we are working on. */
394 static int this_operand = -1;
395
396 /* We support four different modes. FLAG_CODE variable is used to distinguish
397 these. */
398
399 enum flag_code {
400 CODE_32BIT,
401 CODE_16BIT,
402 CODE_64BIT };
403
404 static enum flag_code flag_code;
405 static unsigned int object_64bit;
406 static unsigned int disallow_64bit_reloc;
407 static int use_rela_relocations = 0;
408
409 #if ((defined (OBJ_MAYBE_COFF) && defined (OBJ_MAYBE_AOUT)) \
410 || defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) \
411 || defined (TE_PE) || defined (TE_PEP) || defined (OBJ_MACH_O))
412
413 /* The ELF ABI to use. */
414 enum x86_elf_abi
415 {
416 I386_ABI,
417 X86_64_ABI,
418 X86_64_X32_ABI
419 };
420
421 static enum x86_elf_abi x86_elf_abi = I386_ABI;
422 #endif
423
424 /* The names used to print error messages. */
425 static const char *flag_code_names[] =
426 {
427 "32",
428 "16",
429 "64"
430 };
431
432 /* 1 for intel syntax,
433 0 if att syntax. */
434 static int intel_syntax = 0;
435
436 /* 1 for intel mnemonic,
437 0 if att mnemonic. */
438 static int intel_mnemonic = !SYSV386_COMPAT;
439
440 /* 1 if support old (<= 2.8.1) versions of gcc. */
441 static int old_gcc = OLDGCC_COMPAT;
442
443 /* 1 if pseudo registers are permitted. */
444 static int allow_pseudo_reg = 0;
445
446 /* 1 if register prefix % not required. */
447 static int allow_naked_reg = 0;
448
449 /* 1 if pseudo index register, eiz/riz, is allowed . */
450 static int allow_index_reg = 0;
451
452 static enum
453 {
454 sse_check_none = 0,
455 sse_check_warning,
456 sse_check_error
457 }
458 sse_check;
459
460 /* Register prefix used for error message. */
461 static const char *register_prefix = "%";
462
463 /* Used in 16 bit gcc mode to add an l suffix to call, ret, enter,
464 leave, push, and pop instructions so that gcc has the same stack
465 frame as in 32 bit mode. */
466 static char stackop_size = '\0';
467
468 /* Non-zero to optimize code alignment. */
469 int optimize_align_code = 1;
470
471 /* Non-zero to quieten some warnings. */
472 static int quiet_warnings = 0;
473
474 /* CPU name. */
475 static const char *cpu_arch_name = NULL;
476 static char *cpu_sub_arch_name = NULL;
477
478 /* CPU feature flags. */
479 static i386_cpu_flags cpu_arch_flags = CPU_UNKNOWN_FLAGS;
480
481 /* If we have selected a cpu we are generating instructions for. */
482 static int cpu_arch_tune_set = 0;
483
484 /* Cpu we are generating instructions for. */
485 enum processor_type cpu_arch_tune = PROCESSOR_UNKNOWN;
486
487 /* CPU feature flags of cpu we are generating instructions for. */
488 static i386_cpu_flags cpu_arch_tune_flags;
489
490 /* CPU instruction set architecture used. */
491 enum processor_type cpu_arch_isa = PROCESSOR_UNKNOWN;
492
493 /* CPU feature flags of instruction set architecture used. */
494 i386_cpu_flags cpu_arch_isa_flags;
495
496 /* If set, conditional jumps are not automatically promoted to handle
497 larger than a byte offset. */
498 static unsigned int no_cond_jump_promotion = 0;
499
500 /* Encode SSE instructions with VEX prefix. */
501 static unsigned int sse2avx;
502
503 /* Encode scalar AVX instructions with specific vector length. */
504 static enum
505 {
506 vex128 = 0,
507 vex256
508 } avxscalar;
509
510 /* Pre-defined "_GLOBAL_OFFSET_TABLE_". */
511 static symbolS *GOT_symbol;
512
513 /* The dwarf2 return column, adjusted for 32 or 64 bit. */
514 unsigned int x86_dwarf2_return_column;
515
516 /* The dwarf2 data alignment, adjusted for 32 or 64 bit. */
517 int x86_cie_data_alignment;
518
519 /* Interface to relax_segment.
520 There are 3 major relax states for 386 jump insns because the
521 different types of jumps add different sizes to frags when we're
522 figuring out what sort of jump to choose to reach a given label. */
523
524 /* Types. */
525 #define UNCOND_JUMP 0
526 #define COND_JUMP 1
527 #define COND_JUMP86 2
528
529 /* Sizes. */
530 #define CODE16 1
531 #define SMALL 0
532 #define SMALL16 (SMALL | CODE16)
533 #define BIG 2
534 #define BIG16 (BIG | CODE16)
535
536 #ifndef INLINE
537 #ifdef __GNUC__
538 #define INLINE __inline__
539 #else
540 #define INLINE
541 #endif
542 #endif
543
544 #define ENCODE_RELAX_STATE(type, size) \
545 ((relax_substateT) (((type) << 2) | (size)))
546 #define TYPE_FROM_RELAX_STATE(s) \
547 ((s) >> 2)
548 #define DISP_SIZE_FROM_RELAX_STATE(s) \
549 ((((s) & 3) == BIG ? 4 : (((s) & 3) == BIG16 ? 2 : 1)))
550
551 /* This table is used by relax_frag to promote short jumps to long
552 ones where necessary. SMALL (short) jumps may be promoted to BIG
553 (32 bit long) ones, and SMALL16 jumps to BIG16 (16 bit long). We
554 don't allow a short jump in a 32 bit code segment to be promoted to
555 a 16 bit offset jump because it's slower (requires data size
556 prefix), and doesn't work, unless the destination is in the bottom
557 64k of the code segment (The top 16 bits of eip are zeroed). */
558
559 const relax_typeS md_relax_table[] =
560 {
561 /* The fields are:
562 1) most positive reach of this state,
563 2) most negative reach of this state,
564 3) how many bytes this mode will have in the variable part of the frag
565 4) which index into the table to try if we can't fit into this one. */
566
567 /* UNCOND_JUMP states. */
568 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (UNCOND_JUMP, BIG)},
569 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (UNCOND_JUMP, BIG16)},
570 /* dword jmp adds 4 bytes to frag:
571 0 extra opcode bytes, 4 displacement bytes. */
572 {0, 0, 4, 0},
573 /* word jmp adds 2 byte2 to frag:
574 0 extra opcode bytes, 2 displacement bytes. */
575 {0, 0, 2, 0},
576
577 /* COND_JUMP states. */
578 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (COND_JUMP, BIG)},
579 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (COND_JUMP, BIG16)},
580 /* dword conditionals adds 5 bytes to frag:
581 1 extra opcode byte, 4 displacement bytes. */
582 {0, 0, 5, 0},
583 /* word conditionals add 3 bytes to frag:
584 1 extra opcode byte, 2 displacement bytes. */
585 {0, 0, 3, 0},
586
587 /* COND_JUMP86 states. */
588 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (COND_JUMP86, BIG)},
589 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (COND_JUMP86, BIG16)},
590 /* dword conditionals adds 5 bytes to frag:
591 1 extra opcode byte, 4 displacement bytes. */
592 {0, 0, 5, 0},
593 /* word conditionals add 4 bytes to frag:
594 1 displacement byte and a 3 byte long branch insn. */
595 {0, 0, 4, 0}
596 };
597
598 static const arch_entry cpu_arch[] =
599 {
600 /* Do not replace the first two entries - i386_target_format()
601 relies on them being there in this order. */
602 { STRING_COMMA_LEN ("generic32"), PROCESSOR_GENERIC32,
603 CPU_GENERIC32_FLAGS, 0, 0 },
604 { STRING_COMMA_LEN ("generic64"), PROCESSOR_GENERIC64,
605 CPU_GENERIC64_FLAGS, 0, 0 },
606 { STRING_COMMA_LEN ("i8086"), PROCESSOR_UNKNOWN,
607 CPU_NONE_FLAGS, 0, 0 },
608 { STRING_COMMA_LEN ("i186"), PROCESSOR_UNKNOWN,
609 CPU_I186_FLAGS, 0, 0 },
610 { STRING_COMMA_LEN ("i286"), PROCESSOR_UNKNOWN,
611 CPU_I286_FLAGS, 0, 0 },
612 { STRING_COMMA_LEN ("i386"), PROCESSOR_I386,
613 CPU_I386_FLAGS, 0, 0 },
614 { STRING_COMMA_LEN ("i486"), PROCESSOR_I486,
615 CPU_I486_FLAGS, 0, 0 },
616 { STRING_COMMA_LEN ("i586"), PROCESSOR_PENTIUM,
617 CPU_I586_FLAGS, 0, 0 },
618 { STRING_COMMA_LEN ("i686"), PROCESSOR_PENTIUMPRO,
619 CPU_I686_FLAGS, 0, 0 },
620 { STRING_COMMA_LEN ("pentium"), PROCESSOR_PENTIUM,
621 CPU_I586_FLAGS, 0, 0 },
622 { STRING_COMMA_LEN ("pentiumpro"), PROCESSOR_PENTIUMPRO,
623 CPU_PENTIUMPRO_FLAGS, 0, 0 },
624 { STRING_COMMA_LEN ("pentiumii"), PROCESSOR_PENTIUMPRO,
625 CPU_P2_FLAGS, 0, 0 },
626 { STRING_COMMA_LEN ("pentiumiii"),PROCESSOR_PENTIUMPRO,
627 CPU_P3_FLAGS, 0, 0 },
628 { STRING_COMMA_LEN ("pentium4"), PROCESSOR_PENTIUM4,
629 CPU_P4_FLAGS, 0, 0 },
630 { STRING_COMMA_LEN ("prescott"), PROCESSOR_NOCONA,
631 CPU_CORE_FLAGS, 0, 0 },
632 { STRING_COMMA_LEN ("nocona"), PROCESSOR_NOCONA,
633 CPU_NOCONA_FLAGS, 0, 0 },
634 { STRING_COMMA_LEN ("yonah"), PROCESSOR_CORE,
635 CPU_CORE_FLAGS, 1, 0 },
636 { STRING_COMMA_LEN ("core"), PROCESSOR_CORE,
637 CPU_CORE_FLAGS, 0, 0 },
638 { STRING_COMMA_LEN ("merom"), PROCESSOR_CORE2,
639 CPU_CORE2_FLAGS, 1, 0 },
640 { STRING_COMMA_LEN ("core2"), PROCESSOR_CORE2,
641 CPU_CORE2_FLAGS, 0, 0 },
642 { STRING_COMMA_LEN ("corei7"), PROCESSOR_COREI7,
643 CPU_COREI7_FLAGS, 0, 0 },
644 { STRING_COMMA_LEN ("l1om"), PROCESSOR_L1OM,
645 CPU_L1OM_FLAGS, 0, 0 },
646 { STRING_COMMA_LEN ("k1om"), PROCESSOR_K1OM,
647 CPU_K1OM_FLAGS, 0, 0 },
648 { STRING_COMMA_LEN ("k6"), PROCESSOR_K6,
649 CPU_K6_FLAGS, 0, 0 },
650 { STRING_COMMA_LEN ("k6_2"), PROCESSOR_K6,
651 CPU_K6_2_FLAGS, 0, 0 },
652 { STRING_COMMA_LEN ("athlon"), PROCESSOR_ATHLON,
653 CPU_ATHLON_FLAGS, 0, 0 },
654 { STRING_COMMA_LEN ("sledgehammer"), PROCESSOR_K8,
655 CPU_K8_FLAGS, 1, 0 },
656 { STRING_COMMA_LEN ("opteron"), PROCESSOR_K8,
657 CPU_K8_FLAGS, 0, 0 },
658 { STRING_COMMA_LEN ("k8"), PROCESSOR_K8,
659 CPU_K8_FLAGS, 0, 0 },
660 { STRING_COMMA_LEN ("amdfam10"), PROCESSOR_AMDFAM10,
661 CPU_AMDFAM10_FLAGS, 0, 0 },
662 { STRING_COMMA_LEN ("bdver1"), PROCESSOR_BD,
663 CPU_BDVER1_FLAGS, 0, 0 },
664 { STRING_COMMA_LEN ("bdver2"), PROCESSOR_BD,
665 CPU_BDVER2_FLAGS, 0, 0 },
666 { STRING_COMMA_LEN (".8087"), PROCESSOR_UNKNOWN,
667 CPU_8087_FLAGS, 0, 0 },
668 { STRING_COMMA_LEN (".287"), PROCESSOR_UNKNOWN,
669 CPU_287_FLAGS, 0, 0 },
670 { STRING_COMMA_LEN (".387"), PROCESSOR_UNKNOWN,
671 CPU_387_FLAGS, 0, 0 },
672 { STRING_COMMA_LEN (".no87"), PROCESSOR_UNKNOWN,
673 CPU_ANY87_FLAGS, 0, 1 },
674 { STRING_COMMA_LEN (".mmx"), PROCESSOR_UNKNOWN,
675 CPU_MMX_FLAGS, 0, 0 },
676 { STRING_COMMA_LEN (".nommx"), PROCESSOR_UNKNOWN,
677 CPU_3DNOWA_FLAGS, 0, 1 },
678 { STRING_COMMA_LEN (".sse"), PROCESSOR_UNKNOWN,
679 CPU_SSE_FLAGS, 0, 0 },
680 { STRING_COMMA_LEN (".sse2"), PROCESSOR_UNKNOWN,
681 CPU_SSE2_FLAGS, 0, 0 },
682 { STRING_COMMA_LEN (".sse3"), PROCESSOR_UNKNOWN,
683 CPU_SSE3_FLAGS, 0, 0 },
684 { STRING_COMMA_LEN (".ssse3"), PROCESSOR_UNKNOWN,
685 CPU_SSSE3_FLAGS, 0, 0 },
686 { STRING_COMMA_LEN (".sse4.1"), PROCESSOR_UNKNOWN,
687 CPU_SSE4_1_FLAGS, 0, 0 },
688 { STRING_COMMA_LEN (".sse4.2"), PROCESSOR_UNKNOWN,
689 CPU_SSE4_2_FLAGS, 0, 0 },
690 { STRING_COMMA_LEN (".sse4"), PROCESSOR_UNKNOWN,
691 CPU_SSE4_2_FLAGS, 0, 0 },
692 { STRING_COMMA_LEN (".nosse"), PROCESSOR_UNKNOWN,
693 CPU_ANY_SSE_FLAGS, 0, 1 },
694 { STRING_COMMA_LEN (".avx"), PROCESSOR_UNKNOWN,
695 CPU_AVX_FLAGS, 0, 0 },
696 { STRING_COMMA_LEN (".avx2"), PROCESSOR_UNKNOWN,
697 CPU_AVX2_FLAGS, 0, 0 },
698 { STRING_COMMA_LEN (".noavx"), PROCESSOR_UNKNOWN,
699 CPU_ANY_AVX_FLAGS, 0, 1 },
700 { STRING_COMMA_LEN (".vmx"), PROCESSOR_UNKNOWN,
701 CPU_VMX_FLAGS, 0, 0 },
702 { STRING_COMMA_LEN (".vmfunc"), PROCESSOR_UNKNOWN,
703 CPU_VMFUNC_FLAGS, 0, 0 },
704 { STRING_COMMA_LEN (".smx"), PROCESSOR_UNKNOWN,
705 CPU_SMX_FLAGS, 0, 0 },
706 { STRING_COMMA_LEN (".xsave"), PROCESSOR_UNKNOWN,
707 CPU_XSAVE_FLAGS, 0, 0 },
708 { STRING_COMMA_LEN (".xsaveopt"), PROCESSOR_UNKNOWN,
709 CPU_XSAVEOPT_FLAGS, 0, 0 },
710 { STRING_COMMA_LEN (".aes"), PROCESSOR_UNKNOWN,
711 CPU_AES_FLAGS, 0, 0 },
712 { STRING_COMMA_LEN (".pclmul"), PROCESSOR_UNKNOWN,
713 CPU_PCLMUL_FLAGS, 0, 0 },
714 { STRING_COMMA_LEN (".clmul"), PROCESSOR_UNKNOWN,
715 CPU_PCLMUL_FLAGS, 1, 0 },
716 { STRING_COMMA_LEN (".fsgsbase"), PROCESSOR_UNKNOWN,
717 CPU_FSGSBASE_FLAGS, 0, 0 },
718 { STRING_COMMA_LEN (".rdrnd"), PROCESSOR_UNKNOWN,
719 CPU_RDRND_FLAGS, 0, 0 },
720 { STRING_COMMA_LEN (".f16c"), PROCESSOR_UNKNOWN,
721 CPU_F16C_FLAGS, 0, 0 },
722 { STRING_COMMA_LEN (".bmi2"), PROCESSOR_UNKNOWN,
723 CPU_BMI2_FLAGS, 0, 0 },
724 { STRING_COMMA_LEN (".fma"), PROCESSOR_UNKNOWN,
725 CPU_FMA_FLAGS, 0, 0 },
726 { STRING_COMMA_LEN (".fma4"), PROCESSOR_UNKNOWN,
727 CPU_FMA4_FLAGS, 0, 0 },
728 { STRING_COMMA_LEN (".xop"), PROCESSOR_UNKNOWN,
729 CPU_XOP_FLAGS, 0, 0 },
730 { STRING_COMMA_LEN (".lwp"), PROCESSOR_UNKNOWN,
731 CPU_LWP_FLAGS, 0, 0 },
732 { STRING_COMMA_LEN (".movbe"), PROCESSOR_UNKNOWN,
733 CPU_MOVBE_FLAGS, 0, 0 },
734 { STRING_COMMA_LEN (".ept"), PROCESSOR_UNKNOWN,
735 CPU_EPT_FLAGS, 0, 0 },
736 { STRING_COMMA_LEN (".lzcnt"), PROCESSOR_UNKNOWN,
737 CPU_LZCNT_FLAGS, 0, 0 },
738 { STRING_COMMA_LEN (".hle"), PROCESSOR_UNKNOWN,
739 CPU_HLE_FLAGS, 0, 0 },
740 { STRING_COMMA_LEN (".rtm"), PROCESSOR_UNKNOWN,
741 CPU_RTM_FLAGS, 0, 0 },
742 { STRING_COMMA_LEN (".invpcid"), PROCESSOR_UNKNOWN,
743 CPU_INVPCID_FLAGS, 0, 0 },
744 { STRING_COMMA_LEN (".clflush"), PROCESSOR_UNKNOWN,
745 CPU_CLFLUSH_FLAGS, 0, 0 },
746 { STRING_COMMA_LEN (".nop"), PROCESSOR_UNKNOWN,
747 CPU_NOP_FLAGS, 0, 0 },
748 { STRING_COMMA_LEN (".syscall"), PROCESSOR_UNKNOWN,
749 CPU_SYSCALL_FLAGS, 0, 0 },
750 { STRING_COMMA_LEN (".rdtscp"), PROCESSOR_UNKNOWN,
751 CPU_RDTSCP_FLAGS, 0, 0 },
752 { STRING_COMMA_LEN (".3dnow"), PROCESSOR_UNKNOWN,
753 CPU_3DNOW_FLAGS, 0, 0 },
754 { STRING_COMMA_LEN (".3dnowa"), PROCESSOR_UNKNOWN,
755 CPU_3DNOWA_FLAGS, 0, 0 },
756 { STRING_COMMA_LEN (".padlock"), PROCESSOR_UNKNOWN,
757 CPU_PADLOCK_FLAGS, 0, 0 },
758 { STRING_COMMA_LEN (".pacifica"), PROCESSOR_UNKNOWN,
759 CPU_SVME_FLAGS, 1, 0 },
760 { STRING_COMMA_LEN (".svme"), PROCESSOR_UNKNOWN,
761 CPU_SVME_FLAGS, 0, 0 },
762 { STRING_COMMA_LEN (".sse4a"), PROCESSOR_UNKNOWN,
763 CPU_SSE4A_FLAGS, 0, 0 },
764 { STRING_COMMA_LEN (".abm"), PROCESSOR_UNKNOWN,
765 CPU_ABM_FLAGS, 0, 0 },
766 { STRING_COMMA_LEN (".bmi"), PROCESSOR_UNKNOWN,
767 CPU_BMI_FLAGS, 0, 0 },
768 { STRING_COMMA_LEN (".tbm"), PROCESSOR_UNKNOWN,
769 CPU_TBM_FLAGS, 0, 0 },
770 { STRING_COMMA_LEN (".adx"), PROCESSOR_UNKNOWN,
771 CPU_ADX_FLAGS, 0, 0 },
772 { STRING_COMMA_LEN (".rdseed"), PROCESSOR_UNKNOWN,
773 CPU_RDSEED_FLAGS, 0, 0 },
774 { STRING_COMMA_LEN (".prfchw"), PROCESSOR_UNKNOWN,
775 CPU_PRFCHW_FLAGS, 0, 0 },
776 };
777
778 #ifdef I386COFF
779 /* Like s_lcomm_internal in gas/read.c but the alignment string
780 is allowed to be optional. */
781
782 static symbolS *
783 pe_lcomm_internal (int needs_align, symbolS *symbolP, addressT size)
784 {
785 addressT align = 0;
786
787 SKIP_WHITESPACE ();
788
789 if (needs_align
790 && *input_line_pointer == ',')
791 {
792 align = parse_align (needs_align - 1);
793
794 if (align == (addressT) -1)
795 return NULL;
796 }
797 else
798 {
799 if (size >= 8)
800 align = 3;
801 else if (size >= 4)
802 align = 2;
803 else if (size >= 2)
804 align = 1;
805 else
806 align = 0;
807 }
808
809 bss_alloc (symbolP, size, align);
810 return symbolP;
811 }
812
813 static void
814 pe_lcomm (int needs_align)
815 {
816 s_comm_internal (needs_align * 2, pe_lcomm_internal);
817 }
818 #endif
819
820 const pseudo_typeS md_pseudo_table[] =
821 {
822 #if !defined(OBJ_AOUT) && !defined(USE_ALIGN_PTWO)
823 {"align", s_align_bytes, 0},
824 #else
825 {"align", s_align_ptwo, 0},
826 #endif
827 {"arch", set_cpu_arch, 0},
828 #ifndef I386COFF
829 {"bss", s_bss, 0},
830 #else
831 {"lcomm", pe_lcomm, 1},
832 #endif
833 {"ffloat", float_cons, 'f'},
834 {"dfloat", float_cons, 'd'},
835 {"tfloat", float_cons, 'x'},
836 {"value", cons, 2},
837 {"slong", signed_cons, 4},
838 {"noopt", s_ignore, 0},
839 {"optim", s_ignore, 0},
840 {"code16gcc", set_16bit_gcc_code_flag, CODE_16BIT},
841 {"code16", set_code_flag, CODE_16BIT},
842 {"code32", set_code_flag, CODE_32BIT},
843 {"code64", set_code_flag, CODE_64BIT},
844 {"intel_syntax", set_intel_syntax, 1},
845 {"att_syntax", set_intel_syntax, 0},
846 {"intel_mnemonic", set_intel_mnemonic, 1},
847 {"att_mnemonic", set_intel_mnemonic, 0},
848 {"allow_index_reg", set_allow_index_reg, 1},
849 {"disallow_index_reg", set_allow_index_reg, 0},
850 {"sse_check", set_sse_check, 0},
851 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
852 {"largecomm", handle_large_common, 0},
853 #else
854 {"file", (void (*) (int)) dwarf2_directive_file, 0},
855 {"loc", dwarf2_directive_loc, 0},
856 {"loc_mark_labels", dwarf2_directive_loc_mark_labels, 0},
857 #endif
858 #ifdef TE_PE
859 {"secrel32", pe_directive_secrel, 0},
860 #endif
861 {0, 0, 0}
862 };
863
864 /* For interface with expression (). */
865 extern char *input_line_pointer;
866
867 /* Hash table for instruction mnemonic lookup. */
868 static struct hash_control *op_hash;
869
870 /* Hash table for register lookup. */
871 static struct hash_control *reg_hash;
872 \f
873 void
874 i386_align_code (fragS *fragP, int count)
875 {
876 /* Various efficient no-op patterns for aligning code labels.
877 Note: Don't try to assemble the instructions in the comments.
878 0L and 0w are not legal. */
879 static const char f32_1[] =
880 {0x90}; /* nop */
881 static const char f32_2[] =
882 {0x66,0x90}; /* xchg %ax,%ax */
883 static const char f32_3[] =
884 {0x8d,0x76,0x00}; /* leal 0(%esi),%esi */
885 static const char f32_4[] =
886 {0x8d,0x74,0x26,0x00}; /* leal 0(%esi,1),%esi */
887 static const char f32_5[] =
888 {0x90, /* nop */
889 0x8d,0x74,0x26,0x00}; /* leal 0(%esi,1),%esi */
890 static const char f32_6[] =
891 {0x8d,0xb6,0x00,0x00,0x00,0x00}; /* leal 0L(%esi),%esi */
892 static const char f32_7[] =
893 {0x8d,0xb4,0x26,0x00,0x00,0x00,0x00}; /* leal 0L(%esi,1),%esi */
894 static const char f32_8[] =
895 {0x90, /* nop */
896 0x8d,0xb4,0x26,0x00,0x00,0x00,0x00}; /* leal 0L(%esi,1),%esi */
897 static const char f32_9[] =
898 {0x89,0xf6, /* movl %esi,%esi */
899 0x8d,0xbc,0x27,0x00,0x00,0x00,0x00}; /* leal 0L(%edi,1),%edi */
900 static const char f32_10[] =
901 {0x8d,0x76,0x00, /* leal 0(%esi),%esi */
902 0x8d,0xbc,0x27,0x00,0x00,0x00,0x00}; /* leal 0L(%edi,1),%edi */
903 static const char f32_11[] =
904 {0x8d,0x74,0x26,0x00, /* leal 0(%esi,1),%esi */
905 0x8d,0xbc,0x27,0x00,0x00,0x00,0x00}; /* leal 0L(%edi,1),%edi */
906 static const char f32_12[] =
907 {0x8d,0xb6,0x00,0x00,0x00,0x00, /* leal 0L(%esi),%esi */
908 0x8d,0xbf,0x00,0x00,0x00,0x00}; /* leal 0L(%edi),%edi */
909 static const char f32_13[] =
910 {0x8d,0xb6,0x00,0x00,0x00,0x00, /* leal 0L(%esi),%esi */
911 0x8d,0xbc,0x27,0x00,0x00,0x00,0x00}; /* leal 0L(%edi,1),%edi */
912 static const char f32_14[] =
913 {0x8d,0xb4,0x26,0x00,0x00,0x00,0x00, /* leal 0L(%esi,1),%esi */
914 0x8d,0xbc,0x27,0x00,0x00,0x00,0x00}; /* leal 0L(%edi,1),%edi */
915 static const char f16_3[] =
916 {0x8d,0x74,0x00}; /* lea 0(%esi),%esi */
917 static const char f16_4[] =
918 {0x8d,0xb4,0x00,0x00}; /* lea 0w(%si),%si */
919 static const char f16_5[] =
920 {0x90, /* nop */
921 0x8d,0xb4,0x00,0x00}; /* lea 0w(%si),%si */
922 static const char f16_6[] =
923 {0x89,0xf6, /* mov %si,%si */
924 0x8d,0xbd,0x00,0x00}; /* lea 0w(%di),%di */
925 static const char f16_7[] =
926 {0x8d,0x74,0x00, /* lea 0(%si),%si */
927 0x8d,0xbd,0x00,0x00}; /* lea 0w(%di),%di */
928 static const char f16_8[] =
929 {0x8d,0xb4,0x00,0x00, /* lea 0w(%si),%si */
930 0x8d,0xbd,0x00,0x00}; /* lea 0w(%di),%di */
931 static const char jump_31[] =
932 {0xeb,0x1d,0x90,0x90,0x90,0x90,0x90, /* jmp .+31; lotsa nops */
933 0x90,0x90,0x90,0x90,0x90,0x90,0x90,0x90,
934 0x90,0x90,0x90,0x90,0x90,0x90,0x90,0x90,
935 0x90,0x90,0x90,0x90,0x90,0x90,0x90,0x90};
936 static const char *const f32_patt[] = {
937 f32_1, f32_2, f32_3, f32_4, f32_5, f32_6, f32_7, f32_8,
938 f32_9, f32_10, f32_11, f32_12, f32_13, f32_14
939 };
940 static const char *const f16_patt[] = {
941 f32_1, f32_2, f16_3, f16_4, f16_5, f16_6, f16_7, f16_8
942 };
943 /* nopl (%[re]ax) */
944 static const char alt_3[] =
945 {0x0f,0x1f,0x00};
946 /* nopl 0(%[re]ax) */
947 static const char alt_4[] =
948 {0x0f,0x1f,0x40,0x00};
949 /* nopl 0(%[re]ax,%[re]ax,1) */
950 static const char alt_5[] =
951 {0x0f,0x1f,0x44,0x00,0x00};
952 /* nopw 0(%[re]ax,%[re]ax,1) */
953 static const char alt_6[] =
954 {0x66,0x0f,0x1f,0x44,0x00,0x00};
955 /* nopl 0L(%[re]ax) */
956 static const char alt_7[] =
957 {0x0f,0x1f,0x80,0x00,0x00,0x00,0x00};
958 /* nopl 0L(%[re]ax,%[re]ax,1) */
959 static const char alt_8[] =
960 {0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
961 /* nopw 0L(%[re]ax,%[re]ax,1) */
962 static const char alt_9[] =
963 {0x66,0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
964 /* nopw %cs:0L(%[re]ax,%[re]ax,1) */
965 static const char alt_10[] =
966 {0x66,0x2e,0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
967 /* data16
968 nopw %cs:0L(%[re]ax,%[re]ax,1) */
969 static const char alt_long_11[] =
970 {0x66,
971 0x66,0x2e,0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
972 /* data16
973 data16
974 nopw %cs:0L(%[re]ax,%[re]ax,1) */
975 static const char alt_long_12[] =
976 {0x66,
977 0x66,
978 0x66,0x2e,0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
979 /* data16
980 data16
981 data16
982 nopw %cs:0L(%[re]ax,%[re]ax,1) */
983 static const char alt_long_13[] =
984 {0x66,
985 0x66,
986 0x66,
987 0x66,0x2e,0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
988 /* data16
989 data16
990 data16
991 data16
992 nopw %cs:0L(%[re]ax,%[re]ax,1) */
993 static const char alt_long_14[] =
994 {0x66,
995 0x66,
996 0x66,
997 0x66,
998 0x66,0x2e,0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
999 /* data16
1000 data16
1001 data16
1002 data16
1003 data16
1004 nopw %cs:0L(%[re]ax,%[re]ax,1) */
1005 static const char alt_long_15[] =
1006 {0x66,
1007 0x66,
1008 0x66,
1009 0x66,
1010 0x66,
1011 0x66,0x2e,0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
1012 /* nopl 0(%[re]ax,%[re]ax,1)
1013 nopw 0(%[re]ax,%[re]ax,1) */
1014 static const char alt_short_11[] =
1015 {0x0f,0x1f,0x44,0x00,0x00,
1016 0x66,0x0f,0x1f,0x44,0x00,0x00};
1017 /* nopw 0(%[re]ax,%[re]ax,1)
1018 nopw 0(%[re]ax,%[re]ax,1) */
1019 static const char alt_short_12[] =
1020 {0x66,0x0f,0x1f,0x44,0x00,0x00,
1021 0x66,0x0f,0x1f,0x44,0x00,0x00};
1022 /* nopw 0(%[re]ax,%[re]ax,1)
1023 nopl 0L(%[re]ax) */
1024 static const char alt_short_13[] =
1025 {0x66,0x0f,0x1f,0x44,0x00,0x00,
1026 0x0f,0x1f,0x80,0x00,0x00,0x00,0x00};
1027 /* nopl 0L(%[re]ax)
1028 nopl 0L(%[re]ax) */
1029 static const char alt_short_14[] =
1030 {0x0f,0x1f,0x80,0x00,0x00,0x00,0x00,
1031 0x0f,0x1f,0x80,0x00,0x00,0x00,0x00};
1032 /* nopl 0L(%[re]ax)
1033 nopl 0L(%[re]ax,%[re]ax,1) */
1034 static const char alt_short_15[] =
1035 {0x0f,0x1f,0x80,0x00,0x00,0x00,0x00,
1036 0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
1037 static const char *const alt_short_patt[] = {
1038 f32_1, f32_2, alt_3, alt_4, alt_5, alt_6, alt_7, alt_8,
1039 alt_9, alt_10, alt_short_11, alt_short_12, alt_short_13,
1040 alt_short_14, alt_short_15
1041 };
1042 static const char *const alt_long_patt[] = {
1043 f32_1, f32_2, alt_3, alt_4, alt_5, alt_6, alt_7, alt_8,
1044 alt_9, alt_10, alt_long_11, alt_long_12, alt_long_13,
1045 alt_long_14, alt_long_15
1046 };
1047
1048 /* Only align for at least a positive non-zero boundary. */
1049 if (count <= 0 || count > MAX_MEM_FOR_RS_ALIGN_CODE)
1050 return;
1051
1052 /* We need to decide which NOP sequence to use for 32bit and
1053 64bit. When -mtune= is used:
1054
1055 1. For PROCESSOR_I386, PROCESSOR_I486, PROCESSOR_PENTIUM and
1056 PROCESSOR_GENERIC32, f32_patt will be used.
1057 2. For PROCESSOR_PENTIUMPRO, PROCESSOR_PENTIUM4, PROCESSOR_NOCONA,
1058 PROCESSOR_CORE, PROCESSOR_CORE2, PROCESSOR_COREI7, and
1059 PROCESSOR_GENERIC64, alt_long_patt will be used.
1060 3. For PROCESSOR_ATHLON, PROCESSOR_K6, PROCESSOR_K8 and
1061 PROCESSOR_AMDFAM10, and PROCESSOR_BD, alt_short_patt
1062 will be used.
1063
1064 When -mtune= isn't used, alt_long_patt will be used if
1065 cpu_arch_isa_flags has CpuNop. Otherwise, f32_patt will
1066 be used.
1067
1068 When -march= or .arch is used, we can't use anything beyond
1069 cpu_arch_isa_flags. */
1070
1071 if (flag_code == CODE_16BIT)
1072 {
1073 if (count > 8)
1074 {
1075 memcpy (fragP->fr_literal + fragP->fr_fix,
1076 jump_31, count);
1077 /* Adjust jump offset. */
1078 fragP->fr_literal[fragP->fr_fix + 1] = count - 2;
1079 }
1080 else
1081 memcpy (fragP->fr_literal + fragP->fr_fix,
1082 f16_patt[count - 1], count);
1083 }
1084 else
1085 {
1086 const char *const *patt = NULL;
1087
1088 if (fragP->tc_frag_data.isa == PROCESSOR_UNKNOWN)
1089 {
1090 /* PROCESSOR_UNKNOWN means that all ISAs may be used. */
1091 switch (cpu_arch_tune)
1092 {
1093 case PROCESSOR_UNKNOWN:
1094 /* We use cpu_arch_isa_flags to check if we SHOULD
1095 optimize with nops. */
1096 if (fragP->tc_frag_data.isa_flags.bitfield.cpunop)
1097 patt = alt_long_patt;
1098 else
1099 patt = f32_patt;
1100 break;
1101 case PROCESSOR_PENTIUM4:
1102 case PROCESSOR_NOCONA:
1103 case PROCESSOR_CORE:
1104 case PROCESSOR_CORE2:
1105 case PROCESSOR_COREI7:
1106 case PROCESSOR_L1OM:
1107 case PROCESSOR_K1OM:
1108 case PROCESSOR_GENERIC64:
1109 patt = alt_long_patt;
1110 break;
1111 case PROCESSOR_K6:
1112 case PROCESSOR_ATHLON:
1113 case PROCESSOR_K8:
1114 case PROCESSOR_AMDFAM10:
1115 case PROCESSOR_BD:
1116 patt = alt_short_patt;
1117 break;
1118 case PROCESSOR_I386:
1119 case PROCESSOR_I486:
1120 case PROCESSOR_PENTIUM:
1121 case PROCESSOR_PENTIUMPRO:
1122 case PROCESSOR_GENERIC32:
1123 patt = f32_patt;
1124 break;
1125 }
1126 }
1127 else
1128 {
1129 switch (fragP->tc_frag_data.tune)
1130 {
1131 case PROCESSOR_UNKNOWN:
1132 /* When cpu_arch_isa is set, cpu_arch_tune shouldn't be
1133 PROCESSOR_UNKNOWN. */
1134 abort ();
1135 break;
1136
1137 case PROCESSOR_I386:
1138 case PROCESSOR_I486:
1139 case PROCESSOR_PENTIUM:
1140 case PROCESSOR_K6:
1141 case PROCESSOR_ATHLON:
1142 case PROCESSOR_K8:
1143 case PROCESSOR_AMDFAM10:
1144 case PROCESSOR_BD:
1145 case PROCESSOR_GENERIC32:
1146 /* We use cpu_arch_isa_flags to check if we CAN optimize
1147 with nops. */
1148 if (fragP->tc_frag_data.isa_flags.bitfield.cpunop)
1149 patt = alt_short_patt;
1150 else
1151 patt = f32_patt;
1152 break;
1153 case PROCESSOR_PENTIUMPRO:
1154 case PROCESSOR_PENTIUM4:
1155 case PROCESSOR_NOCONA:
1156 case PROCESSOR_CORE:
1157 case PROCESSOR_CORE2:
1158 case PROCESSOR_COREI7:
1159 case PROCESSOR_L1OM:
1160 case PROCESSOR_K1OM:
1161 if (fragP->tc_frag_data.isa_flags.bitfield.cpunop)
1162 patt = alt_long_patt;
1163 else
1164 patt = f32_patt;
1165 break;
1166 case PROCESSOR_GENERIC64:
1167 patt = alt_long_patt;
1168 break;
1169 }
1170 }
1171
1172 if (patt == f32_patt)
1173 {
1174 /* If the padding is less than 15 bytes, we use the normal
1175 ones. Otherwise, we use a jump instruction and adjust
1176 its offset. */
1177 int limit;
1178
1179 /* For 64bit, the limit is 3 bytes. */
1180 if (flag_code == CODE_64BIT
1181 && fragP->tc_frag_data.isa_flags.bitfield.cpulm)
1182 limit = 3;
1183 else
1184 limit = 15;
1185 if (count < limit)
1186 memcpy (fragP->fr_literal + fragP->fr_fix,
1187 patt[count - 1], count);
1188 else
1189 {
1190 memcpy (fragP->fr_literal + fragP->fr_fix,
1191 jump_31, count);
1192 /* Adjust jump offset. */
1193 fragP->fr_literal[fragP->fr_fix + 1] = count - 2;
1194 }
1195 }
1196 else
1197 {
1198 /* Maximum length of an instruction is 15 byte. If the
1199 padding is greater than 15 bytes and we don't use jump,
1200 we have to break it into smaller pieces. */
1201 int padding = count;
1202 while (padding > 15)
1203 {
1204 padding -= 15;
1205 memcpy (fragP->fr_literal + fragP->fr_fix + padding,
1206 patt [14], 15);
1207 }
1208
1209 if (padding)
1210 memcpy (fragP->fr_literal + fragP->fr_fix,
1211 patt [padding - 1], padding);
1212 }
1213 }
1214 fragP->fr_var = count;
1215 }
1216
1217 static INLINE int
1218 operand_type_all_zero (const union i386_operand_type *x)
1219 {
1220 switch (ARRAY_SIZE(x->array))
1221 {
1222 case 3:
1223 if (x->array[2])
1224 return 0;
1225 case 2:
1226 if (x->array[1])
1227 return 0;
1228 case 1:
1229 return !x->array[0];
1230 default:
1231 abort ();
1232 }
1233 }
1234
1235 static INLINE void
1236 operand_type_set (union i386_operand_type *x, unsigned int v)
1237 {
1238 switch (ARRAY_SIZE(x->array))
1239 {
1240 case 3:
1241 x->array[2] = v;
1242 case 2:
1243 x->array[1] = v;
1244 case 1:
1245 x->array[0] = v;
1246 break;
1247 default:
1248 abort ();
1249 }
1250 }
1251
1252 static INLINE int
1253 operand_type_equal (const union i386_operand_type *x,
1254 const union i386_operand_type *y)
1255 {
1256 switch (ARRAY_SIZE(x->array))
1257 {
1258 case 3:
1259 if (x->array[2] != y->array[2])
1260 return 0;
1261 case 2:
1262 if (x->array[1] != y->array[1])
1263 return 0;
1264 case 1:
1265 return x->array[0] == y->array[0];
1266 break;
1267 default:
1268 abort ();
1269 }
1270 }
1271
1272 static INLINE int
1273 cpu_flags_all_zero (const union i386_cpu_flags *x)
1274 {
1275 switch (ARRAY_SIZE(x->array))
1276 {
1277 case 3:
1278 if (x->array[2])
1279 return 0;
1280 case 2:
1281 if (x->array[1])
1282 return 0;
1283 case 1:
1284 return !x->array[0];
1285 default:
1286 abort ();
1287 }
1288 }
1289
1290 static INLINE void
1291 cpu_flags_set (union i386_cpu_flags *x, unsigned int v)
1292 {
1293 switch (ARRAY_SIZE(x->array))
1294 {
1295 case 3:
1296 x->array[2] = v;
1297 case 2:
1298 x->array[1] = v;
1299 case 1:
1300 x->array[0] = v;
1301 break;
1302 default:
1303 abort ();
1304 }
1305 }
1306
1307 static INLINE int
1308 cpu_flags_equal (const union i386_cpu_flags *x,
1309 const union i386_cpu_flags *y)
1310 {
1311 switch (ARRAY_SIZE(x->array))
1312 {
1313 case 3:
1314 if (x->array[2] != y->array[2])
1315 return 0;
1316 case 2:
1317 if (x->array[1] != y->array[1])
1318 return 0;
1319 case 1:
1320 return x->array[0] == y->array[0];
1321 break;
1322 default:
1323 abort ();
1324 }
1325 }
1326
1327 static INLINE int
1328 cpu_flags_check_cpu64 (i386_cpu_flags f)
1329 {
1330 return !((flag_code == CODE_64BIT && f.bitfield.cpuno64)
1331 || (flag_code != CODE_64BIT && f.bitfield.cpu64));
1332 }
1333
1334 static INLINE i386_cpu_flags
1335 cpu_flags_and (i386_cpu_flags x, i386_cpu_flags y)
1336 {
1337 switch (ARRAY_SIZE (x.array))
1338 {
1339 case 3:
1340 x.array [2] &= y.array [2];
1341 case 2:
1342 x.array [1] &= y.array [1];
1343 case 1:
1344 x.array [0] &= y.array [0];
1345 break;
1346 default:
1347 abort ();
1348 }
1349 return x;
1350 }
1351
1352 static INLINE i386_cpu_flags
1353 cpu_flags_or (i386_cpu_flags x, i386_cpu_flags y)
1354 {
1355 switch (ARRAY_SIZE (x.array))
1356 {
1357 case 3:
1358 x.array [2] |= y.array [2];
1359 case 2:
1360 x.array [1] |= y.array [1];
1361 case 1:
1362 x.array [0] |= y.array [0];
1363 break;
1364 default:
1365 abort ();
1366 }
1367 return x;
1368 }
1369
1370 static INLINE i386_cpu_flags
1371 cpu_flags_and_not (i386_cpu_flags x, i386_cpu_flags y)
1372 {
1373 switch (ARRAY_SIZE (x.array))
1374 {
1375 case 3:
1376 x.array [2] &= ~y.array [2];
1377 case 2:
1378 x.array [1] &= ~y.array [1];
1379 case 1:
1380 x.array [0] &= ~y.array [0];
1381 break;
1382 default:
1383 abort ();
1384 }
1385 return x;
1386 }
1387
1388 #define CPU_FLAGS_ARCH_MATCH 0x1
1389 #define CPU_FLAGS_64BIT_MATCH 0x2
1390 #define CPU_FLAGS_AES_MATCH 0x4
1391 #define CPU_FLAGS_PCLMUL_MATCH 0x8
1392 #define CPU_FLAGS_AVX_MATCH 0x10
1393
1394 #define CPU_FLAGS_32BIT_MATCH \
1395 (CPU_FLAGS_ARCH_MATCH | CPU_FLAGS_AES_MATCH \
1396 | CPU_FLAGS_PCLMUL_MATCH | CPU_FLAGS_AVX_MATCH)
1397 #define CPU_FLAGS_PERFECT_MATCH \
1398 (CPU_FLAGS_32BIT_MATCH | CPU_FLAGS_64BIT_MATCH)
1399
1400 /* Return CPU flags match bits. */
1401
1402 static int
1403 cpu_flags_match (const insn_template *t)
1404 {
1405 i386_cpu_flags x = t->cpu_flags;
1406 int match = cpu_flags_check_cpu64 (x) ? CPU_FLAGS_64BIT_MATCH : 0;
1407
1408 x.bitfield.cpu64 = 0;
1409 x.bitfield.cpuno64 = 0;
1410
1411 if (cpu_flags_all_zero (&x))
1412 {
1413 /* This instruction is available on all archs. */
1414 match |= CPU_FLAGS_32BIT_MATCH;
1415 }
1416 else
1417 {
1418 /* This instruction is available only on some archs. */
1419 i386_cpu_flags cpu = cpu_arch_flags;
1420
1421 cpu.bitfield.cpu64 = 0;
1422 cpu.bitfield.cpuno64 = 0;
1423 cpu = cpu_flags_and (x, cpu);
1424 if (!cpu_flags_all_zero (&cpu))
1425 {
1426 if (x.bitfield.cpuavx)
1427 {
1428 /* We only need to check AES/PCLMUL/SSE2AVX with AVX. */
1429 if (cpu.bitfield.cpuavx)
1430 {
1431 /* Check SSE2AVX. */
1432 if (!t->opcode_modifier.sse2avx|| sse2avx)
1433 {
1434 match |= (CPU_FLAGS_ARCH_MATCH
1435 | CPU_FLAGS_AVX_MATCH);
1436 /* Check AES. */
1437 if (!x.bitfield.cpuaes || cpu.bitfield.cpuaes)
1438 match |= CPU_FLAGS_AES_MATCH;
1439 /* Check PCLMUL. */
1440 if (!x.bitfield.cpupclmul
1441 || cpu.bitfield.cpupclmul)
1442 match |= CPU_FLAGS_PCLMUL_MATCH;
1443 }
1444 }
1445 else
1446 match |= CPU_FLAGS_ARCH_MATCH;
1447 }
1448 else
1449 match |= CPU_FLAGS_32BIT_MATCH;
1450 }
1451 }
1452 return match;
1453 }
1454
1455 static INLINE i386_operand_type
1456 operand_type_and (i386_operand_type x, i386_operand_type y)
1457 {
1458 switch (ARRAY_SIZE (x.array))
1459 {
1460 case 3:
1461 x.array [2] &= y.array [2];
1462 case 2:
1463 x.array [1] &= y.array [1];
1464 case 1:
1465 x.array [0] &= y.array [0];
1466 break;
1467 default:
1468 abort ();
1469 }
1470 return x;
1471 }
1472
1473 static INLINE i386_operand_type
1474 operand_type_or (i386_operand_type x, i386_operand_type y)
1475 {
1476 switch (ARRAY_SIZE (x.array))
1477 {
1478 case 3:
1479 x.array [2] |= y.array [2];
1480 case 2:
1481 x.array [1] |= y.array [1];
1482 case 1:
1483 x.array [0] |= y.array [0];
1484 break;
1485 default:
1486 abort ();
1487 }
1488 return x;
1489 }
1490
1491 static INLINE i386_operand_type
1492 operand_type_xor (i386_operand_type x, i386_operand_type y)
1493 {
1494 switch (ARRAY_SIZE (x.array))
1495 {
1496 case 3:
1497 x.array [2] ^= y.array [2];
1498 case 2:
1499 x.array [1] ^= y.array [1];
1500 case 1:
1501 x.array [0] ^= y.array [0];
1502 break;
1503 default:
1504 abort ();
1505 }
1506 return x;
1507 }
1508
1509 static const i386_operand_type acc32 = OPERAND_TYPE_ACC32;
1510 static const i386_operand_type acc64 = OPERAND_TYPE_ACC64;
1511 static const i386_operand_type control = OPERAND_TYPE_CONTROL;
1512 static const i386_operand_type inoutportreg
1513 = OPERAND_TYPE_INOUTPORTREG;
1514 static const i386_operand_type reg16_inoutportreg
1515 = OPERAND_TYPE_REG16_INOUTPORTREG;
1516 static const i386_operand_type disp16 = OPERAND_TYPE_DISP16;
1517 static const i386_operand_type disp32 = OPERAND_TYPE_DISP32;
1518 static const i386_operand_type disp32s = OPERAND_TYPE_DISP32S;
1519 static const i386_operand_type disp16_32 = OPERAND_TYPE_DISP16_32;
1520 static const i386_operand_type anydisp
1521 = OPERAND_TYPE_ANYDISP;
1522 static const i386_operand_type regxmm = OPERAND_TYPE_REGXMM;
1523 static const i386_operand_type regymm = OPERAND_TYPE_REGYMM;
1524 static const i386_operand_type imm8 = OPERAND_TYPE_IMM8;
1525 static const i386_operand_type imm8s = OPERAND_TYPE_IMM8S;
1526 static const i386_operand_type imm16 = OPERAND_TYPE_IMM16;
1527 static const i386_operand_type imm32 = OPERAND_TYPE_IMM32;
1528 static const i386_operand_type imm32s = OPERAND_TYPE_IMM32S;
1529 static const i386_operand_type imm64 = OPERAND_TYPE_IMM64;
1530 static const i386_operand_type imm16_32 = OPERAND_TYPE_IMM16_32;
1531 static const i386_operand_type imm16_32s = OPERAND_TYPE_IMM16_32S;
1532 static const i386_operand_type imm16_32_32s = OPERAND_TYPE_IMM16_32_32S;
1533 static const i386_operand_type vec_imm4 = OPERAND_TYPE_VEC_IMM4;
1534
1535 enum operand_type
1536 {
1537 reg,
1538 imm,
1539 disp,
1540 anymem
1541 };
1542
1543 static INLINE int
1544 operand_type_check (i386_operand_type t, enum operand_type c)
1545 {
1546 switch (c)
1547 {
1548 case reg:
1549 return (t.bitfield.reg8
1550 || t.bitfield.reg16
1551 || t.bitfield.reg32
1552 || t.bitfield.reg64);
1553
1554 case imm:
1555 return (t.bitfield.imm8
1556 || t.bitfield.imm8s
1557 || t.bitfield.imm16
1558 || t.bitfield.imm32
1559 || t.bitfield.imm32s
1560 || t.bitfield.imm64);
1561
1562 case disp:
1563 return (t.bitfield.disp8
1564 || t.bitfield.disp16
1565 || t.bitfield.disp32
1566 || t.bitfield.disp32s
1567 || t.bitfield.disp64);
1568
1569 case anymem:
1570 return (t.bitfield.disp8
1571 || t.bitfield.disp16
1572 || t.bitfield.disp32
1573 || t.bitfield.disp32s
1574 || t.bitfield.disp64
1575 || t.bitfield.baseindex);
1576
1577 default:
1578 abort ();
1579 }
1580
1581 return 0;
1582 }
1583
1584 /* Return 1 if there is no conflict in 8bit/16bit/32bit/64bit on
1585 operand J for instruction template T. */
1586
1587 static INLINE int
1588 match_reg_size (const insn_template *t, unsigned int j)
1589 {
1590 return !((i.types[j].bitfield.byte
1591 && !t->operand_types[j].bitfield.byte)
1592 || (i.types[j].bitfield.word
1593 && !t->operand_types[j].bitfield.word)
1594 || (i.types[j].bitfield.dword
1595 && !t->operand_types[j].bitfield.dword)
1596 || (i.types[j].bitfield.qword
1597 && !t->operand_types[j].bitfield.qword));
1598 }
1599
1600 /* Return 1 if there is no conflict in any size on operand J for
1601 instruction template T. */
1602
1603 static INLINE int
1604 match_mem_size (const insn_template *t, unsigned int j)
1605 {
1606 return (match_reg_size (t, j)
1607 && !((i.types[j].bitfield.unspecified
1608 && !t->operand_types[j].bitfield.unspecified)
1609 || (i.types[j].bitfield.fword
1610 && !t->operand_types[j].bitfield.fword)
1611 || (i.types[j].bitfield.tbyte
1612 && !t->operand_types[j].bitfield.tbyte)
1613 || (i.types[j].bitfield.xmmword
1614 && !t->operand_types[j].bitfield.xmmword)
1615 || (i.types[j].bitfield.ymmword
1616 && !t->operand_types[j].bitfield.ymmword)));
1617 }
1618
1619 /* Return 1 if there is no size conflict on any operands for
1620 instruction template T. */
1621
1622 static INLINE int
1623 operand_size_match (const insn_template *t)
1624 {
1625 unsigned int j;
1626 int match = 1;
1627
1628 /* Don't check jump instructions. */
1629 if (t->opcode_modifier.jump
1630 || t->opcode_modifier.jumpbyte
1631 || t->opcode_modifier.jumpdword
1632 || t->opcode_modifier.jumpintersegment)
1633 return match;
1634
1635 /* Check memory and accumulator operand size. */
1636 for (j = 0; j < i.operands; j++)
1637 {
1638 if (t->operand_types[j].bitfield.anysize)
1639 continue;
1640
1641 if (t->operand_types[j].bitfield.acc && !match_reg_size (t, j))
1642 {
1643 match = 0;
1644 break;
1645 }
1646
1647 if (i.types[j].bitfield.mem && !match_mem_size (t, j))
1648 {
1649 match = 0;
1650 break;
1651 }
1652 }
1653
1654 if (match)
1655 return match;
1656 else if (!t->opcode_modifier.d && !t->opcode_modifier.floatd)
1657 {
1658 mismatch:
1659 i.error = operand_size_mismatch;
1660 return 0;
1661 }
1662
1663 /* Check reverse. */
1664 gas_assert (i.operands == 2);
1665
1666 match = 1;
1667 for (j = 0; j < 2; j++)
1668 {
1669 if (t->operand_types[j].bitfield.acc
1670 && !match_reg_size (t, j ? 0 : 1))
1671 goto mismatch;
1672
1673 if (i.types[j].bitfield.mem
1674 && !match_mem_size (t, j ? 0 : 1))
1675 goto mismatch;
1676 }
1677
1678 return match;
1679 }
1680
1681 static INLINE int
1682 operand_type_match (i386_operand_type overlap,
1683 i386_operand_type given)
1684 {
1685 i386_operand_type temp = overlap;
1686
1687 temp.bitfield.jumpabsolute = 0;
1688 temp.bitfield.unspecified = 0;
1689 temp.bitfield.byte = 0;
1690 temp.bitfield.word = 0;
1691 temp.bitfield.dword = 0;
1692 temp.bitfield.fword = 0;
1693 temp.bitfield.qword = 0;
1694 temp.bitfield.tbyte = 0;
1695 temp.bitfield.xmmword = 0;
1696 temp.bitfield.ymmword = 0;
1697 if (operand_type_all_zero (&temp))
1698 goto mismatch;
1699
1700 if (given.bitfield.baseindex == overlap.bitfield.baseindex
1701 && given.bitfield.jumpabsolute == overlap.bitfield.jumpabsolute)
1702 return 1;
1703
1704 mismatch:
1705 i.error = operand_type_mismatch;
1706 return 0;
1707 }
1708
1709 /* If given types g0 and g1 are registers they must be of the same type
1710 unless the expected operand type register overlap is null.
1711 Note that Acc in a template matches every size of reg. */
1712
1713 static INLINE int
1714 operand_type_register_match (i386_operand_type m0,
1715 i386_operand_type g0,
1716 i386_operand_type t0,
1717 i386_operand_type m1,
1718 i386_operand_type g1,
1719 i386_operand_type t1)
1720 {
1721 if (!operand_type_check (g0, reg))
1722 return 1;
1723
1724 if (!operand_type_check (g1, reg))
1725 return 1;
1726
1727 if (g0.bitfield.reg8 == g1.bitfield.reg8
1728 && g0.bitfield.reg16 == g1.bitfield.reg16
1729 && g0.bitfield.reg32 == g1.bitfield.reg32
1730 && g0.bitfield.reg64 == g1.bitfield.reg64)
1731 return 1;
1732
1733 if (m0.bitfield.acc)
1734 {
1735 t0.bitfield.reg8 = 1;
1736 t0.bitfield.reg16 = 1;
1737 t0.bitfield.reg32 = 1;
1738 t0.bitfield.reg64 = 1;
1739 }
1740
1741 if (m1.bitfield.acc)
1742 {
1743 t1.bitfield.reg8 = 1;
1744 t1.bitfield.reg16 = 1;
1745 t1.bitfield.reg32 = 1;
1746 t1.bitfield.reg64 = 1;
1747 }
1748
1749 if (!(t0.bitfield.reg8 & t1.bitfield.reg8)
1750 && !(t0.bitfield.reg16 & t1.bitfield.reg16)
1751 && !(t0.bitfield.reg32 & t1.bitfield.reg32)
1752 && !(t0.bitfield.reg64 & t1.bitfield.reg64))
1753 return 1;
1754
1755 i.error = register_type_mismatch;
1756
1757 return 0;
1758 }
1759
1760 static INLINE unsigned int
1761 mode_from_disp_size (i386_operand_type t)
1762 {
1763 if (t.bitfield.disp8)
1764 return 1;
1765 else if (t.bitfield.disp16
1766 || t.bitfield.disp32
1767 || t.bitfield.disp32s)
1768 return 2;
1769 else
1770 return 0;
1771 }
1772
1773 static INLINE int
1774 fits_in_signed_byte (offsetT num)
1775 {
1776 return (num >= -128) && (num <= 127);
1777 }
1778
1779 static INLINE int
1780 fits_in_unsigned_byte (offsetT num)
1781 {
1782 return (num & 0xff) == num;
1783 }
1784
1785 static INLINE int
1786 fits_in_unsigned_word (offsetT num)
1787 {
1788 return (num & 0xffff) == num;
1789 }
1790
1791 static INLINE int
1792 fits_in_signed_word (offsetT num)
1793 {
1794 return (-32768 <= num) && (num <= 32767);
1795 }
1796
1797 static INLINE int
1798 fits_in_signed_long (offsetT num ATTRIBUTE_UNUSED)
1799 {
1800 #ifndef BFD64
1801 return 1;
1802 #else
1803 return (!(((offsetT) -1 << 31) & num)
1804 || (((offsetT) -1 << 31) & num) == ((offsetT) -1 << 31));
1805 #endif
1806 } /* fits_in_signed_long() */
1807
1808 static INLINE int
1809 fits_in_unsigned_long (offsetT num ATTRIBUTE_UNUSED)
1810 {
1811 #ifndef BFD64
1812 return 1;
1813 #else
1814 return (num & (((offsetT) 2 << 31) - 1)) == num;
1815 #endif
1816 } /* fits_in_unsigned_long() */
1817
1818 static INLINE int
1819 fits_in_imm4 (offsetT num)
1820 {
1821 return (num & 0xf) == num;
1822 }
1823
1824 static i386_operand_type
1825 smallest_imm_type (offsetT num)
1826 {
1827 i386_operand_type t;
1828
1829 operand_type_set (&t, 0);
1830 t.bitfield.imm64 = 1;
1831
1832 if (cpu_arch_tune != PROCESSOR_I486 && num == 1)
1833 {
1834 /* This code is disabled on the 486 because all the Imm1 forms
1835 in the opcode table are slower on the i486. They're the
1836 versions with the implicitly specified single-position
1837 displacement, which has another syntax if you really want to
1838 use that form. */
1839 t.bitfield.imm1 = 1;
1840 t.bitfield.imm8 = 1;
1841 t.bitfield.imm8s = 1;
1842 t.bitfield.imm16 = 1;
1843 t.bitfield.imm32 = 1;
1844 t.bitfield.imm32s = 1;
1845 }
1846 else if (fits_in_signed_byte (num))
1847 {
1848 t.bitfield.imm8 = 1;
1849 t.bitfield.imm8s = 1;
1850 t.bitfield.imm16 = 1;
1851 t.bitfield.imm32 = 1;
1852 t.bitfield.imm32s = 1;
1853 }
1854 else if (fits_in_unsigned_byte (num))
1855 {
1856 t.bitfield.imm8 = 1;
1857 t.bitfield.imm16 = 1;
1858 t.bitfield.imm32 = 1;
1859 t.bitfield.imm32s = 1;
1860 }
1861 else if (fits_in_signed_word (num) || fits_in_unsigned_word (num))
1862 {
1863 t.bitfield.imm16 = 1;
1864 t.bitfield.imm32 = 1;
1865 t.bitfield.imm32s = 1;
1866 }
1867 else if (fits_in_signed_long (num))
1868 {
1869 t.bitfield.imm32 = 1;
1870 t.bitfield.imm32s = 1;
1871 }
1872 else if (fits_in_unsigned_long (num))
1873 t.bitfield.imm32 = 1;
1874
1875 return t;
1876 }
1877
1878 static offsetT
1879 offset_in_range (offsetT val, int size)
1880 {
1881 addressT mask;
1882
1883 switch (size)
1884 {
1885 case 1: mask = ((addressT) 1 << 8) - 1; break;
1886 case 2: mask = ((addressT) 1 << 16) - 1; break;
1887 case 4: mask = ((addressT) 2 << 31) - 1; break;
1888 #ifdef BFD64
1889 case 8: mask = ((addressT) 2 << 63) - 1; break;
1890 #endif
1891 default: abort ();
1892 }
1893
1894 #ifdef BFD64
1895 /* If BFD64, sign extend val for 32bit address mode. */
1896 if (flag_code != CODE_64BIT
1897 || i.prefix[ADDR_PREFIX])
1898 if ((val & ~(((addressT) 2 << 31) - 1)) == 0)
1899 val = (val ^ ((addressT) 1 << 31)) - ((addressT) 1 << 31);
1900 #endif
1901
1902 if ((val & ~mask) != 0 && (val & ~mask) != ~mask)
1903 {
1904 char buf1[40], buf2[40];
1905
1906 sprint_value (buf1, val);
1907 sprint_value (buf2, val & mask);
1908 as_warn (_("%s shortened to %s"), buf1, buf2);
1909 }
1910 return val & mask;
1911 }
1912
1913 enum PREFIX_GROUP
1914 {
1915 PREFIX_EXIST = 0,
1916 PREFIX_LOCK,
1917 PREFIX_REP,
1918 PREFIX_OTHER
1919 };
1920
1921 /* Returns
1922 a. PREFIX_EXIST if attempting to add a prefix where one from the
1923 same class already exists.
1924 b. PREFIX_LOCK if lock prefix is added.
1925 c. PREFIX_REP if rep/repne prefix is added.
1926 d. PREFIX_OTHER if other prefix is added.
1927 */
1928
1929 static enum PREFIX_GROUP
1930 add_prefix (unsigned int prefix)
1931 {
1932 enum PREFIX_GROUP ret = PREFIX_OTHER;
1933 unsigned int q;
1934
1935 if (prefix >= REX_OPCODE && prefix < REX_OPCODE + 16
1936 && flag_code == CODE_64BIT)
1937 {
1938 if ((i.prefix[REX_PREFIX] & prefix & REX_W)
1939 || ((i.prefix[REX_PREFIX] & (REX_R | REX_X | REX_B))
1940 && (prefix & (REX_R | REX_X | REX_B))))
1941 ret = PREFIX_EXIST;
1942 q = REX_PREFIX;
1943 }
1944 else
1945 {
1946 switch (prefix)
1947 {
1948 default:
1949 abort ();
1950
1951 case CS_PREFIX_OPCODE:
1952 case DS_PREFIX_OPCODE:
1953 case ES_PREFIX_OPCODE:
1954 case FS_PREFIX_OPCODE:
1955 case GS_PREFIX_OPCODE:
1956 case SS_PREFIX_OPCODE:
1957 q = SEG_PREFIX;
1958 break;
1959
1960 case REPNE_PREFIX_OPCODE:
1961 case REPE_PREFIX_OPCODE:
1962 q = REP_PREFIX;
1963 ret = PREFIX_REP;
1964 break;
1965
1966 case LOCK_PREFIX_OPCODE:
1967 q = LOCK_PREFIX;
1968 ret = PREFIX_LOCK;
1969 break;
1970
1971 case FWAIT_OPCODE:
1972 q = WAIT_PREFIX;
1973 break;
1974
1975 case ADDR_PREFIX_OPCODE:
1976 q = ADDR_PREFIX;
1977 break;
1978
1979 case DATA_PREFIX_OPCODE:
1980 q = DATA_PREFIX;
1981 break;
1982 }
1983 if (i.prefix[q] != 0)
1984 ret = PREFIX_EXIST;
1985 }
1986
1987 if (ret)
1988 {
1989 if (!i.prefix[q])
1990 ++i.prefixes;
1991 i.prefix[q] |= prefix;
1992 }
1993 else
1994 as_bad (_("same type of prefix used twice"));
1995
1996 return ret;
1997 }
1998
1999 static void
2000 update_code_flag (int value, int check)
2001 {
2002 PRINTF_LIKE ((*as_error));
2003
2004 flag_code = (enum flag_code) value;
2005 if (flag_code == CODE_64BIT)
2006 {
2007 cpu_arch_flags.bitfield.cpu64 = 1;
2008 cpu_arch_flags.bitfield.cpuno64 = 0;
2009 }
2010 else
2011 {
2012 cpu_arch_flags.bitfield.cpu64 = 0;
2013 cpu_arch_flags.bitfield.cpuno64 = 1;
2014 }
2015 if (value == CODE_64BIT && !cpu_arch_flags.bitfield.cpulm )
2016 {
2017 if (check)
2018 as_error = as_fatal;
2019 else
2020 as_error = as_bad;
2021 (*as_error) (_("64bit mode not supported on `%s'."),
2022 cpu_arch_name ? cpu_arch_name : default_arch);
2023 }
2024 if (value == CODE_32BIT && !cpu_arch_flags.bitfield.cpui386)
2025 {
2026 if (check)
2027 as_error = as_fatal;
2028 else
2029 as_error = as_bad;
2030 (*as_error) (_("32bit mode not supported on `%s'."),
2031 cpu_arch_name ? cpu_arch_name : default_arch);
2032 }
2033 stackop_size = '\0';
2034 }
2035
2036 static void
2037 set_code_flag (int value)
2038 {
2039 update_code_flag (value, 0);
2040 }
2041
2042 static void
2043 set_16bit_gcc_code_flag (int new_code_flag)
2044 {
2045 flag_code = (enum flag_code) new_code_flag;
2046 if (flag_code != CODE_16BIT)
2047 abort ();
2048 cpu_arch_flags.bitfield.cpu64 = 0;
2049 cpu_arch_flags.bitfield.cpuno64 = 1;
2050 stackop_size = LONG_MNEM_SUFFIX;
2051 }
2052
2053 static void
2054 set_intel_syntax (int syntax_flag)
2055 {
2056 /* Find out if register prefixing is specified. */
2057 int ask_naked_reg = 0;
2058
2059 SKIP_WHITESPACE ();
2060 if (!is_end_of_line[(unsigned char) *input_line_pointer])
2061 {
2062 char *string = input_line_pointer;
2063 int e = get_symbol_end ();
2064
2065 if (strcmp (string, "prefix") == 0)
2066 ask_naked_reg = 1;
2067 else if (strcmp (string, "noprefix") == 0)
2068 ask_naked_reg = -1;
2069 else
2070 as_bad (_("bad argument to syntax directive."));
2071 *input_line_pointer = e;
2072 }
2073 demand_empty_rest_of_line ();
2074
2075 intel_syntax = syntax_flag;
2076
2077 if (ask_naked_reg == 0)
2078 allow_naked_reg = (intel_syntax
2079 && (bfd_get_symbol_leading_char (stdoutput) != '\0'));
2080 else
2081 allow_naked_reg = (ask_naked_reg < 0);
2082
2083 expr_set_rank (O_full_ptr, syntax_flag ? 10 : 0);
2084
2085 identifier_chars['%'] = intel_syntax && allow_naked_reg ? '%' : 0;
2086 identifier_chars['$'] = intel_syntax ? '$' : 0;
2087 register_prefix = allow_naked_reg ? "" : "%";
2088 }
2089
2090 static void
2091 set_intel_mnemonic (int mnemonic_flag)
2092 {
2093 intel_mnemonic = mnemonic_flag;
2094 }
2095
2096 static void
2097 set_allow_index_reg (int flag)
2098 {
2099 allow_index_reg = flag;
2100 }
2101
2102 static void
2103 set_sse_check (int dummy ATTRIBUTE_UNUSED)
2104 {
2105 SKIP_WHITESPACE ();
2106
2107 if (!is_end_of_line[(unsigned char) *input_line_pointer])
2108 {
2109 char *string = input_line_pointer;
2110 int e = get_symbol_end ();
2111
2112 if (strcmp (string, "none") == 0)
2113 sse_check = sse_check_none;
2114 else if (strcmp (string, "warning") == 0)
2115 sse_check = sse_check_warning;
2116 else if (strcmp (string, "error") == 0)
2117 sse_check = sse_check_error;
2118 else
2119 as_bad (_("bad argument to sse_check directive."));
2120 *input_line_pointer = e;
2121 }
2122 else
2123 as_bad (_("missing argument for sse_check directive"));
2124
2125 demand_empty_rest_of_line ();
2126 }
2127
2128 static void
2129 check_cpu_arch_compatible (const char *name ATTRIBUTE_UNUSED,
2130 i386_cpu_flags new_flag ATTRIBUTE_UNUSED)
2131 {
2132 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
2133 static const char *arch;
2134
2135 /* Intel LIOM is only supported on ELF. */
2136 if (!IS_ELF)
2137 return;
2138
2139 if (!arch)
2140 {
2141 /* Use cpu_arch_name if it is set in md_parse_option. Otherwise
2142 use default_arch. */
2143 arch = cpu_arch_name;
2144 if (!arch)
2145 arch = default_arch;
2146 }
2147
2148 /* If we are targeting Intel L1OM, we must enable it. */
2149 if (get_elf_backend_data (stdoutput)->elf_machine_code != EM_L1OM
2150 || new_flag.bitfield.cpul1om)
2151 return;
2152
2153 /* If we are targeting Intel K1OM, we must enable it. */
2154 if (get_elf_backend_data (stdoutput)->elf_machine_code != EM_K1OM
2155 || new_flag.bitfield.cpuk1om)
2156 return;
2157
2158 as_bad (_("`%s' is not supported on `%s'"), name, arch);
2159 #endif
2160 }
2161
2162 static void
2163 set_cpu_arch (int dummy ATTRIBUTE_UNUSED)
2164 {
2165 SKIP_WHITESPACE ();
2166
2167 if (!is_end_of_line[(unsigned char) *input_line_pointer])
2168 {
2169 char *string = input_line_pointer;
2170 int e = get_symbol_end ();
2171 unsigned int j;
2172 i386_cpu_flags flags;
2173
2174 for (j = 0; j < ARRAY_SIZE (cpu_arch); j++)
2175 {
2176 if (strcmp (string, cpu_arch[j].name) == 0)
2177 {
2178 check_cpu_arch_compatible (string, cpu_arch[j].flags);
2179
2180 if (*string != '.')
2181 {
2182 cpu_arch_name = cpu_arch[j].name;
2183 cpu_sub_arch_name = NULL;
2184 cpu_arch_flags = cpu_arch[j].flags;
2185 if (flag_code == CODE_64BIT)
2186 {
2187 cpu_arch_flags.bitfield.cpu64 = 1;
2188 cpu_arch_flags.bitfield.cpuno64 = 0;
2189 }
2190 else
2191 {
2192 cpu_arch_flags.bitfield.cpu64 = 0;
2193 cpu_arch_flags.bitfield.cpuno64 = 1;
2194 }
2195 cpu_arch_isa = cpu_arch[j].type;
2196 cpu_arch_isa_flags = cpu_arch[j].flags;
2197 if (!cpu_arch_tune_set)
2198 {
2199 cpu_arch_tune = cpu_arch_isa;
2200 cpu_arch_tune_flags = cpu_arch_isa_flags;
2201 }
2202 break;
2203 }
2204
2205 if (!cpu_arch[j].negated)
2206 flags = cpu_flags_or (cpu_arch_flags,
2207 cpu_arch[j].flags);
2208 else
2209 flags = cpu_flags_and_not (cpu_arch_flags,
2210 cpu_arch[j].flags);
2211 if (!cpu_flags_equal (&flags, &cpu_arch_flags))
2212 {
2213 if (cpu_sub_arch_name)
2214 {
2215 char *name = cpu_sub_arch_name;
2216 cpu_sub_arch_name = concat (name,
2217 cpu_arch[j].name,
2218 (const char *) NULL);
2219 free (name);
2220 }
2221 else
2222 cpu_sub_arch_name = xstrdup (cpu_arch[j].name);
2223 cpu_arch_flags = flags;
2224 cpu_arch_isa_flags = flags;
2225 }
2226 *input_line_pointer = e;
2227 demand_empty_rest_of_line ();
2228 return;
2229 }
2230 }
2231 if (j >= ARRAY_SIZE (cpu_arch))
2232 as_bad (_("no such architecture: `%s'"), string);
2233
2234 *input_line_pointer = e;
2235 }
2236 else
2237 as_bad (_("missing cpu architecture"));
2238
2239 no_cond_jump_promotion = 0;
2240 if (*input_line_pointer == ','
2241 && !is_end_of_line[(unsigned char) input_line_pointer[1]])
2242 {
2243 char *string = ++input_line_pointer;
2244 int e = get_symbol_end ();
2245
2246 if (strcmp (string, "nojumps") == 0)
2247 no_cond_jump_promotion = 1;
2248 else if (strcmp (string, "jumps") == 0)
2249 ;
2250 else
2251 as_bad (_("no such architecture modifier: `%s'"), string);
2252
2253 *input_line_pointer = e;
2254 }
2255
2256 demand_empty_rest_of_line ();
2257 }
2258
2259 enum bfd_architecture
2260 i386_arch (void)
2261 {
2262 if (cpu_arch_isa == PROCESSOR_L1OM)
2263 {
2264 if (OUTPUT_FLAVOR != bfd_target_elf_flavour
2265 || flag_code != CODE_64BIT)
2266 as_fatal (_("Intel L1OM is 64bit ELF only"));
2267 return bfd_arch_l1om;
2268 }
2269 else if (cpu_arch_isa == PROCESSOR_K1OM)
2270 {
2271 if (OUTPUT_FLAVOR != bfd_target_elf_flavour
2272 || flag_code != CODE_64BIT)
2273 as_fatal (_("Intel K1OM is 64bit ELF only"));
2274 return bfd_arch_k1om;
2275 }
2276 else
2277 return bfd_arch_i386;
2278 }
2279
2280 unsigned long
2281 i386_mach (void)
2282 {
2283 if (!strncmp (default_arch, "x86_64", 6))
2284 {
2285 if (cpu_arch_isa == PROCESSOR_L1OM)
2286 {
2287 if (OUTPUT_FLAVOR != bfd_target_elf_flavour
2288 || default_arch[6] != '\0')
2289 as_fatal (_("Intel L1OM is 64bit ELF only"));
2290 return bfd_mach_l1om;
2291 }
2292 else if (cpu_arch_isa == PROCESSOR_K1OM)
2293 {
2294 if (OUTPUT_FLAVOR != bfd_target_elf_flavour
2295 || default_arch[6] != '\0')
2296 as_fatal (_("Intel K1OM is 64bit ELF only"));
2297 return bfd_mach_k1om;
2298 }
2299 else if (default_arch[6] == '\0')
2300 return bfd_mach_x86_64;
2301 else
2302 return bfd_mach_x64_32;
2303 }
2304 else if (!strcmp (default_arch, "i386"))
2305 return bfd_mach_i386_i386;
2306 else
2307 as_fatal (_("unknown architecture"));
2308 }
2309 \f
2310 void
2311 md_begin (void)
2312 {
2313 const char *hash_err;
2314
2315 /* Initialize op_hash hash table. */
2316 op_hash = hash_new ();
2317
2318 {
2319 const insn_template *optab;
2320 templates *core_optab;
2321
2322 /* Setup for loop. */
2323 optab = i386_optab;
2324 core_optab = (templates *) xmalloc (sizeof (templates));
2325 core_optab->start = optab;
2326
2327 while (1)
2328 {
2329 ++optab;
2330 if (optab->name == NULL
2331 || strcmp (optab->name, (optab - 1)->name) != 0)
2332 {
2333 /* different name --> ship out current template list;
2334 add to hash table; & begin anew. */
2335 core_optab->end = optab;
2336 hash_err = hash_insert (op_hash,
2337 (optab - 1)->name,
2338 (void *) core_optab);
2339 if (hash_err)
2340 {
2341 as_fatal (_("internal Error: Can't hash %s: %s"),
2342 (optab - 1)->name,
2343 hash_err);
2344 }
2345 if (optab->name == NULL)
2346 break;
2347 core_optab = (templates *) xmalloc (sizeof (templates));
2348 core_optab->start = optab;
2349 }
2350 }
2351 }
2352
2353 /* Initialize reg_hash hash table. */
2354 reg_hash = hash_new ();
2355 {
2356 const reg_entry *regtab;
2357 unsigned int regtab_size = i386_regtab_size;
2358
2359 for (regtab = i386_regtab; regtab_size--; regtab++)
2360 {
2361 hash_err = hash_insert (reg_hash, regtab->reg_name, (void *) regtab);
2362 if (hash_err)
2363 as_fatal (_("internal Error: Can't hash %s: %s"),
2364 regtab->reg_name,
2365 hash_err);
2366 }
2367 }
2368
2369 /* Fill in lexical tables: mnemonic_chars, operand_chars. */
2370 {
2371 int c;
2372 char *p;
2373
2374 for (c = 0; c < 256; c++)
2375 {
2376 if (ISDIGIT (c))
2377 {
2378 digit_chars[c] = c;
2379 mnemonic_chars[c] = c;
2380 register_chars[c] = c;
2381 operand_chars[c] = c;
2382 }
2383 else if (ISLOWER (c))
2384 {
2385 mnemonic_chars[c] = c;
2386 register_chars[c] = c;
2387 operand_chars[c] = c;
2388 }
2389 else if (ISUPPER (c))
2390 {
2391 mnemonic_chars[c] = TOLOWER (c);
2392 register_chars[c] = mnemonic_chars[c];
2393 operand_chars[c] = c;
2394 }
2395
2396 if (ISALPHA (c) || ISDIGIT (c))
2397 identifier_chars[c] = c;
2398 else if (c >= 128)
2399 {
2400 identifier_chars[c] = c;
2401 operand_chars[c] = c;
2402 }
2403 }
2404
2405 #ifdef LEX_AT
2406 identifier_chars['@'] = '@';
2407 #endif
2408 #ifdef LEX_QM
2409 identifier_chars['?'] = '?';
2410 operand_chars['?'] = '?';
2411 #endif
2412 digit_chars['-'] = '-';
2413 mnemonic_chars['_'] = '_';
2414 mnemonic_chars['-'] = '-';
2415 mnemonic_chars['.'] = '.';
2416 identifier_chars['_'] = '_';
2417 identifier_chars['.'] = '.';
2418
2419 for (p = operand_special_chars; *p != '\0'; p++)
2420 operand_chars[(unsigned char) *p] = *p;
2421 }
2422
2423 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
2424 if (IS_ELF)
2425 {
2426 record_alignment (text_section, 2);
2427 record_alignment (data_section, 2);
2428 record_alignment (bss_section, 2);
2429 }
2430 #endif
2431
2432 if (flag_code == CODE_64BIT)
2433 {
2434 #if defined (OBJ_COFF) && defined (TE_PE)
2435 x86_dwarf2_return_column = (OUTPUT_FLAVOR == bfd_target_coff_flavour
2436 ? 32 : 16);
2437 #else
2438 x86_dwarf2_return_column = 16;
2439 #endif
2440 x86_cie_data_alignment = -8;
2441 }
2442 else
2443 {
2444 x86_dwarf2_return_column = 8;
2445 x86_cie_data_alignment = -4;
2446 }
2447 }
2448
2449 void
2450 i386_print_statistics (FILE *file)
2451 {
2452 hash_print_statistics (file, "i386 opcode", op_hash);
2453 hash_print_statistics (file, "i386 register", reg_hash);
2454 }
2455 \f
2456 #ifdef DEBUG386
2457
2458 /* Debugging routines for md_assemble. */
2459 static void pte (insn_template *);
2460 static void pt (i386_operand_type);
2461 static void pe (expressionS *);
2462 static void ps (symbolS *);
2463
2464 static void
2465 pi (char *line, i386_insn *x)
2466 {
2467 unsigned int j;
2468
2469 fprintf (stdout, "%s: template ", line);
2470 pte (&x->tm);
2471 fprintf (stdout, " address: base %s index %s scale %x\n",
2472 x->base_reg ? x->base_reg->reg_name : "none",
2473 x->index_reg ? x->index_reg->reg_name : "none",
2474 x->log2_scale_factor);
2475 fprintf (stdout, " modrm: mode %x reg %x reg/mem %x\n",
2476 x->rm.mode, x->rm.reg, x->rm.regmem);
2477 fprintf (stdout, " sib: base %x index %x scale %x\n",
2478 x->sib.base, x->sib.index, x->sib.scale);
2479 fprintf (stdout, " rex: 64bit %x extX %x extY %x extZ %x\n",
2480 (x->rex & REX_W) != 0,
2481 (x->rex & REX_R) != 0,
2482 (x->rex & REX_X) != 0,
2483 (x->rex & REX_B) != 0);
2484 for (j = 0; j < x->operands; j++)
2485 {
2486 fprintf (stdout, " #%d: ", j + 1);
2487 pt (x->types[j]);
2488 fprintf (stdout, "\n");
2489 if (x->types[j].bitfield.reg8
2490 || x->types[j].bitfield.reg16
2491 || x->types[j].bitfield.reg32
2492 || x->types[j].bitfield.reg64
2493 || x->types[j].bitfield.regmmx
2494 || x->types[j].bitfield.regxmm
2495 || x->types[j].bitfield.regymm
2496 || x->types[j].bitfield.sreg2
2497 || x->types[j].bitfield.sreg3
2498 || x->types[j].bitfield.control
2499 || x->types[j].bitfield.debug
2500 || x->types[j].bitfield.test)
2501 fprintf (stdout, "%s\n", x->op[j].regs->reg_name);
2502 if (operand_type_check (x->types[j], imm))
2503 pe (x->op[j].imms);
2504 if (operand_type_check (x->types[j], disp))
2505 pe (x->op[j].disps);
2506 }
2507 }
2508
2509 static void
2510 pte (insn_template *t)
2511 {
2512 unsigned int j;
2513 fprintf (stdout, " %d operands ", t->operands);
2514 fprintf (stdout, "opcode %x ", t->base_opcode);
2515 if (t->extension_opcode != None)
2516 fprintf (stdout, "ext %x ", t->extension_opcode);
2517 if (t->opcode_modifier.d)
2518 fprintf (stdout, "D");
2519 if (t->opcode_modifier.w)
2520 fprintf (stdout, "W");
2521 fprintf (stdout, "\n");
2522 for (j = 0; j < t->operands; j++)
2523 {
2524 fprintf (stdout, " #%d type ", j + 1);
2525 pt (t->operand_types[j]);
2526 fprintf (stdout, "\n");
2527 }
2528 }
2529
2530 static void
2531 pe (expressionS *e)
2532 {
2533 fprintf (stdout, " operation %d\n", e->X_op);
2534 fprintf (stdout, " add_number %ld (%lx)\n",
2535 (long) e->X_add_number, (long) e->X_add_number);
2536 if (e->X_add_symbol)
2537 {
2538 fprintf (stdout, " add_symbol ");
2539 ps (e->X_add_symbol);
2540 fprintf (stdout, "\n");
2541 }
2542 if (e->X_op_symbol)
2543 {
2544 fprintf (stdout, " op_symbol ");
2545 ps (e->X_op_symbol);
2546 fprintf (stdout, "\n");
2547 }
2548 }
2549
2550 static void
2551 ps (symbolS *s)
2552 {
2553 fprintf (stdout, "%s type %s%s",
2554 S_GET_NAME (s),
2555 S_IS_EXTERNAL (s) ? "EXTERNAL " : "",
2556 segment_name (S_GET_SEGMENT (s)));
2557 }
2558
2559 static struct type_name
2560 {
2561 i386_operand_type mask;
2562 const char *name;
2563 }
2564 const type_names[] =
2565 {
2566 { OPERAND_TYPE_REG8, "r8" },
2567 { OPERAND_TYPE_REG16, "r16" },
2568 { OPERAND_TYPE_REG32, "r32" },
2569 { OPERAND_TYPE_REG64, "r64" },
2570 { OPERAND_TYPE_IMM8, "i8" },
2571 { OPERAND_TYPE_IMM8, "i8s" },
2572 { OPERAND_TYPE_IMM16, "i16" },
2573 { OPERAND_TYPE_IMM32, "i32" },
2574 { OPERAND_TYPE_IMM32S, "i32s" },
2575 { OPERAND_TYPE_IMM64, "i64" },
2576 { OPERAND_TYPE_IMM1, "i1" },
2577 { OPERAND_TYPE_BASEINDEX, "BaseIndex" },
2578 { OPERAND_TYPE_DISP8, "d8" },
2579 { OPERAND_TYPE_DISP16, "d16" },
2580 { OPERAND_TYPE_DISP32, "d32" },
2581 { OPERAND_TYPE_DISP32S, "d32s" },
2582 { OPERAND_TYPE_DISP64, "d64" },
2583 { OPERAND_TYPE_INOUTPORTREG, "InOutPortReg" },
2584 { OPERAND_TYPE_SHIFTCOUNT, "ShiftCount" },
2585 { OPERAND_TYPE_CONTROL, "control reg" },
2586 { OPERAND_TYPE_TEST, "test reg" },
2587 { OPERAND_TYPE_DEBUG, "debug reg" },
2588 { OPERAND_TYPE_FLOATREG, "FReg" },
2589 { OPERAND_TYPE_FLOATACC, "FAcc" },
2590 { OPERAND_TYPE_SREG2, "SReg2" },
2591 { OPERAND_TYPE_SREG3, "SReg3" },
2592 { OPERAND_TYPE_ACC, "Acc" },
2593 { OPERAND_TYPE_JUMPABSOLUTE, "Jump Absolute" },
2594 { OPERAND_TYPE_REGMMX, "rMMX" },
2595 { OPERAND_TYPE_REGXMM, "rXMM" },
2596 { OPERAND_TYPE_REGYMM, "rYMM" },
2597 { OPERAND_TYPE_ESSEG, "es" },
2598 };
2599
2600 static void
2601 pt (i386_operand_type t)
2602 {
2603 unsigned int j;
2604 i386_operand_type a;
2605
2606 for (j = 0; j < ARRAY_SIZE (type_names); j++)
2607 {
2608 a = operand_type_and (t, type_names[j].mask);
2609 if (!operand_type_all_zero (&a))
2610 fprintf (stdout, "%s, ", type_names[j].name);
2611 }
2612 fflush (stdout);
2613 }
2614
2615 #endif /* DEBUG386 */
2616 \f
2617 static bfd_reloc_code_real_type
2618 reloc (unsigned int size,
2619 int pcrel,
2620 int sign,
2621 bfd_reloc_code_real_type other)
2622 {
2623 if (other != NO_RELOC)
2624 {
2625 reloc_howto_type *rel;
2626
2627 if (size == 8)
2628 switch (other)
2629 {
2630 case BFD_RELOC_X86_64_GOT32:
2631 return BFD_RELOC_X86_64_GOT64;
2632 break;
2633 case BFD_RELOC_X86_64_PLTOFF64:
2634 return BFD_RELOC_X86_64_PLTOFF64;
2635 break;
2636 case BFD_RELOC_X86_64_GOTPC32:
2637 other = BFD_RELOC_X86_64_GOTPC64;
2638 break;
2639 case BFD_RELOC_X86_64_GOTPCREL:
2640 other = BFD_RELOC_X86_64_GOTPCREL64;
2641 break;
2642 case BFD_RELOC_X86_64_TPOFF32:
2643 other = BFD_RELOC_X86_64_TPOFF64;
2644 break;
2645 case BFD_RELOC_X86_64_DTPOFF32:
2646 other = BFD_RELOC_X86_64_DTPOFF64;
2647 break;
2648 default:
2649 break;
2650 }
2651
2652 /* Sign-checking 4-byte relocations in 16-/32-bit code is pointless. */
2653 if (size == 4 && (flag_code != CODE_64BIT || disallow_64bit_reloc))
2654 sign = -1;
2655
2656 rel = bfd_reloc_type_lookup (stdoutput, other);
2657 if (!rel)
2658 as_bad (_("unknown relocation (%u)"), other);
2659 else if (size != bfd_get_reloc_size (rel))
2660 as_bad (_("%u-byte relocation cannot be applied to %u-byte field"),
2661 bfd_get_reloc_size (rel),
2662 size);
2663 else if (pcrel && !rel->pc_relative)
2664 as_bad (_("non-pc-relative relocation for pc-relative field"));
2665 else if ((rel->complain_on_overflow == complain_overflow_signed
2666 && !sign)
2667 || (rel->complain_on_overflow == complain_overflow_unsigned
2668 && sign > 0))
2669 as_bad (_("relocated field and relocation type differ in signedness"));
2670 else
2671 return other;
2672 return NO_RELOC;
2673 }
2674
2675 if (pcrel)
2676 {
2677 if (!sign)
2678 as_bad (_("there are no unsigned pc-relative relocations"));
2679 switch (size)
2680 {
2681 case 1: return BFD_RELOC_8_PCREL;
2682 case 2: return BFD_RELOC_16_PCREL;
2683 case 4: return BFD_RELOC_32_PCREL;
2684 case 8: return BFD_RELOC_64_PCREL;
2685 }
2686 as_bad (_("cannot do %u byte pc-relative relocation"), size);
2687 }
2688 else
2689 {
2690 if (sign > 0)
2691 switch (size)
2692 {
2693 case 4: return BFD_RELOC_X86_64_32S;
2694 }
2695 else
2696 switch (size)
2697 {
2698 case 1: return BFD_RELOC_8;
2699 case 2: return BFD_RELOC_16;
2700 case 4: return BFD_RELOC_32;
2701 case 8: return BFD_RELOC_64;
2702 }
2703 as_bad (_("cannot do %s %u byte relocation"),
2704 sign > 0 ? "signed" : "unsigned", size);
2705 }
2706
2707 return NO_RELOC;
2708 }
2709
2710 /* Here we decide which fixups can be adjusted to make them relative to
2711 the beginning of the section instead of the symbol. Basically we need
2712 to make sure that the dynamic relocations are done correctly, so in
2713 some cases we force the original symbol to be used. */
2714
2715 int
2716 tc_i386_fix_adjustable (fixS *fixP ATTRIBUTE_UNUSED)
2717 {
2718 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
2719 if (!IS_ELF)
2720 return 1;
2721
2722 /* Don't adjust pc-relative references to merge sections in 64-bit
2723 mode. */
2724 if (use_rela_relocations
2725 && (S_GET_SEGMENT (fixP->fx_addsy)->flags & SEC_MERGE) != 0
2726 && fixP->fx_pcrel)
2727 return 0;
2728
2729 /* The x86_64 GOTPCREL are represented as 32bit PCrel relocations
2730 and changed later by validate_fix. */
2731 if (GOT_symbol && fixP->fx_subsy == GOT_symbol
2732 && fixP->fx_r_type == BFD_RELOC_32_PCREL)
2733 return 0;
2734
2735 /* adjust_reloc_syms doesn't know about the GOT. */
2736 if (fixP->fx_r_type == BFD_RELOC_386_GOTOFF
2737 || fixP->fx_r_type == BFD_RELOC_386_PLT32
2738 || fixP->fx_r_type == BFD_RELOC_386_GOT32
2739 || fixP->fx_r_type == BFD_RELOC_386_TLS_GD
2740 || fixP->fx_r_type == BFD_RELOC_386_TLS_LDM
2741 || fixP->fx_r_type == BFD_RELOC_386_TLS_LDO_32
2742 || fixP->fx_r_type == BFD_RELOC_386_TLS_IE_32
2743 || fixP->fx_r_type == BFD_RELOC_386_TLS_IE
2744 || fixP->fx_r_type == BFD_RELOC_386_TLS_GOTIE
2745 || fixP->fx_r_type == BFD_RELOC_386_TLS_LE_32
2746 || fixP->fx_r_type == BFD_RELOC_386_TLS_LE
2747 || fixP->fx_r_type == BFD_RELOC_386_TLS_GOTDESC
2748 || fixP->fx_r_type == BFD_RELOC_386_TLS_DESC_CALL
2749 || fixP->fx_r_type == BFD_RELOC_X86_64_PLT32
2750 || fixP->fx_r_type == BFD_RELOC_X86_64_GOT32
2751 || fixP->fx_r_type == BFD_RELOC_X86_64_GOTPCREL
2752 || fixP->fx_r_type == BFD_RELOC_X86_64_TLSGD
2753 || fixP->fx_r_type == BFD_RELOC_X86_64_TLSLD
2754 || fixP->fx_r_type == BFD_RELOC_X86_64_DTPOFF32
2755 || fixP->fx_r_type == BFD_RELOC_X86_64_DTPOFF64
2756 || fixP->fx_r_type == BFD_RELOC_X86_64_GOTTPOFF
2757 || fixP->fx_r_type == BFD_RELOC_X86_64_TPOFF32
2758 || fixP->fx_r_type == BFD_RELOC_X86_64_TPOFF64
2759 || fixP->fx_r_type == BFD_RELOC_X86_64_GOTOFF64
2760 || fixP->fx_r_type == BFD_RELOC_X86_64_GOTPC32_TLSDESC
2761 || fixP->fx_r_type == BFD_RELOC_X86_64_TLSDESC_CALL
2762 || fixP->fx_r_type == BFD_RELOC_VTABLE_INHERIT
2763 || fixP->fx_r_type == BFD_RELOC_VTABLE_ENTRY)
2764 return 0;
2765 #endif
2766 return 1;
2767 }
2768
2769 static int
2770 intel_float_operand (const char *mnemonic)
2771 {
2772 /* Note that the value returned is meaningful only for opcodes with (memory)
2773 operands, hence the code here is free to improperly handle opcodes that
2774 have no operands (for better performance and smaller code). */
2775
2776 if (mnemonic[0] != 'f')
2777 return 0; /* non-math */
2778
2779 switch (mnemonic[1])
2780 {
2781 /* fclex, fdecstp, fdisi, femms, feni, fincstp, finit, fsetpm, and
2782 the fs segment override prefix not currently handled because no
2783 call path can make opcodes without operands get here */
2784 case 'i':
2785 return 2 /* integer op */;
2786 case 'l':
2787 if (mnemonic[2] == 'd' && (mnemonic[3] == 'c' || mnemonic[3] == 'e'))
2788 return 3; /* fldcw/fldenv */
2789 break;
2790 case 'n':
2791 if (mnemonic[2] != 'o' /* fnop */)
2792 return 3; /* non-waiting control op */
2793 break;
2794 case 'r':
2795 if (mnemonic[2] == 's')
2796 return 3; /* frstor/frstpm */
2797 break;
2798 case 's':
2799 if (mnemonic[2] == 'a')
2800 return 3; /* fsave */
2801 if (mnemonic[2] == 't')
2802 {
2803 switch (mnemonic[3])
2804 {
2805 case 'c': /* fstcw */
2806 case 'd': /* fstdw */
2807 case 'e': /* fstenv */
2808 case 's': /* fsts[gw] */
2809 return 3;
2810 }
2811 }
2812 break;
2813 case 'x':
2814 if (mnemonic[2] == 'r' || mnemonic[2] == 's')
2815 return 0; /* fxsave/fxrstor are not really math ops */
2816 break;
2817 }
2818
2819 return 1;
2820 }
2821
2822 /* Build the VEX prefix. */
2823
2824 static void
2825 build_vex_prefix (const insn_template *t)
2826 {
2827 unsigned int register_specifier;
2828 unsigned int implied_prefix;
2829 unsigned int vector_length;
2830
2831 /* Check register specifier. */
2832 if (i.vex.register_specifier)
2833 {
2834 register_specifier = i.vex.register_specifier->reg_num;
2835 if ((i.vex.register_specifier->reg_flags & RegRex))
2836 register_specifier += 8;
2837 register_specifier = ~register_specifier & 0xf;
2838 }
2839 else
2840 register_specifier = 0xf;
2841
2842 /* Use 2-byte VEX prefix by swappping destination and source
2843 operand. */
2844 if (!i.swap_operand
2845 && i.operands == i.reg_operands
2846 && i.tm.opcode_modifier.vexopcode == VEX0F
2847 && i.tm.opcode_modifier.s
2848 && i.rex == REX_B)
2849 {
2850 unsigned int xchg = i.operands - 1;
2851 union i386_op temp_op;
2852 i386_operand_type temp_type;
2853
2854 temp_type = i.types[xchg];
2855 i.types[xchg] = i.types[0];
2856 i.types[0] = temp_type;
2857 temp_op = i.op[xchg];
2858 i.op[xchg] = i.op[0];
2859 i.op[0] = temp_op;
2860
2861 gas_assert (i.rm.mode == 3);
2862
2863 i.rex = REX_R;
2864 xchg = i.rm.regmem;
2865 i.rm.regmem = i.rm.reg;
2866 i.rm.reg = xchg;
2867
2868 /* Use the next insn. */
2869 i.tm = t[1];
2870 }
2871
2872 if (i.tm.opcode_modifier.vex == VEXScalar)
2873 vector_length = avxscalar;
2874 else
2875 vector_length = i.tm.opcode_modifier.vex == VEX256 ? 1 : 0;
2876
2877 switch ((i.tm.base_opcode >> 8) & 0xff)
2878 {
2879 case 0:
2880 implied_prefix = 0;
2881 break;
2882 case DATA_PREFIX_OPCODE:
2883 implied_prefix = 1;
2884 break;
2885 case REPE_PREFIX_OPCODE:
2886 implied_prefix = 2;
2887 break;
2888 case REPNE_PREFIX_OPCODE:
2889 implied_prefix = 3;
2890 break;
2891 default:
2892 abort ();
2893 }
2894
2895 /* Use 2-byte VEX prefix if possible. */
2896 if (i.tm.opcode_modifier.vexopcode == VEX0F
2897 && i.tm.opcode_modifier.vexw != VEXW1
2898 && (i.rex & (REX_W | REX_X | REX_B)) == 0)
2899 {
2900 /* 2-byte VEX prefix. */
2901 unsigned int r;
2902
2903 i.vex.length = 2;
2904 i.vex.bytes[0] = 0xc5;
2905
2906 /* Check the REX.R bit. */
2907 r = (i.rex & REX_R) ? 0 : 1;
2908 i.vex.bytes[1] = (r << 7
2909 | register_specifier << 3
2910 | vector_length << 2
2911 | implied_prefix);
2912 }
2913 else
2914 {
2915 /* 3-byte VEX prefix. */
2916 unsigned int m, w;
2917
2918 i.vex.length = 3;
2919
2920 switch (i.tm.opcode_modifier.vexopcode)
2921 {
2922 case VEX0F:
2923 m = 0x1;
2924 i.vex.bytes[0] = 0xc4;
2925 break;
2926 case VEX0F38:
2927 m = 0x2;
2928 i.vex.bytes[0] = 0xc4;
2929 break;
2930 case VEX0F3A:
2931 m = 0x3;
2932 i.vex.bytes[0] = 0xc4;
2933 break;
2934 case XOP08:
2935 m = 0x8;
2936 i.vex.bytes[0] = 0x8f;
2937 break;
2938 case XOP09:
2939 m = 0x9;
2940 i.vex.bytes[0] = 0x8f;
2941 break;
2942 case XOP0A:
2943 m = 0xa;
2944 i.vex.bytes[0] = 0x8f;
2945 break;
2946 default:
2947 abort ();
2948 }
2949
2950 /* The high 3 bits of the second VEX byte are 1's compliment
2951 of RXB bits from REX. */
2952 i.vex.bytes[1] = (~i.rex & 0x7) << 5 | m;
2953
2954 /* Check the REX.W bit. */
2955 w = (i.rex & REX_W) ? 1 : 0;
2956 if (i.tm.opcode_modifier.vexw)
2957 {
2958 if (w)
2959 abort ();
2960
2961 if (i.tm.opcode_modifier.vexw == VEXW1)
2962 w = 1;
2963 }
2964
2965 i.vex.bytes[2] = (w << 7
2966 | register_specifier << 3
2967 | vector_length << 2
2968 | implied_prefix);
2969 }
2970 }
2971
2972 static void
2973 process_immext (void)
2974 {
2975 expressionS *exp;
2976
2977 if (i.tm.cpu_flags.bitfield.cpusse3 && i.operands > 0)
2978 {
2979 /* SSE3 Instructions have the fixed operands with an opcode
2980 suffix which is coded in the same place as an 8-bit immediate
2981 field would be. Here we check those operands and remove them
2982 afterwards. */
2983 unsigned int x;
2984
2985 for (x = 0; x < i.operands; x++)
2986 if (i.op[x].regs->reg_num != x)
2987 as_bad (_("can't use register '%s%s' as operand %d in '%s'."),
2988 register_prefix, i.op[x].regs->reg_name, x + 1,
2989 i.tm.name);
2990
2991 i.operands = 0;
2992 }
2993
2994 /* These AMD 3DNow! and SSE2 instructions have an opcode suffix
2995 which is coded in the same place as an 8-bit immediate field
2996 would be. Here we fake an 8-bit immediate operand from the
2997 opcode suffix stored in tm.extension_opcode.
2998
2999 AVX instructions also use this encoding, for some of
3000 3 argument instructions. */
3001
3002 gas_assert (i.imm_operands == 0
3003 && (i.operands <= 2
3004 || (i.tm.opcode_modifier.vex
3005 && i.operands <= 4)));
3006
3007 exp = &im_expressions[i.imm_operands++];
3008 i.op[i.operands].imms = exp;
3009 i.types[i.operands] = imm8;
3010 i.operands++;
3011 exp->X_op = O_constant;
3012 exp->X_add_number = i.tm.extension_opcode;
3013 i.tm.extension_opcode = None;
3014 }
3015
3016
3017 static int
3018 check_hle (void)
3019 {
3020 switch (i.tm.opcode_modifier.hleprefixok)
3021 {
3022 default:
3023 abort ();
3024 case HLEPrefixNone:
3025 if (i.prefix[HLE_PREFIX] == XACQUIRE_PREFIX_OPCODE)
3026 as_bad (_("invalid instruction `%s' after `xacquire'"),
3027 i.tm.name);
3028 else
3029 as_bad (_("invalid instruction `%s' after `xrelease'"),
3030 i.tm.name);
3031 return 0;
3032 case HLEPrefixLock:
3033 if (i.prefix[LOCK_PREFIX])
3034 return 1;
3035 if (i.prefix[HLE_PREFIX] == XACQUIRE_PREFIX_OPCODE)
3036 as_bad (_("missing `lock' with `xacquire'"));
3037 else
3038 as_bad (_("missing `lock' with `xrelease'"));
3039 return 0;
3040 case HLEPrefixAny:
3041 return 1;
3042 case HLEPrefixRelease:
3043 if (i.prefix[HLE_PREFIX] != XRELEASE_PREFIX_OPCODE)
3044 {
3045 as_bad (_("instruction `%s' after `xacquire' not allowed"),
3046 i.tm.name);
3047 return 0;
3048 }
3049 if (i.mem_operands == 0
3050 || !operand_type_check (i.types[i.operands - 1], anymem))
3051 {
3052 as_bad (_("memory destination needed for instruction `%s'"
3053 " after `xrelease'"), i.tm.name);
3054 return 0;
3055 }
3056 return 1;
3057 }
3058 }
3059
3060 /* This is the guts of the machine-dependent assembler. LINE points to a
3061 machine dependent instruction. This function is supposed to emit
3062 the frags/bytes it assembles to. */
3063
3064 void
3065 md_assemble (char *line)
3066 {
3067 unsigned int j;
3068 char mnemonic[MAX_MNEM_SIZE];
3069 const insn_template *t;
3070
3071 /* Initialize globals. */
3072 memset (&i, '\0', sizeof (i));
3073 for (j = 0; j < MAX_OPERANDS; j++)
3074 i.reloc[j] = NO_RELOC;
3075 memset (disp_expressions, '\0', sizeof (disp_expressions));
3076 memset (im_expressions, '\0', sizeof (im_expressions));
3077 save_stack_p = save_stack;
3078
3079 /* First parse an instruction mnemonic & call i386_operand for the operands.
3080 We assume that the scrubber has arranged it so that line[0] is the valid
3081 start of a (possibly prefixed) mnemonic. */
3082
3083 line = parse_insn (line, mnemonic);
3084 if (line == NULL)
3085 return;
3086
3087 line = parse_operands (line, mnemonic);
3088 this_operand = -1;
3089 if (line == NULL)
3090 return;
3091
3092 /* Now we've parsed the mnemonic into a set of templates, and have the
3093 operands at hand. */
3094
3095 /* All intel opcodes have reversed operands except for "bound" and
3096 "enter". We also don't reverse intersegment "jmp" and "call"
3097 instructions with 2 immediate operands so that the immediate segment
3098 precedes the offset, as it does when in AT&T mode. */
3099 if (intel_syntax
3100 && i.operands > 1
3101 && (strcmp (mnemonic, "bound") != 0)
3102 && (strcmp (mnemonic, "invlpga") != 0)
3103 && !(operand_type_check (i.types[0], imm)
3104 && operand_type_check (i.types[1], imm)))
3105 swap_operands ();
3106
3107 /* The order of the immediates should be reversed
3108 for 2 immediates extrq and insertq instructions */
3109 if (i.imm_operands == 2
3110 && (strcmp (mnemonic, "extrq") == 0
3111 || strcmp (mnemonic, "insertq") == 0))
3112 swap_2_operands (0, 1);
3113
3114 if (i.imm_operands)
3115 optimize_imm ();
3116
3117 /* Don't optimize displacement for movabs since it only takes 64bit
3118 displacement. */
3119 if (i.disp_operands
3120 && i.disp_encoding != disp_encoding_32bit
3121 && (flag_code != CODE_64BIT
3122 || strcmp (mnemonic, "movabs") != 0))
3123 optimize_disp ();
3124
3125 /* Next, we find a template that matches the given insn,
3126 making sure the overlap of the given operands types is consistent
3127 with the template operand types. */
3128
3129 if (!(t = match_template ()))
3130 return;
3131
3132 if (sse_check != sse_check_none
3133 && !i.tm.opcode_modifier.noavx
3134 && (i.tm.cpu_flags.bitfield.cpusse
3135 || i.tm.cpu_flags.bitfield.cpusse2
3136 || i.tm.cpu_flags.bitfield.cpusse3
3137 || i.tm.cpu_flags.bitfield.cpussse3
3138 || i.tm.cpu_flags.bitfield.cpusse4_1
3139 || i.tm.cpu_flags.bitfield.cpusse4_2))
3140 {
3141 (sse_check == sse_check_warning
3142 ? as_warn
3143 : as_bad) (_("SSE instruction `%s' is used"), i.tm.name);
3144 }
3145
3146 /* Zap movzx and movsx suffix. The suffix has been set from
3147 "word ptr" or "byte ptr" on the source operand in Intel syntax
3148 or extracted from mnemonic in AT&T syntax. But we'll use
3149 the destination register to choose the suffix for encoding. */
3150 if ((i.tm.base_opcode & ~9) == 0x0fb6)
3151 {
3152 /* In Intel syntax, there must be a suffix. In AT&T syntax, if
3153 there is no suffix, the default will be byte extension. */
3154 if (i.reg_operands != 2
3155 && !i.suffix
3156 && intel_syntax)
3157 as_bad (_("ambiguous operand size for `%s'"), i.tm.name);
3158
3159 i.suffix = 0;
3160 }
3161
3162 if (i.tm.opcode_modifier.fwait)
3163 if (!add_prefix (FWAIT_OPCODE))
3164 return;
3165
3166 /* Check for lock without a lockable instruction. Destination operand
3167 must be memory unless it is xchg (0x86). */
3168 if (i.prefix[LOCK_PREFIX]
3169 && (!i.tm.opcode_modifier.islockable
3170 || i.mem_operands == 0
3171 || (i.tm.base_opcode != 0x86
3172 && !operand_type_check (i.types[i.operands - 1], anymem))))
3173 {
3174 as_bad (_("expecting lockable instruction after `lock'"));
3175 return;
3176 }
3177
3178 /* Check if HLE prefix is OK. */
3179 if (i.have_hle && !check_hle ())
3180 return;
3181
3182 /* Check string instruction segment overrides. */
3183 if (i.tm.opcode_modifier.isstring && i.mem_operands != 0)
3184 {
3185 if (!check_string ())
3186 return;
3187 i.disp_operands = 0;
3188 }
3189
3190 if (!process_suffix ())
3191 return;
3192
3193 /* Update operand types. */
3194 for (j = 0; j < i.operands; j++)
3195 i.types[j] = operand_type_and (i.types[j], i.tm.operand_types[j]);
3196
3197 /* Make still unresolved immediate matches conform to size of immediate
3198 given in i.suffix. */
3199 if (!finalize_imm ())
3200 return;
3201
3202 if (i.types[0].bitfield.imm1)
3203 i.imm_operands = 0; /* kludge for shift insns. */
3204
3205 /* We only need to check those implicit registers for instructions
3206 with 3 operands or less. */
3207 if (i.operands <= 3)
3208 for (j = 0; j < i.operands; j++)
3209 if (i.types[j].bitfield.inoutportreg
3210 || i.types[j].bitfield.shiftcount
3211 || i.types[j].bitfield.acc
3212 || i.types[j].bitfield.floatacc)
3213 i.reg_operands--;
3214
3215 /* ImmExt should be processed after SSE2AVX. */
3216 if (!i.tm.opcode_modifier.sse2avx
3217 && i.tm.opcode_modifier.immext)
3218 process_immext ();
3219
3220 /* For insns with operands there are more diddles to do to the opcode. */
3221 if (i.operands)
3222 {
3223 if (!process_operands ())
3224 return;
3225 }
3226 else if (!quiet_warnings && i.tm.opcode_modifier.ugh)
3227 {
3228 /* UnixWare fsub no args is alias for fsubp, fadd -> faddp, etc. */
3229 as_warn (_("translating to `%sp'"), i.tm.name);
3230 }
3231
3232 if (i.tm.opcode_modifier.vex)
3233 build_vex_prefix (t);
3234
3235 /* Handle conversion of 'int $3' --> special int3 insn. XOP or FMA4
3236 instructions may define INT_OPCODE as well, so avoid this corner
3237 case for those instructions that use MODRM. */
3238 if (i.tm.base_opcode == INT_OPCODE
3239 && !i.tm.opcode_modifier.modrm
3240 && i.op[0].imms->X_add_number == 3)
3241 {
3242 i.tm.base_opcode = INT3_OPCODE;
3243 i.imm_operands = 0;
3244 }
3245
3246 if ((i.tm.opcode_modifier.jump
3247 || i.tm.opcode_modifier.jumpbyte
3248 || i.tm.opcode_modifier.jumpdword)
3249 && i.op[0].disps->X_op == O_constant)
3250 {
3251 /* Convert "jmp constant" (and "call constant") to a jump (call) to
3252 the absolute address given by the constant. Since ix86 jumps and
3253 calls are pc relative, we need to generate a reloc. */
3254 i.op[0].disps->X_add_symbol = &abs_symbol;
3255 i.op[0].disps->X_op = O_symbol;
3256 }
3257
3258 if (i.tm.opcode_modifier.rex64)
3259 i.rex |= REX_W;
3260
3261 /* For 8 bit registers we need an empty rex prefix. Also if the
3262 instruction already has a prefix, we need to convert old
3263 registers to new ones. */
3264
3265 if ((i.types[0].bitfield.reg8
3266 && (i.op[0].regs->reg_flags & RegRex64) != 0)
3267 || (i.types[1].bitfield.reg8
3268 && (i.op[1].regs->reg_flags & RegRex64) != 0)
3269 || ((i.types[0].bitfield.reg8
3270 || i.types[1].bitfield.reg8)
3271 && i.rex != 0))
3272 {
3273 int x;
3274
3275 i.rex |= REX_OPCODE;
3276 for (x = 0; x < 2; x++)
3277 {
3278 /* Look for 8 bit operand that uses old registers. */
3279 if (i.types[x].bitfield.reg8
3280 && (i.op[x].regs->reg_flags & RegRex64) == 0)
3281 {
3282 /* In case it is "hi" register, give up. */
3283 if (i.op[x].regs->reg_num > 3)
3284 as_bad (_("can't encode register '%s%s' in an "
3285 "instruction requiring REX prefix."),
3286 register_prefix, i.op[x].regs->reg_name);
3287
3288 /* Otherwise it is equivalent to the extended register.
3289 Since the encoding doesn't change this is merely
3290 cosmetic cleanup for debug output. */
3291
3292 i.op[x].regs = i.op[x].regs + 8;
3293 }
3294 }
3295 }
3296
3297 if (i.rex != 0)
3298 add_prefix (REX_OPCODE | i.rex);
3299
3300 /* We are ready to output the insn. */
3301 output_insn ();
3302 }
3303
3304 static char *
3305 parse_insn (char *line, char *mnemonic)
3306 {
3307 char *l = line;
3308 char *token_start = l;
3309 char *mnem_p;
3310 int supported;
3311 const insn_template *t;
3312 char *dot_p = NULL;
3313
3314 /* Non-zero if we found a prefix only acceptable with string insns. */
3315 const char *expecting_string_instruction = NULL;
3316
3317 while (1)
3318 {
3319 mnem_p = mnemonic;
3320 while ((*mnem_p = mnemonic_chars[(unsigned char) *l]) != 0)
3321 {
3322 if (*mnem_p == '.')
3323 dot_p = mnem_p;
3324 mnem_p++;
3325 if (mnem_p >= mnemonic + MAX_MNEM_SIZE)
3326 {
3327 as_bad (_("no such instruction: `%s'"), token_start);
3328 return NULL;
3329 }
3330 l++;
3331 }
3332 if (!is_space_char (*l)
3333 && *l != END_OF_INSN
3334 && (intel_syntax
3335 || (*l != PREFIX_SEPARATOR
3336 && *l != ',')))
3337 {
3338 as_bad (_("invalid character %s in mnemonic"),
3339 output_invalid (*l));
3340 return NULL;
3341 }
3342 if (token_start == l)
3343 {
3344 if (!intel_syntax && *l == PREFIX_SEPARATOR)
3345 as_bad (_("expecting prefix; got nothing"));
3346 else
3347 as_bad (_("expecting mnemonic; got nothing"));
3348 return NULL;
3349 }
3350
3351 /* Look up instruction (or prefix) via hash table. */
3352 current_templates = (const templates *) hash_find (op_hash, mnemonic);
3353
3354 if (*l != END_OF_INSN
3355 && (!is_space_char (*l) || l[1] != END_OF_INSN)
3356 && current_templates
3357 && current_templates->start->opcode_modifier.isprefix)
3358 {
3359 if (!cpu_flags_check_cpu64 (current_templates->start->cpu_flags))
3360 {
3361 as_bad ((flag_code != CODE_64BIT
3362 ? _("`%s' is only supported in 64-bit mode")
3363 : _("`%s' is not supported in 64-bit mode")),
3364 current_templates->start->name);
3365 return NULL;
3366 }
3367 /* If we are in 16-bit mode, do not allow addr16 or data16.
3368 Similarly, in 32-bit mode, do not allow addr32 or data32. */
3369 if ((current_templates->start->opcode_modifier.size16
3370 || current_templates->start->opcode_modifier.size32)
3371 && flag_code != CODE_64BIT
3372 && (current_templates->start->opcode_modifier.size32
3373 ^ (flag_code == CODE_16BIT)))
3374 {
3375 as_bad (_("redundant %s prefix"),
3376 current_templates->start->name);
3377 return NULL;
3378 }
3379 /* Add prefix, checking for repeated prefixes. */
3380 switch (add_prefix (current_templates->start->base_opcode))
3381 {
3382 case PREFIX_EXIST:
3383 return NULL;
3384 case PREFIX_REP:
3385 if (current_templates->start->cpu_flags.bitfield.cpuhle)
3386 i.have_hle = 1;
3387 else
3388 expecting_string_instruction = current_templates->start->name;
3389 break;
3390 default:
3391 break;
3392 }
3393 /* Skip past PREFIX_SEPARATOR and reset token_start. */
3394 token_start = ++l;
3395 }
3396 else
3397 break;
3398 }
3399
3400 if (!current_templates)
3401 {
3402 /* Check if we should swap operand or force 32bit displacement in
3403 encoding. */
3404 if (mnem_p - 2 == dot_p && dot_p[1] == 's')
3405 i.swap_operand = 1;
3406 else if (mnem_p - 3 == dot_p
3407 && dot_p[1] == 'd'
3408 && dot_p[2] == '8')
3409 i.disp_encoding = disp_encoding_8bit;
3410 else if (mnem_p - 4 == dot_p
3411 && dot_p[1] == 'd'
3412 && dot_p[2] == '3'
3413 && dot_p[3] == '2')
3414 i.disp_encoding = disp_encoding_32bit;
3415 else
3416 goto check_suffix;
3417 mnem_p = dot_p;
3418 *dot_p = '\0';
3419 current_templates = (const templates *) hash_find (op_hash, mnemonic);
3420 }
3421
3422 if (!current_templates)
3423 {
3424 check_suffix:
3425 /* See if we can get a match by trimming off a suffix. */
3426 switch (mnem_p[-1])
3427 {
3428 case WORD_MNEM_SUFFIX:
3429 if (intel_syntax && (intel_float_operand (mnemonic) & 2))
3430 i.suffix = SHORT_MNEM_SUFFIX;
3431 else
3432 case BYTE_MNEM_SUFFIX:
3433 case QWORD_MNEM_SUFFIX:
3434 i.suffix = mnem_p[-1];
3435 mnem_p[-1] = '\0';
3436 current_templates = (const templates *) hash_find (op_hash,
3437 mnemonic);
3438 break;
3439 case SHORT_MNEM_SUFFIX:
3440 case LONG_MNEM_SUFFIX:
3441 if (!intel_syntax)
3442 {
3443 i.suffix = mnem_p[-1];
3444 mnem_p[-1] = '\0';
3445 current_templates = (const templates *) hash_find (op_hash,
3446 mnemonic);
3447 }
3448 break;
3449
3450 /* Intel Syntax. */
3451 case 'd':
3452 if (intel_syntax)
3453 {
3454 if (intel_float_operand (mnemonic) == 1)
3455 i.suffix = SHORT_MNEM_SUFFIX;
3456 else
3457 i.suffix = LONG_MNEM_SUFFIX;
3458 mnem_p[-1] = '\0';
3459 current_templates = (const templates *) hash_find (op_hash,
3460 mnemonic);
3461 }
3462 break;
3463 }
3464 if (!current_templates)
3465 {
3466 as_bad (_("no such instruction: `%s'"), token_start);
3467 return NULL;
3468 }
3469 }
3470
3471 if (current_templates->start->opcode_modifier.jump
3472 || current_templates->start->opcode_modifier.jumpbyte)
3473 {
3474 /* Check for a branch hint. We allow ",pt" and ",pn" for
3475 predict taken and predict not taken respectively.
3476 I'm not sure that branch hints actually do anything on loop
3477 and jcxz insns (JumpByte) for current Pentium4 chips. They
3478 may work in the future and it doesn't hurt to accept them
3479 now. */
3480 if (l[0] == ',' && l[1] == 'p')
3481 {
3482 if (l[2] == 't')
3483 {
3484 if (!add_prefix (DS_PREFIX_OPCODE))
3485 return NULL;
3486 l += 3;
3487 }
3488 else if (l[2] == 'n')
3489 {
3490 if (!add_prefix (CS_PREFIX_OPCODE))
3491 return NULL;
3492 l += 3;
3493 }
3494 }
3495 }
3496 /* Any other comma loses. */
3497 if (*l == ',')
3498 {
3499 as_bad (_("invalid character %s in mnemonic"),
3500 output_invalid (*l));
3501 return NULL;
3502 }
3503
3504 /* Check if instruction is supported on specified architecture. */
3505 supported = 0;
3506 for (t = current_templates->start; t < current_templates->end; ++t)
3507 {
3508 supported |= cpu_flags_match (t);
3509 if (supported == CPU_FLAGS_PERFECT_MATCH)
3510 goto skip;
3511 }
3512
3513 if (!(supported & CPU_FLAGS_64BIT_MATCH))
3514 {
3515 as_bad (flag_code == CODE_64BIT
3516 ? _("`%s' is not supported in 64-bit mode")
3517 : _("`%s' is only supported in 64-bit mode"),
3518 current_templates->start->name);
3519 return NULL;
3520 }
3521 if (supported != CPU_FLAGS_PERFECT_MATCH)
3522 {
3523 as_bad (_("`%s' is not supported on `%s%s'"),
3524 current_templates->start->name,
3525 cpu_arch_name ? cpu_arch_name : default_arch,
3526 cpu_sub_arch_name ? cpu_sub_arch_name : "");
3527 return NULL;
3528 }
3529
3530 skip:
3531 if (!cpu_arch_flags.bitfield.cpui386
3532 && (flag_code != CODE_16BIT))
3533 {
3534 as_warn (_("use .code16 to ensure correct addressing mode"));
3535 }
3536
3537 /* Check for rep/repne without a string (or other allowed) instruction. */
3538 if (expecting_string_instruction)
3539 {
3540 static templates override;
3541
3542 for (t = current_templates->start; t < current_templates->end; ++t)
3543 if (t->opcode_modifier.repprefixok)
3544 break;
3545 if (t >= current_templates->end)
3546 {
3547 as_bad (_("expecting string instruction after `%s'"),
3548 expecting_string_instruction);
3549 return NULL;
3550 }
3551 for (override.start = t; t < current_templates->end; ++t)
3552 if (!t->opcode_modifier.repprefixok)
3553 break;
3554 override.end = t;
3555 current_templates = &override;
3556 }
3557
3558 return l;
3559 }
3560
3561 static char *
3562 parse_operands (char *l, const char *mnemonic)
3563 {
3564 char *token_start;
3565
3566 /* 1 if operand is pending after ','. */
3567 unsigned int expecting_operand = 0;
3568
3569 /* Non-zero if operand parens not balanced. */
3570 unsigned int paren_not_balanced;
3571
3572 while (*l != END_OF_INSN)
3573 {
3574 /* Skip optional white space before operand. */
3575 if (is_space_char (*l))
3576 ++l;
3577 if (!is_operand_char (*l) && *l != END_OF_INSN)
3578 {
3579 as_bad (_("invalid character %s before operand %d"),
3580 output_invalid (*l),
3581 i.operands + 1);
3582 return NULL;
3583 }
3584 token_start = l; /* after white space */
3585 paren_not_balanced = 0;
3586 while (paren_not_balanced || *l != ',')
3587 {
3588 if (*l == END_OF_INSN)
3589 {
3590 if (paren_not_balanced)
3591 {
3592 if (!intel_syntax)
3593 as_bad (_("unbalanced parenthesis in operand %d."),
3594 i.operands + 1);
3595 else
3596 as_bad (_("unbalanced brackets in operand %d."),
3597 i.operands + 1);
3598 return NULL;
3599 }
3600 else
3601 break; /* we are done */
3602 }
3603 else if (!is_operand_char (*l) && !is_space_char (*l))
3604 {
3605 as_bad (_("invalid character %s in operand %d"),
3606 output_invalid (*l),
3607 i.operands + 1);
3608 return NULL;
3609 }
3610 if (!intel_syntax)
3611 {
3612 if (*l == '(')
3613 ++paren_not_balanced;
3614 if (*l == ')')
3615 --paren_not_balanced;
3616 }
3617 else
3618 {
3619 if (*l == '[')
3620 ++paren_not_balanced;
3621 if (*l == ']')
3622 --paren_not_balanced;
3623 }
3624 l++;
3625 }
3626 if (l != token_start)
3627 { /* Yes, we've read in another operand. */
3628 unsigned int operand_ok;
3629 this_operand = i.operands++;
3630 i.types[this_operand].bitfield.unspecified = 1;
3631 if (i.operands > MAX_OPERANDS)
3632 {
3633 as_bad (_("spurious operands; (%d operands/instruction max)"),
3634 MAX_OPERANDS);
3635 return NULL;
3636 }
3637 /* Now parse operand adding info to 'i' as we go along. */
3638 END_STRING_AND_SAVE (l);
3639
3640 if (intel_syntax)
3641 operand_ok =
3642 i386_intel_operand (token_start,
3643 intel_float_operand (mnemonic));
3644 else
3645 operand_ok = i386_att_operand (token_start);
3646
3647 RESTORE_END_STRING (l);
3648 if (!operand_ok)
3649 return NULL;
3650 }
3651 else
3652 {
3653 if (expecting_operand)
3654 {
3655 expecting_operand_after_comma:
3656 as_bad (_("expecting operand after ','; got nothing"));
3657 return NULL;
3658 }
3659 if (*l == ',')
3660 {
3661 as_bad (_("expecting operand before ','; got nothing"));
3662 return NULL;
3663 }
3664 }
3665
3666 /* Now *l must be either ',' or END_OF_INSN. */
3667 if (*l == ',')
3668 {
3669 if (*++l == END_OF_INSN)
3670 {
3671 /* Just skip it, if it's \n complain. */
3672 goto expecting_operand_after_comma;
3673 }
3674 expecting_operand = 1;
3675 }
3676 }
3677 return l;
3678 }
3679
3680 static void
3681 swap_2_operands (int xchg1, int xchg2)
3682 {
3683 union i386_op temp_op;
3684 i386_operand_type temp_type;
3685 enum bfd_reloc_code_real temp_reloc;
3686
3687 temp_type = i.types[xchg2];
3688 i.types[xchg2] = i.types[xchg1];
3689 i.types[xchg1] = temp_type;
3690 temp_op = i.op[xchg2];
3691 i.op[xchg2] = i.op[xchg1];
3692 i.op[xchg1] = temp_op;
3693 temp_reloc = i.reloc[xchg2];
3694 i.reloc[xchg2] = i.reloc[xchg1];
3695 i.reloc[xchg1] = temp_reloc;
3696 }
3697
3698 static void
3699 swap_operands (void)
3700 {
3701 switch (i.operands)
3702 {
3703 case 5:
3704 case 4:
3705 swap_2_operands (1, i.operands - 2);
3706 case 3:
3707 case 2:
3708 swap_2_operands (0, i.operands - 1);
3709 break;
3710 default:
3711 abort ();
3712 }
3713
3714 if (i.mem_operands == 2)
3715 {
3716 const seg_entry *temp_seg;
3717 temp_seg = i.seg[0];
3718 i.seg[0] = i.seg[1];
3719 i.seg[1] = temp_seg;
3720 }
3721 }
3722
3723 /* Try to ensure constant immediates are represented in the smallest
3724 opcode possible. */
3725 static void
3726 optimize_imm (void)
3727 {
3728 char guess_suffix = 0;
3729 int op;
3730
3731 if (i.suffix)
3732 guess_suffix = i.suffix;
3733 else if (i.reg_operands)
3734 {
3735 /* Figure out a suffix from the last register operand specified.
3736 We can't do this properly yet, ie. excluding InOutPortReg,
3737 but the following works for instructions with immediates.
3738 In any case, we can't set i.suffix yet. */
3739 for (op = i.operands; --op >= 0;)
3740 if (i.types[op].bitfield.reg8)
3741 {
3742 guess_suffix = BYTE_MNEM_SUFFIX;
3743 break;
3744 }
3745 else if (i.types[op].bitfield.reg16)
3746 {
3747 guess_suffix = WORD_MNEM_SUFFIX;
3748 break;
3749 }
3750 else if (i.types[op].bitfield.reg32)
3751 {
3752 guess_suffix = LONG_MNEM_SUFFIX;
3753 break;
3754 }
3755 else if (i.types[op].bitfield.reg64)
3756 {
3757 guess_suffix = QWORD_MNEM_SUFFIX;
3758 break;
3759 }
3760 }
3761 else if ((flag_code == CODE_16BIT) ^ (i.prefix[DATA_PREFIX] != 0))
3762 guess_suffix = WORD_MNEM_SUFFIX;
3763
3764 for (op = i.operands; --op >= 0;)
3765 if (operand_type_check (i.types[op], imm))
3766 {
3767 switch (i.op[op].imms->X_op)
3768 {
3769 case O_constant:
3770 /* If a suffix is given, this operand may be shortened. */
3771 switch (guess_suffix)
3772 {
3773 case LONG_MNEM_SUFFIX:
3774 i.types[op].bitfield.imm32 = 1;
3775 i.types[op].bitfield.imm64 = 1;
3776 break;
3777 case WORD_MNEM_SUFFIX:
3778 i.types[op].bitfield.imm16 = 1;
3779 i.types[op].bitfield.imm32 = 1;
3780 i.types[op].bitfield.imm32s = 1;
3781 i.types[op].bitfield.imm64 = 1;
3782 break;
3783 case BYTE_MNEM_SUFFIX:
3784 i.types[op].bitfield.imm8 = 1;
3785 i.types[op].bitfield.imm8s = 1;
3786 i.types[op].bitfield.imm16 = 1;
3787 i.types[op].bitfield.imm32 = 1;
3788 i.types[op].bitfield.imm32s = 1;
3789 i.types[op].bitfield.imm64 = 1;
3790 break;
3791 }
3792
3793 /* If this operand is at most 16 bits, convert it
3794 to a signed 16 bit number before trying to see
3795 whether it will fit in an even smaller size.
3796 This allows a 16-bit operand such as $0xffe0 to
3797 be recognised as within Imm8S range. */
3798 if ((i.types[op].bitfield.imm16)
3799 && (i.op[op].imms->X_add_number & ~(offsetT) 0xffff) == 0)
3800 {
3801 i.op[op].imms->X_add_number =
3802 (((i.op[op].imms->X_add_number & 0xffff) ^ 0x8000) - 0x8000);
3803 }
3804 if ((i.types[op].bitfield.imm32)
3805 && ((i.op[op].imms->X_add_number & ~(((offsetT) 2 << 31) - 1))
3806 == 0))
3807 {
3808 i.op[op].imms->X_add_number = ((i.op[op].imms->X_add_number
3809 ^ ((offsetT) 1 << 31))
3810 - ((offsetT) 1 << 31));
3811 }
3812 i.types[op]
3813 = operand_type_or (i.types[op],
3814 smallest_imm_type (i.op[op].imms->X_add_number));
3815
3816 /* We must avoid matching of Imm32 templates when 64bit
3817 only immediate is available. */
3818 if (guess_suffix == QWORD_MNEM_SUFFIX)
3819 i.types[op].bitfield.imm32 = 0;
3820 break;
3821
3822 case O_absent:
3823 case O_register:
3824 abort ();
3825
3826 /* Symbols and expressions. */
3827 default:
3828 /* Convert symbolic operand to proper sizes for matching, but don't
3829 prevent matching a set of insns that only supports sizes other
3830 than those matching the insn suffix. */
3831 {
3832 i386_operand_type mask, allowed;
3833 const insn_template *t;
3834
3835 operand_type_set (&mask, 0);
3836 operand_type_set (&allowed, 0);
3837
3838 for (t = current_templates->start;
3839 t < current_templates->end;
3840 ++t)
3841 allowed = operand_type_or (allowed,
3842 t->operand_types[op]);
3843 switch (guess_suffix)
3844 {
3845 case QWORD_MNEM_SUFFIX:
3846 mask.bitfield.imm64 = 1;
3847 mask.bitfield.imm32s = 1;
3848 break;
3849 case LONG_MNEM_SUFFIX:
3850 mask.bitfield.imm32 = 1;
3851 break;
3852 case WORD_MNEM_SUFFIX:
3853 mask.bitfield.imm16 = 1;
3854 break;
3855 case BYTE_MNEM_SUFFIX:
3856 mask.bitfield.imm8 = 1;
3857 break;
3858 default:
3859 break;
3860 }
3861 allowed = operand_type_and (mask, allowed);
3862 if (!operand_type_all_zero (&allowed))
3863 i.types[op] = operand_type_and (i.types[op], mask);
3864 }
3865 break;
3866 }
3867 }
3868 }
3869
3870 /* Try to use the smallest displacement type too. */
3871 static void
3872 optimize_disp (void)
3873 {
3874 int op;
3875
3876 for (op = i.operands; --op >= 0;)
3877 if (operand_type_check (i.types[op], disp))
3878 {
3879 if (i.op[op].disps->X_op == O_constant)
3880 {
3881 offsetT op_disp = i.op[op].disps->X_add_number;
3882
3883 if (i.types[op].bitfield.disp16
3884 && (op_disp & ~(offsetT) 0xffff) == 0)
3885 {
3886 /* If this operand is at most 16 bits, convert
3887 to a signed 16 bit number and don't use 64bit
3888 displacement. */
3889 op_disp = (((op_disp & 0xffff) ^ 0x8000) - 0x8000);
3890 i.types[op].bitfield.disp64 = 0;
3891 }
3892 if (i.types[op].bitfield.disp32
3893 && (op_disp & ~(((offsetT) 2 << 31) - 1)) == 0)
3894 {
3895 /* If this operand is at most 32 bits, convert
3896 to a signed 32 bit number and don't use 64bit
3897 displacement. */
3898 op_disp &= (((offsetT) 2 << 31) - 1);
3899 op_disp = (op_disp ^ ((offsetT) 1 << 31)) - ((addressT) 1 << 31);
3900 i.types[op].bitfield.disp64 = 0;
3901 }
3902 if (!op_disp && i.types[op].bitfield.baseindex)
3903 {
3904 i.types[op].bitfield.disp8 = 0;
3905 i.types[op].bitfield.disp16 = 0;
3906 i.types[op].bitfield.disp32 = 0;
3907 i.types[op].bitfield.disp32s = 0;
3908 i.types[op].bitfield.disp64 = 0;
3909 i.op[op].disps = 0;
3910 i.disp_operands--;
3911 }
3912 else if (flag_code == CODE_64BIT)
3913 {
3914 if (fits_in_signed_long (op_disp))
3915 {
3916 i.types[op].bitfield.disp64 = 0;
3917 i.types[op].bitfield.disp32s = 1;
3918 }
3919 if (i.prefix[ADDR_PREFIX]
3920 && fits_in_unsigned_long (op_disp))
3921 i.types[op].bitfield.disp32 = 1;
3922 }
3923 if ((i.types[op].bitfield.disp32
3924 || i.types[op].bitfield.disp32s
3925 || i.types[op].bitfield.disp16)
3926 && fits_in_signed_byte (op_disp))
3927 i.types[op].bitfield.disp8 = 1;
3928 }
3929 else if (i.reloc[op] == BFD_RELOC_386_TLS_DESC_CALL
3930 || i.reloc[op] == BFD_RELOC_X86_64_TLSDESC_CALL)
3931 {
3932 fix_new_exp (frag_now, frag_more (0) - frag_now->fr_literal, 0,
3933 i.op[op].disps, 0, i.reloc[op]);
3934 i.types[op].bitfield.disp8 = 0;
3935 i.types[op].bitfield.disp16 = 0;
3936 i.types[op].bitfield.disp32 = 0;
3937 i.types[op].bitfield.disp32s = 0;
3938 i.types[op].bitfield.disp64 = 0;
3939 }
3940 else
3941 /* We only support 64bit displacement on constants. */
3942 i.types[op].bitfield.disp64 = 0;
3943 }
3944 }
3945
3946 /* Check if operands are valid for the instruction. */
3947
3948 static int
3949 check_VecOperands (const insn_template *t)
3950 {
3951 /* Without VSIB byte, we can't have a vector register for index. */
3952 if (!t->opcode_modifier.vecsib
3953 && i.index_reg
3954 && (i.index_reg->reg_type.bitfield.regxmm
3955 || i.index_reg->reg_type.bitfield.regymm))
3956 {
3957 i.error = unsupported_vector_index_register;
3958 return 1;
3959 }
3960
3961 /* For VSIB byte, we need a vector register for index and no PC
3962 relative addressing is allowed. */
3963 if (t->opcode_modifier.vecsib
3964 && (!i.index_reg
3965 || !((t->opcode_modifier.vecsib == VecSIB128
3966 && i.index_reg->reg_type.bitfield.regxmm)
3967 || (t->opcode_modifier.vecsib == VecSIB256
3968 && i.index_reg->reg_type.bitfield.regymm))
3969 || (i.base_reg && i.base_reg->reg_num == RegRip)))
3970 {
3971 i.error = invalid_vsib_address;
3972 return 1;
3973 }
3974
3975 return 0;
3976 }
3977
3978 /* Check if operands are valid for the instruction. Update VEX
3979 operand types. */
3980
3981 static int
3982 VEX_check_operands (const insn_template *t)
3983 {
3984 if (!t->opcode_modifier.vex)
3985 return 0;
3986
3987 /* Only check VEX_Imm4, which must be the first operand. */
3988 if (t->operand_types[0].bitfield.vec_imm4)
3989 {
3990 if (i.op[0].imms->X_op != O_constant
3991 || !fits_in_imm4 (i.op[0].imms->X_add_number))
3992 {
3993 i.error = bad_imm4;
3994 return 1;
3995 }
3996
3997 /* Turn off Imm8 so that update_imm won't complain. */
3998 i.types[0] = vec_imm4;
3999 }
4000
4001 return 0;
4002 }
4003
4004 static const insn_template *
4005 match_template (void)
4006 {
4007 /* Points to template once we've found it. */
4008 const insn_template *t;
4009 i386_operand_type overlap0, overlap1, overlap2, overlap3;
4010 i386_operand_type overlap4;
4011 unsigned int found_reverse_match;
4012 i386_opcode_modifier suffix_check;
4013 i386_operand_type operand_types [MAX_OPERANDS];
4014 int addr_prefix_disp;
4015 unsigned int j;
4016 unsigned int found_cpu_match;
4017 unsigned int check_register;
4018 enum i386_error specific_error = 0;
4019
4020 #if MAX_OPERANDS != 5
4021 # error "MAX_OPERANDS must be 5."
4022 #endif
4023
4024 found_reverse_match = 0;
4025 addr_prefix_disp = -1;
4026
4027 memset (&suffix_check, 0, sizeof (suffix_check));
4028 if (i.suffix == BYTE_MNEM_SUFFIX)
4029 suffix_check.no_bsuf = 1;
4030 else if (i.suffix == WORD_MNEM_SUFFIX)
4031 suffix_check.no_wsuf = 1;
4032 else if (i.suffix == SHORT_MNEM_SUFFIX)
4033 suffix_check.no_ssuf = 1;
4034 else if (i.suffix == LONG_MNEM_SUFFIX)
4035 suffix_check.no_lsuf = 1;
4036 else if (i.suffix == QWORD_MNEM_SUFFIX)
4037 suffix_check.no_qsuf = 1;
4038 else if (i.suffix == LONG_DOUBLE_MNEM_SUFFIX)
4039 suffix_check.no_ldsuf = 1;
4040
4041 /* Must have right number of operands. */
4042 i.error = number_of_operands_mismatch;
4043
4044 for (t = current_templates->start; t < current_templates->end; t++)
4045 {
4046 addr_prefix_disp = -1;
4047
4048 if (i.operands != t->operands)
4049 continue;
4050
4051 /* Check processor support. */
4052 i.error = unsupported;
4053 found_cpu_match = (cpu_flags_match (t)
4054 == CPU_FLAGS_PERFECT_MATCH);
4055 if (!found_cpu_match)
4056 continue;
4057
4058 /* Check old gcc support. */
4059 i.error = old_gcc_only;
4060 if (!old_gcc && t->opcode_modifier.oldgcc)
4061 continue;
4062
4063 /* Check AT&T mnemonic. */
4064 i.error = unsupported_with_intel_mnemonic;
4065 if (intel_mnemonic && t->opcode_modifier.attmnemonic)
4066 continue;
4067
4068 /* Check AT&T/Intel syntax. */
4069 i.error = unsupported_syntax;
4070 if ((intel_syntax && t->opcode_modifier.attsyntax)
4071 || (!intel_syntax && t->opcode_modifier.intelsyntax))
4072 continue;
4073
4074 /* Check the suffix, except for some instructions in intel mode. */
4075 i.error = invalid_instruction_suffix;
4076 if ((!intel_syntax || !t->opcode_modifier.ignoresize)
4077 && ((t->opcode_modifier.no_bsuf && suffix_check.no_bsuf)
4078 || (t->opcode_modifier.no_wsuf && suffix_check.no_wsuf)
4079 || (t->opcode_modifier.no_lsuf && suffix_check.no_lsuf)
4080 || (t->opcode_modifier.no_ssuf && suffix_check.no_ssuf)
4081 || (t->opcode_modifier.no_qsuf && suffix_check.no_qsuf)
4082 || (t->opcode_modifier.no_ldsuf && suffix_check.no_ldsuf)))
4083 continue;
4084
4085 if (!operand_size_match (t))
4086 continue;
4087
4088 for (j = 0; j < MAX_OPERANDS; j++)
4089 operand_types[j] = t->operand_types[j];
4090
4091 /* In general, don't allow 64-bit operands in 32-bit mode. */
4092 if (i.suffix == QWORD_MNEM_SUFFIX
4093 && flag_code != CODE_64BIT
4094 && (intel_syntax
4095 ? (!t->opcode_modifier.ignoresize
4096 && !intel_float_operand (t->name))
4097 : intel_float_operand (t->name) != 2)
4098 && ((!operand_types[0].bitfield.regmmx
4099 && !operand_types[0].bitfield.regxmm
4100 && !operand_types[0].bitfield.regymm)
4101 || (!operand_types[t->operands > 1].bitfield.regmmx
4102 && !!operand_types[t->operands > 1].bitfield.regxmm
4103 && !!operand_types[t->operands > 1].bitfield.regymm))
4104 && (t->base_opcode != 0x0fc7
4105 || t->extension_opcode != 1 /* cmpxchg8b */))
4106 continue;
4107
4108 /* In general, don't allow 32-bit operands on pre-386. */
4109 else if (i.suffix == LONG_MNEM_SUFFIX
4110 && !cpu_arch_flags.bitfield.cpui386
4111 && (intel_syntax
4112 ? (!t->opcode_modifier.ignoresize
4113 && !intel_float_operand (t->name))
4114 : intel_float_operand (t->name) != 2)
4115 && ((!operand_types[0].bitfield.regmmx
4116 && !operand_types[0].bitfield.regxmm)
4117 || (!operand_types[t->operands > 1].bitfield.regmmx
4118 && !!operand_types[t->operands > 1].bitfield.regxmm)))
4119 continue;
4120
4121 /* Do not verify operands when there are none. */
4122 else
4123 {
4124 if (!t->operands)
4125 /* We've found a match; break out of loop. */
4126 break;
4127 }
4128
4129 /* Address size prefix will turn Disp64/Disp32/Disp16 operand
4130 into Disp32/Disp16/Disp32 operand. */
4131 if (i.prefix[ADDR_PREFIX] != 0)
4132 {
4133 /* There should be only one Disp operand. */
4134 switch (flag_code)
4135 {
4136 case CODE_16BIT:
4137 for (j = 0; j < MAX_OPERANDS; j++)
4138 {
4139 if (operand_types[j].bitfield.disp16)
4140 {
4141 addr_prefix_disp = j;
4142 operand_types[j].bitfield.disp32 = 1;
4143 operand_types[j].bitfield.disp16 = 0;
4144 break;
4145 }
4146 }
4147 break;
4148 case CODE_32BIT:
4149 for (j = 0; j < MAX_OPERANDS; j++)
4150 {
4151 if (operand_types[j].bitfield.disp32)
4152 {
4153 addr_prefix_disp = j;
4154 operand_types[j].bitfield.disp32 = 0;
4155 operand_types[j].bitfield.disp16 = 1;
4156 break;
4157 }
4158 }
4159 break;
4160 case CODE_64BIT:
4161 for (j = 0; j < MAX_OPERANDS; j++)
4162 {
4163 if (operand_types[j].bitfield.disp64)
4164 {
4165 addr_prefix_disp = j;
4166 operand_types[j].bitfield.disp64 = 0;
4167 operand_types[j].bitfield.disp32 = 1;
4168 break;
4169 }
4170 }
4171 break;
4172 }
4173 }
4174
4175 /* We check register size if needed. */
4176 check_register = t->opcode_modifier.checkregsize;
4177 overlap0 = operand_type_and (i.types[0], operand_types[0]);
4178 switch (t->operands)
4179 {
4180 case 1:
4181 if (!operand_type_match (overlap0, i.types[0]))
4182 continue;
4183 break;
4184 case 2:
4185 /* xchg %eax, %eax is a special case. It is an aliase for nop
4186 only in 32bit mode and we can use opcode 0x90. In 64bit
4187 mode, we can't use 0x90 for xchg %eax, %eax since it should
4188 zero-extend %eax to %rax. */
4189 if (flag_code == CODE_64BIT
4190 && t->base_opcode == 0x90
4191 && operand_type_equal (&i.types [0], &acc32)
4192 && operand_type_equal (&i.types [1], &acc32))
4193 continue;
4194 if (i.swap_operand)
4195 {
4196 /* If we swap operand in encoding, we either match
4197 the next one or reverse direction of operands. */
4198 if (t->opcode_modifier.s)
4199 continue;
4200 else if (t->opcode_modifier.d)
4201 goto check_reverse;
4202 }
4203
4204 case 3:
4205 /* If we swap operand in encoding, we match the next one. */
4206 if (i.swap_operand && t->opcode_modifier.s)
4207 continue;
4208 case 4:
4209 case 5:
4210 overlap1 = operand_type_and (i.types[1], operand_types[1]);
4211 if (!operand_type_match (overlap0, i.types[0])
4212 || !operand_type_match (overlap1, i.types[1])
4213 || (check_register
4214 && !operand_type_register_match (overlap0, i.types[0],
4215 operand_types[0],
4216 overlap1, i.types[1],
4217 operand_types[1])))
4218 {
4219 /* Check if other direction is valid ... */
4220 if (!t->opcode_modifier.d && !t->opcode_modifier.floatd)
4221 continue;
4222
4223 check_reverse:
4224 /* Try reversing direction of operands. */
4225 overlap0 = operand_type_and (i.types[0], operand_types[1]);
4226 overlap1 = operand_type_and (i.types[1], operand_types[0]);
4227 if (!operand_type_match (overlap0, i.types[0])
4228 || !operand_type_match (overlap1, i.types[1])
4229 || (check_register
4230 && !operand_type_register_match (overlap0,
4231 i.types[0],
4232 operand_types[1],
4233 overlap1,
4234 i.types[1],
4235 operand_types[0])))
4236 {
4237 /* Does not match either direction. */
4238 continue;
4239 }
4240 /* found_reverse_match holds which of D or FloatDR
4241 we've found. */
4242 if (t->opcode_modifier.d)
4243 found_reverse_match = Opcode_D;
4244 else if (t->opcode_modifier.floatd)
4245 found_reverse_match = Opcode_FloatD;
4246 else
4247 found_reverse_match = 0;
4248 if (t->opcode_modifier.floatr)
4249 found_reverse_match |= Opcode_FloatR;
4250 }
4251 else
4252 {
4253 /* Found a forward 2 operand match here. */
4254 switch (t->operands)
4255 {
4256 case 5:
4257 overlap4 = operand_type_and (i.types[4],
4258 operand_types[4]);
4259 case 4:
4260 overlap3 = operand_type_and (i.types[3],
4261 operand_types[3]);
4262 case 3:
4263 overlap2 = operand_type_and (i.types[2],
4264 operand_types[2]);
4265 break;
4266 }
4267
4268 switch (t->operands)
4269 {
4270 case 5:
4271 if (!operand_type_match (overlap4, i.types[4])
4272 || !operand_type_register_match (overlap3,
4273 i.types[3],
4274 operand_types[3],
4275 overlap4,
4276 i.types[4],
4277 operand_types[4]))
4278 continue;
4279 case 4:
4280 if (!operand_type_match (overlap3, i.types[3])
4281 || (check_register
4282 && !operand_type_register_match (overlap2,
4283 i.types[2],
4284 operand_types[2],
4285 overlap3,
4286 i.types[3],
4287 operand_types[3])))
4288 continue;
4289 case 3:
4290 /* Here we make use of the fact that there are no
4291 reverse match 3 operand instructions, and all 3
4292 operand instructions only need to be checked for
4293 register consistency between operands 2 and 3. */
4294 if (!operand_type_match (overlap2, i.types[2])
4295 || (check_register
4296 && !operand_type_register_match (overlap1,
4297 i.types[1],
4298 operand_types[1],
4299 overlap2,
4300 i.types[2],
4301 operand_types[2])))
4302 continue;
4303 break;
4304 }
4305 }
4306 /* Found either forward/reverse 2, 3 or 4 operand match here:
4307 slip through to break. */
4308 }
4309 if (!found_cpu_match)
4310 {
4311 found_reverse_match = 0;
4312 continue;
4313 }
4314
4315 /* Check if vector and VEX operands are valid. */
4316 if (check_VecOperands (t) || VEX_check_operands (t))
4317 {
4318 specific_error = i.error;
4319 continue;
4320 }
4321
4322 /* We've found a match; break out of loop. */
4323 break;
4324 }
4325
4326 if (t == current_templates->end)
4327 {
4328 /* We found no match. */
4329 const char *err_msg;
4330 switch (specific_error ? specific_error : i.error)
4331 {
4332 default:
4333 abort ();
4334 case operand_size_mismatch:
4335 err_msg = _("operand size mismatch");
4336 break;
4337 case operand_type_mismatch:
4338 err_msg = _("operand type mismatch");
4339 break;
4340 case register_type_mismatch:
4341 err_msg = _("register type mismatch");
4342 break;
4343 case number_of_operands_mismatch:
4344 err_msg = _("number of operands mismatch");
4345 break;
4346 case invalid_instruction_suffix:
4347 err_msg = _("invalid instruction suffix");
4348 break;
4349 case bad_imm4:
4350 err_msg = _("constant doesn't fit in 4 bits");
4351 break;
4352 case old_gcc_only:
4353 err_msg = _("only supported with old gcc");
4354 break;
4355 case unsupported_with_intel_mnemonic:
4356 err_msg = _("unsupported with Intel mnemonic");
4357 break;
4358 case unsupported_syntax:
4359 err_msg = _("unsupported syntax");
4360 break;
4361 case unsupported:
4362 as_bad (_("unsupported instruction `%s'"),
4363 current_templates->start->name);
4364 return NULL;
4365 case invalid_vsib_address:
4366 err_msg = _("invalid VSIB address");
4367 break;
4368 case unsupported_vector_index_register:
4369 err_msg = _("unsupported vector index register");
4370 break;
4371 }
4372 as_bad (_("%s for `%s'"), err_msg,
4373 current_templates->start->name);
4374 return NULL;
4375 }
4376
4377 if (!quiet_warnings)
4378 {
4379 if (!intel_syntax
4380 && (i.types[0].bitfield.jumpabsolute
4381 != operand_types[0].bitfield.jumpabsolute))
4382 {
4383 as_warn (_("indirect %s without `*'"), t->name);
4384 }
4385
4386 if (t->opcode_modifier.isprefix
4387 && t->opcode_modifier.ignoresize)
4388 {
4389 /* Warn them that a data or address size prefix doesn't
4390 affect assembly of the next line of code. */
4391 as_warn (_("stand-alone `%s' prefix"), t->name);
4392 }
4393 }
4394
4395 /* Copy the template we found. */
4396 i.tm = *t;
4397
4398 if (addr_prefix_disp != -1)
4399 i.tm.operand_types[addr_prefix_disp]
4400 = operand_types[addr_prefix_disp];
4401
4402 if (found_reverse_match)
4403 {
4404 /* If we found a reverse match we must alter the opcode
4405 direction bit. found_reverse_match holds bits to change
4406 (different for int & float insns). */
4407
4408 i.tm.base_opcode ^= found_reverse_match;
4409
4410 i.tm.operand_types[0] = operand_types[1];
4411 i.tm.operand_types[1] = operand_types[0];
4412 }
4413
4414 return t;
4415 }
4416
4417 static int
4418 check_string (void)
4419 {
4420 int mem_op = operand_type_check (i.types[0], anymem) ? 0 : 1;
4421 if (i.tm.operand_types[mem_op].bitfield.esseg)
4422 {
4423 if (i.seg[0] != NULL && i.seg[0] != &es)
4424 {
4425 as_bad (_("`%s' operand %d must use `%ses' segment"),
4426 i.tm.name,
4427 mem_op + 1,
4428 register_prefix);
4429 return 0;
4430 }
4431 /* There's only ever one segment override allowed per instruction.
4432 This instruction possibly has a legal segment override on the
4433 second operand, so copy the segment to where non-string
4434 instructions store it, allowing common code. */
4435 i.seg[0] = i.seg[1];
4436 }
4437 else if (i.tm.operand_types[mem_op + 1].bitfield.esseg)
4438 {
4439 if (i.seg[1] != NULL && i.seg[1] != &es)
4440 {
4441 as_bad (_("`%s' operand %d must use `%ses' segment"),
4442 i.tm.name,
4443 mem_op + 2,
4444 register_prefix);
4445 return 0;
4446 }
4447 }
4448 return 1;
4449 }
4450
4451 static int
4452 process_suffix (void)
4453 {
4454 /* If matched instruction specifies an explicit instruction mnemonic
4455 suffix, use it. */
4456 if (i.tm.opcode_modifier.size16)
4457 i.suffix = WORD_MNEM_SUFFIX;
4458 else if (i.tm.opcode_modifier.size32)
4459 i.suffix = LONG_MNEM_SUFFIX;
4460 else if (i.tm.opcode_modifier.size64)
4461 i.suffix = QWORD_MNEM_SUFFIX;
4462 else if (i.reg_operands)
4463 {
4464 /* If there's no instruction mnemonic suffix we try to invent one
4465 based on register operands. */
4466 if (!i.suffix)
4467 {
4468 /* We take i.suffix from the last register operand specified,
4469 Destination register type is more significant than source
4470 register type. crc32 in SSE4.2 prefers source register
4471 type. */
4472 if (i.tm.base_opcode == 0xf20f38f1)
4473 {
4474 if (i.types[0].bitfield.reg16)
4475 i.suffix = WORD_MNEM_SUFFIX;
4476 else if (i.types[0].bitfield.reg32)
4477 i.suffix = LONG_MNEM_SUFFIX;
4478 else if (i.types[0].bitfield.reg64)
4479 i.suffix = QWORD_MNEM_SUFFIX;
4480 }
4481 else if (i.tm.base_opcode == 0xf20f38f0)
4482 {
4483 if (i.types[0].bitfield.reg8)
4484 i.suffix = BYTE_MNEM_SUFFIX;
4485 }
4486
4487 if (!i.suffix)
4488 {
4489 int op;
4490
4491 if (i.tm.base_opcode == 0xf20f38f1
4492 || i.tm.base_opcode == 0xf20f38f0)
4493 {
4494 /* We have to know the operand size for crc32. */
4495 as_bad (_("ambiguous memory operand size for `%s`"),
4496 i.tm.name);
4497 return 0;
4498 }
4499
4500 for (op = i.operands; --op >= 0;)
4501 if (!i.tm.operand_types[op].bitfield.inoutportreg)
4502 {
4503 if (i.types[op].bitfield.reg8)
4504 {
4505 i.suffix = BYTE_MNEM_SUFFIX;
4506 break;
4507 }
4508 else if (i.types[op].bitfield.reg16)
4509 {
4510 i.suffix = WORD_MNEM_SUFFIX;
4511 break;
4512 }
4513 else if (i.types[op].bitfield.reg32)
4514 {
4515 i.suffix = LONG_MNEM_SUFFIX;
4516 break;
4517 }
4518 else if (i.types[op].bitfield.reg64)
4519 {
4520 i.suffix = QWORD_MNEM_SUFFIX;
4521 break;
4522 }
4523 }
4524 }
4525 }
4526 else if (i.suffix == BYTE_MNEM_SUFFIX)
4527 {
4528 if (intel_syntax
4529 && i.tm.opcode_modifier.ignoresize
4530 && i.tm.opcode_modifier.no_bsuf)
4531 i.suffix = 0;
4532 else if (!check_byte_reg ())
4533 return 0;
4534 }
4535 else if (i.suffix == LONG_MNEM_SUFFIX)
4536 {
4537 if (intel_syntax
4538 && i.tm.opcode_modifier.ignoresize
4539 && i.tm.opcode_modifier.no_lsuf)
4540 i.suffix = 0;
4541 else if (!check_long_reg ())
4542 return 0;
4543 }
4544 else if (i.suffix == QWORD_MNEM_SUFFIX)
4545 {
4546 if (intel_syntax
4547 && i.tm.opcode_modifier.ignoresize
4548 && i.tm.opcode_modifier.no_qsuf)
4549 i.suffix = 0;
4550 else if (!check_qword_reg ())
4551 return 0;
4552 }
4553 else if (i.suffix == WORD_MNEM_SUFFIX)
4554 {
4555 if (intel_syntax
4556 && i.tm.opcode_modifier.ignoresize
4557 && i.tm.opcode_modifier.no_wsuf)
4558 i.suffix = 0;
4559 else if (!check_word_reg ())
4560 return 0;
4561 }
4562 else if (i.suffix == XMMWORD_MNEM_SUFFIX
4563 || i.suffix == YMMWORD_MNEM_SUFFIX)
4564 {
4565 /* Skip if the instruction has x/y suffix. match_template
4566 should check if it is a valid suffix. */
4567 }
4568 else if (intel_syntax && i.tm.opcode_modifier.ignoresize)
4569 /* Do nothing if the instruction is going to ignore the prefix. */
4570 ;
4571 else
4572 abort ();
4573 }
4574 else if (i.tm.opcode_modifier.defaultsize
4575 && !i.suffix
4576 /* exclude fldenv/frstor/fsave/fstenv */
4577 && i.tm.opcode_modifier.no_ssuf)
4578 {
4579 i.suffix = stackop_size;
4580 }
4581 else if (intel_syntax
4582 && !i.suffix
4583 && (i.tm.operand_types[0].bitfield.jumpabsolute
4584 || i.tm.opcode_modifier.jumpbyte
4585 || i.tm.opcode_modifier.jumpintersegment
4586 || (i.tm.base_opcode == 0x0f01 /* [ls][gi]dt */
4587 && i.tm.extension_opcode <= 3)))
4588 {
4589 switch (flag_code)
4590 {
4591 case CODE_64BIT:
4592 if (!i.tm.opcode_modifier.no_qsuf)
4593 {
4594 i.suffix = QWORD_MNEM_SUFFIX;
4595 break;
4596 }
4597 case CODE_32BIT:
4598 if (!i.tm.opcode_modifier.no_lsuf)
4599 i.suffix = LONG_MNEM_SUFFIX;
4600 break;
4601 case CODE_16BIT:
4602 if (!i.tm.opcode_modifier.no_wsuf)
4603 i.suffix = WORD_MNEM_SUFFIX;
4604 break;
4605 }
4606 }
4607
4608 if (!i.suffix)
4609 {
4610 if (!intel_syntax)
4611 {
4612 if (i.tm.opcode_modifier.w)
4613 {
4614 as_bad (_("no instruction mnemonic suffix given and "
4615 "no register operands; can't size instruction"));
4616 return 0;
4617 }
4618 }
4619 else
4620 {
4621 unsigned int suffixes;
4622
4623 suffixes = !i.tm.opcode_modifier.no_bsuf;
4624 if (!i.tm.opcode_modifier.no_wsuf)
4625 suffixes |= 1 << 1;
4626 if (!i.tm.opcode_modifier.no_lsuf)
4627 suffixes |= 1 << 2;
4628 if (!i.tm.opcode_modifier.no_ldsuf)
4629 suffixes |= 1 << 3;
4630 if (!i.tm.opcode_modifier.no_ssuf)
4631 suffixes |= 1 << 4;
4632 if (!i.tm.opcode_modifier.no_qsuf)
4633 suffixes |= 1 << 5;
4634
4635 /* There are more than suffix matches. */
4636 if (i.tm.opcode_modifier.w
4637 || ((suffixes & (suffixes - 1))
4638 && !i.tm.opcode_modifier.defaultsize
4639 && !i.tm.opcode_modifier.ignoresize))
4640 {
4641 as_bad (_("ambiguous operand size for `%s'"), i.tm.name);
4642 return 0;
4643 }
4644 }
4645 }
4646
4647 /* Change the opcode based on the operand size given by i.suffix;
4648 We don't need to change things for byte insns. */
4649
4650 if (i.suffix
4651 && i.suffix != BYTE_MNEM_SUFFIX
4652 && i.suffix != XMMWORD_MNEM_SUFFIX
4653 && i.suffix != YMMWORD_MNEM_SUFFIX)
4654 {
4655 /* It's not a byte, select word/dword operation. */
4656 if (i.tm.opcode_modifier.w)
4657 {
4658 if (i.tm.opcode_modifier.shortform)
4659 i.tm.base_opcode |= 8;
4660 else
4661 i.tm.base_opcode |= 1;
4662 }
4663
4664 /* Now select between word & dword operations via the operand
4665 size prefix, except for instructions that will ignore this
4666 prefix anyway. */
4667 if (i.tm.opcode_modifier.addrprefixop0)
4668 {
4669 /* The address size override prefix changes the size of the
4670 first operand. */
4671 if ((flag_code == CODE_32BIT
4672 && i.op->regs[0].reg_type.bitfield.reg16)
4673 || (flag_code != CODE_32BIT
4674 && i.op->regs[0].reg_type.bitfield.reg32))
4675 if (!add_prefix (ADDR_PREFIX_OPCODE))
4676 return 0;
4677 }
4678 else if (i.suffix != QWORD_MNEM_SUFFIX
4679 && i.suffix != LONG_DOUBLE_MNEM_SUFFIX
4680 && !i.tm.opcode_modifier.ignoresize
4681 && !i.tm.opcode_modifier.floatmf
4682 && ((i.suffix == LONG_MNEM_SUFFIX) == (flag_code == CODE_16BIT)
4683 || (flag_code == CODE_64BIT
4684 && i.tm.opcode_modifier.jumpbyte)))
4685 {
4686 unsigned int prefix = DATA_PREFIX_OPCODE;
4687
4688 if (i.tm.opcode_modifier.jumpbyte) /* jcxz, loop */
4689 prefix = ADDR_PREFIX_OPCODE;
4690
4691 if (!add_prefix (prefix))
4692 return 0;
4693 }
4694
4695 /* Set mode64 for an operand. */
4696 if (i.suffix == QWORD_MNEM_SUFFIX
4697 && flag_code == CODE_64BIT
4698 && !i.tm.opcode_modifier.norex64)
4699 {
4700 /* Special case for xchg %rax,%rax. It is NOP and doesn't
4701 need rex64. cmpxchg8b is also a special case. */
4702 if (! (i.operands == 2
4703 && i.tm.base_opcode == 0x90
4704 && i.tm.extension_opcode == None
4705 && operand_type_equal (&i.types [0], &acc64)
4706 && operand_type_equal (&i.types [1], &acc64))
4707 && ! (i.operands == 1
4708 && i.tm.base_opcode == 0xfc7
4709 && i.tm.extension_opcode == 1
4710 && !operand_type_check (i.types [0], reg)
4711 && operand_type_check (i.types [0], anymem)))
4712 i.rex |= REX_W;
4713 }
4714
4715 /* Size floating point instruction. */
4716 if (i.suffix == LONG_MNEM_SUFFIX)
4717 if (i.tm.opcode_modifier.floatmf)
4718 i.tm.base_opcode ^= 4;
4719 }
4720
4721 return 1;
4722 }
4723
4724 static int
4725 check_byte_reg (void)
4726 {
4727 int op;
4728
4729 for (op = i.operands; --op >= 0;)
4730 {
4731 /* If this is an eight bit register, it's OK. If it's the 16 or
4732 32 bit version of an eight bit register, we will just use the
4733 low portion, and that's OK too. */
4734 if (i.types[op].bitfield.reg8)
4735 continue;
4736
4737 /* I/O port address operands are OK too. */
4738 if (i.tm.operand_types[op].bitfield.inoutportreg)
4739 continue;
4740
4741 /* crc32 doesn't generate this warning. */
4742 if (i.tm.base_opcode == 0xf20f38f0)
4743 continue;
4744
4745 if ((i.types[op].bitfield.reg16
4746 || i.types[op].bitfield.reg32
4747 || i.types[op].bitfield.reg64)
4748 && i.op[op].regs->reg_num < 4
4749 /* Prohibit these changes in 64bit mode, since the lowering
4750 would be more complicated. */
4751 && flag_code != CODE_64BIT)
4752 {
4753 #if REGISTER_WARNINGS
4754 if (!quiet_warnings)
4755 as_warn (_("using `%s%s' instead of `%s%s' due to `%c' suffix"),
4756 register_prefix,
4757 (i.op[op].regs + (i.types[op].bitfield.reg16
4758 ? REGNAM_AL - REGNAM_AX
4759 : REGNAM_AL - REGNAM_EAX))->reg_name,
4760 register_prefix,
4761 i.op[op].regs->reg_name,
4762 i.suffix);
4763 #endif
4764 continue;
4765 }
4766 /* Any other register is bad. */
4767 if (i.types[op].bitfield.reg16
4768 || i.types[op].bitfield.reg32
4769 || i.types[op].bitfield.reg64
4770 || i.types[op].bitfield.regmmx
4771 || i.types[op].bitfield.regxmm
4772 || i.types[op].bitfield.regymm
4773 || i.types[op].bitfield.sreg2
4774 || i.types[op].bitfield.sreg3
4775 || i.types[op].bitfield.control
4776 || i.types[op].bitfield.debug
4777 || i.types[op].bitfield.test
4778 || i.types[op].bitfield.floatreg
4779 || i.types[op].bitfield.floatacc)
4780 {
4781 as_bad (_("`%s%s' not allowed with `%s%c'"),
4782 register_prefix,
4783 i.op[op].regs->reg_name,
4784 i.tm.name,
4785 i.suffix);
4786 return 0;
4787 }
4788 }
4789 return 1;
4790 }
4791
4792 static int
4793 check_long_reg (void)
4794 {
4795 int op;
4796
4797 for (op = i.operands; --op >= 0;)
4798 /* Reject eight bit registers, except where the template requires
4799 them. (eg. movzb) */
4800 if (i.types[op].bitfield.reg8
4801 && (i.tm.operand_types[op].bitfield.reg16
4802 || i.tm.operand_types[op].bitfield.reg32
4803 || i.tm.operand_types[op].bitfield.acc))
4804 {
4805 as_bad (_("`%s%s' not allowed with `%s%c'"),
4806 register_prefix,
4807 i.op[op].regs->reg_name,
4808 i.tm.name,
4809 i.suffix);
4810 return 0;
4811 }
4812 /* Warn if the e prefix on a general reg is missing. */
4813 else if ((!quiet_warnings || flag_code == CODE_64BIT)
4814 && i.types[op].bitfield.reg16
4815 && (i.tm.operand_types[op].bitfield.reg32
4816 || i.tm.operand_types[op].bitfield.acc))
4817 {
4818 /* Prohibit these changes in the 64bit mode, since the
4819 lowering is more complicated. */
4820 if (flag_code == CODE_64BIT)
4821 {
4822 as_bad (_("incorrect register `%s%s' used with `%c' suffix"),
4823 register_prefix, i.op[op].regs->reg_name,
4824 i.suffix);
4825 return 0;
4826 }
4827 #if REGISTER_WARNINGS
4828 else
4829 as_warn (_("using `%s%s' instead of `%s%s' due to `%c' suffix"),
4830 register_prefix,
4831 (i.op[op].regs + REGNAM_EAX - REGNAM_AX)->reg_name,
4832 register_prefix,
4833 i.op[op].regs->reg_name,
4834 i.suffix);
4835 #endif
4836 }
4837 /* Warn if the r prefix on a general reg is missing. */
4838 else if (i.types[op].bitfield.reg64
4839 && (i.tm.operand_types[op].bitfield.reg32
4840 || i.tm.operand_types[op].bitfield.acc))
4841 {
4842 if (intel_syntax
4843 && i.tm.opcode_modifier.toqword
4844 && !i.types[0].bitfield.regxmm)
4845 {
4846 /* Convert to QWORD. We want REX byte. */
4847 i.suffix = QWORD_MNEM_SUFFIX;
4848 }
4849 else
4850 {
4851 as_bad (_("incorrect register `%s%s' used with `%c' suffix"),
4852 register_prefix, i.op[op].regs->reg_name,
4853 i.suffix);
4854 return 0;
4855 }
4856 }
4857 return 1;
4858 }
4859
4860 static int
4861 check_qword_reg (void)
4862 {
4863 int op;
4864
4865 for (op = i.operands; --op >= 0; )
4866 /* Reject eight bit registers, except where the template requires
4867 them. (eg. movzb) */
4868 if (i.types[op].bitfield.reg8
4869 && (i.tm.operand_types[op].bitfield.reg16
4870 || i.tm.operand_types[op].bitfield.reg32
4871 || i.tm.operand_types[op].bitfield.acc))
4872 {
4873 as_bad (_("`%s%s' not allowed with `%s%c'"),
4874 register_prefix,
4875 i.op[op].regs->reg_name,
4876 i.tm.name,
4877 i.suffix);
4878 return 0;
4879 }
4880 /* Warn if the e prefix on a general reg is missing. */
4881 else if ((i.types[op].bitfield.reg16
4882 || i.types[op].bitfield.reg32)
4883 && (i.tm.operand_types[op].bitfield.reg32
4884 || i.tm.operand_types[op].bitfield.acc))
4885 {
4886 /* Prohibit these changes in the 64bit mode, since the
4887 lowering is more complicated. */
4888 if (intel_syntax
4889 && i.tm.opcode_modifier.todword
4890 && !i.types[0].bitfield.regxmm)
4891 {
4892 /* Convert to DWORD. We don't want REX byte. */
4893 i.suffix = LONG_MNEM_SUFFIX;
4894 }
4895 else
4896 {
4897 as_bad (_("incorrect register `%s%s' used with `%c' suffix"),
4898 register_prefix, i.op[op].regs->reg_name,
4899 i.suffix);
4900 return 0;
4901 }
4902 }
4903 return 1;
4904 }
4905
4906 static int
4907 check_word_reg (void)
4908 {
4909 int op;
4910 for (op = i.operands; --op >= 0;)
4911 /* Reject eight bit registers, except where the template requires
4912 them. (eg. movzb) */
4913 if (i.types[op].bitfield.reg8
4914 && (i.tm.operand_types[op].bitfield.reg16
4915 || i.tm.operand_types[op].bitfield.reg32
4916 || i.tm.operand_types[op].bitfield.acc))
4917 {
4918 as_bad (_("`%s%s' not allowed with `%s%c'"),
4919 register_prefix,
4920 i.op[op].regs->reg_name,
4921 i.tm.name,
4922 i.suffix);
4923 return 0;
4924 }
4925 /* Warn if the e prefix on a general reg is present. */
4926 else if ((!quiet_warnings || flag_code == CODE_64BIT)
4927 && i.types[op].bitfield.reg32
4928 && (i.tm.operand_types[op].bitfield.reg16
4929 || i.tm.operand_types[op].bitfield.acc))
4930 {
4931 /* Prohibit these changes in the 64bit mode, since the
4932 lowering is more complicated. */
4933 if (flag_code == CODE_64BIT)
4934 {
4935 as_bad (_("incorrect register `%s%s' used with `%c' suffix"),
4936 register_prefix, i.op[op].regs->reg_name,
4937 i.suffix);
4938 return 0;
4939 }
4940 else
4941 #if REGISTER_WARNINGS
4942 as_warn (_("using `%s%s' instead of `%s%s' due to `%c' suffix"),
4943 register_prefix,
4944 (i.op[op].regs + REGNAM_AX - REGNAM_EAX)->reg_name,
4945 register_prefix,
4946 i.op[op].regs->reg_name,
4947 i.suffix);
4948 #endif
4949 }
4950 return 1;
4951 }
4952
4953 static int
4954 update_imm (unsigned int j)
4955 {
4956 i386_operand_type overlap = i.types[j];
4957 if ((overlap.bitfield.imm8
4958 || overlap.bitfield.imm8s
4959 || overlap.bitfield.imm16
4960 || overlap.bitfield.imm32
4961 || overlap.bitfield.imm32s
4962 || overlap.bitfield.imm64)
4963 && !operand_type_equal (&overlap, &imm8)
4964 && !operand_type_equal (&overlap, &imm8s)
4965 && !operand_type_equal (&overlap, &imm16)
4966 && !operand_type_equal (&overlap, &imm32)
4967 && !operand_type_equal (&overlap, &imm32s)
4968 && !operand_type_equal (&overlap, &imm64))
4969 {
4970 if (i.suffix)
4971 {
4972 i386_operand_type temp;
4973
4974 operand_type_set (&temp, 0);
4975 if (i.suffix == BYTE_MNEM_SUFFIX)
4976 {
4977 temp.bitfield.imm8 = overlap.bitfield.imm8;
4978 temp.bitfield.imm8s = overlap.bitfield.imm8s;
4979 }
4980 else if (i.suffix == WORD_MNEM_SUFFIX)
4981 temp.bitfield.imm16 = overlap.bitfield.imm16;
4982 else if (i.suffix == QWORD_MNEM_SUFFIX)
4983 {
4984 temp.bitfield.imm64 = overlap.bitfield.imm64;
4985 temp.bitfield.imm32s = overlap.bitfield.imm32s;
4986 }
4987 else
4988 temp.bitfield.imm32 = overlap.bitfield.imm32;
4989 overlap = temp;
4990 }
4991 else if (operand_type_equal (&overlap, &imm16_32_32s)
4992 || operand_type_equal (&overlap, &imm16_32)
4993 || operand_type_equal (&overlap, &imm16_32s))
4994 {
4995 if ((flag_code == CODE_16BIT) ^ (i.prefix[DATA_PREFIX] != 0))
4996 overlap = imm16;
4997 else
4998 overlap = imm32s;
4999 }
5000 if (!operand_type_equal (&overlap, &imm8)
5001 && !operand_type_equal (&overlap, &imm8s)
5002 && !operand_type_equal (&overlap, &imm16)
5003 && !operand_type_equal (&overlap, &imm32)
5004 && !operand_type_equal (&overlap, &imm32s)
5005 && !operand_type_equal (&overlap, &imm64))
5006 {
5007 as_bad (_("no instruction mnemonic suffix given; "
5008 "can't determine immediate size"));
5009 return 0;
5010 }
5011 }
5012 i.types[j] = overlap;
5013
5014 return 1;
5015 }
5016
5017 static int
5018 finalize_imm (void)
5019 {
5020 unsigned int j, n;
5021
5022 /* Update the first 2 immediate operands. */
5023 n = i.operands > 2 ? 2 : i.operands;
5024 if (n)
5025 {
5026 for (j = 0; j < n; j++)
5027 if (update_imm (j) == 0)
5028 return 0;
5029
5030 /* The 3rd operand can't be immediate operand. */
5031 gas_assert (operand_type_check (i.types[2], imm) == 0);
5032 }
5033
5034 return 1;
5035 }
5036
5037 static int
5038 bad_implicit_operand (int xmm)
5039 {
5040 const char *ireg = xmm ? "xmm0" : "ymm0";
5041
5042 if (intel_syntax)
5043 as_bad (_("the last operand of `%s' must be `%s%s'"),
5044 i.tm.name, register_prefix, ireg);
5045 else
5046 as_bad (_("the first operand of `%s' must be `%s%s'"),
5047 i.tm.name, register_prefix, ireg);
5048 return 0;
5049 }
5050
5051 static int
5052 process_operands (void)
5053 {
5054 /* Default segment register this instruction will use for memory
5055 accesses. 0 means unknown. This is only for optimizing out
5056 unnecessary segment overrides. */
5057 const seg_entry *default_seg = 0;
5058
5059 if (i.tm.opcode_modifier.sse2avx && i.tm.opcode_modifier.vexvvvv)
5060 {
5061 unsigned int dupl = i.operands;
5062 unsigned int dest = dupl - 1;
5063 unsigned int j;
5064
5065 /* The destination must be an xmm register. */
5066 gas_assert (i.reg_operands
5067 && MAX_OPERANDS > dupl
5068 && operand_type_equal (&i.types[dest], &regxmm));
5069
5070 if (i.tm.opcode_modifier.firstxmm0)
5071 {
5072 /* The first operand is implicit and must be xmm0. */
5073 gas_assert (operand_type_equal (&i.types[0], &regxmm));
5074 if (i.op[0].regs->reg_num != 0)
5075 return bad_implicit_operand (1);
5076
5077 if (i.tm.opcode_modifier.vexsources == VEX3SOURCES)
5078 {
5079 /* Keep xmm0 for instructions with VEX prefix and 3
5080 sources. */
5081 goto duplicate;
5082 }
5083 else
5084 {
5085 /* We remove the first xmm0 and keep the number of
5086 operands unchanged, which in fact duplicates the
5087 destination. */
5088 for (j = 1; j < i.operands; j++)
5089 {
5090 i.op[j - 1] = i.op[j];
5091 i.types[j - 1] = i.types[j];
5092 i.tm.operand_types[j - 1] = i.tm.operand_types[j];
5093 }
5094 }
5095 }
5096 else if (i.tm.opcode_modifier.implicit1stxmm0)
5097 {
5098 gas_assert ((MAX_OPERANDS - 1) > dupl
5099 && (i.tm.opcode_modifier.vexsources
5100 == VEX3SOURCES));
5101
5102 /* Add the implicit xmm0 for instructions with VEX prefix
5103 and 3 sources. */
5104 for (j = i.operands; j > 0; j--)
5105 {
5106 i.op[j] = i.op[j - 1];
5107 i.types[j] = i.types[j - 1];
5108 i.tm.operand_types[j] = i.tm.operand_types[j - 1];
5109 }
5110 i.op[0].regs
5111 = (const reg_entry *) hash_find (reg_hash, "xmm0");
5112 i.types[0] = regxmm;
5113 i.tm.operand_types[0] = regxmm;
5114
5115 i.operands += 2;
5116 i.reg_operands += 2;
5117 i.tm.operands += 2;
5118
5119 dupl++;
5120 dest++;
5121 i.op[dupl] = i.op[dest];
5122 i.types[dupl] = i.types[dest];
5123 i.tm.operand_types[dupl] = i.tm.operand_types[dest];
5124 }
5125 else
5126 {
5127 duplicate:
5128 i.operands++;
5129 i.reg_operands++;
5130 i.tm.operands++;
5131
5132 i.op[dupl] = i.op[dest];
5133 i.types[dupl] = i.types[dest];
5134 i.tm.operand_types[dupl] = i.tm.operand_types[dest];
5135 }
5136
5137 if (i.tm.opcode_modifier.immext)
5138 process_immext ();
5139 }
5140 else if (i.tm.opcode_modifier.firstxmm0)
5141 {
5142 unsigned int j;
5143
5144 /* The first operand is implicit and must be xmm0/ymm0. */
5145 gas_assert (i.reg_operands
5146 && (operand_type_equal (&i.types[0], &regxmm)
5147 || operand_type_equal (&i.types[0], &regymm)));
5148 if (i.op[0].regs->reg_num != 0)
5149 return bad_implicit_operand (i.types[0].bitfield.regxmm);
5150
5151 for (j = 1; j < i.operands; j++)
5152 {
5153 i.op[j - 1] = i.op[j];
5154 i.types[j - 1] = i.types[j];
5155
5156 /* We need to adjust fields in i.tm since they are used by
5157 build_modrm_byte. */
5158 i.tm.operand_types [j - 1] = i.tm.operand_types [j];
5159 }
5160
5161 i.operands--;
5162 i.reg_operands--;
5163 i.tm.operands--;
5164 }
5165 else if (i.tm.opcode_modifier.regkludge)
5166 {
5167 /* The imul $imm, %reg instruction is converted into
5168 imul $imm, %reg, %reg, and the clr %reg instruction
5169 is converted into xor %reg, %reg. */
5170
5171 unsigned int first_reg_op;
5172
5173 if (operand_type_check (i.types[0], reg))
5174 first_reg_op = 0;
5175 else
5176 first_reg_op = 1;
5177 /* Pretend we saw the extra register operand. */
5178 gas_assert (i.reg_operands == 1
5179 && i.op[first_reg_op + 1].regs == 0);
5180 i.op[first_reg_op + 1].regs = i.op[first_reg_op].regs;
5181 i.types[first_reg_op + 1] = i.types[first_reg_op];
5182 i.operands++;
5183 i.reg_operands++;
5184 }
5185
5186 if (i.tm.opcode_modifier.shortform)
5187 {
5188 if (i.types[0].bitfield.sreg2
5189 || i.types[0].bitfield.sreg3)
5190 {
5191 if (i.tm.base_opcode == POP_SEG_SHORT
5192 && i.op[0].regs->reg_num == 1)
5193 {
5194 as_bad (_("you can't `pop %scs'"), register_prefix);
5195 return 0;
5196 }
5197 i.tm.base_opcode |= (i.op[0].regs->reg_num << 3);
5198 if ((i.op[0].regs->reg_flags & RegRex) != 0)
5199 i.rex |= REX_B;
5200 }
5201 else
5202 {
5203 /* The register or float register operand is in operand
5204 0 or 1. */
5205 unsigned int op;
5206
5207 if (i.types[0].bitfield.floatreg
5208 || operand_type_check (i.types[0], reg))
5209 op = 0;
5210 else
5211 op = 1;
5212 /* Register goes in low 3 bits of opcode. */
5213 i.tm.base_opcode |= i.op[op].regs->reg_num;
5214 if ((i.op[op].regs->reg_flags & RegRex) != 0)
5215 i.rex |= REX_B;
5216 if (!quiet_warnings && i.tm.opcode_modifier.ugh)
5217 {
5218 /* Warn about some common errors, but press on regardless.
5219 The first case can be generated by gcc (<= 2.8.1). */
5220 if (i.operands == 2)
5221 {
5222 /* Reversed arguments on faddp, fsubp, etc. */
5223 as_warn (_("translating to `%s %s%s,%s%s'"), i.tm.name,
5224 register_prefix, i.op[!intel_syntax].regs->reg_name,
5225 register_prefix, i.op[intel_syntax].regs->reg_name);
5226 }
5227 else
5228 {
5229 /* Extraneous `l' suffix on fp insn. */
5230 as_warn (_("translating to `%s %s%s'"), i.tm.name,
5231 register_prefix, i.op[0].regs->reg_name);
5232 }
5233 }
5234 }
5235 }
5236 else if (i.tm.opcode_modifier.modrm)
5237 {
5238 /* The opcode is completed (modulo i.tm.extension_opcode which
5239 must be put into the modrm byte). Now, we make the modrm and
5240 index base bytes based on all the info we've collected. */
5241
5242 default_seg = build_modrm_byte ();
5243 }
5244 else if ((i.tm.base_opcode & ~0x3) == MOV_AX_DISP32)
5245 {
5246 default_seg = &ds;
5247 }
5248 else if (i.tm.opcode_modifier.isstring)
5249 {
5250 /* For the string instructions that allow a segment override
5251 on one of their operands, the default segment is ds. */
5252 default_seg = &ds;
5253 }
5254
5255 if (i.tm.base_opcode == 0x8d /* lea */
5256 && i.seg[0]
5257 && !quiet_warnings)
5258 as_warn (_("segment override on `%s' is ineffectual"), i.tm.name);
5259
5260 /* If a segment was explicitly specified, and the specified segment
5261 is not the default, use an opcode prefix to select it. If we
5262 never figured out what the default segment is, then default_seg
5263 will be zero at this point, and the specified segment prefix will
5264 always be used. */
5265 if ((i.seg[0]) && (i.seg[0] != default_seg))
5266 {
5267 if (!add_prefix (i.seg[0]->seg_prefix))
5268 return 0;
5269 }
5270 return 1;
5271 }
5272
5273 static const seg_entry *
5274 build_modrm_byte (void)
5275 {
5276 const seg_entry *default_seg = 0;
5277 unsigned int source, dest;
5278 int vex_3_sources;
5279
5280 /* The first operand of instructions with VEX prefix and 3 sources
5281 must be VEX_Imm4. */
5282 vex_3_sources = i.tm.opcode_modifier.vexsources == VEX3SOURCES;
5283 if (vex_3_sources)
5284 {
5285 unsigned int nds, reg_slot;
5286 expressionS *exp;
5287
5288 if (i.tm.opcode_modifier.veximmext
5289 && i.tm.opcode_modifier.immext)
5290 {
5291 dest = i.operands - 2;
5292 gas_assert (dest == 3);
5293 }
5294 else
5295 dest = i.operands - 1;
5296 nds = dest - 1;
5297
5298 /* There are 2 kinds of instructions:
5299 1. 5 operands: 4 register operands or 3 register operands
5300 plus 1 memory operand plus one Vec_Imm4 operand, VexXDS, and
5301 VexW0 or VexW1. The destination must be either XMM or YMM
5302 register.
5303 2. 4 operands: 4 register operands or 3 register operands
5304 plus 1 memory operand, VexXDS, and VexImmExt */
5305 gas_assert ((i.reg_operands == 4
5306 || (i.reg_operands == 3 && i.mem_operands == 1))
5307 && i.tm.opcode_modifier.vexvvvv == VEXXDS
5308 && (i.tm.opcode_modifier.veximmext
5309 || (i.imm_operands == 1
5310 && i.types[0].bitfield.vec_imm4
5311 && (i.tm.opcode_modifier.vexw == VEXW0
5312 || i.tm.opcode_modifier.vexw == VEXW1)
5313 && (operand_type_equal (&i.tm.operand_types[dest], &regxmm)
5314 || operand_type_equal (&i.tm.operand_types[dest], &regymm)))));
5315
5316 if (i.imm_operands == 0)
5317 {
5318 /* When there is no immediate operand, generate an 8bit
5319 immediate operand to encode the first operand. */
5320 exp = &im_expressions[i.imm_operands++];
5321 i.op[i.operands].imms = exp;
5322 i.types[i.operands] = imm8;
5323 i.operands++;
5324 /* If VexW1 is set, the first operand is the source and
5325 the second operand is encoded in the immediate operand. */
5326 if (i.tm.opcode_modifier.vexw == VEXW1)
5327 {
5328 source = 0;
5329 reg_slot = 1;
5330 }
5331 else
5332 {
5333 source = 1;
5334 reg_slot = 0;
5335 }
5336
5337 /* FMA swaps REG and NDS. */
5338 if (i.tm.cpu_flags.bitfield.cpufma)
5339 {
5340 unsigned int tmp;
5341 tmp = reg_slot;
5342 reg_slot = nds;
5343 nds = tmp;
5344 }
5345
5346 gas_assert (operand_type_equal (&i.tm.operand_types[reg_slot],
5347 &regxmm)
5348 || operand_type_equal (&i.tm.operand_types[reg_slot],
5349 &regymm));
5350 exp->X_op = O_constant;
5351 exp->X_add_number
5352 = ((i.op[reg_slot].regs->reg_num
5353 + ((i.op[reg_slot].regs->reg_flags & RegRex) ? 8 : 0))
5354 << 4);
5355 }
5356 else
5357 {
5358 unsigned int imm_slot;
5359
5360 if (i.tm.opcode_modifier.vexw == VEXW0)
5361 {
5362 /* If VexW0 is set, the third operand is the source and
5363 the second operand is encoded in the immediate
5364 operand. */
5365 source = 2;
5366 reg_slot = 1;
5367 }
5368 else
5369 {
5370 /* VexW1 is set, the second operand is the source and
5371 the third operand is encoded in the immediate
5372 operand. */
5373 source = 1;
5374 reg_slot = 2;
5375 }
5376
5377 if (i.tm.opcode_modifier.immext)
5378 {
5379 /* When ImmExt is set, the immdiate byte is the last
5380 operand. */
5381 imm_slot = i.operands - 1;
5382 source--;
5383 reg_slot--;
5384 }
5385 else
5386 {
5387 imm_slot = 0;
5388
5389 /* Turn on Imm8 so that output_imm will generate it. */
5390 i.types[imm_slot].bitfield.imm8 = 1;
5391 }
5392
5393 gas_assert (operand_type_equal (&i.tm.operand_types[reg_slot],
5394 &regxmm)
5395 || operand_type_equal (&i.tm.operand_types[reg_slot],
5396 &regymm));
5397 i.op[imm_slot].imms->X_add_number
5398 |= ((i.op[reg_slot].regs->reg_num
5399 + ((i.op[reg_slot].regs->reg_flags & RegRex) ? 8 : 0))
5400 << 4);
5401 }
5402
5403 gas_assert (operand_type_equal (&i.tm.operand_types[nds], &regxmm)
5404 || operand_type_equal (&i.tm.operand_types[nds],
5405 &regymm));
5406 i.vex.register_specifier = i.op[nds].regs;
5407 }
5408 else
5409 source = dest = 0;
5410
5411 /* i.reg_operands MUST be the number of real register operands;
5412 implicit registers do not count. If there are 3 register
5413 operands, it must be a instruction with VexNDS. For a
5414 instruction with VexNDD, the destination register is encoded
5415 in VEX prefix. If there are 4 register operands, it must be
5416 a instruction with VEX prefix and 3 sources. */
5417 if (i.mem_operands == 0
5418 && ((i.reg_operands == 2
5419 && i.tm.opcode_modifier.vexvvvv <= VEXXDS)
5420 || (i.reg_operands == 3
5421 && i.tm.opcode_modifier.vexvvvv == VEXXDS)
5422 || (i.reg_operands == 4 && vex_3_sources)))
5423 {
5424 switch (i.operands)
5425 {
5426 case 2:
5427 source = 0;
5428 break;
5429 case 3:
5430 /* When there are 3 operands, one of them may be immediate,
5431 which may be the first or the last operand. Otherwise,
5432 the first operand must be shift count register (cl) or it
5433 is an instruction with VexNDS. */
5434 gas_assert (i.imm_operands == 1
5435 || (i.imm_operands == 0
5436 && (i.tm.opcode_modifier.vexvvvv == VEXXDS
5437 || i.types[0].bitfield.shiftcount)));
5438 if (operand_type_check (i.types[0], imm)
5439 || i.types[0].bitfield.shiftcount)
5440 source = 1;
5441 else
5442 source = 0;
5443 break;
5444 case 4:
5445 /* When there are 4 operands, the first two must be 8bit
5446 immediate operands. The source operand will be the 3rd
5447 one.
5448
5449 For instructions with VexNDS, if the first operand
5450 an imm8, the source operand is the 2nd one. If the last
5451 operand is imm8, the source operand is the first one. */
5452 gas_assert ((i.imm_operands == 2
5453 && i.types[0].bitfield.imm8
5454 && i.types[1].bitfield.imm8)
5455 || (i.tm.opcode_modifier.vexvvvv == VEXXDS
5456 && i.imm_operands == 1
5457 && (i.types[0].bitfield.imm8
5458 || i.types[i.operands - 1].bitfield.imm8)));
5459 if (i.imm_operands == 2)
5460 source = 2;
5461 else
5462 {
5463 if (i.types[0].bitfield.imm8)
5464 source = 1;
5465 else
5466 source = 0;
5467 }
5468 break;
5469 case 5:
5470 break;
5471 default:
5472 abort ();
5473 }
5474
5475 if (!vex_3_sources)
5476 {
5477 dest = source + 1;
5478
5479 if (i.tm.opcode_modifier.vexvvvv == VEXXDS)
5480 {
5481 /* For instructions with VexNDS, the register-only
5482 source operand must be 32/64bit integer, XMM or
5483 YMM register. It is encoded in VEX prefix. We
5484 need to clear RegMem bit before calling
5485 operand_type_equal. */
5486
5487 i386_operand_type op;
5488 unsigned int vvvv;
5489
5490 /* Check register-only source operand when two source
5491 operands are swapped. */
5492 if (!i.tm.operand_types[source].bitfield.baseindex
5493 && i.tm.operand_types[dest].bitfield.baseindex)
5494 {
5495 vvvv = source;
5496 source = dest;
5497 }
5498 else
5499 vvvv = dest;
5500
5501 op = i.tm.operand_types[vvvv];
5502 op.bitfield.regmem = 0;
5503 if ((dest + 1) >= i.operands
5504 || (op.bitfield.reg32 != 1
5505 && !op.bitfield.reg64 != 1
5506 && !operand_type_equal (&op, &regxmm)
5507 && !operand_type_equal (&op, &regymm)))
5508 abort ();
5509 i.vex.register_specifier = i.op[vvvv].regs;
5510 dest++;
5511 }
5512 }
5513
5514 i.rm.mode = 3;
5515 /* One of the register operands will be encoded in the i.tm.reg
5516 field, the other in the combined i.tm.mode and i.tm.regmem
5517 fields. If no form of this instruction supports a memory
5518 destination operand, then we assume the source operand may
5519 sometimes be a memory operand and so we need to store the
5520 destination in the i.rm.reg field. */
5521 if (!i.tm.operand_types[dest].bitfield.regmem
5522 && operand_type_check (i.tm.operand_types[dest], anymem) == 0)
5523 {
5524 i.rm.reg = i.op[dest].regs->reg_num;
5525 i.rm.regmem = i.op[source].regs->reg_num;
5526 if ((i.op[dest].regs->reg_flags & RegRex) != 0)
5527 i.rex |= REX_R;
5528 if ((i.op[source].regs->reg_flags & RegRex) != 0)
5529 i.rex |= REX_B;
5530 }
5531 else
5532 {
5533 i.rm.reg = i.op[source].regs->reg_num;
5534 i.rm.regmem = i.op[dest].regs->reg_num;
5535 if ((i.op[dest].regs->reg_flags & RegRex) != 0)
5536 i.rex |= REX_B;
5537 if ((i.op[source].regs->reg_flags & RegRex) != 0)
5538 i.rex |= REX_R;
5539 }
5540 if (flag_code != CODE_64BIT && (i.rex & (REX_R | REX_B)))
5541 {
5542 if (!i.types[0].bitfield.control
5543 && !i.types[1].bitfield.control)
5544 abort ();
5545 i.rex &= ~(REX_R | REX_B);
5546 add_prefix (LOCK_PREFIX_OPCODE);
5547 }
5548 }
5549 else
5550 { /* If it's not 2 reg operands... */
5551 unsigned int mem;
5552
5553 if (i.mem_operands)
5554 {
5555 unsigned int fake_zero_displacement = 0;
5556 unsigned int op;
5557
5558 for (op = 0; op < i.operands; op++)
5559 if (operand_type_check (i.types[op], anymem))
5560 break;
5561 gas_assert (op < i.operands);
5562
5563 if (i.tm.opcode_modifier.vecsib)
5564 {
5565 if (i.index_reg->reg_num == RegEiz
5566 || i.index_reg->reg_num == RegRiz)
5567 abort ();
5568
5569 i.rm.regmem = ESCAPE_TO_TWO_BYTE_ADDRESSING;
5570 if (!i.base_reg)
5571 {
5572 i.sib.base = NO_BASE_REGISTER;
5573 i.sib.scale = i.log2_scale_factor;
5574 i.types[op].bitfield.disp8 = 0;
5575 i.types[op].bitfield.disp16 = 0;
5576 i.types[op].bitfield.disp64 = 0;
5577 if (flag_code != CODE_64BIT)
5578 {
5579 /* Must be 32 bit */
5580 i.types[op].bitfield.disp32 = 1;
5581 i.types[op].bitfield.disp32s = 0;
5582 }
5583 else
5584 {
5585 i.types[op].bitfield.disp32 = 0;
5586 i.types[op].bitfield.disp32s = 1;
5587 }
5588 }
5589 i.sib.index = i.index_reg->reg_num;
5590 if ((i.index_reg->reg_flags & RegRex) != 0)
5591 i.rex |= REX_X;
5592 }
5593
5594 default_seg = &ds;
5595
5596 if (i.base_reg == 0)
5597 {
5598 i.rm.mode = 0;
5599 if (!i.disp_operands)
5600 {
5601 fake_zero_displacement = 1;
5602 /* Instructions with VSIB byte need 32bit displacement
5603 if there is no base register. */
5604 if (i.tm.opcode_modifier.vecsib)
5605 i.types[op].bitfield.disp32 = 1;
5606 }
5607 if (i.index_reg == 0)
5608 {
5609 gas_assert (!i.tm.opcode_modifier.vecsib);
5610 /* Operand is just <disp> */
5611 if (flag_code == CODE_64BIT)
5612 {
5613 /* 64bit mode overwrites the 32bit absolute
5614 addressing by RIP relative addressing and
5615 absolute addressing is encoded by one of the
5616 redundant SIB forms. */
5617 i.rm.regmem = ESCAPE_TO_TWO_BYTE_ADDRESSING;
5618 i.sib.base = NO_BASE_REGISTER;
5619 i.sib.index = NO_INDEX_REGISTER;
5620 i.types[op] = ((i.prefix[ADDR_PREFIX] == 0)
5621 ? disp32s : disp32);
5622 }
5623 else if ((flag_code == CODE_16BIT)
5624 ^ (i.prefix[ADDR_PREFIX] != 0))
5625 {
5626 i.rm.regmem = NO_BASE_REGISTER_16;
5627 i.types[op] = disp16;
5628 }
5629 else
5630 {
5631 i.rm.regmem = NO_BASE_REGISTER;
5632 i.types[op] = disp32;
5633 }
5634 }
5635 else if (!i.tm.opcode_modifier.vecsib)
5636 {
5637 /* !i.base_reg && i.index_reg */
5638 if (i.index_reg->reg_num == RegEiz
5639 || i.index_reg->reg_num == RegRiz)
5640 i.sib.index = NO_INDEX_REGISTER;
5641 else
5642 i.sib.index = i.index_reg->reg_num;
5643 i.sib.base = NO_BASE_REGISTER;
5644 i.sib.scale = i.log2_scale_factor;
5645 i.rm.regmem = ESCAPE_TO_TWO_BYTE_ADDRESSING;
5646 i.types[op].bitfield.disp8 = 0;
5647 i.types[op].bitfield.disp16 = 0;
5648 i.types[op].bitfield.disp64 = 0;
5649 if (flag_code != CODE_64BIT)
5650 {
5651 /* Must be 32 bit */
5652 i.types[op].bitfield.disp32 = 1;
5653 i.types[op].bitfield.disp32s = 0;
5654 }
5655 else
5656 {
5657 i.types[op].bitfield.disp32 = 0;
5658 i.types[op].bitfield.disp32s = 1;
5659 }
5660 if ((i.index_reg->reg_flags & RegRex) != 0)
5661 i.rex |= REX_X;
5662 }
5663 }
5664 /* RIP addressing for 64bit mode. */
5665 else if (i.base_reg->reg_num == RegRip ||
5666 i.base_reg->reg_num == RegEip)
5667 {
5668 gas_assert (!i.tm.opcode_modifier.vecsib);
5669 i.rm.regmem = NO_BASE_REGISTER;
5670 i.types[op].bitfield.disp8 = 0;
5671 i.types[op].bitfield.disp16 = 0;
5672 i.types[op].bitfield.disp32 = 0;
5673 i.types[op].bitfield.disp32s = 1;
5674 i.types[op].bitfield.disp64 = 0;
5675 i.flags[op] |= Operand_PCrel;
5676 if (! i.disp_operands)
5677 fake_zero_displacement = 1;
5678 }
5679 else if (i.base_reg->reg_type.bitfield.reg16)
5680 {
5681 gas_assert (!i.tm.opcode_modifier.vecsib);
5682 switch (i.base_reg->reg_num)
5683 {
5684 case 3: /* (%bx) */
5685 if (i.index_reg == 0)
5686 i.rm.regmem = 7;
5687 else /* (%bx,%si) -> 0, or (%bx,%di) -> 1 */
5688 i.rm.regmem = i.index_reg->reg_num - 6;
5689 break;
5690 case 5: /* (%bp) */
5691 default_seg = &ss;
5692 if (i.index_reg == 0)
5693 {
5694 i.rm.regmem = 6;
5695 if (operand_type_check (i.types[op], disp) == 0)
5696 {
5697 /* fake (%bp) into 0(%bp) */
5698 i.types[op].bitfield.disp8 = 1;
5699 fake_zero_displacement = 1;
5700 }
5701 }
5702 else /* (%bp,%si) -> 2, or (%bp,%di) -> 3 */
5703 i.rm.regmem = i.index_reg->reg_num - 6 + 2;
5704 break;
5705 default: /* (%si) -> 4 or (%di) -> 5 */
5706 i.rm.regmem = i.base_reg->reg_num - 6 + 4;
5707 }
5708 i.rm.mode = mode_from_disp_size (i.types[op]);
5709 }
5710 else /* i.base_reg and 32/64 bit mode */
5711 {
5712 if (flag_code == CODE_64BIT
5713 && operand_type_check (i.types[op], disp))
5714 {
5715 i386_operand_type temp;
5716 operand_type_set (&temp, 0);
5717 temp.bitfield.disp8 = i.types[op].bitfield.disp8;
5718 i.types[op] = temp;
5719 if (i.prefix[ADDR_PREFIX] == 0)
5720 i.types[op].bitfield.disp32s = 1;
5721 else
5722 i.types[op].bitfield.disp32 = 1;
5723 }
5724
5725 if (!i.tm.opcode_modifier.vecsib)
5726 i.rm.regmem = i.base_reg->reg_num;
5727 if ((i.base_reg->reg_flags & RegRex) != 0)
5728 i.rex |= REX_B;
5729 i.sib.base = i.base_reg->reg_num;
5730 /* x86-64 ignores REX prefix bit here to avoid decoder
5731 complications. */
5732 if ((i.base_reg->reg_num & 7) == EBP_REG_NUM)
5733 {
5734 default_seg = &ss;
5735 if (i.disp_operands == 0)
5736 {
5737 fake_zero_displacement = 1;
5738 i.types[op].bitfield.disp8 = 1;
5739 }
5740 }
5741 else if (i.base_reg->reg_num == ESP_REG_NUM)
5742 {
5743 default_seg = &ss;
5744 }
5745 i.sib.scale = i.log2_scale_factor;
5746 if (i.index_reg == 0)
5747 {
5748 gas_assert (!i.tm.opcode_modifier.vecsib);
5749 /* <disp>(%esp) becomes two byte modrm with no index
5750 register. We've already stored the code for esp
5751 in i.rm.regmem ie. ESCAPE_TO_TWO_BYTE_ADDRESSING.
5752 Any base register besides %esp will not use the
5753 extra modrm byte. */
5754 i.sib.index = NO_INDEX_REGISTER;
5755 }
5756 else if (!i.tm.opcode_modifier.vecsib)
5757 {
5758 if (i.index_reg->reg_num == RegEiz
5759 || i.index_reg->reg_num == RegRiz)
5760 i.sib.index = NO_INDEX_REGISTER;
5761 else
5762 i.sib.index = i.index_reg->reg_num;
5763 i.rm.regmem = ESCAPE_TO_TWO_BYTE_ADDRESSING;
5764 if ((i.index_reg->reg_flags & RegRex) != 0)
5765 i.rex |= REX_X;
5766 }
5767
5768 if (i.disp_operands
5769 && (i.reloc[op] == BFD_RELOC_386_TLS_DESC_CALL
5770 || i.reloc[op] == BFD_RELOC_X86_64_TLSDESC_CALL))
5771 i.rm.mode = 0;
5772 else
5773 {
5774 if (!fake_zero_displacement
5775 && !i.disp_operands
5776 && i.disp_encoding)
5777 {
5778 fake_zero_displacement = 1;
5779 if (i.disp_encoding == disp_encoding_8bit)
5780 i.types[op].bitfield.disp8 = 1;
5781 else
5782 i.types[op].bitfield.disp32 = 1;
5783 }
5784 i.rm.mode = mode_from_disp_size (i.types[op]);
5785 }
5786 }
5787
5788 if (fake_zero_displacement)
5789 {
5790 /* Fakes a zero displacement assuming that i.types[op]
5791 holds the correct displacement size. */
5792 expressionS *exp;
5793
5794 gas_assert (i.op[op].disps == 0);
5795 exp = &disp_expressions[i.disp_operands++];
5796 i.op[op].disps = exp;
5797 exp->X_op = O_constant;
5798 exp->X_add_number = 0;
5799 exp->X_add_symbol = (symbolS *) 0;
5800 exp->X_op_symbol = (symbolS *) 0;
5801 }
5802
5803 mem = op;
5804 }
5805 else
5806 mem = ~0;
5807
5808 if (i.tm.opcode_modifier.vexsources == XOP2SOURCES)
5809 {
5810 if (operand_type_check (i.types[0], imm))
5811 i.vex.register_specifier = NULL;
5812 else
5813 {
5814 /* VEX.vvvv encodes one of the sources when the first
5815 operand is not an immediate. */
5816 if (i.tm.opcode_modifier.vexw == VEXW0)
5817 i.vex.register_specifier = i.op[0].regs;
5818 else
5819 i.vex.register_specifier = i.op[1].regs;
5820 }
5821
5822 /* Destination is a XMM register encoded in the ModRM.reg
5823 and VEX.R bit. */
5824 i.rm.reg = i.op[2].regs->reg_num;
5825 if ((i.op[2].regs->reg_flags & RegRex) != 0)
5826 i.rex |= REX_R;
5827
5828 /* ModRM.rm and VEX.B encodes the other source. */
5829 if (!i.mem_operands)
5830 {
5831 i.rm.mode = 3;
5832
5833 if (i.tm.opcode_modifier.vexw == VEXW0)
5834 i.rm.regmem = i.op[1].regs->reg_num;
5835 else
5836 i.rm.regmem = i.op[0].regs->reg_num;
5837
5838 if ((i.op[1].regs->reg_flags & RegRex) != 0)
5839 i.rex |= REX_B;
5840 }
5841 }
5842 else if (i.tm.opcode_modifier.vexvvvv == VEXLWP)
5843 {
5844 i.vex.register_specifier = i.op[2].regs;
5845 if (!i.mem_operands)
5846 {
5847 i.rm.mode = 3;
5848 i.rm.regmem = i.op[1].regs->reg_num;
5849 if ((i.op[1].regs->reg_flags & RegRex) != 0)
5850 i.rex |= REX_B;
5851 }
5852 }
5853 /* Fill in i.rm.reg or i.rm.regmem field with register operand
5854 (if any) based on i.tm.extension_opcode. Again, we must be
5855 careful to make sure that segment/control/debug/test/MMX
5856 registers are coded into the i.rm.reg field. */
5857 else if (i.reg_operands)
5858 {
5859 unsigned int op;
5860 unsigned int vex_reg = ~0;
5861
5862 for (op = 0; op < i.operands; op++)
5863 if (i.types[op].bitfield.reg8
5864 || i.types[op].bitfield.reg16
5865 || i.types[op].bitfield.reg32
5866 || i.types[op].bitfield.reg64
5867 || i.types[op].bitfield.regmmx
5868 || i.types[op].bitfield.regxmm
5869 || i.types[op].bitfield.regymm
5870 || i.types[op].bitfield.sreg2
5871 || i.types[op].bitfield.sreg3
5872 || i.types[op].bitfield.control
5873 || i.types[op].bitfield.debug
5874 || i.types[op].bitfield.test)
5875 break;
5876
5877 if (vex_3_sources)
5878 op = dest;
5879 else if (i.tm.opcode_modifier.vexvvvv == VEXXDS)
5880 {
5881 /* For instructions with VexNDS, the register-only
5882 source operand is encoded in VEX prefix. */
5883 gas_assert (mem != (unsigned int) ~0);
5884
5885 if (op > mem)
5886 {
5887 vex_reg = op++;
5888 gas_assert (op < i.operands);
5889 }
5890 else
5891 {
5892 /* Check register-only source operand when two source
5893 operands are swapped. */
5894 if (!i.tm.operand_types[op].bitfield.baseindex
5895 && i.tm.operand_types[op + 1].bitfield.baseindex)
5896 {
5897 vex_reg = op;
5898 op += 2;
5899 gas_assert (mem == (vex_reg + 1)
5900 && op < i.operands);
5901 }
5902 else
5903 {
5904 vex_reg = op + 1;
5905 gas_assert (vex_reg < i.operands);
5906 }
5907 }
5908 }
5909 else if (i.tm.opcode_modifier.vexvvvv == VEXNDD)
5910 {
5911 /* For instructions with VexNDD, the register destination
5912 is encoded in VEX prefix. */
5913 if (i.mem_operands == 0)
5914 {
5915 /* There is no memory operand. */
5916 gas_assert ((op + 2) == i.operands);
5917 vex_reg = op + 1;
5918 }
5919 else
5920 {
5921 /* There are only 2 operands. */
5922 gas_assert (op < 2 && i.operands == 2);
5923 vex_reg = 1;
5924 }
5925 }
5926 else
5927 gas_assert (op < i.operands);
5928
5929 if (vex_reg != (unsigned int) ~0)
5930 {
5931 i386_operand_type *type = &i.tm.operand_types[vex_reg];
5932
5933 if (type->bitfield.reg32 != 1
5934 && type->bitfield.reg64 != 1
5935 && !operand_type_equal (type, &regxmm)
5936 && !operand_type_equal (type, &regymm))
5937 abort ();
5938
5939 i.vex.register_specifier = i.op[vex_reg].regs;
5940 }
5941
5942 /* Don't set OP operand twice. */
5943 if (vex_reg != op)
5944 {
5945 /* If there is an extension opcode to put here, the
5946 register number must be put into the regmem field. */
5947 if (i.tm.extension_opcode != None)
5948 {
5949 i.rm.regmem = i.op[op].regs->reg_num;
5950 if ((i.op[op].regs->reg_flags & RegRex) != 0)
5951 i.rex |= REX_B;
5952 }
5953 else
5954 {
5955 i.rm.reg = i.op[op].regs->reg_num;
5956 if ((i.op[op].regs->reg_flags & RegRex) != 0)
5957 i.rex |= REX_R;
5958 }
5959 }
5960
5961 /* Now, if no memory operand has set i.rm.mode = 0, 1, 2 we
5962 must set it to 3 to indicate this is a register operand
5963 in the regmem field. */
5964 if (!i.mem_operands)
5965 i.rm.mode = 3;
5966 }
5967
5968 /* Fill in i.rm.reg field with extension opcode (if any). */
5969 if (i.tm.extension_opcode != None)
5970 i.rm.reg = i.tm.extension_opcode;
5971 }
5972 return default_seg;
5973 }
5974
5975 static void
5976 output_branch (void)
5977 {
5978 char *p;
5979 int size;
5980 int code16;
5981 int prefix;
5982 relax_substateT subtype;
5983 symbolS *sym;
5984 offsetT off;
5985
5986 code16 = flag_code == CODE_16BIT ? CODE16 : 0;
5987 size = i.disp_encoding == disp_encoding_32bit ? BIG : SMALL;
5988
5989 prefix = 0;
5990 if (i.prefix[DATA_PREFIX] != 0)
5991 {
5992 prefix = 1;
5993 i.prefixes -= 1;
5994 code16 ^= CODE16;
5995 }
5996 /* Pentium4 branch hints. */
5997 if (i.prefix[SEG_PREFIX] == CS_PREFIX_OPCODE /* not taken */
5998 || i.prefix[SEG_PREFIX] == DS_PREFIX_OPCODE /* taken */)
5999 {
6000 prefix++;
6001 i.prefixes--;
6002 }
6003 if (i.prefix[REX_PREFIX] != 0)
6004 {
6005 prefix++;
6006 i.prefixes--;
6007 }
6008
6009 if (i.prefixes != 0 && !intel_syntax)
6010 as_warn (_("skipping prefixes on this instruction"));
6011
6012 /* It's always a symbol; End frag & setup for relax.
6013 Make sure there is enough room in this frag for the largest
6014 instruction we may generate in md_convert_frag. This is 2
6015 bytes for the opcode and room for the prefix and largest
6016 displacement. */
6017 frag_grow (prefix + 2 + 4);
6018 /* Prefix and 1 opcode byte go in fr_fix. */
6019 p = frag_more (prefix + 1);
6020 if (i.prefix[DATA_PREFIX] != 0)
6021 *p++ = DATA_PREFIX_OPCODE;
6022 if (i.prefix[SEG_PREFIX] == CS_PREFIX_OPCODE
6023 || i.prefix[SEG_PREFIX] == DS_PREFIX_OPCODE)
6024 *p++ = i.prefix[SEG_PREFIX];
6025 if (i.prefix[REX_PREFIX] != 0)
6026 *p++ = i.prefix[REX_PREFIX];
6027 *p = i.tm.base_opcode;
6028
6029 if ((unsigned char) *p == JUMP_PC_RELATIVE)
6030 subtype = ENCODE_RELAX_STATE (UNCOND_JUMP, size);
6031 else if (cpu_arch_flags.bitfield.cpui386)
6032 subtype = ENCODE_RELAX_STATE (COND_JUMP, size);
6033 else
6034 subtype = ENCODE_RELAX_STATE (COND_JUMP86, size);
6035 subtype |= code16;
6036
6037 sym = i.op[0].disps->X_add_symbol;
6038 off = i.op[0].disps->X_add_number;
6039
6040 if (i.op[0].disps->X_op != O_constant
6041 && i.op[0].disps->X_op != O_symbol)
6042 {
6043 /* Handle complex expressions. */
6044 sym = make_expr_symbol (i.op[0].disps);
6045 off = 0;
6046 }
6047
6048 /* 1 possible extra opcode + 4 byte displacement go in var part.
6049 Pass reloc in fr_var. */
6050 frag_var (rs_machine_dependent, 5, i.reloc[0], subtype, sym, off, p);
6051 }
6052
6053 static void
6054 output_jump (void)
6055 {
6056 char *p;
6057 int size;
6058 fixS *fixP;
6059
6060 if (i.tm.opcode_modifier.jumpbyte)
6061 {
6062 /* This is a loop or jecxz type instruction. */
6063 size = 1;
6064 if (i.prefix[ADDR_PREFIX] != 0)
6065 {
6066 FRAG_APPEND_1_CHAR (ADDR_PREFIX_OPCODE);
6067 i.prefixes -= 1;
6068 }
6069 /* Pentium4 branch hints. */
6070 if (i.prefix[SEG_PREFIX] == CS_PREFIX_OPCODE /* not taken */
6071 || i.prefix[SEG_PREFIX] == DS_PREFIX_OPCODE /* taken */)
6072 {
6073 FRAG_APPEND_1_CHAR (i.prefix[SEG_PREFIX]);
6074 i.prefixes--;
6075 }
6076 }
6077 else
6078 {
6079 int code16;
6080
6081 code16 = 0;
6082 if (flag_code == CODE_16BIT)
6083 code16 = CODE16;
6084
6085 if (i.prefix[DATA_PREFIX] != 0)
6086 {
6087 FRAG_APPEND_1_CHAR (DATA_PREFIX_OPCODE);
6088 i.prefixes -= 1;
6089 code16 ^= CODE16;
6090 }
6091
6092 size = 4;
6093 if (code16)
6094 size = 2;
6095 }
6096
6097 if (i.prefix[REX_PREFIX] != 0)
6098 {
6099 FRAG_APPEND_1_CHAR (i.prefix[REX_PREFIX]);
6100 i.prefixes -= 1;
6101 }
6102
6103 if (i.prefixes != 0 && !intel_syntax)
6104 as_warn (_("skipping prefixes on this instruction"));
6105
6106 p = frag_more (i.tm.opcode_length + size);
6107 switch (i.tm.opcode_length)
6108 {
6109 case 2:
6110 *p++ = i.tm.base_opcode >> 8;
6111 case 1:
6112 *p++ = i.tm.base_opcode;
6113 break;
6114 default:
6115 abort ();
6116 }
6117
6118 fixP = fix_new_exp (frag_now, p - frag_now->fr_literal, size,
6119 i.op[0].disps, 1, reloc (size, 1, 1, i.reloc[0]));
6120
6121 /* All jumps handled here are signed, but don't use a signed limit
6122 check for 32 and 16 bit jumps as we want to allow wrap around at
6123 4G and 64k respectively. */
6124 if (size == 1)
6125 fixP->fx_signed = 1;
6126 }
6127
6128 static void
6129 output_interseg_jump (void)
6130 {
6131 char *p;
6132 int size;
6133 int prefix;
6134 int code16;
6135
6136 code16 = 0;
6137 if (flag_code == CODE_16BIT)
6138 code16 = CODE16;
6139
6140 prefix = 0;
6141 if (i.prefix[DATA_PREFIX] != 0)
6142 {
6143 prefix = 1;
6144 i.prefixes -= 1;
6145 code16 ^= CODE16;
6146 }
6147 if (i.prefix[REX_PREFIX] != 0)
6148 {
6149 prefix++;
6150 i.prefixes -= 1;
6151 }
6152
6153 size = 4;
6154 if (code16)
6155 size = 2;
6156
6157 if (i.prefixes != 0 && !intel_syntax)
6158 as_warn (_("skipping prefixes on this instruction"));
6159
6160 /* 1 opcode; 2 segment; offset */
6161 p = frag_more (prefix + 1 + 2 + size);
6162
6163 if (i.prefix[DATA_PREFIX] != 0)
6164 *p++ = DATA_PREFIX_OPCODE;
6165
6166 if (i.prefix[REX_PREFIX] != 0)
6167 *p++ = i.prefix[REX_PREFIX];
6168
6169 *p++ = i.tm.base_opcode;
6170 if (i.op[1].imms->X_op == O_constant)
6171 {
6172 offsetT n = i.op[1].imms->X_add_number;
6173
6174 if (size == 2
6175 && !fits_in_unsigned_word (n)
6176 && !fits_in_signed_word (n))
6177 {
6178 as_bad (_("16-bit jump out of range"));
6179 return;
6180 }
6181 md_number_to_chars (p, n, size);
6182 }
6183 else
6184 fix_new_exp (frag_now, p - frag_now->fr_literal, size,
6185 i.op[1].imms, 0, reloc (size, 0, 0, i.reloc[1]));
6186 if (i.op[0].imms->X_op != O_constant)
6187 as_bad (_("can't handle non absolute segment in `%s'"),
6188 i.tm.name);
6189 md_number_to_chars (p + size, (valueT) i.op[0].imms->X_add_number, 2);
6190 }
6191
6192 static void
6193 output_insn (void)
6194 {
6195 fragS *insn_start_frag;
6196 offsetT insn_start_off;
6197
6198 /* Tie dwarf2 debug info to the address at the start of the insn.
6199 We can't do this after the insn has been output as the current
6200 frag may have been closed off. eg. by frag_var. */
6201 dwarf2_emit_insn (0);
6202
6203 insn_start_frag = frag_now;
6204 insn_start_off = frag_now_fix ();
6205
6206 /* Output jumps. */
6207 if (i.tm.opcode_modifier.jump)
6208 output_branch ();
6209 else if (i.tm.opcode_modifier.jumpbyte
6210 || i.tm.opcode_modifier.jumpdword)
6211 output_jump ();
6212 else if (i.tm.opcode_modifier.jumpintersegment)
6213 output_interseg_jump ();
6214 else
6215 {
6216 /* Output normal instructions here. */
6217 char *p;
6218 unsigned char *q;
6219 unsigned int j;
6220 unsigned int prefix;
6221
6222 /* Since the VEX prefix contains the implicit prefix, we don't
6223 need the explicit prefix. */
6224 if (!i.tm.opcode_modifier.vex)
6225 {
6226 switch (i.tm.opcode_length)
6227 {
6228 case 3:
6229 if (i.tm.base_opcode & 0xff000000)
6230 {
6231 prefix = (i.tm.base_opcode >> 24) & 0xff;
6232 goto check_prefix;
6233 }
6234 break;
6235 case 2:
6236 if ((i.tm.base_opcode & 0xff0000) != 0)
6237 {
6238 prefix = (i.tm.base_opcode >> 16) & 0xff;
6239 if (i.tm.cpu_flags.bitfield.cpupadlock)
6240 {
6241 check_prefix:
6242 if (prefix != REPE_PREFIX_OPCODE
6243 || (i.prefix[REP_PREFIX]
6244 != REPE_PREFIX_OPCODE))
6245 add_prefix (prefix);
6246 }
6247 else
6248 add_prefix (prefix);
6249 }
6250 break;
6251 case 1:
6252 break;
6253 default:
6254 abort ();
6255 }
6256
6257 /* The prefix bytes. */
6258 for (j = ARRAY_SIZE (i.prefix), q = i.prefix; j > 0; j--, q++)
6259 if (*q)
6260 FRAG_APPEND_1_CHAR (*q);
6261 }
6262 else
6263 {
6264 for (j = 0, q = i.prefix; j < ARRAY_SIZE (i.prefix); j++, q++)
6265 if (*q)
6266 switch (j)
6267 {
6268 case REX_PREFIX:
6269 /* REX byte is encoded in VEX prefix. */
6270 break;
6271 case SEG_PREFIX:
6272 case ADDR_PREFIX:
6273 FRAG_APPEND_1_CHAR (*q);
6274 break;
6275 default:
6276 /* There should be no other prefixes for instructions
6277 with VEX prefix. */
6278 abort ();
6279 }
6280
6281 /* Now the VEX prefix. */
6282 p = frag_more (i.vex.length);
6283 for (j = 0; j < i.vex.length; j++)
6284 p[j] = i.vex.bytes[j];
6285 }
6286
6287 /* Now the opcode; be careful about word order here! */
6288 if (i.tm.opcode_length == 1)
6289 {
6290 FRAG_APPEND_1_CHAR (i.tm.base_opcode);
6291 }
6292 else
6293 {
6294 switch (i.tm.opcode_length)
6295 {
6296 case 3:
6297 p = frag_more (3);
6298 *p++ = (i.tm.base_opcode >> 16) & 0xff;
6299 break;
6300 case 2:
6301 p = frag_more (2);
6302 break;
6303 default:
6304 abort ();
6305 break;
6306 }
6307
6308 /* Put out high byte first: can't use md_number_to_chars! */
6309 *p++ = (i.tm.base_opcode >> 8) & 0xff;
6310 *p = i.tm.base_opcode & 0xff;
6311 }
6312
6313 /* Now the modrm byte and sib byte (if present). */
6314 if (i.tm.opcode_modifier.modrm)
6315 {
6316 FRAG_APPEND_1_CHAR ((i.rm.regmem << 0
6317 | i.rm.reg << 3
6318 | i.rm.mode << 6));
6319 /* If i.rm.regmem == ESP (4)
6320 && i.rm.mode != (Register mode)
6321 && not 16 bit
6322 ==> need second modrm byte. */
6323 if (i.rm.regmem == ESCAPE_TO_TWO_BYTE_ADDRESSING
6324 && i.rm.mode != 3
6325 && !(i.base_reg && i.base_reg->reg_type.bitfield.reg16))
6326 FRAG_APPEND_1_CHAR ((i.sib.base << 0
6327 | i.sib.index << 3
6328 | i.sib.scale << 6));
6329 }
6330
6331 if (i.disp_operands)
6332 output_disp (insn_start_frag, insn_start_off);
6333
6334 if (i.imm_operands)
6335 output_imm (insn_start_frag, insn_start_off);
6336 }
6337
6338 #ifdef DEBUG386
6339 if (flag_debug)
6340 {
6341 pi ("" /*line*/, &i);
6342 }
6343 #endif /* DEBUG386 */
6344 }
6345
6346 /* Return the size of the displacement operand N. */
6347
6348 static int
6349 disp_size (unsigned int n)
6350 {
6351 int size = 4;
6352 if (i.types[n].bitfield.disp64)
6353 size = 8;
6354 else if (i.types[n].bitfield.disp8)
6355 size = 1;
6356 else if (i.types[n].bitfield.disp16)
6357 size = 2;
6358 return size;
6359 }
6360
6361 /* Return the size of the immediate operand N. */
6362
6363 static int
6364 imm_size (unsigned int n)
6365 {
6366 int size = 4;
6367 if (i.types[n].bitfield.imm64)
6368 size = 8;
6369 else if (i.types[n].bitfield.imm8 || i.types[n].bitfield.imm8s)
6370 size = 1;
6371 else if (i.types[n].bitfield.imm16)
6372 size = 2;
6373 return size;
6374 }
6375
6376 static void
6377 output_disp (fragS *insn_start_frag, offsetT insn_start_off)
6378 {
6379 char *p;
6380 unsigned int n;
6381
6382 for (n = 0; n < i.operands; n++)
6383 {
6384 if (operand_type_check (i.types[n], disp))
6385 {
6386 if (i.op[n].disps->X_op == O_constant)
6387 {
6388 int size = disp_size (n);
6389 offsetT val;
6390
6391 val = offset_in_range (i.op[n].disps->X_add_number,
6392 size);
6393 p = frag_more (size);
6394 md_number_to_chars (p, val, size);
6395 }
6396 else
6397 {
6398 enum bfd_reloc_code_real reloc_type;
6399 int size = disp_size (n);
6400 int sign = i.types[n].bitfield.disp32s;
6401 int pcrel = (i.flags[n] & Operand_PCrel) != 0;
6402
6403 /* We can't have 8 bit displacement here. */
6404 gas_assert (!i.types[n].bitfield.disp8);
6405
6406 /* The PC relative address is computed relative
6407 to the instruction boundary, so in case immediate
6408 fields follows, we need to adjust the value. */
6409 if (pcrel && i.imm_operands)
6410 {
6411 unsigned int n1;
6412 int sz = 0;
6413
6414 for (n1 = 0; n1 < i.operands; n1++)
6415 if (operand_type_check (i.types[n1], imm))
6416 {
6417 /* Only one immediate is allowed for PC
6418 relative address. */
6419 gas_assert (sz == 0);
6420 sz = imm_size (n1);
6421 i.op[n].disps->X_add_number -= sz;
6422 }
6423 /* We should find the immediate. */
6424 gas_assert (sz != 0);
6425 }
6426
6427 p = frag_more (size);
6428 reloc_type = reloc (size, pcrel, sign, i.reloc[n]);
6429 if (GOT_symbol
6430 && GOT_symbol == i.op[n].disps->X_add_symbol
6431 && (((reloc_type == BFD_RELOC_32
6432 || reloc_type == BFD_RELOC_X86_64_32S
6433 || (reloc_type == BFD_RELOC_64
6434 && object_64bit))
6435 && (i.op[n].disps->X_op == O_symbol
6436 || (i.op[n].disps->X_op == O_add
6437 && ((symbol_get_value_expression
6438 (i.op[n].disps->X_op_symbol)->X_op)
6439 == O_subtract))))
6440 || reloc_type == BFD_RELOC_32_PCREL))
6441 {
6442 offsetT add;
6443
6444 if (insn_start_frag == frag_now)
6445 add = (p - frag_now->fr_literal) - insn_start_off;
6446 else
6447 {
6448 fragS *fr;
6449
6450 add = insn_start_frag->fr_fix - insn_start_off;
6451 for (fr = insn_start_frag->fr_next;
6452 fr && fr != frag_now; fr = fr->fr_next)
6453 add += fr->fr_fix;
6454 add += p - frag_now->fr_literal;
6455 }
6456
6457 if (!object_64bit)
6458 {
6459 reloc_type = BFD_RELOC_386_GOTPC;
6460 i.op[n].imms->X_add_number += add;
6461 }
6462 else if (reloc_type == BFD_RELOC_64)
6463 reloc_type = BFD_RELOC_X86_64_GOTPC64;
6464 else
6465 /* Don't do the adjustment for x86-64, as there
6466 the pcrel addressing is relative to the _next_
6467 insn, and that is taken care of in other code. */
6468 reloc_type = BFD_RELOC_X86_64_GOTPC32;
6469 }
6470 fix_new_exp (frag_now, p - frag_now->fr_literal, size,
6471 i.op[n].disps, pcrel, reloc_type);
6472 }
6473 }
6474 }
6475 }
6476
6477 static void
6478 output_imm (fragS *insn_start_frag, offsetT insn_start_off)
6479 {
6480 char *p;
6481 unsigned int n;
6482
6483 for (n = 0; n < i.operands; n++)
6484 {
6485 if (operand_type_check (i.types[n], imm))
6486 {
6487 if (i.op[n].imms->X_op == O_constant)
6488 {
6489 int size = imm_size (n);
6490 offsetT val;
6491
6492 val = offset_in_range (i.op[n].imms->X_add_number,
6493 size);
6494 p = frag_more (size);
6495 md_number_to_chars (p, val, size);
6496 }
6497 else
6498 {
6499 /* Not absolute_section.
6500 Need a 32-bit fixup (don't support 8bit
6501 non-absolute imms). Try to support other
6502 sizes ... */
6503 enum bfd_reloc_code_real reloc_type;
6504 int size = imm_size (n);
6505 int sign;
6506
6507 if (i.types[n].bitfield.imm32s
6508 && (i.suffix == QWORD_MNEM_SUFFIX
6509 || (!i.suffix && i.tm.opcode_modifier.no_lsuf)))
6510 sign = 1;
6511 else
6512 sign = 0;
6513
6514 p = frag_more (size);
6515 reloc_type = reloc (size, 0, sign, i.reloc[n]);
6516
6517 /* This is tough to explain. We end up with this one if we
6518 * have operands that look like
6519 * "_GLOBAL_OFFSET_TABLE_+[.-.L284]". The goal here is to
6520 * obtain the absolute address of the GOT, and it is strongly
6521 * preferable from a performance point of view to avoid using
6522 * a runtime relocation for this. The actual sequence of
6523 * instructions often look something like:
6524 *
6525 * call .L66
6526 * .L66:
6527 * popl %ebx
6528 * addl $_GLOBAL_OFFSET_TABLE_+[.-.L66],%ebx
6529 *
6530 * The call and pop essentially return the absolute address
6531 * of the label .L66 and store it in %ebx. The linker itself
6532 * will ultimately change the first operand of the addl so
6533 * that %ebx points to the GOT, but to keep things simple, the
6534 * .o file must have this operand set so that it generates not
6535 * the absolute address of .L66, but the absolute address of
6536 * itself. This allows the linker itself simply treat a GOTPC
6537 * relocation as asking for a pcrel offset to the GOT to be
6538 * added in, and the addend of the relocation is stored in the
6539 * operand field for the instruction itself.
6540 *
6541 * Our job here is to fix the operand so that it would add
6542 * the correct offset so that %ebx would point to itself. The
6543 * thing that is tricky is that .-.L66 will point to the
6544 * beginning of the instruction, so we need to further modify
6545 * the operand so that it will point to itself. There are
6546 * other cases where you have something like:
6547 *
6548 * .long $_GLOBAL_OFFSET_TABLE_+[.-.L66]
6549 *
6550 * and here no correction would be required. Internally in
6551 * the assembler we treat operands of this form as not being
6552 * pcrel since the '.' is explicitly mentioned, and I wonder
6553 * whether it would simplify matters to do it this way. Who
6554 * knows. In earlier versions of the PIC patches, the
6555 * pcrel_adjust field was used to store the correction, but
6556 * since the expression is not pcrel, I felt it would be
6557 * confusing to do it this way. */
6558
6559 if ((reloc_type == BFD_RELOC_32
6560 || reloc_type == BFD_RELOC_X86_64_32S
6561 || reloc_type == BFD_RELOC_64)
6562 && GOT_symbol
6563 && GOT_symbol == i.op[n].imms->X_add_symbol
6564 && (i.op[n].imms->X_op == O_symbol
6565 || (i.op[n].imms->X_op == O_add
6566 && ((symbol_get_value_expression
6567 (i.op[n].imms->X_op_symbol)->X_op)
6568 == O_subtract))))
6569 {
6570 offsetT add;
6571
6572 if (insn_start_frag == frag_now)
6573 add = (p - frag_now->fr_literal) - insn_start_off;
6574 else
6575 {
6576 fragS *fr;
6577
6578 add = insn_start_frag->fr_fix - insn_start_off;
6579 for (fr = insn_start_frag->fr_next;
6580 fr && fr != frag_now; fr = fr->fr_next)
6581 add += fr->fr_fix;
6582 add += p - frag_now->fr_literal;
6583 }
6584
6585 if (!object_64bit)
6586 reloc_type = BFD_RELOC_386_GOTPC;
6587 else if (size == 4)
6588 reloc_type = BFD_RELOC_X86_64_GOTPC32;
6589 else if (size == 8)
6590 reloc_type = BFD_RELOC_X86_64_GOTPC64;
6591 i.op[n].imms->X_add_number += add;
6592 }
6593 fix_new_exp (frag_now, p - frag_now->fr_literal, size,
6594 i.op[n].imms, 0, reloc_type);
6595 }
6596 }
6597 }
6598 }
6599 \f
6600 /* x86_cons_fix_new is called via the expression parsing code when a
6601 reloc is needed. We use this hook to get the correct .got reloc. */
6602 static enum bfd_reloc_code_real got_reloc = NO_RELOC;
6603 static int cons_sign = -1;
6604
6605 void
6606 x86_cons_fix_new (fragS *frag, unsigned int off, unsigned int len,
6607 expressionS *exp)
6608 {
6609 enum bfd_reloc_code_real r = reloc (len, 0, cons_sign, got_reloc);
6610
6611 got_reloc = NO_RELOC;
6612
6613 #ifdef TE_PE
6614 if (exp->X_op == O_secrel)
6615 {
6616 exp->X_op = O_symbol;
6617 r = BFD_RELOC_32_SECREL;
6618 }
6619 #endif
6620
6621 fix_new_exp (frag, off, len, exp, 0, r);
6622 }
6623
6624 /* Export the ABI address size for use by TC_ADDRESS_BYTES for the
6625 purpose of the `.dc.a' internal pseudo-op. */
6626
6627 int
6628 x86_address_bytes (void)
6629 {
6630 if ((stdoutput->arch_info->mach & bfd_mach_x64_32))
6631 return 4;
6632 return stdoutput->arch_info->bits_per_address / 8;
6633 }
6634
6635 #if !(defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) || defined (OBJ_MACH_O)) \
6636 || defined (LEX_AT)
6637 # define lex_got(reloc, adjust, types) NULL
6638 #else
6639 /* Parse operands of the form
6640 <symbol>@GOTOFF+<nnn>
6641 and similar .plt or .got references.
6642
6643 If we find one, set up the correct relocation in RELOC and copy the
6644 input string, minus the `@GOTOFF' into a malloc'd buffer for
6645 parsing by the calling routine. Return this buffer, and if ADJUST
6646 is non-null set it to the length of the string we removed from the
6647 input line. Otherwise return NULL. */
6648 static char *
6649 lex_got (enum bfd_reloc_code_real *rel,
6650 int *adjust,
6651 i386_operand_type *types)
6652 {
6653 /* Some of the relocations depend on the size of what field is to
6654 be relocated. But in our callers i386_immediate and i386_displacement
6655 we don't yet know the operand size (this will be set by insn
6656 matching). Hence we record the word32 relocation here,
6657 and adjust the reloc according to the real size in reloc(). */
6658 static const struct {
6659 const char *str;
6660 int len;
6661 const enum bfd_reloc_code_real rel[2];
6662 const i386_operand_type types64;
6663 } gotrel[] = {
6664 { STRING_COMMA_LEN ("PLTOFF"), { _dummy_first_bfd_reloc_code_real,
6665 BFD_RELOC_X86_64_PLTOFF64 },
6666 OPERAND_TYPE_IMM64 },
6667 { STRING_COMMA_LEN ("PLT"), { BFD_RELOC_386_PLT32,
6668 BFD_RELOC_X86_64_PLT32 },
6669 OPERAND_TYPE_IMM32_32S_DISP32 },
6670 { STRING_COMMA_LEN ("GOTPLT"), { _dummy_first_bfd_reloc_code_real,
6671 BFD_RELOC_X86_64_GOTPLT64 },
6672 OPERAND_TYPE_IMM64_DISP64 },
6673 { STRING_COMMA_LEN ("GOTOFF"), { BFD_RELOC_386_GOTOFF,
6674 BFD_RELOC_X86_64_GOTOFF64 },
6675 OPERAND_TYPE_IMM64_DISP64 },
6676 { STRING_COMMA_LEN ("GOTPCREL"), { _dummy_first_bfd_reloc_code_real,
6677 BFD_RELOC_X86_64_GOTPCREL },
6678 OPERAND_TYPE_IMM32_32S_DISP32 },
6679 { STRING_COMMA_LEN ("TLSGD"), { BFD_RELOC_386_TLS_GD,
6680 BFD_RELOC_X86_64_TLSGD },
6681 OPERAND_TYPE_IMM32_32S_DISP32 },
6682 { STRING_COMMA_LEN ("TLSLDM"), { BFD_RELOC_386_TLS_LDM,
6683 _dummy_first_bfd_reloc_code_real },
6684 OPERAND_TYPE_NONE },
6685 { STRING_COMMA_LEN ("TLSLD"), { _dummy_first_bfd_reloc_code_real,
6686 BFD_RELOC_X86_64_TLSLD },
6687 OPERAND_TYPE_IMM32_32S_DISP32 },
6688 { STRING_COMMA_LEN ("GOTTPOFF"), { BFD_RELOC_386_TLS_IE_32,
6689 BFD_RELOC_X86_64_GOTTPOFF },
6690 OPERAND_TYPE_IMM32_32S_DISP32 },
6691 { STRING_COMMA_LEN ("TPOFF"), { BFD_RELOC_386_TLS_LE_32,
6692 BFD_RELOC_X86_64_TPOFF32 },
6693 OPERAND_TYPE_IMM32_32S_64_DISP32_64 },
6694 { STRING_COMMA_LEN ("NTPOFF"), { BFD_RELOC_386_TLS_LE,
6695 _dummy_first_bfd_reloc_code_real },
6696 OPERAND_TYPE_NONE },
6697 { STRING_COMMA_LEN ("DTPOFF"), { BFD_RELOC_386_TLS_LDO_32,
6698 BFD_RELOC_X86_64_DTPOFF32 },
6699 OPERAND_TYPE_IMM32_32S_64_DISP32_64 },
6700 { STRING_COMMA_LEN ("GOTNTPOFF"),{ BFD_RELOC_386_TLS_GOTIE,
6701 _dummy_first_bfd_reloc_code_real },
6702 OPERAND_TYPE_NONE },
6703 { STRING_COMMA_LEN ("INDNTPOFF"),{ BFD_RELOC_386_TLS_IE,
6704 _dummy_first_bfd_reloc_code_real },
6705 OPERAND_TYPE_NONE },
6706 { STRING_COMMA_LEN ("GOT"), { BFD_RELOC_386_GOT32,
6707 BFD_RELOC_X86_64_GOT32 },
6708 OPERAND_TYPE_IMM32_32S_64_DISP32 },
6709 { STRING_COMMA_LEN ("TLSDESC"), { BFD_RELOC_386_TLS_GOTDESC,
6710 BFD_RELOC_X86_64_GOTPC32_TLSDESC },
6711 OPERAND_TYPE_IMM32_32S_DISP32 },
6712 { STRING_COMMA_LEN ("TLSCALL"), { BFD_RELOC_386_TLS_DESC_CALL,
6713 BFD_RELOC_X86_64_TLSDESC_CALL },
6714 OPERAND_TYPE_IMM32_32S_DISP32 },
6715 };
6716 char *cp;
6717 unsigned int j;
6718
6719 #if defined (OBJ_MAYBE_ELF)
6720 if (!IS_ELF)
6721 return NULL;
6722 #endif
6723
6724 for (cp = input_line_pointer; *cp != '@'; cp++)
6725 if (is_end_of_line[(unsigned char) *cp] || *cp == ',')
6726 return NULL;
6727
6728 for (j = 0; j < ARRAY_SIZE (gotrel); j++)
6729 {
6730 int len = gotrel[j].len;
6731 if (strncasecmp (cp + 1, gotrel[j].str, len) == 0)
6732 {
6733 if (gotrel[j].rel[object_64bit] != 0)
6734 {
6735 int first, second;
6736 char *tmpbuf, *past_reloc;
6737
6738 *rel = gotrel[j].rel[object_64bit];
6739 if (adjust)
6740 *adjust = len;
6741
6742 if (types)
6743 {
6744 if (flag_code != CODE_64BIT)
6745 {
6746 types->bitfield.imm32 = 1;
6747 types->bitfield.disp32 = 1;
6748 }
6749 else
6750 *types = gotrel[j].types64;
6751 }
6752
6753 if (GOT_symbol == NULL)
6754 GOT_symbol = symbol_find_or_make (GLOBAL_OFFSET_TABLE_NAME);
6755
6756 /* The length of the first part of our input line. */
6757 first = cp - input_line_pointer;
6758
6759 /* The second part goes from after the reloc token until
6760 (and including) an end_of_line char or comma. */
6761 past_reloc = cp + 1 + len;
6762 cp = past_reloc;
6763 while (!is_end_of_line[(unsigned char) *cp] && *cp != ',')
6764 ++cp;
6765 second = cp + 1 - past_reloc;
6766
6767 /* Allocate and copy string. The trailing NUL shouldn't
6768 be necessary, but be safe. */
6769 tmpbuf = (char *) xmalloc (first + second + 2);
6770 memcpy (tmpbuf, input_line_pointer, first);
6771 if (second != 0 && *past_reloc != ' ')
6772 /* Replace the relocation token with ' ', so that
6773 errors like foo@GOTOFF1 will be detected. */
6774 tmpbuf[first++] = ' ';
6775 memcpy (tmpbuf + first, past_reloc, second);
6776 tmpbuf[first + second] = '\0';
6777 return tmpbuf;
6778 }
6779
6780 as_bad (_("@%s reloc is not supported with %d-bit output format"),
6781 gotrel[j].str, 1 << (5 + object_64bit));
6782 return NULL;
6783 }
6784 }
6785
6786 /* Might be a symbol version string. Don't as_bad here. */
6787 return NULL;
6788 }
6789 #endif
6790
6791 #ifdef TE_PE
6792 #ifdef lex_got
6793 #undef lex_got
6794 #endif
6795 /* Parse operands of the form
6796 <symbol>@SECREL32+<nnn>
6797
6798 If we find one, set up the correct relocation in RELOC and copy the
6799 input string, minus the `@SECREL32' into a malloc'd buffer for
6800 parsing by the calling routine. Return this buffer, and if ADJUST
6801 is non-null set it to the length of the string we removed from the
6802 input line. Otherwise return NULL.
6803
6804 This function is copied from the ELF version above adjusted for PE targets. */
6805
6806 static char *
6807 lex_got (enum bfd_reloc_code_real *rel ATTRIBUTE_UNUSED,
6808 int *adjust ATTRIBUTE_UNUSED,
6809 i386_operand_type *types ATTRIBUTE_UNUSED)
6810 {
6811 static const struct
6812 {
6813 const char *str;
6814 int len;
6815 const enum bfd_reloc_code_real rel[2];
6816 const i386_operand_type types64;
6817 }
6818 gotrel[] =
6819 {
6820 { STRING_COMMA_LEN ("SECREL32"), { BFD_RELOC_32_SECREL,
6821 BFD_RELOC_32_SECREL },
6822 OPERAND_TYPE_IMM32_32S_64_DISP32_64 },
6823 };
6824
6825 char *cp;
6826 unsigned j;
6827
6828 for (cp = input_line_pointer; *cp != '@'; cp++)
6829 if (is_end_of_line[(unsigned char) *cp] || *cp == ',')
6830 return NULL;
6831
6832 for (j = 0; j < ARRAY_SIZE (gotrel); j++)
6833 {
6834 int len = gotrel[j].len;
6835
6836 if (strncasecmp (cp + 1, gotrel[j].str, len) == 0)
6837 {
6838 if (gotrel[j].rel[object_64bit] != 0)
6839 {
6840 int first, second;
6841 char *tmpbuf, *past_reloc;
6842
6843 *rel = gotrel[j].rel[object_64bit];
6844 if (adjust)
6845 *adjust = len;
6846
6847 if (types)
6848 {
6849 if (flag_code != CODE_64BIT)
6850 {
6851 types->bitfield.imm32 = 1;
6852 types->bitfield.disp32 = 1;
6853 }
6854 else
6855 *types = gotrel[j].types64;
6856 }
6857
6858 /* The length of the first part of our input line. */
6859 first = cp - input_line_pointer;
6860
6861 /* The second part goes from after the reloc token until
6862 (and including) an end_of_line char or comma. */
6863 past_reloc = cp + 1 + len;
6864 cp = past_reloc;
6865 while (!is_end_of_line[(unsigned char) *cp] && *cp != ',')
6866 ++cp;
6867 second = cp + 1 - past_reloc;
6868
6869 /* Allocate and copy string. The trailing NUL shouldn't
6870 be necessary, but be safe. */
6871 tmpbuf = (char *) xmalloc (first + second + 2);
6872 memcpy (tmpbuf, input_line_pointer, first);
6873 if (second != 0 && *past_reloc != ' ')
6874 /* Replace the relocation token with ' ', so that
6875 errors like foo@SECLREL321 will be detected. */
6876 tmpbuf[first++] = ' ';
6877 memcpy (tmpbuf + first, past_reloc, second);
6878 tmpbuf[first + second] = '\0';
6879 return tmpbuf;
6880 }
6881
6882 as_bad (_("@%s reloc is not supported with %d-bit output format"),
6883 gotrel[j].str, 1 << (5 + object_64bit));
6884 return NULL;
6885 }
6886 }
6887
6888 /* Might be a symbol version string. Don't as_bad here. */
6889 return NULL;
6890 }
6891
6892 #endif /* TE_PE */
6893
6894 void
6895 x86_cons (expressionS *exp, int size)
6896 {
6897 intel_syntax = -intel_syntax;
6898
6899 exp->X_md = 0;
6900 if (size == 4 || (object_64bit && size == 8))
6901 {
6902 /* Handle @GOTOFF and the like in an expression. */
6903 char *save;
6904 char *gotfree_input_line;
6905 int adjust = 0;
6906
6907 save = input_line_pointer;
6908 gotfree_input_line = lex_got (&got_reloc, &adjust, NULL);
6909 if (gotfree_input_line)
6910 input_line_pointer = gotfree_input_line;
6911
6912 expression (exp);
6913
6914 if (gotfree_input_line)
6915 {
6916 /* expression () has merrily parsed up to the end of line,
6917 or a comma - in the wrong buffer. Transfer how far
6918 input_line_pointer has moved to the right buffer. */
6919 input_line_pointer = (save
6920 + (input_line_pointer - gotfree_input_line)
6921 + adjust);
6922 free (gotfree_input_line);
6923 if (exp->X_op == O_constant
6924 || exp->X_op == O_absent
6925 || exp->X_op == O_illegal
6926 || exp->X_op == O_register
6927 || exp->X_op == O_big)
6928 {
6929 char c = *input_line_pointer;
6930 *input_line_pointer = 0;
6931 as_bad (_("missing or invalid expression `%s'"), save);
6932 *input_line_pointer = c;
6933 }
6934 }
6935 }
6936 else
6937 expression (exp);
6938
6939 intel_syntax = -intel_syntax;
6940
6941 if (intel_syntax)
6942 i386_intel_simplify (exp);
6943 }
6944
6945 static void
6946 signed_cons (int size)
6947 {
6948 if (flag_code == CODE_64BIT)
6949 cons_sign = 1;
6950 cons (size);
6951 cons_sign = -1;
6952 }
6953
6954 #ifdef TE_PE
6955 static void
6956 pe_directive_secrel (int dummy ATTRIBUTE_UNUSED)
6957 {
6958 expressionS exp;
6959
6960 do
6961 {
6962 expression (&exp);
6963 if (exp.X_op == O_symbol)
6964 exp.X_op = O_secrel;
6965
6966 emit_expr (&exp, 4);
6967 }
6968 while (*input_line_pointer++ == ',');
6969
6970 input_line_pointer--;
6971 demand_empty_rest_of_line ();
6972 }
6973 #endif
6974
6975 static int
6976 i386_immediate (char *imm_start)
6977 {
6978 char *save_input_line_pointer;
6979 char *gotfree_input_line;
6980 segT exp_seg = 0;
6981 expressionS *exp;
6982 i386_operand_type types;
6983
6984 operand_type_set (&types, ~0);
6985
6986 if (i.imm_operands == MAX_IMMEDIATE_OPERANDS)
6987 {
6988 as_bad (_("at most %d immediate operands are allowed"),
6989 MAX_IMMEDIATE_OPERANDS);
6990 return 0;
6991 }
6992
6993 exp = &im_expressions[i.imm_operands++];
6994 i.op[this_operand].imms = exp;
6995
6996 if (is_space_char (*imm_start))
6997 ++imm_start;
6998
6999 save_input_line_pointer = input_line_pointer;
7000 input_line_pointer = imm_start;
7001
7002 gotfree_input_line = lex_got (&i.reloc[this_operand], NULL, &types);
7003 if (gotfree_input_line)
7004 input_line_pointer = gotfree_input_line;
7005
7006 exp_seg = expression (exp);
7007
7008 SKIP_WHITESPACE ();
7009 if (*input_line_pointer)
7010 as_bad (_("junk `%s' after expression"), input_line_pointer);
7011
7012 input_line_pointer = save_input_line_pointer;
7013 if (gotfree_input_line)
7014 {
7015 free (gotfree_input_line);
7016
7017 if (exp->X_op == O_constant || exp->X_op == O_register)
7018 exp->X_op = O_illegal;
7019 }
7020
7021 return i386_finalize_immediate (exp_seg, exp, types, imm_start);
7022 }
7023
7024 static int
7025 i386_finalize_immediate (segT exp_seg ATTRIBUTE_UNUSED, expressionS *exp,
7026 i386_operand_type types, const char *imm_start)
7027 {
7028 if (exp->X_op == O_absent || exp->X_op == O_illegal || exp->X_op == O_big)
7029 {
7030 if (imm_start)
7031 as_bad (_("missing or invalid immediate expression `%s'"),
7032 imm_start);
7033 return 0;
7034 }
7035 else if (exp->X_op == O_constant)
7036 {
7037 /* Size it properly later. */
7038 i.types[this_operand].bitfield.imm64 = 1;
7039 /* If not 64bit, sign extend val. */
7040 if (flag_code != CODE_64BIT
7041 && (exp->X_add_number & ~(((addressT) 2 << 31) - 1)) == 0)
7042 exp->X_add_number
7043 = (exp->X_add_number ^ ((addressT) 1 << 31)) - ((addressT) 1 << 31);
7044 }
7045 #if (defined (OBJ_AOUT) || defined (OBJ_MAYBE_AOUT))
7046 else if (OUTPUT_FLAVOR == bfd_target_aout_flavour
7047 && exp_seg != absolute_section
7048 && exp_seg != text_section
7049 && exp_seg != data_section
7050 && exp_seg != bss_section
7051 && exp_seg != undefined_section
7052 && !bfd_is_com_section (exp_seg))
7053 {
7054 as_bad (_("unimplemented segment %s in operand"), exp_seg->name);
7055 return 0;
7056 }
7057 #endif
7058 else if (!intel_syntax && exp->X_op == O_register)
7059 {
7060 if (imm_start)
7061 as_bad (_("illegal immediate register operand %s"), imm_start);
7062 return 0;
7063 }
7064 else
7065 {
7066 /* This is an address. The size of the address will be
7067 determined later, depending on destination register,
7068 suffix, or the default for the section. */
7069 i.types[this_operand].bitfield.imm8 = 1;
7070 i.types[this_operand].bitfield.imm16 = 1;
7071 i.types[this_operand].bitfield.imm32 = 1;
7072 i.types[this_operand].bitfield.imm32s = 1;
7073 i.types[this_operand].bitfield.imm64 = 1;
7074 i.types[this_operand] = operand_type_and (i.types[this_operand],
7075 types);
7076 }
7077
7078 return 1;
7079 }
7080
7081 static char *
7082 i386_scale (char *scale)
7083 {
7084 offsetT val;
7085 char *save = input_line_pointer;
7086
7087 input_line_pointer = scale;
7088 val = get_absolute_expression ();
7089
7090 switch (val)
7091 {
7092 case 1:
7093 i.log2_scale_factor = 0;
7094 break;
7095 case 2:
7096 i.log2_scale_factor = 1;
7097 break;
7098 case 4:
7099 i.log2_scale_factor = 2;
7100 break;
7101 case 8:
7102 i.log2_scale_factor = 3;
7103 break;
7104 default:
7105 {
7106 char sep = *input_line_pointer;
7107
7108 *input_line_pointer = '\0';
7109 as_bad (_("expecting scale factor of 1, 2, 4, or 8: got `%s'"),
7110 scale);
7111 *input_line_pointer = sep;
7112 input_line_pointer = save;
7113 return NULL;
7114 }
7115 }
7116 if (i.log2_scale_factor != 0 && i.index_reg == 0)
7117 {
7118 as_warn (_("scale factor of %d without an index register"),
7119 1 << i.log2_scale_factor);
7120 i.log2_scale_factor = 0;
7121 }
7122 scale = input_line_pointer;
7123 input_line_pointer = save;
7124 return scale;
7125 }
7126
7127 static int
7128 i386_displacement (char *disp_start, char *disp_end)
7129 {
7130 expressionS *exp;
7131 segT exp_seg = 0;
7132 char *save_input_line_pointer;
7133 char *gotfree_input_line;
7134 int override;
7135 i386_operand_type bigdisp, types = anydisp;
7136 int ret;
7137
7138 if (i.disp_operands == MAX_MEMORY_OPERANDS)
7139 {
7140 as_bad (_("at most %d displacement operands are allowed"),
7141 MAX_MEMORY_OPERANDS);
7142 return 0;
7143 }
7144
7145 operand_type_set (&bigdisp, 0);
7146 if ((i.types[this_operand].bitfield.jumpabsolute)
7147 || (!current_templates->start->opcode_modifier.jump
7148 && !current_templates->start->opcode_modifier.jumpdword))
7149 {
7150 bigdisp.bitfield.disp32 = 1;
7151 override = (i.prefix[ADDR_PREFIX] != 0);
7152 if (flag_code == CODE_64BIT)
7153 {
7154 if (!override)
7155 {
7156 bigdisp.bitfield.disp32s = 1;
7157 bigdisp.bitfield.disp64 = 1;
7158 }
7159 }
7160 else if ((flag_code == CODE_16BIT) ^ override)
7161 {
7162 bigdisp.bitfield.disp32 = 0;
7163 bigdisp.bitfield.disp16 = 1;
7164 }
7165 }
7166 else
7167 {
7168 /* For PC-relative branches, the width of the displacement
7169 is dependent upon data size, not address size. */
7170 override = (i.prefix[DATA_PREFIX] != 0);
7171 if (flag_code == CODE_64BIT)
7172 {
7173 if (override || i.suffix == WORD_MNEM_SUFFIX)
7174 bigdisp.bitfield.disp16 = 1;
7175 else
7176 {
7177 bigdisp.bitfield.disp32 = 1;
7178 bigdisp.bitfield.disp32s = 1;
7179 }
7180 }
7181 else
7182 {
7183 if (!override)
7184 override = (i.suffix == (flag_code != CODE_16BIT
7185 ? WORD_MNEM_SUFFIX
7186 : LONG_MNEM_SUFFIX));
7187 bigdisp.bitfield.disp32 = 1;
7188 if ((flag_code == CODE_16BIT) ^ override)
7189 {
7190 bigdisp.bitfield.disp32 = 0;
7191 bigdisp.bitfield.disp16 = 1;
7192 }
7193 }
7194 }
7195 i.types[this_operand] = operand_type_or (i.types[this_operand],
7196 bigdisp);
7197
7198 exp = &disp_expressions[i.disp_operands];
7199 i.op[this_operand].disps = exp;
7200 i.disp_operands++;
7201 save_input_line_pointer = input_line_pointer;
7202 input_line_pointer = disp_start;
7203 END_STRING_AND_SAVE (disp_end);
7204
7205 #ifndef GCC_ASM_O_HACK
7206 #define GCC_ASM_O_HACK 0
7207 #endif
7208 #if GCC_ASM_O_HACK
7209 END_STRING_AND_SAVE (disp_end + 1);
7210 if (i.types[this_operand].bitfield.baseIndex
7211 && displacement_string_end[-1] == '+')
7212 {
7213 /* This hack is to avoid a warning when using the "o"
7214 constraint within gcc asm statements.
7215 For instance:
7216
7217 #define _set_tssldt_desc(n,addr,limit,type) \
7218 __asm__ __volatile__ ( \
7219 "movw %w2,%0\n\t" \
7220 "movw %w1,2+%0\n\t" \
7221 "rorl $16,%1\n\t" \
7222 "movb %b1,4+%0\n\t" \
7223 "movb %4,5+%0\n\t" \
7224 "movb $0,6+%0\n\t" \
7225 "movb %h1,7+%0\n\t" \
7226 "rorl $16,%1" \
7227 : "=o"(*(n)) : "q" (addr), "ri"(limit), "i"(type))
7228
7229 This works great except that the output assembler ends
7230 up looking a bit weird if it turns out that there is
7231 no offset. You end up producing code that looks like:
7232
7233 #APP
7234 movw $235,(%eax)
7235 movw %dx,2+(%eax)
7236 rorl $16,%edx
7237 movb %dl,4+(%eax)
7238 movb $137,5+(%eax)
7239 movb $0,6+(%eax)
7240 movb %dh,7+(%eax)
7241 rorl $16,%edx
7242 #NO_APP
7243
7244 So here we provide the missing zero. */
7245
7246 *displacement_string_end = '0';
7247 }
7248 #endif
7249 gotfree_input_line = lex_got (&i.reloc[this_operand], NULL, &types);
7250 if (gotfree_input_line)
7251 input_line_pointer = gotfree_input_line;
7252
7253 exp_seg = expression (exp);
7254
7255 SKIP_WHITESPACE ();
7256 if (*input_line_pointer)
7257 as_bad (_("junk `%s' after expression"), input_line_pointer);
7258 #if GCC_ASM_O_HACK
7259 RESTORE_END_STRING (disp_end + 1);
7260 #endif
7261 input_line_pointer = save_input_line_pointer;
7262 if (gotfree_input_line)
7263 {
7264 free (gotfree_input_line);
7265
7266 if (exp->X_op == O_constant || exp->X_op == O_register)
7267 exp->X_op = O_illegal;
7268 }
7269
7270 ret = i386_finalize_displacement (exp_seg, exp, types, disp_start);
7271
7272 RESTORE_END_STRING (disp_end);
7273
7274 return ret;
7275 }
7276
7277 static int
7278 i386_finalize_displacement (segT exp_seg ATTRIBUTE_UNUSED, expressionS *exp,
7279 i386_operand_type types, const char *disp_start)
7280 {
7281 i386_operand_type bigdisp;
7282 int ret = 1;
7283
7284 /* We do this to make sure that the section symbol is in
7285 the symbol table. We will ultimately change the relocation
7286 to be relative to the beginning of the section. */
7287 if (i.reloc[this_operand] == BFD_RELOC_386_GOTOFF
7288 || i.reloc[this_operand] == BFD_RELOC_X86_64_GOTPCREL
7289 || i.reloc[this_operand] == BFD_RELOC_X86_64_GOTOFF64)
7290 {
7291 if (exp->X_op != O_symbol)
7292 goto inv_disp;
7293
7294 if (S_IS_LOCAL (exp->X_add_symbol)
7295 && S_GET_SEGMENT (exp->X_add_symbol) != undefined_section
7296 && S_GET_SEGMENT (exp->X_add_symbol) != expr_section)
7297 section_symbol (S_GET_SEGMENT (exp->X_add_symbol));
7298 exp->X_op = O_subtract;
7299 exp->X_op_symbol = GOT_symbol;
7300 if (i.reloc[this_operand] == BFD_RELOC_X86_64_GOTPCREL)
7301 i.reloc[this_operand] = BFD_RELOC_32_PCREL;
7302 else if (i.reloc[this_operand] == BFD_RELOC_X86_64_GOTOFF64)
7303 i.reloc[this_operand] = BFD_RELOC_64;
7304 else
7305 i.reloc[this_operand] = BFD_RELOC_32;
7306 }
7307
7308 else if (exp->X_op == O_absent
7309 || exp->X_op == O_illegal
7310 || exp->X_op == O_big)
7311 {
7312 inv_disp:
7313 as_bad (_("missing or invalid displacement expression `%s'"),
7314 disp_start);
7315 ret = 0;
7316 }
7317
7318 else if (flag_code == CODE_64BIT
7319 && !i.prefix[ADDR_PREFIX]
7320 && exp->X_op == O_constant)
7321 {
7322 /* Since displacement is signed extended to 64bit, don't allow
7323 disp32 and turn off disp32s if they are out of range. */
7324 i.types[this_operand].bitfield.disp32 = 0;
7325 if (!fits_in_signed_long (exp->X_add_number))
7326 {
7327 i.types[this_operand].bitfield.disp32s = 0;
7328 if (i.types[this_operand].bitfield.baseindex)
7329 {
7330 as_bad (_("0x%lx out range of signed 32bit displacement"),
7331 (long) exp->X_add_number);
7332 ret = 0;
7333 }
7334 }
7335 }
7336
7337 #if (defined (OBJ_AOUT) || defined (OBJ_MAYBE_AOUT))
7338 else if (exp->X_op != O_constant
7339 && OUTPUT_FLAVOR == bfd_target_aout_flavour
7340 && exp_seg != absolute_section
7341 && exp_seg != text_section
7342 && exp_seg != data_section
7343 && exp_seg != bss_section
7344 && exp_seg != undefined_section
7345 && !bfd_is_com_section (exp_seg))
7346 {
7347 as_bad (_("unimplemented segment %s in operand"), exp_seg->name);
7348 ret = 0;
7349 }
7350 #endif
7351
7352 /* Check if this is a displacement only operand. */
7353 bigdisp = i.types[this_operand];
7354 bigdisp.bitfield.disp8 = 0;
7355 bigdisp.bitfield.disp16 = 0;
7356 bigdisp.bitfield.disp32 = 0;
7357 bigdisp.bitfield.disp32s = 0;
7358 bigdisp.bitfield.disp64 = 0;
7359 if (operand_type_all_zero (&bigdisp))
7360 i.types[this_operand] = operand_type_and (i.types[this_operand],
7361 types);
7362
7363 return ret;
7364 }
7365
7366 /* Make sure the memory operand we've been dealt is valid.
7367 Return 1 on success, 0 on a failure. */
7368
7369 static int
7370 i386_index_check (const char *operand_string)
7371 {
7372 int ok;
7373 const char *kind = "base/index";
7374 #if INFER_ADDR_PREFIX
7375 int fudged = 0;
7376
7377 tryprefix:
7378 #endif
7379 ok = 1;
7380 if (current_templates->start->opcode_modifier.isstring
7381 && !current_templates->start->opcode_modifier.immext
7382 && (current_templates->end[-1].opcode_modifier.isstring
7383 || i.mem_operands))
7384 {
7385 /* Memory operands of string insns are special in that they only allow
7386 a single register (rDI, rSI, or rBX) as their memory address. */
7387 unsigned int expected;
7388
7389 kind = "string address";
7390
7391 if (current_templates->start->opcode_modifier.w)
7392 {
7393 i386_operand_type type = current_templates->end[-1].operand_types[0];
7394
7395 if (!type.bitfield.baseindex
7396 || ((!i.mem_operands != !intel_syntax)
7397 && current_templates->end[-1].operand_types[1]
7398 .bitfield.baseindex))
7399 type = current_templates->end[-1].operand_types[1];
7400 expected = type.bitfield.esseg ? 7 /* rDI */ : 6 /* rSI */;
7401 }
7402 else
7403 expected = 3 /* rBX */;
7404
7405 if (!i.base_reg || i.index_reg
7406 || operand_type_check (i.types[this_operand], disp))
7407 ok = -1;
7408 else if (!(flag_code == CODE_64BIT
7409 ? i.prefix[ADDR_PREFIX]
7410 ? i.base_reg->reg_type.bitfield.reg32
7411 : i.base_reg->reg_type.bitfield.reg64
7412 : (flag_code == CODE_16BIT) ^ !i.prefix[ADDR_PREFIX]
7413 ? i.base_reg->reg_type.bitfield.reg32
7414 : i.base_reg->reg_type.bitfield.reg16))
7415 ok = 0;
7416 else if (i.base_reg->reg_num != expected)
7417 ok = -1;
7418
7419 if (ok < 0)
7420 {
7421 unsigned int j;
7422
7423 for (j = 0; j < i386_regtab_size; ++j)
7424 if ((flag_code == CODE_64BIT
7425 ? i.prefix[ADDR_PREFIX]
7426 ? i386_regtab[j].reg_type.bitfield.reg32
7427 : i386_regtab[j].reg_type.bitfield.reg64
7428 : (flag_code == CODE_16BIT) ^ !i.prefix[ADDR_PREFIX]
7429 ? i386_regtab[j].reg_type.bitfield.reg32
7430 : i386_regtab[j].reg_type.bitfield.reg16)
7431 && i386_regtab[j].reg_num == expected)
7432 break;
7433 gas_assert (j < i386_regtab_size);
7434 as_warn (_("`%s' is not valid here (expected `%c%s%s%c')"),
7435 operand_string,
7436 intel_syntax ? '[' : '(',
7437 register_prefix,
7438 i386_regtab[j].reg_name,
7439 intel_syntax ? ']' : ')');
7440 ok = 1;
7441 }
7442 }
7443 else if (flag_code == CODE_64BIT)
7444 {
7445 if ((i.base_reg
7446 && ((i.prefix[ADDR_PREFIX] == 0
7447 && !i.base_reg->reg_type.bitfield.reg64)
7448 || (i.prefix[ADDR_PREFIX]
7449 && !i.base_reg->reg_type.bitfield.reg32))
7450 && (i.index_reg
7451 || i.base_reg->reg_num !=
7452 (i.prefix[ADDR_PREFIX] == 0 ? RegRip : RegEip)))
7453 || (i.index_reg
7454 && !(i.index_reg->reg_type.bitfield.regxmm
7455 || i.index_reg->reg_type.bitfield.regymm)
7456 && (!i.index_reg->reg_type.bitfield.baseindex
7457 || (i.prefix[ADDR_PREFIX] == 0
7458 && i.index_reg->reg_num != RegRiz
7459 && !i.index_reg->reg_type.bitfield.reg64
7460 )
7461 || (i.prefix[ADDR_PREFIX]
7462 && i.index_reg->reg_num != RegEiz
7463 && !i.index_reg->reg_type.bitfield.reg32))))
7464 ok = 0;
7465 }
7466 else
7467 {
7468 if ((flag_code == CODE_16BIT) ^ (i.prefix[ADDR_PREFIX] != 0))
7469 {
7470 /* 16bit checks. */
7471 if ((i.base_reg
7472 && (!i.base_reg->reg_type.bitfield.reg16
7473 || !i.base_reg->reg_type.bitfield.baseindex))
7474 || (i.index_reg
7475 && (!i.index_reg->reg_type.bitfield.reg16
7476 || !i.index_reg->reg_type.bitfield.baseindex
7477 || !(i.base_reg
7478 && i.base_reg->reg_num < 6
7479 && i.index_reg->reg_num >= 6
7480 && i.log2_scale_factor == 0))))
7481 ok = 0;
7482 }
7483 else
7484 {
7485 /* 32bit checks. */
7486 if ((i.base_reg
7487 && !i.base_reg->reg_type.bitfield.reg32)
7488 || (i.index_reg
7489 && !i.index_reg->reg_type.bitfield.regxmm
7490 && !i.index_reg->reg_type.bitfield.regymm
7491 && ((!i.index_reg->reg_type.bitfield.reg32
7492 && i.index_reg->reg_num != RegEiz)
7493 || !i.index_reg->reg_type.bitfield.baseindex)))
7494 ok = 0;
7495 }
7496 }
7497 if (!ok)
7498 {
7499 #if INFER_ADDR_PREFIX
7500 if (!i.mem_operands && !i.prefix[ADDR_PREFIX])
7501 {
7502 i.prefix[ADDR_PREFIX] = ADDR_PREFIX_OPCODE;
7503 i.prefixes += 1;
7504 /* Change the size of any displacement too. At most one of
7505 Disp16 or Disp32 is set.
7506 FIXME. There doesn't seem to be any real need for separate
7507 Disp16 and Disp32 flags. The same goes for Imm16 and Imm32.
7508 Removing them would probably clean up the code quite a lot. */
7509 if (flag_code != CODE_64BIT
7510 && (i.types[this_operand].bitfield.disp16
7511 || i.types[this_operand].bitfield.disp32))
7512 i.types[this_operand]
7513 = operand_type_xor (i.types[this_operand], disp16_32);
7514 fudged = 1;
7515 goto tryprefix;
7516 }
7517 if (fudged)
7518 as_bad (_("`%s' is not a valid %s expression"),
7519 operand_string,
7520 kind);
7521 else
7522 #endif
7523 as_bad (_("`%s' is not a valid %s-bit %s expression"),
7524 operand_string,
7525 flag_code_names[i.prefix[ADDR_PREFIX]
7526 ? flag_code == CODE_32BIT
7527 ? CODE_16BIT
7528 : CODE_32BIT
7529 : flag_code],
7530 kind);
7531 }
7532 return ok;
7533 }
7534
7535 /* Parse OPERAND_STRING into the i386_insn structure I. Returns zero
7536 on error. */
7537
7538 static int
7539 i386_att_operand (char *operand_string)
7540 {
7541 const reg_entry *r;
7542 char *end_op;
7543 char *op_string = operand_string;
7544
7545 if (is_space_char (*op_string))
7546 ++op_string;
7547
7548 /* We check for an absolute prefix (differentiating,
7549 for example, 'jmp pc_relative_label' from 'jmp *absolute_label'. */
7550 if (*op_string == ABSOLUTE_PREFIX)
7551 {
7552 ++op_string;
7553 if (is_space_char (*op_string))
7554 ++op_string;
7555 i.types[this_operand].bitfield.jumpabsolute = 1;
7556 }
7557
7558 /* Check if operand is a register. */
7559 if ((r = parse_register (op_string, &end_op)) != NULL)
7560 {
7561 i386_operand_type temp;
7562
7563 /* Check for a segment override by searching for ':' after a
7564 segment register. */
7565 op_string = end_op;
7566 if (is_space_char (*op_string))
7567 ++op_string;
7568 if (*op_string == ':'
7569 && (r->reg_type.bitfield.sreg2
7570 || r->reg_type.bitfield.sreg3))
7571 {
7572 switch (r->reg_num)
7573 {
7574 case 0:
7575 i.seg[i.mem_operands] = &es;
7576 break;
7577 case 1:
7578 i.seg[i.mem_operands] = &cs;
7579 break;
7580 case 2:
7581 i.seg[i.mem_operands] = &ss;
7582 break;
7583 case 3:
7584 i.seg[i.mem_operands] = &ds;
7585 break;
7586 case 4:
7587 i.seg[i.mem_operands] = &fs;
7588 break;
7589 case 5:
7590 i.seg[i.mem_operands] = &gs;
7591 break;
7592 }
7593
7594 /* Skip the ':' and whitespace. */
7595 ++op_string;
7596 if (is_space_char (*op_string))
7597 ++op_string;
7598
7599 if (!is_digit_char (*op_string)
7600 && !is_identifier_char (*op_string)
7601 && *op_string != '('
7602 && *op_string != ABSOLUTE_PREFIX)
7603 {
7604 as_bad (_("bad memory operand `%s'"), op_string);
7605 return 0;
7606 }
7607 /* Handle case of %es:*foo. */
7608 if (*op_string == ABSOLUTE_PREFIX)
7609 {
7610 ++op_string;
7611 if (is_space_char (*op_string))
7612 ++op_string;
7613 i.types[this_operand].bitfield.jumpabsolute = 1;
7614 }
7615 goto do_memory_reference;
7616 }
7617 if (*op_string)
7618 {
7619 as_bad (_("junk `%s' after register"), op_string);
7620 return 0;
7621 }
7622 temp = r->reg_type;
7623 temp.bitfield.baseindex = 0;
7624 i.types[this_operand] = operand_type_or (i.types[this_operand],
7625 temp);
7626 i.types[this_operand].bitfield.unspecified = 0;
7627 i.op[this_operand].regs = r;
7628 i.reg_operands++;
7629 }
7630 else if (*op_string == REGISTER_PREFIX)
7631 {
7632 as_bad (_("bad register name `%s'"), op_string);
7633 return 0;
7634 }
7635 else if (*op_string == IMMEDIATE_PREFIX)
7636 {
7637 ++op_string;
7638 if (i.types[this_operand].bitfield.jumpabsolute)
7639 {
7640 as_bad (_("immediate operand illegal with absolute jump"));
7641 return 0;
7642 }
7643 if (!i386_immediate (op_string))
7644 return 0;
7645 }
7646 else if (is_digit_char (*op_string)
7647 || is_identifier_char (*op_string)
7648 || *op_string == '(')
7649 {
7650 /* This is a memory reference of some sort. */
7651 char *base_string;
7652
7653 /* Start and end of displacement string expression (if found). */
7654 char *displacement_string_start;
7655 char *displacement_string_end;
7656
7657 do_memory_reference:
7658 if ((i.mem_operands == 1
7659 && !current_templates->start->opcode_modifier.isstring)
7660 || i.mem_operands == 2)
7661 {
7662 as_bad (_("too many memory references for `%s'"),
7663 current_templates->start->name);
7664 return 0;
7665 }
7666
7667 /* Check for base index form. We detect the base index form by
7668 looking for an ')' at the end of the operand, searching
7669 for the '(' matching it, and finding a REGISTER_PREFIX or ','
7670 after the '('. */
7671 base_string = op_string + strlen (op_string);
7672
7673 --base_string;
7674 if (is_space_char (*base_string))
7675 --base_string;
7676
7677 /* If we only have a displacement, set-up for it to be parsed later. */
7678 displacement_string_start = op_string;
7679 displacement_string_end = base_string + 1;
7680
7681 if (*base_string == ')')
7682 {
7683 char *temp_string;
7684 unsigned int parens_balanced = 1;
7685 /* We've already checked that the number of left & right ()'s are
7686 equal, so this loop will not be infinite. */
7687 do
7688 {
7689 base_string--;
7690 if (*base_string == ')')
7691 parens_balanced++;
7692 if (*base_string == '(')
7693 parens_balanced--;
7694 }
7695 while (parens_balanced);
7696
7697 temp_string = base_string;
7698
7699 /* Skip past '(' and whitespace. */
7700 ++base_string;
7701 if (is_space_char (*base_string))
7702 ++base_string;
7703
7704 if (*base_string == ','
7705 || ((i.base_reg = parse_register (base_string, &end_op))
7706 != NULL))
7707 {
7708 displacement_string_end = temp_string;
7709
7710 i.types[this_operand].bitfield.baseindex = 1;
7711
7712 if (i.base_reg)
7713 {
7714 base_string = end_op;
7715 if (is_space_char (*base_string))
7716 ++base_string;
7717 }
7718
7719 /* There may be an index reg or scale factor here. */
7720 if (*base_string == ',')
7721 {
7722 ++base_string;
7723 if (is_space_char (*base_string))
7724 ++base_string;
7725
7726 if ((i.index_reg = parse_register (base_string, &end_op))
7727 != NULL)
7728 {
7729 base_string = end_op;
7730 if (is_space_char (*base_string))
7731 ++base_string;
7732 if (*base_string == ',')
7733 {
7734 ++base_string;
7735 if (is_space_char (*base_string))
7736 ++base_string;
7737 }
7738 else if (*base_string != ')')
7739 {
7740 as_bad (_("expecting `,' or `)' "
7741 "after index register in `%s'"),
7742 operand_string);
7743 return 0;
7744 }
7745 }
7746 else if (*base_string == REGISTER_PREFIX)
7747 {
7748 as_bad (_("bad register name `%s'"), base_string);
7749 return 0;
7750 }
7751
7752 /* Check for scale factor. */
7753 if (*base_string != ')')
7754 {
7755 char *end_scale = i386_scale (base_string);
7756
7757 if (!end_scale)
7758 return 0;
7759
7760 base_string = end_scale;
7761 if (is_space_char (*base_string))
7762 ++base_string;
7763 if (*base_string != ')')
7764 {
7765 as_bad (_("expecting `)' "
7766 "after scale factor in `%s'"),
7767 operand_string);
7768 return 0;
7769 }
7770 }
7771 else if (!i.index_reg)
7772 {
7773 as_bad (_("expecting index register or scale factor "
7774 "after `,'; got '%c'"),
7775 *base_string);
7776 return 0;
7777 }
7778 }
7779 else if (*base_string != ')')
7780 {
7781 as_bad (_("expecting `,' or `)' "
7782 "after base register in `%s'"),
7783 operand_string);
7784 return 0;
7785 }
7786 }
7787 else if (*base_string == REGISTER_PREFIX)
7788 {
7789 as_bad (_("bad register name `%s'"), base_string);
7790 return 0;
7791 }
7792 }
7793
7794 /* If there's an expression beginning the operand, parse it,
7795 assuming displacement_string_start and
7796 displacement_string_end are meaningful. */
7797 if (displacement_string_start != displacement_string_end)
7798 {
7799 if (!i386_displacement (displacement_string_start,
7800 displacement_string_end))
7801 return 0;
7802 }
7803
7804 /* Special case for (%dx) while doing input/output op. */
7805 if (i.base_reg
7806 && operand_type_equal (&i.base_reg->reg_type,
7807 &reg16_inoutportreg)
7808 && i.index_reg == 0
7809 && i.log2_scale_factor == 0
7810 && i.seg[i.mem_operands] == 0
7811 && !operand_type_check (i.types[this_operand], disp))
7812 {
7813 i.types[this_operand] = inoutportreg;
7814 return 1;
7815 }
7816
7817 if (i386_index_check (operand_string) == 0)
7818 return 0;
7819 i.types[this_operand].bitfield.mem = 1;
7820 i.mem_operands++;
7821 }
7822 else
7823 {
7824 /* It's not a memory operand; argh! */
7825 as_bad (_("invalid char %s beginning operand %d `%s'"),
7826 output_invalid (*op_string),
7827 this_operand + 1,
7828 op_string);
7829 return 0;
7830 }
7831 return 1; /* Normal return. */
7832 }
7833 \f
7834 /* Calculate the maximum variable size (i.e., excluding fr_fix)
7835 that an rs_machine_dependent frag may reach. */
7836
7837 unsigned int
7838 i386_frag_max_var (fragS *frag)
7839 {
7840 /* The only relaxable frags are for jumps.
7841 Unconditional jumps can grow by 4 bytes and others by 5 bytes. */
7842 gas_assert (frag->fr_type == rs_machine_dependent);
7843 return TYPE_FROM_RELAX_STATE (frag->fr_subtype) == UNCOND_JUMP ? 4 : 5;
7844 }
7845
7846 /* md_estimate_size_before_relax()
7847
7848 Called just before relax() for rs_machine_dependent frags. The x86
7849 assembler uses these frags to handle variable size jump
7850 instructions.
7851
7852 Any symbol that is now undefined will not become defined.
7853 Return the correct fr_subtype in the frag.
7854 Return the initial "guess for variable size of frag" to caller.
7855 The guess is actually the growth beyond the fixed part. Whatever
7856 we do to grow the fixed or variable part contributes to our
7857 returned value. */
7858
7859 int
7860 md_estimate_size_before_relax (fragS *fragP, segT segment)
7861 {
7862 /* We've already got fragP->fr_subtype right; all we have to do is
7863 check for un-relaxable symbols. On an ELF system, we can't relax
7864 an externally visible symbol, because it may be overridden by a
7865 shared library. */
7866 if (S_GET_SEGMENT (fragP->fr_symbol) != segment
7867 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
7868 || (IS_ELF
7869 && (S_IS_EXTERNAL (fragP->fr_symbol)
7870 || S_IS_WEAK (fragP->fr_symbol)
7871 || ((symbol_get_bfdsym (fragP->fr_symbol)->flags
7872 & BSF_GNU_INDIRECT_FUNCTION))))
7873 #endif
7874 #if defined (OBJ_COFF) && defined (TE_PE)
7875 || (OUTPUT_FLAVOR == bfd_target_coff_flavour
7876 && S_IS_WEAK (fragP->fr_symbol))
7877 #endif
7878 )
7879 {
7880 /* Symbol is undefined in this segment, or we need to keep a
7881 reloc so that weak symbols can be overridden. */
7882 int size = (fragP->fr_subtype & CODE16) ? 2 : 4;
7883 enum bfd_reloc_code_real reloc_type;
7884 unsigned char *opcode;
7885 int old_fr_fix;
7886
7887 if (fragP->fr_var != NO_RELOC)
7888 reloc_type = (enum bfd_reloc_code_real) fragP->fr_var;
7889 else if (size == 2)
7890 reloc_type = BFD_RELOC_16_PCREL;
7891 else
7892 reloc_type = BFD_RELOC_32_PCREL;
7893
7894 old_fr_fix = fragP->fr_fix;
7895 opcode = (unsigned char *) fragP->fr_opcode;
7896
7897 switch (TYPE_FROM_RELAX_STATE (fragP->fr_subtype))
7898 {
7899 case UNCOND_JUMP:
7900 /* Make jmp (0xeb) a (d)word displacement jump. */
7901 opcode[0] = 0xe9;
7902 fragP->fr_fix += size;
7903 fix_new (fragP, old_fr_fix, size,
7904 fragP->fr_symbol,
7905 fragP->fr_offset, 1,
7906 reloc_type);
7907 break;
7908
7909 case COND_JUMP86:
7910 if (size == 2
7911 && (!no_cond_jump_promotion || fragP->fr_var != NO_RELOC))
7912 {
7913 /* Negate the condition, and branch past an
7914 unconditional jump. */
7915 opcode[0] ^= 1;
7916 opcode[1] = 3;
7917 /* Insert an unconditional jump. */
7918 opcode[2] = 0xe9;
7919 /* We added two extra opcode bytes, and have a two byte
7920 offset. */
7921 fragP->fr_fix += 2 + 2;
7922 fix_new (fragP, old_fr_fix + 2, 2,
7923 fragP->fr_symbol,
7924 fragP->fr_offset, 1,
7925 reloc_type);
7926 break;
7927 }
7928 /* Fall through. */
7929
7930 case COND_JUMP:
7931 if (no_cond_jump_promotion && fragP->fr_var == NO_RELOC)
7932 {
7933 fixS *fixP;
7934
7935 fragP->fr_fix += 1;
7936 fixP = fix_new (fragP, old_fr_fix, 1,
7937 fragP->fr_symbol,
7938 fragP->fr_offset, 1,
7939 BFD_RELOC_8_PCREL);
7940 fixP->fx_signed = 1;
7941 break;
7942 }
7943
7944 /* This changes the byte-displacement jump 0x7N
7945 to the (d)word-displacement jump 0x0f,0x8N. */
7946 opcode[1] = opcode[0] + 0x10;
7947 opcode[0] = TWO_BYTE_OPCODE_ESCAPE;
7948 /* We've added an opcode byte. */
7949 fragP->fr_fix += 1 + size;
7950 fix_new (fragP, old_fr_fix + 1, size,
7951 fragP->fr_symbol,
7952 fragP->fr_offset, 1,
7953 reloc_type);
7954 break;
7955
7956 default:
7957 BAD_CASE (fragP->fr_subtype);
7958 break;
7959 }
7960 frag_wane (fragP);
7961 return fragP->fr_fix - old_fr_fix;
7962 }
7963
7964 /* Guess size depending on current relax state. Initially the relax
7965 state will correspond to a short jump and we return 1, because
7966 the variable part of the frag (the branch offset) is one byte
7967 long. However, we can relax a section more than once and in that
7968 case we must either set fr_subtype back to the unrelaxed state,
7969 or return the value for the appropriate branch. */
7970 return md_relax_table[fragP->fr_subtype].rlx_length;
7971 }
7972
7973 /* Called after relax() is finished.
7974
7975 In: Address of frag.
7976 fr_type == rs_machine_dependent.
7977 fr_subtype is what the address relaxed to.
7978
7979 Out: Any fixSs and constants are set up.
7980 Caller will turn frag into a ".space 0". */
7981
7982 void
7983 md_convert_frag (bfd *abfd ATTRIBUTE_UNUSED, segT sec ATTRIBUTE_UNUSED,
7984 fragS *fragP)
7985 {
7986 unsigned char *opcode;
7987 unsigned char *where_to_put_displacement = NULL;
7988 offsetT target_address;
7989 offsetT opcode_address;
7990 unsigned int extension = 0;
7991 offsetT displacement_from_opcode_start;
7992
7993 opcode = (unsigned char *) fragP->fr_opcode;
7994
7995 /* Address we want to reach in file space. */
7996 target_address = S_GET_VALUE (fragP->fr_symbol) + fragP->fr_offset;
7997
7998 /* Address opcode resides at in file space. */
7999 opcode_address = fragP->fr_address + fragP->fr_fix;
8000
8001 /* Displacement from opcode start to fill into instruction. */
8002 displacement_from_opcode_start = target_address - opcode_address;
8003
8004 if ((fragP->fr_subtype & BIG) == 0)
8005 {
8006 /* Don't have to change opcode. */
8007 extension = 1; /* 1 opcode + 1 displacement */
8008 where_to_put_displacement = &opcode[1];
8009 }
8010 else
8011 {
8012 if (no_cond_jump_promotion
8013 && TYPE_FROM_RELAX_STATE (fragP->fr_subtype) != UNCOND_JUMP)
8014 as_warn_where (fragP->fr_file, fragP->fr_line,
8015 _("long jump required"));
8016
8017 switch (fragP->fr_subtype)
8018 {
8019 case ENCODE_RELAX_STATE (UNCOND_JUMP, BIG):
8020 extension = 4; /* 1 opcode + 4 displacement */
8021 opcode[0] = 0xe9;
8022 where_to_put_displacement = &opcode[1];
8023 break;
8024
8025 case ENCODE_RELAX_STATE (UNCOND_JUMP, BIG16):
8026 extension = 2; /* 1 opcode + 2 displacement */
8027 opcode[0] = 0xe9;
8028 where_to_put_displacement = &opcode[1];
8029 break;
8030
8031 case ENCODE_RELAX_STATE (COND_JUMP, BIG):
8032 case ENCODE_RELAX_STATE (COND_JUMP86, BIG):
8033 extension = 5; /* 2 opcode + 4 displacement */
8034 opcode[1] = opcode[0] + 0x10;
8035 opcode[0] = TWO_BYTE_OPCODE_ESCAPE;
8036 where_to_put_displacement = &opcode[2];
8037 break;
8038
8039 case ENCODE_RELAX_STATE (COND_JUMP, BIG16):
8040 extension = 3; /* 2 opcode + 2 displacement */
8041 opcode[1] = opcode[0] + 0x10;
8042 opcode[0] = TWO_BYTE_OPCODE_ESCAPE;
8043 where_to_put_displacement = &opcode[2];
8044 break;
8045
8046 case ENCODE_RELAX_STATE (COND_JUMP86, BIG16):
8047 extension = 4;
8048 opcode[0] ^= 1;
8049 opcode[1] = 3;
8050 opcode[2] = 0xe9;
8051 where_to_put_displacement = &opcode[3];
8052 break;
8053
8054 default:
8055 BAD_CASE (fragP->fr_subtype);
8056 break;
8057 }
8058 }
8059
8060 /* If size if less then four we are sure that the operand fits,
8061 but if it's 4, then it could be that the displacement is larger
8062 then -/+ 2GB. */
8063 if (DISP_SIZE_FROM_RELAX_STATE (fragP->fr_subtype) == 4
8064 && object_64bit
8065 && ((addressT) (displacement_from_opcode_start - extension
8066 + ((addressT) 1 << 31))
8067 > (((addressT) 2 << 31) - 1)))
8068 {
8069 as_bad_where (fragP->fr_file, fragP->fr_line,
8070 _("jump target out of range"));
8071 /* Make us emit 0. */
8072 displacement_from_opcode_start = extension;
8073 }
8074 /* Now put displacement after opcode. */
8075 md_number_to_chars ((char *) where_to_put_displacement,
8076 (valueT) (displacement_from_opcode_start - extension),
8077 DISP_SIZE_FROM_RELAX_STATE (fragP->fr_subtype));
8078 fragP->fr_fix += extension;
8079 }
8080 \f
8081 /* Apply a fixup (fixP) to segment data, once it has been determined
8082 by our caller that we have all the info we need to fix it up.
8083
8084 Parameter valP is the pointer to the value of the bits.
8085
8086 On the 386, immediates, displacements, and data pointers are all in
8087 the same (little-endian) format, so we don't need to care about which
8088 we are handling. */
8089
8090 void
8091 md_apply_fix (fixS *fixP, valueT *valP, segT seg ATTRIBUTE_UNUSED)
8092 {
8093 char *p = fixP->fx_where + fixP->fx_frag->fr_literal;
8094 valueT value = *valP;
8095
8096 #if !defined (TE_Mach)
8097 if (fixP->fx_pcrel)
8098 {
8099 switch (fixP->fx_r_type)
8100 {
8101 default:
8102 break;
8103
8104 case BFD_RELOC_64:
8105 fixP->fx_r_type = BFD_RELOC_64_PCREL;
8106 break;
8107 case BFD_RELOC_32:
8108 case BFD_RELOC_X86_64_32S:
8109 fixP->fx_r_type = BFD_RELOC_32_PCREL;
8110 break;
8111 case BFD_RELOC_16:
8112 fixP->fx_r_type = BFD_RELOC_16_PCREL;
8113 break;
8114 case BFD_RELOC_8:
8115 fixP->fx_r_type = BFD_RELOC_8_PCREL;
8116 break;
8117 }
8118 }
8119
8120 if (fixP->fx_addsy != NULL
8121 && (fixP->fx_r_type == BFD_RELOC_32_PCREL
8122 || fixP->fx_r_type == BFD_RELOC_64_PCREL
8123 || fixP->fx_r_type == BFD_RELOC_16_PCREL
8124 || fixP->fx_r_type == BFD_RELOC_8_PCREL)
8125 && !use_rela_relocations)
8126 {
8127 /* This is a hack. There should be a better way to handle this.
8128 This covers for the fact that bfd_install_relocation will
8129 subtract the current location (for partial_inplace, PC relative
8130 relocations); see more below. */
8131 #ifndef OBJ_AOUT
8132 if (IS_ELF
8133 #ifdef TE_PE
8134 || OUTPUT_FLAVOR == bfd_target_coff_flavour
8135 #endif
8136 )
8137 value += fixP->fx_where + fixP->fx_frag->fr_address;
8138 #endif
8139 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
8140 if (IS_ELF)
8141 {
8142 segT sym_seg = S_GET_SEGMENT (fixP->fx_addsy);
8143
8144 if ((sym_seg == seg
8145 || (symbol_section_p (fixP->fx_addsy)
8146 && sym_seg != absolute_section))
8147 && !generic_force_reloc (fixP))
8148 {
8149 /* Yes, we add the values in twice. This is because
8150 bfd_install_relocation subtracts them out again. I think
8151 bfd_install_relocation is broken, but I don't dare change
8152 it. FIXME. */
8153 value += fixP->fx_where + fixP->fx_frag->fr_address;
8154 }
8155 }
8156 #endif
8157 #if defined (OBJ_COFF) && defined (TE_PE)
8158 /* For some reason, the PE format does not store a
8159 section address offset for a PC relative symbol. */
8160 if (S_GET_SEGMENT (fixP->fx_addsy) != seg
8161 || S_IS_WEAK (fixP->fx_addsy))
8162 value += md_pcrel_from (fixP);
8163 #endif
8164 }
8165 #if defined (OBJ_COFF) && defined (TE_PE)
8166 if (fixP->fx_addsy != NULL && S_IS_WEAK (fixP->fx_addsy))
8167 {
8168 value -= S_GET_VALUE (fixP->fx_addsy);
8169 }
8170 #endif
8171
8172 /* Fix a few things - the dynamic linker expects certain values here,
8173 and we must not disappoint it. */
8174 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
8175 if (IS_ELF && fixP->fx_addsy)
8176 switch (fixP->fx_r_type)
8177 {
8178 case BFD_RELOC_386_PLT32:
8179 case BFD_RELOC_X86_64_PLT32:
8180 /* Make the jump instruction point to the address of the operand. At
8181 runtime we merely add the offset to the actual PLT entry. */
8182 value = -4;
8183 break;
8184
8185 case BFD_RELOC_386_TLS_GD:
8186 case BFD_RELOC_386_TLS_LDM:
8187 case BFD_RELOC_386_TLS_IE_32:
8188 case BFD_RELOC_386_TLS_IE:
8189 case BFD_RELOC_386_TLS_GOTIE:
8190 case BFD_RELOC_386_TLS_GOTDESC:
8191 case BFD_RELOC_X86_64_TLSGD:
8192 case BFD_RELOC_X86_64_TLSLD:
8193 case BFD_RELOC_X86_64_GOTTPOFF:
8194 case BFD_RELOC_X86_64_GOTPC32_TLSDESC:
8195 value = 0; /* Fully resolved at runtime. No addend. */
8196 /* Fallthrough */
8197 case BFD_RELOC_386_TLS_LE:
8198 case BFD_RELOC_386_TLS_LDO_32:
8199 case BFD_RELOC_386_TLS_LE_32:
8200 case BFD_RELOC_X86_64_DTPOFF32:
8201 case BFD_RELOC_X86_64_DTPOFF64:
8202 case BFD_RELOC_X86_64_TPOFF32:
8203 case BFD_RELOC_X86_64_TPOFF64:
8204 S_SET_THREAD_LOCAL (fixP->fx_addsy);
8205 break;
8206
8207 case BFD_RELOC_386_TLS_DESC_CALL:
8208 case BFD_RELOC_X86_64_TLSDESC_CALL:
8209 value = 0; /* Fully resolved at runtime. No addend. */
8210 S_SET_THREAD_LOCAL (fixP->fx_addsy);
8211 fixP->fx_done = 0;
8212 return;
8213
8214 case BFD_RELOC_386_GOT32:
8215 case BFD_RELOC_X86_64_GOT32:
8216 value = 0; /* Fully resolved at runtime. No addend. */
8217 break;
8218
8219 case BFD_RELOC_VTABLE_INHERIT:
8220 case BFD_RELOC_VTABLE_ENTRY:
8221 fixP->fx_done = 0;
8222 return;
8223
8224 default:
8225 break;
8226 }
8227 #endif /* defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) */
8228 *valP = value;
8229 #endif /* !defined (TE_Mach) */
8230
8231 /* Are we finished with this relocation now? */
8232 if (fixP->fx_addsy == NULL)
8233 fixP->fx_done = 1;
8234 #if defined (OBJ_COFF) && defined (TE_PE)
8235 else if (fixP->fx_addsy != NULL && S_IS_WEAK (fixP->fx_addsy))
8236 {
8237 fixP->fx_done = 0;
8238 /* Remember value for tc_gen_reloc. */
8239 fixP->fx_addnumber = value;
8240 /* Clear out the frag for now. */
8241 value = 0;
8242 }
8243 #endif
8244 else if (use_rela_relocations)
8245 {
8246 fixP->fx_no_overflow = 1;
8247 /* Remember value for tc_gen_reloc. */
8248 fixP->fx_addnumber = value;
8249 value = 0;
8250 }
8251
8252 md_number_to_chars (p, value, fixP->fx_size);
8253 }
8254 \f
8255 char *
8256 md_atof (int type, char *litP, int *sizeP)
8257 {
8258 /* This outputs the LITTLENUMs in REVERSE order;
8259 in accord with the bigendian 386. */
8260 return ieee_md_atof (type, litP, sizeP, FALSE);
8261 }
8262 \f
8263 static char output_invalid_buf[sizeof (unsigned char) * 2 + 6];
8264
8265 static char *
8266 output_invalid (int c)
8267 {
8268 if (ISPRINT (c))
8269 snprintf (output_invalid_buf, sizeof (output_invalid_buf),
8270 "'%c'", c);
8271 else
8272 snprintf (output_invalid_buf, sizeof (output_invalid_buf),
8273 "(0x%x)", (unsigned char) c);
8274 return output_invalid_buf;
8275 }
8276
8277 /* REG_STRING starts *before* REGISTER_PREFIX. */
8278
8279 static const reg_entry *
8280 parse_real_register (char *reg_string, char **end_op)
8281 {
8282 char *s = reg_string;
8283 char *p;
8284 char reg_name_given[MAX_REG_NAME_SIZE + 1];
8285 const reg_entry *r;
8286
8287 /* Skip possible REGISTER_PREFIX and possible whitespace. */
8288 if (*s == REGISTER_PREFIX)
8289 ++s;
8290
8291 if (is_space_char (*s))
8292 ++s;
8293
8294 p = reg_name_given;
8295 while ((*p++ = register_chars[(unsigned char) *s]) != '\0')
8296 {
8297 if (p >= reg_name_given + MAX_REG_NAME_SIZE)
8298 return (const reg_entry *) NULL;
8299 s++;
8300 }
8301
8302 /* For naked regs, make sure that we are not dealing with an identifier.
8303 This prevents confusing an identifier like `eax_var' with register
8304 `eax'. */
8305 if (allow_naked_reg && identifier_chars[(unsigned char) *s])
8306 return (const reg_entry *) NULL;
8307
8308 *end_op = s;
8309
8310 r = (const reg_entry *) hash_find (reg_hash, reg_name_given);
8311
8312 /* Handle floating point regs, allowing spaces in the (i) part. */
8313 if (r == i386_regtab /* %st is first entry of table */)
8314 {
8315 if (is_space_char (*s))
8316 ++s;
8317 if (*s == '(')
8318 {
8319 ++s;
8320 if (is_space_char (*s))
8321 ++s;
8322 if (*s >= '0' && *s <= '7')
8323 {
8324 int fpr = *s - '0';
8325 ++s;
8326 if (is_space_char (*s))
8327 ++s;
8328 if (*s == ')')
8329 {
8330 *end_op = s + 1;
8331 r = (const reg_entry *) hash_find (reg_hash, "st(0)");
8332 know (r);
8333 return r + fpr;
8334 }
8335 }
8336 /* We have "%st(" then garbage. */
8337 return (const reg_entry *) NULL;
8338 }
8339 }
8340
8341 if (r == NULL || allow_pseudo_reg)
8342 return r;
8343
8344 if (operand_type_all_zero (&r->reg_type))
8345 return (const reg_entry *) NULL;
8346
8347 if ((r->reg_type.bitfield.reg32
8348 || r->reg_type.bitfield.sreg3
8349 || r->reg_type.bitfield.control
8350 || r->reg_type.bitfield.debug
8351 || r->reg_type.bitfield.test)
8352 && !cpu_arch_flags.bitfield.cpui386)
8353 return (const reg_entry *) NULL;
8354
8355 if (r->reg_type.bitfield.floatreg
8356 && !cpu_arch_flags.bitfield.cpu8087
8357 && !cpu_arch_flags.bitfield.cpu287
8358 && !cpu_arch_flags.bitfield.cpu387)
8359 return (const reg_entry *) NULL;
8360
8361 if (r->reg_type.bitfield.regmmx && !cpu_arch_flags.bitfield.cpummx)
8362 return (const reg_entry *) NULL;
8363
8364 if (r->reg_type.bitfield.regxmm && !cpu_arch_flags.bitfield.cpusse)
8365 return (const reg_entry *) NULL;
8366
8367 if (r->reg_type.bitfield.regymm && !cpu_arch_flags.bitfield.cpuavx)
8368 return (const reg_entry *) NULL;
8369
8370 /* Don't allow fake index register unless allow_index_reg isn't 0. */
8371 if (!allow_index_reg
8372 && (r->reg_num == RegEiz || r->reg_num == RegRiz))
8373 return (const reg_entry *) NULL;
8374
8375 if (((r->reg_flags & (RegRex64 | RegRex))
8376 || r->reg_type.bitfield.reg64)
8377 && (!cpu_arch_flags.bitfield.cpulm
8378 || !operand_type_equal (&r->reg_type, &control))
8379 && flag_code != CODE_64BIT)
8380 return (const reg_entry *) NULL;
8381
8382 if (r->reg_type.bitfield.sreg3 && r->reg_num == RegFlat && !intel_syntax)
8383 return (const reg_entry *) NULL;
8384
8385 return r;
8386 }
8387
8388 /* REG_STRING starts *before* REGISTER_PREFIX. */
8389
8390 static const reg_entry *
8391 parse_register (char *reg_string, char **end_op)
8392 {
8393 const reg_entry *r;
8394
8395 if (*reg_string == REGISTER_PREFIX || allow_naked_reg)
8396 r = parse_real_register (reg_string, end_op);
8397 else
8398 r = NULL;
8399 if (!r)
8400 {
8401 char *save = input_line_pointer;
8402 char c;
8403 symbolS *symbolP;
8404
8405 input_line_pointer = reg_string;
8406 c = get_symbol_end ();
8407 symbolP = symbol_find (reg_string);
8408 if (symbolP && S_GET_SEGMENT (symbolP) == reg_section)
8409 {
8410 const expressionS *e = symbol_get_value_expression (symbolP);
8411
8412 know (e->X_op == O_register);
8413 know (e->X_add_number >= 0
8414 && (valueT) e->X_add_number < i386_regtab_size);
8415 r = i386_regtab + e->X_add_number;
8416 *end_op = input_line_pointer;
8417 }
8418 *input_line_pointer = c;
8419 input_line_pointer = save;
8420 }
8421 return r;
8422 }
8423
8424 int
8425 i386_parse_name (char *name, expressionS *e, char *nextcharP)
8426 {
8427 const reg_entry *r;
8428 char *end = input_line_pointer;
8429
8430 *end = *nextcharP;
8431 r = parse_register (name, &input_line_pointer);
8432 if (r && end <= input_line_pointer)
8433 {
8434 *nextcharP = *input_line_pointer;
8435 *input_line_pointer = 0;
8436 e->X_op = O_register;
8437 e->X_add_number = r - i386_regtab;
8438 return 1;
8439 }
8440 input_line_pointer = end;
8441 *end = 0;
8442 return intel_syntax ? i386_intel_parse_name (name, e) : 0;
8443 }
8444
8445 void
8446 md_operand (expressionS *e)
8447 {
8448 char *end;
8449 const reg_entry *r;
8450
8451 switch (*input_line_pointer)
8452 {
8453 case REGISTER_PREFIX:
8454 r = parse_real_register (input_line_pointer, &end);
8455 if (r)
8456 {
8457 e->X_op = O_register;
8458 e->X_add_number = r - i386_regtab;
8459 input_line_pointer = end;
8460 }
8461 break;
8462
8463 case '[':
8464 gas_assert (intel_syntax);
8465 end = input_line_pointer++;
8466 expression (e);
8467 if (*input_line_pointer == ']')
8468 {
8469 ++input_line_pointer;
8470 e->X_op_symbol = make_expr_symbol (e);
8471 e->X_add_symbol = NULL;
8472 e->X_add_number = 0;
8473 e->X_op = O_index;
8474 }
8475 else
8476 {
8477 e->X_op = O_absent;
8478 input_line_pointer = end;
8479 }
8480 break;
8481 }
8482 }
8483
8484 \f
8485 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
8486 const char *md_shortopts = "kVQ:sqn";
8487 #else
8488 const char *md_shortopts = "qn";
8489 #endif
8490
8491 #define OPTION_32 (OPTION_MD_BASE + 0)
8492 #define OPTION_64 (OPTION_MD_BASE + 1)
8493 #define OPTION_DIVIDE (OPTION_MD_BASE + 2)
8494 #define OPTION_MARCH (OPTION_MD_BASE + 3)
8495 #define OPTION_MTUNE (OPTION_MD_BASE + 4)
8496 #define OPTION_MMNEMONIC (OPTION_MD_BASE + 5)
8497 #define OPTION_MSYNTAX (OPTION_MD_BASE + 6)
8498 #define OPTION_MINDEX_REG (OPTION_MD_BASE + 7)
8499 #define OPTION_MNAKED_REG (OPTION_MD_BASE + 8)
8500 #define OPTION_MOLD_GCC (OPTION_MD_BASE + 9)
8501 #define OPTION_MSSE2AVX (OPTION_MD_BASE + 10)
8502 #define OPTION_MSSE_CHECK (OPTION_MD_BASE + 11)
8503 #define OPTION_MAVXSCALAR (OPTION_MD_BASE + 12)
8504 #define OPTION_X32 (OPTION_MD_BASE + 13)
8505
8506 struct option md_longopts[] =
8507 {
8508 {"32", no_argument, NULL, OPTION_32},
8509 #if (defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) \
8510 || defined (TE_PE) || defined (TE_PEP) || defined (OBJ_MACH_O))
8511 {"64", no_argument, NULL, OPTION_64},
8512 #endif
8513 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
8514 {"x32", no_argument, NULL, OPTION_X32},
8515 #endif
8516 {"divide", no_argument, NULL, OPTION_DIVIDE},
8517 {"march", required_argument, NULL, OPTION_MARCH},
8518 {"mtune", required_argument, NULL, OPTION_MTUNE},
8519 {"mmnemonic", required_argument, NULL, OPTION_MMNEMONIC},
8520 {"msyntax", required_argument, NULL, OPTION_MSYNTAX},
8521 {"mindex-reg", no_argument, NULL, OPTION_MINDEX_REG},
8522 {"mnaked-reg", no_argument, NULL, OPTION_MNAKED_REG},
8523 {"mold-gcc", no_argument, NULL, OPTION_MOLD_GCC},
8524 {"msse2avx", no_argument, NULL, OPTION_MSSE2AVX},
8525 {"msse-check", required_argument, NULL, OPTION_MSSE_CHECK},
8526 {"mavxscalar", required_argument, NULL, OPTION_MAVXSCALAR},
8527 {NULL, no_argument, NULL, 0}
8528 };
8529 size_t md_longopts_size = sizeof (md_longopts);
8530
8531 int
8532 md_parse_option (int c, char *arg)
8533 {
8534 unsigned int j;
8535 char *arch, *next;
8536
8537 switch (c)
8538 {
8539 case 'n':
8540 optimize_align_code = 0;
8541 break;
8542
8543 case 'q':
8544 quiet_warnings = 1;
8545 break;
8546
8547 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
8548 /* -Qy, -Qn: SVR4 arguments controlling whether a .comment section
8549 should be emitted or not. FIXME: Not implemented. */
8550 case 'Q':
8551 break;
8552
8553 /* -V: SVR4 argument to print version ID. */
8554 case 'V':
8555 print_version_id ();
8556 break;
8557
8558 /* -k: Ignore for FreeBSD compatibility. */
8559 case 'k':
8560 break;
8561
8562 case 's':
8563 /* -s: On i386 Solaris, this tells the native assembler to use
8564 .stab instead of .stab.excl. We always use .stab anyhow. */
8565 break;
8566 #endif
8567 #if (defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) \
8568 || defined (TE_PE) || defined (TE_PEP) || defined (OBJ_MACH_O))
8569 case OPTION_64:
8570 {
8571 const char **list, **l;
8572
8573 list = bfd_target_list ();
8574 for (l = list; *l != NULL; l++)
8575 if (CONST_STRNEQ (*l, "elf64-x86-64")
8576 || strcmp (*l, "coff-x86-64") == 0
8577 || strcmp (*l, "pe-x86-64") == 0
8578 || strcmp (*l, "pei-x86-64") == 0
8579 || strcmp (*l, "mach-o-x86-64") == 0)
8580 {
8581 default_arch = "x86_64";
8582 break;
8583 }
8584 if (*l == NULL)
8585 as_fatal (_("no compiled in support for x86_64"));
8586 free (list);
8587 }
8588 break;
8589 #endif
8590
8591 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
8592 case OPTION_X32:
8593 if (IS_ELF)
8594 {
8595 const char **list, **l;
8596
8597 list = bfd_target_list ();
8598 for (l = list; *l != NULL; l++)
8599 if (CONST_STRNEQ (*l, "elf32-x86-64"))
8600 {
8601 default_arch = "x86_64:32";
8602 break;
8603 }
8604 if (*l == NULL)
8605 as_fatal (_("no compiled in support for 32bit x86_64"));
8606 free (list);
8607 }
8608 else
8609 as_fatal (_("32bit x86_64 is only supported for ELF"));
8610 break;
8611 #endif
8612
8613 case OPTION_32:
8614 default_arch = "i386";
8615 break;
8616
8617 case OPTION_DIVIDE:
8618 #ifdef SVR4_COMMENT_CHARS
8619 {
8620 char *n, *t;
8621 const char *s;
8622
8623 n = (char *) xmalloc (strlen (i386_comment_chars) + 1);
8624 t = n;
8625 for (s = i386_comment_chars; *s != '\0'; s++)
8626 if (*s != '/')
8627 *t++ = *s;
8628 *t = '\0';
8629 i386_comment_chars = n;
8630 }
8631 #endif
8632 break;
8633
8634 case OPTION_MARCH:
8635 arch = xstrdup (arg);
8636 do
8637 {
8638 if (*arch == '.')
8639 as_fatal (_("invalid -march= option: `%s'"), arg);
8640 next = strchr (arch, '+');
8641 if (next)
8642 *next++ = '\0';
8643 for (j = 0; j < ARRAY_SIZE (cpu_arch); j++)
8644 {
8645 if (strcmp (arch, cpu_arch [j].name) == 0)
8646 {
8647 /* Processor. */
8648 if (! cpu_arch[j].flags.bitfield.cpui386)
8649 continue;
8650
8651 cpu_arch_name = cpu_arch[j].name;
8652 cpu_sub_arch_name = NULL;
8653 cpu_arch_flags = cpu_arch[j].flags;
8654 cpu_arch_isa = cpu_arch[j].type;
8655 cpu_arch_isa_flags = cpu_arch[j].flags;
8656 if (!cpu_arch_tune_set)
8657 {
8658 cpu_arch_tune = cpu_arch_isa;
8659 cpu_arch_tune_flags = cpu_arch_isa_flags;
8660 }
8661 break;
8662 }
8663 else if (*cpu_arch [j].name == '.'
8664 && strcmp (arch, cpu_arch [j].name + 1) == 0)
8665 {
8666 /* ISA entension. */
8667 i386_cpu_flags flags;
8668
8669 if (!cpu_arch[j].negated)
8670 flags = cpu_flags_or (cpu_arch_flags,
8671 cpu_arch[j].flags);
8672 else
8673 flags = cpu_flags_and_not (cpu_arch_flags,
8674 cpu_arch[j].flags);
8675 if (!cpu_flags_equal (&flags, &cpu_arch_flags))
8676 {
8677 if (cpu_sub_arch_name)
8678 {
8679 char *name = cpu_sub_arch_name;
8680 cpu_sub_arch_name = concat (name,
8681 cpu_arch[j].name,
8682 (const char *) NULL);
8683 free (name);
8684 }
8685 else
8686 cpu_sub_arch_name = xstrdup (cpu_arch[j].name);
8687 cpu_arch_flags = flags;
8688 cpu_arch_isa_flags = flags;
8689 }
8690 break;
8691 }
8692 }
8693
8694 if (j >= ARRAY_SIZE (cpu_arch))
8695 as_fatal (_("invalid -march= option: `%s'"), arg);
8696
8697 arch = next;
8698 }
8699 while (next != NULL );
8700 break;
8701
8702 case OPTION_MTUNE:
8703 if (*arg == '.')
8704 as_fatal (_("invalid -mtune= option: `%s'"), arg);
8705 for (j = 0; j < ARRAY_SIZE (cpu_arch); j++)
8706 {
8707 if (strcmp (arg, cpu_arch [j].name) == 0)
8708 {
8709 cpu_arch_tune_set = 1;
8710 cpu_arch_tune = cpu_arch [j].type;
8711 cpu_arch_tune_flags = cpu_arch[j].flags;
8712 break;
8713 }
8714 }
8715 if (j >= ARRAY_SIZE (cpu_arch))
8716 as_fatal (_("invalid -mtune= option: `%s'"), arg);
8717 break;
8718
8719 case OPTION_MMNEMONIC:
8720 if (strcasecmp (arg, "att") == 0)
8721 intel_mnemonic = 0;
8722 else if (strcasecmp (arg, "intel") == 0)
8723 intel_mnemonic = 1;
8724 else
8725 as_fatal (_("invalid -mmnemonic= option: `%s'"), arg);
8726 break;
8727
8728 case OPTION_MSYNTAX:
8729 if (strcasecmp (arg, "att") == 0)
8730 intel_syntax = 0;
8731 else if (strcasecmp (arg, "intel") == 0)
8732 intel_syntax = 1;
8733 else
8734 as_fatal (_("invalid -msyntax= option: `%s'"), arg);
8735 break;
8736
8737 case OPTION_MINDEX_REG:
8738 allow_index_reg = 1;
8739 break;
8740
8741 case OPTION_MNAKED_REG:
8742 allow_naked_reg = 1;
8743 break;
8744
8745 case OPTION_MOLD_GCC:
8746 old_gcc = 1;
8747 break;
8748
8749 case OPTION_MSSE2AVX:
8750 sse2avx = 1;
8751 break;
8752
8753 case OPTION_MSSE_CHECK:
8754 if (strcasecmp (arg, "error") == 0)
8755 sse_check = sse_check_error;
8756 else if (strcasecmp (arg, "warning") == 0)
8757 sse_check = sse_check_warning;
8758 else if (strcasecmp (arg, "none") == 0)
8759 sse_check = sse_check_none;
8760 else
8761 as_fatal (_("invalid -msse-check= option: `%s'"), arg);
8762 break;
8763
8764 case OPTION_MAVXSCALAR:
8765 if (strcasecmp (arg, "128") == 0)
8766 avxscalar = vex128;
8767 else if (strcasecmp (arg, "256") == 0)
8768 avxscalar = vex256;
8769 else
8770 as_fatal (_("invalid -mavxscalar= option: `%s'"), arg);
8771 break;
8772
8773 default:
8774 return 0;
8775 }
8776 return 1;
8777 }
8778
8779 #define MESSAGE_TEMPLATE \
8780 " "
8781
8782 static void
8783 show_arch (FILE *stream, int ext, int check)
8784 {
8785 static char message[] = MESSAGE_TEMPLATE;
8786 char *start = message + 27;
8787 char *p;
8788 int size = sizeof (MESSAGE_TEMPLATE);
8789 int left;
8790 const char *name;
8791 int len;
8792 unsigned int j;
8793
8794 p = start;
8795 left = size - (start - message);
8796 for (j = 0; j < ARRAY_SIZE (cpu_arch); j++)
8797 {
8798 /* Should it be skipped? */
8799 if (cpu_arch [j].skip)
8800 continue;
8801
8802 name = cpu_arch [j].name;
8803 len = cpu_arch [j].len;
8804 if (*name == '.')
8805 {
8806 /* It is an extension. Skip if we aren't asked to show it. */
8807 if (ext)
8808 {
8809 name++;
8810 len--;
8811 }
8812 else
8813 continue;
8814 }
8815 else if (ext)
8816 {
8817 /* It is an processor. Skip if we show only extension. */
8818 continue;
8819 }
8820 else if (check && ! cpu_arch[j].flags.bitfield.cpui386)
8821 {
8822 /* It is an impossible processor - skip. */
8823 continue;
8824 }
8825
8826 /* Reserve 2 spaces for ", " or ",\0" */
8827 left -= len + 2;
8828
8829 /* Check if there is any room. */
8830 if (left >= 0)
8831 {
8832 if (p != start)
8833 {
8834 *p++ = ',';
8835 *p++ = ' ';
8836 }
8837 p = mempcpy (p, name, len);
8838 }
8839 else
8840 {
8841 /* Output the current message now and start a new one. */
8842 *p++ = ',';
8843 *p = '\0';
8844 fprintf (stream, "%s\n", message);
8845 p = start;
8846 left = size - (start - message) - len - 2;
8847
8848 gas_assert (left >= 0);
8849
8850 p = mempcpy (p, name, len);
8851 }
8852 }
8853
8854 *p = '\0';
8855 fprintf (stream, "%s\n", message);
8856 }
8857
8858 void
8859 md_show_usage (FILE *stream)
8860 {
8861 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
8862 fprintf (stream, _("\
8863 -Q ignored\n\
8864 -V print assembler version number\n\
8865 -k ignored\n"));
8866 #endif
8867 fprintf (stream, _("\
8868 -n Do not optimize code alignment\n\
8869 -q quieten some warnings\n"));
8870 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
8871 fprintf (stream, _("\
8872 -s ignored\n"));
8873 #endif
8874 #if (defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) \
8875 || defined (TE_PE) || defined (TE_PEP))
8876 fprintf (stream, _("\
8877 --32/--64/--x32 generate 32bit/64bit/x32 code\n"));
8878 #endif
8879 #ifdef SVR4_COMMENT_CHARS
8880 fprintf (stream, _("\
8881 --divide do not treat `/' as a comment character\n"));
8882 #else
8883 fprintf (stream, _("\
8884 --divide ignored\n"));
8885 #endif
8886 fprintf (stream, _("\
8887 -march=CPU[,+EXTENSION...]\n\
8888 generate code for CPU and EXTENSION, CPU is one of:\n"));
8889 show_arch (stream, 0, 1);
8890 fprintf (stream, _("\
8891 EXTENSION is combination of:\n"));
8892 show_arch (stream, 1, 0);
8893 fprintf (stream, _("\
8894 -mtune=CPU optimize for CPU, CPU is one of:\n"));
8895 show_arch (stream, 0, 0);
8896 fprintf (stream, _("\
8897 -msse2avx encode SSE instructions with VEX prefix\n"));
8898 fprintf (stream, _("\
8899 -msse-check=[none|error|warning]\n\
8900 check SSE instructions\n"));
8901 fprintf (stream, _("\
8902 -mavxscalar=[128|256] encode scalar AVX instructions with specific vector\n\
8903 length\n"));
8904 fprintf (stream, _("\
8905 -mmnemonic=[att|intel] use AT&T/Intel mnemonic\n"));
8906 fprintf (stream, _("\
8907 -msyntax=[att|intel] use AT&T/Intel syntax\n"));
8908 fprintf (stream, _("\
8909 -mindex-reg support pseudo index registers\n"));
8910 fprintf (stream, _("\
8911 -mnaked-reg don't require `%%' prefix for registers\n"));
8912 fprintf (stream, _("\
8913 -mold-gcc support old (<= 2.8.1) versions of gcc\n"));
8914 }
8915
8916 #if ((defined (OBJ_MAYBE_COFF) && defined (OBJ_MAYBE_AOUT)) \
8917 || defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) \
8918 || defined (TE_PE) || defined (TE_PEP) || defined (OBJ_MACH_O))
8919
8920 /* Pick the target format to use. */
8921
8922 const char *
8923 i386_target_format (void)
8924 {
8925 if (!strncmp (default_arch, "x86_64", 6))
8926 {
8927 update_code_flag (CODE_64BIT, 1);
8928 if (default_arch[6] == '\0')
8929 x86_elf_abi = X86_64_ABI;
8930 else
8931 x86_elf_abi = X86_64_X32_ABI;
8932 }
8933 else if (!strcmp (default_arch, "i386"))
8934 update_code_flag (CODE_32BIT, 1);
8935 else
8936 as_fatal (_("unknown architecture"));
8937
8938 if (cpu_flags_all_zero (&cpu_arch_isa_flags))
8939 cpu_arch_isa_flags = cpu_arch[flag_code == CODE_64BIT].flags;
8940 if (cpu_flags_all_zero (&cpu_arch_tune_flags))
8941 cpu_arch_tune_flags = cpu_arch[flag_code == CODE_64BIT].flags;
8942
8943 switch (OUTPUT_FLAVOR)
8944 {
8945 #if defined (OBJ_MAYBE_AOUT) || defined (OBJ_AOUT)
8946 case bfd_target_aout_flavour:
8947 return AOUT_TARGET_FORMAT;
8948 #endif
8949 #if defined (OBJ_MAYBE_COFF) || defined (OBJ_COFF)
8950 # if defined (TE_PE) || defined (TE_PEP)
8951 case bfd_target_coff_flavour:
8952 return flag_code == CODE_64BIT ? "pe-x86-64" : "pe-i386";
8953 # elif defined (TE_GO32)
8954 case bfd_target_coff_flavour:
8955 return "coff-go32";
8956 # else
8957 case bfd_target_coff_flavour:
8958 return "coff-i386";
8959 # endif
8960 #endif
8961 #if defined (OBJ_MAYBE_ELF) || defined (OBJ_ELF)
8962 case bfd_target_elf_flavour:
8963 {
8964 const char *format;
8965
8966 switch (x86_elf_abi)
8967 {
8968 default:
8969 format = ELF_TARGET_FORMAT;
8970 break;
8971 case X86_64_ABI:
8972 use_rela_relocations = 1;
8973 object_64bit = 1;
8974 format = ELF_TARGET_FORMAT64;
8975 break;
8976 case X86_64_X32_ABI:
8977 use_rela_relocations = 1;
8978 object_64bit = 1;
8979 disallow_64bit_reloc = 1;
8980 format = ELF_TARGET_FORMAT32;
8981 break;
8982 }
8983 if (cpu_arch_isa == PROCESSOR_L1OM)
8984 {
8985 if (x86_elf_abi != X86_64_ABI)
8986 as_fatal (_("Intel L1OM is 64bit only"));
8987 return ELF_TARGET_L1OM_FORMAT;
8988 }
8989 if (cpu_arch_isa == PROCESSOR_K1OM)
8990 {
8991 if (x86_elf_abi != X86_64_ABI)
8992 as_fatal (_("Intel K1OM is 64bit only"));
8993 return ELF_TARGET_K1OM_FORMAT;
8994 }
8995 else
8996 return format;
8997 }
8998 #endif
8999 #if defined (OBJ_MACH_O)
9000 case bfd_target_mach_o_flavour:
9001 if (flag_code == CODE_64BIT)
9002 {
9003 use_rela_relocations = 1;
9004 object_64bit = 1;
9005 return "mach-o-x86-64";
9006 }
9007 else
9008 return "mach-o-i386";
9009 #endif
9010 default:
9011 abort ();
9012 return NULL;
9013 }
9014 }
9015
9016 #endif /* OBJ_MAYBE_ more than one */
9017
9018 #if (defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF))
9019 void
9020 i386_elf_emit_arch_note (void)
9021 {
9022 if (IS_ELF && cpu_arch_name != NULL)
9023 {
9024 char *p;
9025 asection *seg = now_seg;
9026 subsegT subseg = now_subseg;
9027 Elf_Internal_Note i_note;
9028 Elf_External_Note e_note;
9029 asection *note_secp;
9030 int len;
9031
9032 /* Create the .note section. */
9033 note_secp = subseg_new (".note", 0);
9034 bfd_set_section_flags (stdoutput,
9035 note_secp,
9036 SEC_HAS_CONTENTS | SEC_READONLY);
9037
9038 /* Process the arch string. */
9039 len = strlen (cpu_arch_name);
9040
9041 i_note.namesz = len + 1;
9042 i_note.descsz = 0;
9043 i_note.type = NT_ARCH;
9044 p = frag_more (sizeof (e_note.namesz));
9045 md_number_to_chars (p, (valueT) i_note.namesz, sizeof (e_note.namesz));
9046 p = frag_more (sizeof (e_note.descsz));
9047 md_number_to_chars (p, (valueT) i_note.descsz, sizeof (e_note.descsz));
9048 p = frag_more (sizeof (e_note.type));
9049 md_number_to_chars (p, (valueT) i_note.type, sizeof (e_note.type));
9050 p = frag_more (len + 1);
9051 strcpy (p, cpu_arch_name);
9052
9053 frag_align (2, 0, 0);
9054
9055 subseg_set (seg, subseg);
9056 }
9057 }
9058 #endif
9059 \f
9060 symbolS *
9061 md_undefined_symbol (char *name)
9062 {
9063 if (name[0] == GLOBAL_OFFSET_TABLE_NAME[0]
9064 && name[1] == GLOBAL_OFFSET_TABLE_NAME[1]
9065 && name[2] == GLOBAL_OFFSET_TABLE_NAME[2]
9066 && strcmp (name, GLOBAL_OFFSET_TABLE_NAME) == 0)
9067 {
9068 if (!GOT_symbol)
9069 {
9070 if (symbol_find (name))
9071 as_bad (_("GOT already in symbol table"));
9072 GOT_symbol = symbol_new (name, undefined_section,
9073 (valueT) 0, &zero_address_frag);
9074 };
9075 return GOT_symbol;
9076 }
9077 return 0;
9078 }
9079
9080 /* Round up a section size to the appropriate boundary. */
9081
9082 valueT
9083 md_section_align (segT segment ATTRIBUTE_UNUSED, valueT size)
9084 {
9085 #if (defined (OBJ_AOUT) || defined (OBJ_MAYBE_AOUT))
9086 if (OUTPUT_FLAVOR == bfd_target_aout_flavour)
9087 {
9088 /* For a.out, force the section size to be aligned. If we don't do
9089 this, BFD will align it for us, but it will not write out the
9090 final bytes of the section. This may be a bug in BFD, but it is
9091 easier to fix it here since that is how the other a.out targets
9092 work. */
9093 int align;
9094
9095 align = bfd_get_section_alignment (stdoutput, segment);
9096 size = ((size + (1 << align) - 1) & ((valueT) -1 << align));
9097 }
9098 #endif
9099
9100 return size;
9101 }
9102
9103 /* On the i386, PC-relative offsets are relative to the start of the
9104 next instruction. That is, the address of the offset, plus its
9105 size, since the offset is always the last part of the insn. */
9106
9107 long
9108 md_pcrel_from (fixS *fixP)
9109 {
9110 return fixP->fx_size + fixP->fx_where + fixP->fx_frag->fr_address;
9111 }
9112
9113 #ifndef I386COFF
9114
9115 static void
9116 s_bss (int ignore ATTRIBUTE_UNUSED)
9117 {
9118 int temp;
9119
9120 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
9121 if (IS_ELF)
9122 obj_elf_section_change_hook ();
9123 #endif
9124 temp = get_absolute_expression ();
9125 subseg_set (bss_section, (subsegT) temp);
9126 demand_empty_rest_of_line ();
9127 }
9128
9129 #endif
9130
9131 void
9132 i386_validate_fix (fixS *fixp)
9133 {
9134 if (fixp->fx_subsy && fixp->fx_subsy == GOT_symbol)
9135 {
9136 if (fixp->fx_r_type == BFD_RELOC_32_PCREL)
9137 {
9138 if (!object_64bit)
9139 abort ();
9140 fixp->fx_r_type = BFD_RELOC_X86_64_GOTPCREL;
9141 }
9142 else
9143 {
9144 if (!object_64bit)
9145 fixp->fx_r_type = BFD_RELOC_386_GOTOFF;
9146 else
9147 fixp->fx_r_type = BFD_RELOC_X86_64_GOTOFF64;
9148 }
9149 fixp->fx_subsy = 0;
9150 }
9151 }
9152
9153 arelent *
9154 tc_gen_reloc (asection *section ATTRIBUTE_UNUSED, fixS *fixp)
9155 {
9156 arelent *rel;
9157 bfd_reloc_code_real_type code;
9158
9159 switch (fixp->fx_r_type)
9160 {
9161 case BFD_RELOC_X86_64_PLT32:
9162 case BFD_RELOC_X86_64_GOT32:
9163 case BFD_RELOC_X86_64_GOTPCREL:
9164 case BFD_RELOC_386_PLT32:
9165 case BFD_RELOC_386_GOT32:
9166 case BFD_RELOC_386_GOTOFF:
9167 case BFD_RELOC_386_GOTPC:
9168 case BFD_RELOC_386_TLS_GD:
9169 case BFD_RELOC_386_TLS_LDM:
9170 case BFD_RELOC_386_TLS_LDO_32:
9171 case BFD_RELOC_386_TLS_IE_32:
9172 case BFD_RELOC_386_TLS_IE:
9173 case BFD_RELOC_386_TLS_GOTIE:
9174 case BFD_RELOC_386_TLS_LE_32:
9175 case BFD_RELOC_386_TLS_LE:
9176 case BFD_RELOC_386_TLS_GOTDESC:
9177 case BFD_RELOC_386_TLS_DESC_CALL:
9178 case BFD_RELOC_X86_64_TLSGD:
9179 case BFD_RELOC_X86_64_TLSLD:
9180 case BFD_RELOC_X86_64_DTPOFF32:
9181 case BFD_RELOC_X86_64_DTPOFF64:
9182 case BFD_RELOC_X86_64_GOTTPOFF:
9183 case BFD_RELOC_X86_64_TPOFF32:
9184 case BFD_RELOC_X86_64_TPOFF64:
9185 case BFD_RELOC_X86_64_GOTOFF64:
9186 case BFD_RELOC_X86_64_GOTPC32:
9187 case BFD_RELOC_X86_64_GOT64:
9188 case BFD_RELOC_X86_64_GOTPCREL64:
9189 case BFD_RELOC_X86_64_GOTPC64:
9190 case BFD_RELOC_X86_64_GOTPLT64:
9191 case BFD_RELOC_X86_64_PLTOFF64:
9192 case BFD_RELOC_X86_64_GOTPC32_TLSDESC:
9193 case BFD_RELOC_X86_64_TLSDESC_CALL:
9194 case BFD_RELOC_RVA:
9195 case BFD_RELOC_VTABLE_ENTRY:
9196 case BFD_RELOC_VTABLE_INHERIT:
9197 #ifdef TE_PE
9198 case BFD_RELOC_32_SECREL:
9199 #endif
9200 code = fixp->fx_r_type;
9201 break;
9202 case BFD_RELOC_X86_64_32S:
9203 if (!fixp->fx_pcrel)
9204 {
9205 /* Don't turn BFD_RELOC_X86_64_32S into BFD_RELOC_32. */
9206 code = fixp->fx_r_type;
9207 break;
9208 }
9209 default:
9210 if (fixp->fx_pcrel)
9211 {
9212 switch (fixp->fx_size)
9213 {
9214 default:
9215 as_bad_where (fixp->fx_file, fixp->fx_line,
9216 _("can not do %d byte pc-relative relocation"),
9217 fixp->fx_size);
9218 code = BFD_RELOC_32_PCREL;
9219 break;
9220 case 1: code = BFD_RELOC_8_PCREL; break;
9221 case 2: code = BFD_RELOC_16_PCREL; break;
9222 case 4: code = BFD_RELOC_32_PCREL; break;
9223 #ifdef BFD64
9224 case 8: code = BFD_RELOC_64_PCREL; break;
9225 #endif
9226 }
9227 }
9228 else
9229 {
9230 switch (fixp->fx_size)
9231 {
9232 default:
9233 as_bad_where (fixp->fx_file, fixp->fx_line,
9234 _("can not do %d byte relocation"),
9235 fixp->fx_size);
9236 code = BFD_RELOC_32;
9237 break;
9238 case 1: code = BFD_RELOC_8; break;
9239 case 2: code = BFD_RELOC_16; break;
9240 case 4: code = BFD_RELOC_32; break;
9241 #ifdef BFD64
9242 case 8: code = BFD_RELOC_64; break;
9243 #endif
9244 }
9245 }
9246 break;
9247 }
9248
9249 if ((code == BFD_RELOC_32
9250 || code == BFD_RELOC_32_PCREL
9251 || code == BFD_RELOC_X86_64_32S)
9252 && GOT_symbol
9253 && fixp->fx_addsy == GOT_symbol)
9254 {
9255 if (!object_64bit)
9256 code = BFD_RELOC_386_GOTPC;
9257 else
9258 code = BFD_RELOC_X86_64_GOTPC32;
9259 }
9260 if ((code == BFD_RELOC_64 || code == BFD_RELOC_64_PCREL)
9261 && GOT_symbol
9262 && fixp->fx_addsy == GOT_symbol)
9263 {
9264 code = BFD_RELOC_X86_64_GOTPC64;
9265 }
9266
9267 rel = (arelent *) xmalloc (sizeof (arelent));
9268 rel->sym_ptr_ptr = (asymbol **) xmalloc (sizeof (asymbol *));
9269 *rel->sym_ptr_ptr = symbol_get_bfdsym (fixp->fx_addsy);
9270
9271 rel->address = fixp->fx_frag->fr_address + fixp->fx_where;
9272
9273 if (!use_rela_relocations)
9274 {
9275 /* HACK: Since i386 ELF uses Rel instead of Rela, encode the
9276 vtable entry to be used in the relocation's section offset. */
9277 if (fixp->fx_r_type == BFD_RELOC_VTABLE_ENTRY)
9278 rel->address = fixp->fx_offset;
9279 #if defined (OBJ_COFF) && defined (TE_PE)
9280 else if (fixp->fx_addsy && S_IS_WEAK (fixp->fx_addsy))
9281 rel->addend = fixp->fx_addnumber - (S_GET_VALUE (fixp->fx_addsy) * 2);
9282 else
9283 #endif
9284 rel->addend = 0;
9285 }
9286 /* Use the rela in 64bit mode. */
9287 else
9288 {
9289 if (disallow_64bit_reloc)
9290 switch (code)
9291 {
9292 case BFD_RELOC_X86_64_DTPOFF64:
9293 case BFD_RELOC_X86_64_TPOFF64:
9294 case BFD_RELOC_64_PCREL:
9295 case BFD_RELOC_X86_64_GOTOFF64:
9296 case BFD_RELOC_X86_64_GOT64:
9297 case BFD_RELOC_X86_64_GOTPCREL64:
9298 case BFD_RELOC_X86_64_GOTPC64:
9299 case BFD_RELOC_X86_64_GOTPLT64:
9300 case BFD_RELOC_X86_64_PLTOFF64:
9301 as_bad_where (fixp->fx_file, fixp->fx_line,
9302 _("cannot represent relocation type %s in x32 mode"),
9303 bfd_get_reloc_code_name (code));
9304 break;
9305 default:
9306 break;
9307 }
9308
9309 if (!fixp->fx_pcrel)
9310 rel->addend = fixp->fx_offset;
9311 else
9312 switch (code)
9313 {
9314 case BFD_RELOC_X86_64_PLT32:
9315 case BFD_RELOC_X86_64_GOT32:
9316 case BFD_RELOC_X86_64_GOTPCREL:
9317 case BFD_RELOC_X86_64_TLSGD:
9318 case BFD_RELOC_X86_64_TLSLD:
9319 case BFD_RELOC_X86_64_GOTTPOFF:
9320 case BFD_RELOC_X86_64_GOTPC32_TLSDESC:
9321 case BFD_RELOC_X86_64_TLSDESC_CALL:
9322 rel->addend = fixp->fx_offset - fixp->fx_size;
9323 break;
9324 default:
9325 rel->addend = (section->vma
9326 - fixp->fx_size
9327 + fixp->fx_addnumber
9328 + md_pcrel_from (fixp));
9329 break;
9330 }
9331 }
9332
9333 rel->howto = bfd_reloc_type_lookup (stdoutput, code);
9334 if (rel->howto == NULL)
9335 {
9336 as_bad_where (fixp->fx_file, fixp->fx_line,
9337 _("cannot represent relocation type %s"),
9338 bfd_get_reloc_code_name (code));
9339 /* Set howto to a garbage value so that we can keep going. */
9340 rel->howto = bfd_reloc_type_lookup (stdoutput, BFD_RELOC_32);
9341 gas_assert (rel->howto != NULL);
9342 }
9343
9344 return rel;
9345 }
9346
9347 #include "tc-i386-intel.c"
9348
9349 void
9350 tc_x86_parse_to_dw2regnum (expressionS *exp)
9351 {
9352 int saved_naked_reg;
9353 char saved_register_dot;
9354
9355 saved_naked_reg = allow_naked_reg;
9356 allow_naked_reg = 1;
9357 saved_register_dot = register_chars['.'];
9358 register_chars['.'] = '.';
9359 allow_pseudo_reg = 1;
9360 expression_and_evaluate (exp);
9361 allow_pseudo_reg = 0;
9362 register_chars['.'] = saved_register_dot;
9363 allow_naked_reg = saved_naked_reg;
9364
9365 if (exp->X_op == O_register && exp->X_add_number >= 0)
9366 {
9367 if ((addressT) exp->X_add_number < i386_regtab_size)
9368 {
9369 exp->X_op = O_constant;
9370 exp->X_add_number = i386_regtab[exp->X_add_number]
9371 .dw2_regnum[flag_code >> 1];
9372 }
9373 else
9374 exp->X_op = O_illegal;
9375 }
9376 }
9377
9378 void
9379 tc_x86_frame_initial_instructions (void)
9380 {
9381 static unsigned int sp_regno[2];
9382
9383 if (!sp_regno[flag_code >> 1])
9384 {
9385 char *saved_input = input_line_pointer;
9386 char sp[][4] = {"esp", "rsp"};
9387 expressionS exp;
9388
9389 input_line_pointer = sp[flag_code >> 1];
9390 tc_x86_parse_to_dw2regnum (&exp);
9391 gas_assert (exp.X_op == O_constant);
9392 sp_regno[flag_code >> 1] = exp.X_add_number;
9393 input_line_pointer = saved_input;
9394 }
9395
9396 cfi_add_CFA_def_cfa (sp_regno[flag_code >> 1], -x86_cie_data_alignment);
9397 cfi_add_CFA_offset (x86_dwarf2_return_column, x86_cie_data_alignment);
9398 }
9399
9400 int
9401 x86_dwarf2_addr_size (void)
9402 {
9403 #if defined (OBJ_MAYBE_ELF) || defined (OBJ_ELF)
9404 if (x86_elf_abi == X86_64_X32_ABI)
9405 return 4;
9406 #endif
9407 return bfd_arch_bits_per_address (stdoutput) / 8;
9408 }
9409
9410 int
9411 i386_elf_section_type (const char *str, size_t len)
9412 {
9413 if (flag_code == CODE_64BIT
9414 && len == sizeof ("unwind") - 1
9415 && strncmp (str, "unwind", 6) == 0)
9416 return SHT_X86_64_UNWIND;
9417
9418 return -1;
9419 }
9420
9421 #ifdef TE_SOLARIS
9422 void
9423 i386_solaris_fix_up_eh_frame (segT sec)
9424 {
9425 if (flag_code == CODE_64BIT)
9426 elf_section_type (sec) = SHT_X86_64_UNWIND;
9427 }
9428 #endif
9429
9430 #ifdef TE_PE
9431 void
9432 tc_pe_dwarf2_emit_offset (symbolS *symbol, unsigned int size)
9433 {
9434 expressionS exp;
9435
9436 exp.X_op = O_secrel;
9437 exp.X_add_symbol = symbol;
9438 exp.X_add_number = 0;
9439 emit_expr (&exp, size);
9440 }
9441 #endif
9442
9443 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
9444 /* For ELF on x86-64, add support for SHF_X86_64_LARGE. */
9445
9446 bfd_vma
9447 x86_64_section_letter (int letter, char **ptr_msg)
9448 {
9449 if (flag_code == CODE_64BIT)
9450 {
9451 if (letter == 'l')
9452 return SHF_X86_64_LARGE;
9453
9454 *ptr_msg = _("bad .section directive: want a,l,w,x,M,S,G,T in string");
9455 }
9456 else
9457 *ptr_msg = _("bad .section directive: want a,w,x,M,S,G,T in string");
9458 return -1;
9459 }
9460
9461 bfd_vma
9462 x86_64_section_word (char *str, size_t len)
9463 {
9464 if (len == 5 && flag_code == CODE_64BIT && CONST_STRNEQ (str, "large"))
9465 return SHF_X86_64_LARGE;
9466
9467 return -1;
9468 }
9469
9470 static void
9471 handle_large_common (int small ATTRIBUTE_UNUSED)
9472 {
9473 if (flag_code != CODE_64BIT)
9474 {
9475 s_comm_internal (0, elf_common_parse);
9476 as_warn (_(".largecomm supported only in 64bit mode, producing .comm"));
9477 }
9478 else
9479 {
9480 static segT lbss_section;
9481 asection *saved_com_section_ptr = elf_com_section_ptr;
9482 asection *saved_bss_section = bss_section;
9483
9484 if (lbss_section == NULL)
9485 {
9486 flagword applicable;
9487 segT seg = now_seg;
9488 subsegT subseg = now_subseg;
9489
9490 /* The .lbss section is for local .largecomm symbols. */
9491 lbss_section = subseg_new (".lbss", 0);
9492 applicable = bfd_applicable_section_flags (stdoutput);
9493 bfd_set_section_flags (stdoutput, lbss_section,
9494 applicable & SEC_ALLOC);
9495 seg_info (lbss_section)->bss = 1;
9496
9497 subseg_set (seg, subseg);
9498 }
9499
9500 elf_com_section_ptr = &_bfd_elf_large_com_section;
9501 bss_section = lbss_section;
9502
9503 s_comm_internal (0, elf_common_parse);
9504
9505 elf_com_section_ptr = saved_com_section_ptr;
9506 bss_section = saved_bss_section;
9507 }
9508 }
9509 #endif /* OBJ_ELF || OBJ_MAYBE_ELF */
This page took 0.386509 seconds and 5 git commands to generate.