MIPS/BFD: Report `bfd_reloc_outofrange' errors as such
[deliverable/binutils-gdb.git] / gas / config / tc-i386.c
1 /* tc-i386.c -- Assemble code for the Intel 80386
2 Copyright (C) 1989-2016 Free Software Foundation, Inc.
3
4 This file is part of GAS, the GNU Assembler.
5
6 GAS is free software; you can redistribute it and/or modify
7 it under the terms of the GNU General Public License as published by
8 the Free Software Foundation; either version 3, or (at your option)
9 any later version.
10
11 GAS is distributed in the hope that it will be useful,
12 but WITHOUT ANY WARRANTY; without even the implied warranty of
13 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 GNU General Public License for more details.
15
16 You should have received a copy of the GNU General Public License
17 along with GAS; see the file COPYING. If not, write to the Free
18 Software Foundation, 51 Franklin Street - Fifth Floor, Boston, MA
19 02110-1301, USA. */
20
21 /* Intel 80386 machine specific gas.
22 Written by Eliot Dresselhaus (eliot@mgm.mit.edu).
23 x86_64 support by Jan Hubicka (jh@suse.cz)
24 VIA PadLock support by Michal Ludvig (mludvig@suse.cz)
25 Bugs & suggestions are completely welcome. This is free software.
26 Please help us make it better. */
27
28 #include "as.h"
29 #include "safe-ctype.h"
30 #include "subsegs.h"
31 #include "dwarf2dbg.h"
32 #include "dw2gencfi.h"
33 #include "elf/x86-64.h"
34 #include "opcodes/i386-init.h"
35
36 #ifndef REGISTER_WARNINGS
37 #define REGISTER_WARNINGS 1
38 #endif
39
40 #ifndef INFER_ADDR_PREFIX
41 #define INFER_ADDR_PREFIX 1
42 #endif
43
44 #ifndef DEFAULT_ARCH
45 #define DEFAULT_ARCH "i386"
46 #endif
47
48 #ifndef INLINE
49 #if __GNUC__ >= 2
50 #define INLINE __inline__
51 #else
52 #define INLINE
53 #endif
54 #endif
55
56 /* Prefixes will be emitted in the order defined below.
57 WAIT_PREFIX must be the first prefix since FWAIT is really is an
58 instruction, and so must come before any prefixes.
59 The preferred prefix order is SEG_PREFIX, ADDR_PREFIX, DATA_PREFIX,
60 REP_PREFIX/HLE_PREFIX, LOCK_PREFIX. */
61 #define WAIT_PREFIX 0
62 #define SEG_PREFIX 1
63 #define ADDR_PREFIX 2
64 #define DATA_PREFIX 3
65 #define REP_PREFIX 4
66 #define HLE_PREFIX REP_PREFIX
67 #define BND_PREFIX REP_PREFIX
68 #define LOCK_PREFIX 5
69 #define REX_PREFIX 6 /* must come last. */
70 #define MAX_PREFIXES 7 /* max prefixes per opcode */
71
72 /* we define the syntax here (modulo base,index,scale syntax) */
73 #define REGISTER_PREFIX '%'
74 #define IMMEDIATE_PREFIX '$'
75 #define ABSOLUTE_PREFIX '*'
76
77 /* these are the instruction mnemonic suffixes in AT&T syntax or
78 memory operand size in Intel syntax. */
79 #define WORD_MNEM_SUFFIX 'w'
80 #define BYTE_MNEM_SUFFIX 'b'
81 #define SHORT_MNEM_SUFFIX 's'
82 #define LONG_MNEM_SUFFIX 'l'
83 #define QWORD_MNEM_SUFFIX 'q'
84 #define XMMWORD_MNEM_SUFFIX 'x'
85 #define YMMWORD_MNEM_SUFFIX 'y'
86 #define ZMMWORD_MNEM_SUFFIX 'z'
87 /* Intel Syntax. Use a non-ascii letter since since it never appears
88 in instructions. */
89 #define LONG_DOUBLE_MNEM_SUFFIX '\1'
90
91 #define END_OF_INSN '\0'
92
93 /*
94 'templates' is for grouping together 'template' structures for opcodes
95 of the same name. This is only used for storing the insns in the grand
96 ole hash table of insns.
97 The templates themselves start at START and range up to (but not including)
98 END.
99 */
100 typedef struct
101 {
102 const insn_template *start;
103 const insn_template *end;
104 }
105 templates;
106
107 /* 386 operand encoding bytes: see 386 book for details of this. */
108 typedef struct
109 {
110 unsigned int regmem; /* codes register or memory operand */
111 unsigned int reg; /* codes register operand (or extended opcode) */
112 unsigned int mode; /* how to interpret regmem & reg */
113 }
114 modrm_byte;
115
116 /* x86-64 extension prefix. */
117 typedef int rex_byte;
118
119 /* 386 opcode byte to code indirect addressing. */
120 typedef struct
121 {
122 unsigned base;
123 unsigned index;
124 unsigned scale;
125 }
126 sib_byte;
127
128 /* x86 arch names, types and features */
129 typedef struct
130 {
131 const char *name; /* arch name */
132 unsigned int len; /* arch string length */
133 enum processor_type type; /* arch type */
134 i386_cpu_flags flags; /* cpu feature flags */
135 unsigned int skip; /* show_arch should skip this. */
136 }
137 arch_entry;
138
139 /* Used to turn off indicated flags. */
140 typedef struct
141 {
142 const char *name; /* arch name */
143 unsigned int len; /* arch string length */
144 i386_cpu_flags flags; /* cpu feature flags */
145 }
146 noarch_entry;
147
148 static void update_code_flag (int, int);
149 static void set_code_flag (int);
150 static void set_16bit_gcc_code_flag (int);
151 static void set_intel_syntax (int);
152 static void set_intel_mnemonic (int);
153 static void set_allow_index_reg (int);
154 static void set_check (int);
155 static void set_cpu_arch (int);
156 #ifdef TE_PE
157 static void pe_directive_secrel (int);
158 #endif
159 static void signed_cons (int);
160 static char *output_invalid (int c);
161 static int i386_finalize_immediate (segT, expressionS *, i386_operand_type,
162 const char *);
163 static int i386_finalize_displacement (segT, expressionS *, i386_operand_type,
164 const char *);
165 static int i386_att_operand (char *);
166 static int i386_intel_operand (char *, int);
167 static int i386_intel_simplify (expressionS *);
168 static int i386_intel_parse_name (const char *, expressionS *);
169 static const reg_entry *parse_register (char *, char **);
170 static char *parse_insn (char *, char *);
171 static char *parse_operands (char *, const char *);
172 static void swap_operands (void);
173 static void swap_2_operands (int, int);
174 static void optimize_imm (void);
175 static void optimize_disp (void);
176 static const insn_template *match_template (void);
177 static int check_string (void);
178 static int process_suffix (void);
179 static int check_byte_reg (void);
180 static int check_long_reg (void);
181 static int check_qword_reg (void);
182 static int check_word_reg (void);
183 static int finalize_imm (void);
184 static int process_operands (void);
185 static const seg_entry *build_modrm_byte (void);
186 static void output_insn (void);
187 static void output_imm (fragS *, offsetT);
188 static void output_disp (fragS *, offsetT);
189 #ifndef I386COFF
190 static void s_bss (int);
191 #endif
192 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
193 static void handle_large_common (int small ATTRIBUTE_UNUSED);
194 #endif
195
196 static const char *default_arch = DEFAULT_ARCH;
197
198 /* This struct describes rounding control and SAE in the instruction. */
199 struct RC_Operation
200 {
201 enum rc_type
202 {
203 rne = 0,
204 rd,
205 ru,
206 rz,
207 saeonly
208 } type;
209 int operand;
210 };
211
212 static struct RC_Operation rc_op;
213
214 /* The struct describes masking, applied to OPERAND in the instruction.
215 MASK is a pointer to the corresponding mask register. ZEROING tells
216 whether merging or zeroing mask is used. */
217 struct Mask_Operation
218 {
219 const reg_entry *mask;
220 unsigned int zeroing;
221 /* The operand where this operation is associated. */
222 int operand;
223 };
224
225 static struct Mask_Operation mask_op;
226
227 /* The struct describes broadcasting, applied to OPERAND. FACTOR is
228 broadcast factor. */
229 struct Broadcast_Operation
230 {
231 /* Type of broadcast: no broadcast, {1to8}, or {1to16}. */
232 int type;
233
234 /* Index of broadcasted operand. */
235 int operand;
236 };
237
238 static struct Broadcast_Operation broadcast_op;
239
240 /* VEX prefix. */
241 typedef struct
242 {
243 /* VEX prefix is either 2 byte or 3 byte. EVEX is 4 byte. */
244 unsigned char bytes[4];
245 unsigned int length;
246 /* Destination or source register specifier. */
247 const reg_entry *register_specifier;
248 } vex_prefix;
249
250 /* 'md_assemble ()' gathers together information and puts it into a
251 i386_insn. */
252
253 union i386_op
254 {
255 expressionS *disps;
256 expressionS *imms;
257 const reg_entry *regs;
258 };
259
260 enum i386_error
261 {
262 operand_size_mismatch,
263 operand_type_mismatch,
264 register_type_mismatch,
265 number_of_operands_mismatch,
266 invalid_instruction_suffix,
267 bad_imm4,
268 old_gcc_only,
269 unsupported_with_intel_mnemonic,
270 unsupported_syntax,
271 unsupported,
272 invalid_vsib_address,
273 invalid_vector_register_set,
274 unsupported_vector_index_register,
275 unsupported_broadcast,
276 broadcast_not_on_src_operand,
277 broadcast_needed,
278 unsupported_masking,
279 mask_not_on_destination,
280 no_default_mask,
281 unsupported_rc_sae,
282 rc_sae_operand_not_last_imm,
283 invalid_register_operand,
284 try_vector_disp8
285 };
286
287 struct _i386_insn
288 {
289 /* TM holds the template for the insn were currently assembling. */
290 insn_template tm;
291
292 /* SUFFIX holds the instruction size suffix for byte, word, dword
293 or qword, if given. */
294 char suffix;
295
296 /* OPERANDS gives the number of given operands. */
297 unsigned int operands;
298
299 /* REG_OPERANDS, DISP_OPERANDS, MEM_OPERANDS, IMM_OPERANDS give the number
300 of given register, displacement, memory operands and immediate
301 operands. */
302 unsigned int reg_operands, disp_operands, mem_operands, imm_operands;
303
304 /* TYPES [i] is the type (see above #defines) which tells us how to
305 use OP[i] for the corresponding operand. */
306 i386_operand_type types[MAX_OPERANDS];
307
308 /* Displacement expression, immediate expression, or register for each
309 operand. */
310 union i386_op op[MAX_OPERANDS];
311
312 /* Flags for operands. */
313 unsigned int flags[MAX_OPERANDS];
314 #define Operand_PCrel 1
315
316 /* Relocation type for operand */
317 enum bfd_reloc_code_real reloc[MAX_OPERANDS];
318
319 /* BASE_REG, INDEX_REG, and LOG2_SCALE_FACTOR are used to encode
320 the base index byte below. */
321 const reg_entry *base_reg;
322 const reg_entry *index_reg;
323 unsigned int log2_scale_factor;
324
325 /* SEG gives the seg_entries of this insn. They are zero unless
326 explicit segment overrides are given. */
327 const seg_entry *seg[2];
328
329 /* PREFIX holds all the given prefix opcodes (usually null).
330 PREFIXES is the number of prefix opcodes. */
331 unsigned int prefixes;
332 unsigned char prefix[MAX_PREFIXES];
333
334 /* RM and SIB are the modrm byte and the sib byte where the
335 addressing modes of this insn are encoded. */
336 modrm_byte rm;
337 rex_byte rex;
338 rex_byte vrex;
339 sib_byte sib;
340 vex_prefix vex;
341
342 /* Masking attributes. */
343 struct Mask_Operation *mask;
344
345 /* Rounding control and SAE attributes. */
346 struct RC_Operation *rounding;
347
348 /* Broadcasting attributes. */
349 struct Broadcast_Operation *broadcast;
350
351 /* Compressed disp8*N attribute. */
352 unsigned int memshift;
353
354 /* Swap operand in encoding. */
355 unsigned int swap_operand;
356
357 /* Prefer 8bit or 32bit displacement in encoding. */
358 enum
359 {
360 disp_encoding_default = 0,
361 disp_encoding_8bit,
362 disp_encoding_32bit
363 } disp_encoding;
364
365 /* REP prefix. */
366 const char *rep_prefix;
367
368 /* HLE prefix. */
369 const char *hle_prefix;
370
371 /* Have BND prefix. */
372 const char *bnd_prefix;
373
374 /* Need VREX to support upper 16 registers. */
375 int need_vrex;
376
377 /* Error message. */
378 enum i386_error error;
379 };
380
381 typedef struct _i386_insn i386_insn;
382
383 /* Link RC type with corresponding string, that'll be looked for in
384 asm. */
385 struct RC_name
386 {
387 enum rc_type type;
388 const char *name;
389 unsigned int len;
390 };
391
392 static const struct RC_name RC_NamesTable[] =
393 {
394 { rne, STRING_COMMA_LEN ("rn-sae") },
395 { rd, STRING_COMMA_LEN ("rd-sae") },
396 { ru, STRING_COMMA_LEN ("ru-sae") },
397 { rz, STRING_COMMA_LEN ("rz-sae") },
398 { saeonly, STRING_COMMA_LEN ("sae") },
399 };
400
401 /* List of chars besides those in app.c:symbol_chars that can start an
402 operand. Used to prevent the scrubber eating vital white-space. */
403 const char extra_symbol_chars[] = "*%-([{"
404 #ifdef LEX_AT
405 "@"
406 #endif
407 #ifdef LEX_QM
408 "?"
409 #endif
410 ;
411
412 #if (defined (TE_I386AIX) \
413 || ((defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)) \
414 && !defined (TE_GNU) \
415 && !defined (TE_LINUX) \
416 && !defined (TE_NACL) \
417 && !defined (TE_NETWARE) \
418 && !defined (TE_FreeBSD) \
419 && !defined (TE_DragonFly) \
420 && !defined (TE_NetBSD)))
421 /* This array holds the chars that always start a comment. If the
422 pre-processor is disabled, these aren't very useful. The option
423 --divide will remove '/' from this list. */
424 const char *i386_comment_chars = "#/";
425 #define SVR4_COMMENT_CHARS 1
426 #define PREFIX_SEPARATOR '\\'
427
428 #else
429 const char *i386_comment_chars = "#";
430 #define PREFIX_SEPARATOR '/'
431 #endif
432
433 /* This array holds the chars that only start a comment at the beginning of
434 a line. If the line seems to have the form '# 123 filename'
435 .line and .file directives will appear in the pre-processed output.
436 Note that input_file.c hand checks for '#' at the beginning of the
437 first line of the input file. This is because the compiler outputs
438 #NO_APP at the beginning of its output.
439 Also note that comments started like this one will always work if
440 '/' isn't otherwise defined. */
441 const char line_comment_chars[] = "#/";
442
443 const char line_separator_chars[] = ";";
444
445 /* Chars that can be used to separate mant from exp in floating point
446 nums. */
447 const char EXP_CHARS[] = "eE";
448
449 /* Chars that mean this number is a floating point constant
450 As in 0f12.456
451 or 0d1.2345e12. */
452 const char FLT_CHARS[] = "fFdDxX";
453
454 /* Tables for lexical analysis. */
455 static char mnemonic_chars[256];
456 static char register_chars[256];
457 static char operand_chars[256];
458 static char identifier_chars[256];
459 static char digit_chars[256];
460
461 /* Lexical macros. */
462 #define is_mnemonic_char(x) (mnemonic_chars[(unsigned char) x])
463 #define is_operand_char(x) (operand_chars[(unsigned char) x])
464 #define is_register_char(x) (register_chars[(unsigned char) x])
465 #define is_space_char(x) ((x) == ' ')
466 #define is_identifier_char(x) (identifier_chars[(unsigned char) x])
467 #define is_digit_char(x) (digit_chars[(unsigned char) x])
468
469 /* All non-digit non-letter characters that may occur in an operand. */
470 static char operand_special_chars[] = "%$-+(,)*._~/<>|&^!:[@]";
471
472 /* md_assemble() always leaves the strings it's passed unaltered. To
473 effect this we maintain a stack of saved characters that we've smashed
474 with '\0's (indicating end of strings for various sub-fields of the
475 assembler instruction). */
476 static char save_stack[32];
477 static char *save_stack_p;
478 #define END_STRING_AND_SAVE(s) \
479 do { *save_stack_p++ = *(s); *(s) = '\0'; } while (0)
480 #define RESTORE_END_STRING(s) \
481 do { *(s) = *--save_stack_p; } while (0)
482
483 /* The instruction we're assembling. */
484 static i386_insn i;
485
486 /* Possible templates for current insn. */
487 static const templates *current_templates;
488
489 /* Per instruction expressionS buffers: max displacements & immediates. */
490 static expressionS disp_expressions[MAX_MEMORY_OPERANDS];
491 static expressionS im_expressions[MAX_IMMEDIATE_OPERANDS];
492
493 /* Current operand we are working on. */
494 static int this_operand = -1;
495
496 /* We support four different modes. FLAG_CODE variable is used to distinguish
497 these. */
498
499 enum flag_code {
500 CODE_32BIT,
501 CODE_16BIT,
502 CODE_64BIT };
503
504 static enum flag_code flag_code;
505 static unsigned int object_64bit;
506 static unsigned int disallow_64bit_reloc;
507 static int use_rela_relocations = 0;
508
509 #if ((defined (OBJ_MAYBE_COFF) && defined (OBJ_MAYBE_AOUT)) \
510 || defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) \
511 || defined (TE_PE) || defined (TE_PEP) || defined (OBJ_MACH_O))
512
513 /* The ELF ABI to use. */
514 enum x86_elf_abi
515 {
516 I386_ABI,
517 X86_64_ABI,
518 X86_64_X32_ABI
519 };
520
521 static enum x86_elf_abi x86_elf_abi = I386_ABI;
522 #endif
523
524 #if defined (TE_PE) || defined (TE_PEP)
525 /* Use big object file format. */
526 static int use_big_obj = 0;
527 #endif
528
529 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
530 /* 1 if generating code for a shared library. */
531 static int shared = 0;
532 #endif
533
534 /* 1 for intel syntax,
535 0 if att syntax. */
536 static int intel_syntax = 0;
537
538 /* 1 for intel mnemonic,
539 0 if att mnemonic. */
540 static int intel_mnemonic = !SYSV386_COMPAT;
541
542 /* 1 if support old (<= 2.8.1) versions of gcc. */
543 static int old_gcc = OLDGCC_COMPAT;
544
545 /* 1 if pseudo registers are permitted. */
546 static int allow_pseudo_reg = 0;
547
548 /* 1 if register prefix % not required. */
549 static int allow_naked_reg = 0;
550
551 /* 1 if the assembler should add BND prefix for all control-tranferring
552 instructions supporting it, even if this prefix wasn't specified
553 explicitly. */
554 static int add_bnd_prefix = 0;
555
556 /* 1 if pseudo index register, eiz/riz, is allowed . */
557 static int allow_index_reg = 0;
558
559 /* 1 if the assembler should ignore LOCK prefix, even if it was
560 specified explicitly. */
561 static int omit_lock_prefix = 0;
562
563 /* 1 if the assembler should encode lfence, mfence, and sfence as
564 "lock addl $0, (%{re}sp)". */
565 static int avoid_fence = 0;
566
567 /* 1 if the assembler should generate relax relocations. */
568
569 static int generate_relax_relocations
570 = DEFAULT_GENERATE_X86_RELAX_RELOCATIONS;
571
572 static enum check_kind
573 {
574 check_none = 0,
575 check_warning,
576 check_error
577 }
578 sse_check, operand_check = check_warning;
579
580 /* Register prefix used for error message. */
581 static const char *register_prefix = "%";
582
583 /* Used in 16 bit gcc mode to add an l suffix to call, ret, enter,
584 leave, push, and pop instructions so that gcc has the same stack
585 frame as in 32 bit mode. */
586 static char stackop_size = '\0';
587
588 /* Non-zero to optimize code alignment. */
589 int optimize_align_code = 1;
590
591 /* Non-zero to quieten some warnings. */
592 static int quiet_warnings = 0;
593
594 /* CPU name. */
595 static const char *cpu_arch_name = NULL;
596 static char *cpu_sub_arch_name = NULL;
597
598 /* CPU feature flags. */
599 static i386_cpu_flags cpu_arch_flags = CPU_UNKNOWN_FLAGS;
600
601 /* If we have selected a cpu we are generating instructions for. */
602 static int cpu_arch_tune_set = 0;
603
604 /* Cpu we are generating instructions for. */
605 enum processor_type cpu_arch_tune = PROCESSOR_UNKNOWN;
606
607 /* CPU feature flags of cpu we are generating instructions for. */
608 static i386_cpu_flags cpu_arch_tune_flags;
609
610 /* CPU instruction set architecture used. */
611 enum processor_type cpu_arch_isa = PROCESSOR_UNKNOWN;
612
613 /* CPU feature flags of instruction set architecture used. */
614 i386_cpu_flags cpu_arch_isa_flags;
615
616 /* If set, conditional jumps are not automatically promoted to handle
617 larger than a byte offset. */
618 static unsigned int no_cond_jump_promotion = 0;
619
620 /* Encode SSE instructions with VEX prefix. */
621 static unsigned int sse2avx;
622
623 /* Encode scalar AVX instructions with specific vector length. */
624 static enum
625 {
626 vex128 = 0,
627 vex256
628 } avxscalar;
629
630 /* Encode scalar EVEX LIG instructions with specific vector length. */
631 static enum
632 {
633 evexl128 = 0,
634 evexl256,
635 evexl512
636 } evexlig;
637
638 /* Encode EVEX WIG instructions with specific evex.w. */
639 static enum
640 {
641 evexw0 = 0,
642 evexw1
643 } evexwig;
644
645 /* Value to encode in EVEX RC bits, for SAE-only instructions. */
646 static enum rc_type evexrcig = rne;
647
648 /* Pre-defined "_GLOBAL_OFFSET_TABLE_". */
649 static symbolS *GOT_symbol;
650
651 /* The dwarf2 return column, adjusted for 32 or 64 bit. */
652 unsigned int x86_dwarf2_return_column;
653
654 /* The dwarf2 data alignment, adjusted for 32 or 64 bit. */
655 int x86_cie_data_alignment;
656
657 /* Interface to relax_segment.
658 There are 3 major relax states for 386 jump insns because the
659 different types of jumps add different sizes to frags when we're
660 figuring out what sort of jump to choose to reach a given label. */
661
662 /* Types. */
663 #define UNCOND_JUMP 0
664 #define COND_JUMP 1
665 #define COND_JUMP86 2
666
667 /* Sizes. */
668 #define CODE16 1
669 #define SMALL 0
670 #define SMALL16 (SMALL | CODE16)
671 #define BIG 2
672 #define BIG16 (BIG | CODE16)
673
674 #ifndef INLINE
675 #ifdef __GNUC__
676 #define INLINE __inline__
677 #else
678 #define INLINE
679 #endif
680 #endif
681
682 #define ENCODE_RELAX_STATE(type, size) \
683 ((relax_substateT) (((type) << 2) | (size)))
684 #define TYPE_FROM_RELAX_STATE(s) \
685 ((s) >> 2)
686 #define DISP_SIZE_FROM_RELAX_STATE(s) \
687 ((((s) & 3) == BIG ? 4 : (((s) & 3) == BIG16 ? 2 : 1)))
688
689 /* This table is used by relax_frag to promote short jumps to long
690 ones where necessary. SMALL (short) jumps may be promoted to BIG
691 (32 bit long) ones, and SMALL16 jumps to BIG16 (16 bit long). We
692 don't allow a short jump in a 32 bit code segment to be promoted to
693 a 16 bit offset jump because it's slower (requires data size
694 prefix), and doesn't work, unless the destination is in the bottom
695 64k of the code segment (The top 16 bits of eip are zeroed). */
696
697 const relax_typeS md_relax_table[] =
698 {
699 /* The fields are:
700 1) most positive reach of this state,
701 2) most negative reach of this state,
702 3) how many bytes this mode will have in the variable part of the frag
703 4) which index into the table to try if we can't fit into this one. */
704
705 /* UNCOND_JUMP states. */
706 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (UNCOND_JUMP, BIG)},
707 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (UNCOND_JUMP, BIG16)},
708 /* dword jmp adds 4 bytes to frag:
709 0 extra opcode bytes, 4 displacement bytes. */
710 {0, 0, 4, 0},
711 /* word jmp adds 2 byte2 to frag:
712 0 extra opcode bytes, 2 displacement bytes. */
713 {0, 0, 2, 0},
714
715 /* COND_JUMP states. */
716 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (COND_JUMP, BIG)},
717 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (COND_JUMP, BIG16)},
718 /* dword conditionals adds 5 bytes to frag:
719 1 extra opcode byte, 4 displacement bytes. */
720 {0, 0, 5, 0},
721 /* word conditionals add 3 bytes to frag:
722 1 extra opcode byte, 2 displacement bytes. */
723 {0, 0, 3, 0},
724
725 /* COND_JUMP86 states. */
726 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (COND_JUMP86, BIG)},
727 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (COND_JUMP86, BIG16)},
728 /* dword conditionals adds 5 bytes to frag:
729 1 extra opcode byte, 4 displacement bytes. */
730 {0, 0, 5, 0},
731 /* word conditionals add 4 bytes to frag:
732 1 displacement byte and a 3 byte long branch insn. */
733 {0, 0, 4, 0}
734 };
735
736 static const arch_entry cpu_arch[] =
737 {
738 /* Do not replace the first two entries - i386_target_format()
739 relies on them being there in this order. */
740 { STRING_COMMA_LEN ("generic32"), PROCESSOR_GENERIC32,
741 CPU_GENERIC32_FLAGS, 0 },
742 { STRING_COMMA_LEN ("generic64"), PROCESSOR_GENERIC64,
743 CPU_GENERIC64_FLAGS, 0 },
744 { STRING_COMMA_LEN ("i8086"), PROCESSOR_UNKNOWN,
745 CPU_NONE_FLAGS, 0 },
746 { STRING_COMMA_LEN ("i186"), PROCESSOR_UNKNOWN,
747 CPU_I186_FLAGS, 0 },
748 { STRING_COMMA_LEN ("i286"), PROCESSOR_UNKNOWN,
749 CPU_I286_FLAGS, 0 },
750 { STRING_COMMA_LEN ("i386"), PROCESSOR_I386,
751 CPU_I386_FLAGS, 0 },
752 { STRING_COMMA_LEN ("i486"), PROCESSOR_I486,
753 CPU_I486_FLAGS, 0 },
754 { STRING_COMMA_LEN ("i586"), PROCESSOR_PENTIUM,
755 CPU_I586_FLAGS, 0 },
756 { STRING_COMMA_LEN ("i686"), PROCESSOR_PENTIUMPRO,
757 CPU_I686_FLAGS, 0 },
758 { STRING_COMMA_LEN ("pentium"), PROCESSOR_PENTIUM,
759 CPU_I586_FLAGS, 0 },
760 { STRING_COMMA_LEN ("pentiumpro"), PROCESSOR_PENTIUMPRO,
761 CPU_PENTIUMPRO_FLAGS, 0 },
762 { STRING_COMMA_LEN ("pentiumii"), PROCESSOR_PENTIUMPRO,
763 CPU_P2_FLAGS, 0 },
764 { STRING_COMMA_LEN ("pentiumiii"),PROCESSOR_PENTIUMPRO,
765 CPU_P3_FLAGS, 0 },
766 { STRING_COMMA_LEN ("pentium4"), PROCESSOR_PENTIUM4,
767 CPU_P4_FLAGS, 0 },
768 { STRING_COMMA_LEN ("prescott"), PROCESSOR_NOCONA,
769 CPU_CORE_FLAGS, 0 },
770 { STRING_COMMA_LEN ("nocona"), PROCESSOR_NOCONA,
771 CPU_NOCONA_FLAGS, 0 },
772 { STRING_COMMA_LEN ("yonah"), PROCESSOR_CORE,
773 CPU_CORE_FLAGS, 1 },
774 { STRING_COMMA_LEN ("core"), PROCESSOR_CORE,
775 CPU_CORE_FLAGS, 0 },
776 { STRING_COMMA_LEN ("merom"), PROCESSOR_CORE2,
777 CPU_CORE2_FLAGS, 1 },
778 { STRING_COMMA_LEN ("core2"), PROCESSOR_CORE2,
779 CPU_CORE2_FLAGS, 0 },
780 { STRING_COMMA_LEN ("corei7"), PROCESSOR_COREI7,
781 CPU_COREI7_FLAGS, 0 },
782 { STRING_COMMA_LEN ("l1om"), PROCESSOR_L1OM,
783 CPU_L1OM_FLAGS, 0 },
784 { STRING_COMMA_LEN ("k1om"), PROCESSOR_K1OM,
785 CPU_K1OM_FLAGS, 0 },
786 { STRING_COMMA_LEN ("iamcu"), PROCESSOR_IAMCU,
787 CPU_IAMCU_FLAGS, 0 },
788 { STRING_COMMA_LEN ("k6"), PROCESSOR_K6,
789 CPU_K6_FLAGS, 0 },
790 { STRING_COMMA_LEN ("k6_2"), PROCESSOR_K6,
791 CPU_K6_2_FLAGS, 0 },
792 { STRING_COMMA_LEN ("athlon"), PROCESSOR_ATHLON,
793 CPU_ATHLON_FLAGS, 0 },
794 { STRING_COMMA_LEN ("sledgehammer"), PROCESSOR_K8,
795 CPU_K8_FLAGS, 1 },
796 { STRING_COMMA_LEN ("opteron"), PROCESSOR_K8,
797 CPU_K8_FLAGS, 0 },
798 { STRING_COMMA_LEN ("k8"), PROCESSOR_K8,
799 CPU_K8_FLAGS, 0 },
800 { STRING_COMMA_LEN ("amdfam10"), PROCESSOR_AMDFAM10,
801 CPU_AMDFAM10_FLAGS, 0 },
802 { STRING_COMMA_LEN ("bdver1"), PROCESSOR_BD,
803 CPU_BDVER1_FLAGS, 0 },
804 { STRING_COMMA_LEN ("bdver2"), PROCESSOR_BD,
805 CPU_BDVER2_FLAGS, 0 },
806 { STRING_COMMA_LEN ("bdver3"), PROCESSOR_BD,
807 CPU_BDVER3_FLAGS, 0 },
808 { STRING_COMMA_LEN ("bdver4"), PROCESSOR_BD,
809 CPU_BDVER4_FLAGS, 0 },
810 { STRING_COMMA_LEN ("znver1"), PROCESSOR_ZNVER,
811 CPU_ZNVER1_FLAGS, 0 },
812 { STRING_COMMA_LEN ("btver1"), PROCESSOR_BT,
813 CPU_BTVER1_FLAGS, 0 },
814 { STRING_COMMA_LEN ("btver2"), PROCESSOR_BT,
815 CPU_BTVER2_FLAGS, 0 },
816 { STRING_COMMA_LEN (".8087"), PROCESSOR_UNKNOWN,
817 CPU_8087_FLAGS, 0 },
818 { STRING_COMMA_LEN (".287"), PROCESSOR_UNKNOWN,
819 CPU_287_FLAGS, 0 },
820 { STRING_COMMA_LEN (".387"), PROCESSOR_UNKNOWN,
821 CPU_387_FLAGS, 0 },
822 { STRING_COMMA_LEN (".mmx"), PROCESSOR_UNKNOWN,
823 CPU_MMX_FLAGS, 0 },
824 { STRING_COMMA_LEN (".sse"), PROCESSOR_UNKNOWN,
825 CPU_SSE_FLAGS, 0 },
826 { STRING_COMMA_LEN (".sse2"), PROCESSOR_UNKNOWN,
827 CPU_SSE2_FLAGS, 0 },
828 { STRING_COMMA_LEN (".sse3"), PROCESSOR_UNKNOWN,
829 CPU_SSE3_FLAGS, 0 },
830 { STRING_COMMA_LEN (".ssse3"), PROCESSOR_UNKNOWN,
831 CPU_SSSE3_FLAGS, 0 },
832 { STRING_COMMA_LEN (".sse4.1"), PROCESSOR_UNKNOWN,
833 CPU_SSE4_1_FLAGS, 0 },
834 { STRING_COMMA_LEN (".sse4.2"), PROCESSOR_UNKNOWN,
835 CPU_SSE4_2_FLAGS, 0 },
836 { STRING_COMMA_LEN (".sse4"), PROCESSOR_UNKNOWN,
837 CPU_SSE4_2_FLAGS, 0 },
838 { STRING_COMMA_LEN (".avx"), PROCESSOR_UNKNOWN,
839 CPU_AVX_FLAGS, 0 },
840 { STRING_COMMA_LEN (".avx2"), PROCESSOR_UNKNOWN,
841 CPU_AVX2_FLAGS, 0 },
842 { STRING_COMMA_LEN (".avx512f"), PROCESSOR_UNKNOWN,
843 CPU_AVX512F_FLAGS, 0 },
844 { STRING_COMMA_LEN (".avx512cd"), PROCESSOR_UNKNOWN,
845 CPU_AVX512CD_FLAGS, 0 },
846 { STRING_COMMA_LEN (".avx512er"), PROCESSOR_UNKNOWN,
847 CPU_AVX512ER_FLAGS, 0 },
848 { STRING_COMMA_LEN (".avx512pf"), PROCESSOR_UNKNOWN,
849 CPU_AVX512PF_FLAGS, 0 },
850 { STRING_COMMA_LEN (".avx512dq"), PROCESSOR_UNKNOWN,
851 CPU_AVX512DQ_FLAGS, 0 },
852 { STRING_COMMA_LEN (".avx512bw"), PROCESSOR_UNKNOWN,
853 CPU_AVX512BW_FLAGS, 0 },
854 { STRING_COMMA_LEN (".avx512vl"), PROCESSOR_UNKNOWN,
855 CPU_AVX512VL_FLAGS, 0 },
856 { STRING_COMMA_LEN (".vmx"), PROCESSOR_UNKNOWN,
857 CPU_VMX_FLAGS, 0 },
858 { STRING_COMMA_LEN (".vmfunc"), PROCESSOR_UNKNOWN,
859 CPU_VMFUNC_FLAGS, 0 },
860 { STRING_COMMA_LEN (".smx"), PROCESSOR_UNKNOWN,
861 CPU_SMX_FLAGS, 0 },
862 { STRING_COMMA_LEN (".xsave"), PROCESSOR_UNKNOWN,
863 CPU_XSAVE_FLAGS, 0 },
864 { STRING_COMMA_LEN (".xsaveopt"), PROCESSOR_UNKNOWN,
865 CPU_XSAVEOPT_FLAGS, 0 },
866 { STRING_COMMA_LEN (".xsavec"), PROCESSOR_UNKNOWN,
867 CPU_XSAVEC_FLAGS, 0 },
868 { STRING_COMMA_LEN (".xsaves"), PROCESSOR_UNKNOWN,
869 CPU_XSAVES_FLAGS, 0 },
870 { STRING_COMMA_LEN (".aes"), PROCESSOR_UNKNOWN,
871 CPU_AES_FLAGS, 0 },
872 { STRING_COMMA_LEN (".pclmul"), PROCESSOR_UNKNOWN,
873 CPU_PCLMUL_FLAGS, 0 },
874 { STRING_COMMA_LEN (".clmul"), PROCESSOR_UNKNOWN,
875 CPU_PCLMUL_FLAGS, 1 },
876 { STRING_COMMA_LEN (".fsgsbase"), PROCESSOR_UNKNOWN,
877 CPU_FSGSBASE_FLAGS, 0 },
878 { STRING_COMMA_LEN (".rdrnd"), PROCESSOR_UNKNOWN,
879 CPU_RDRND_FLAGS, 0 },
880 { STRING_COMMA_LEN (".f16c"), PROCESSOR_UNKNOWN,
881 CPU_F16C_FLAGS, 0 },
882 { STRING_COMMA_LEN (".bmi2"), PROCESSOR_UNKNOWN,
883 CPU_BMI2_FLAGS, 0 },
884 { STRING_COMMA_LEN (".fma"), PROCESSOR_UNKNOWN,
885 CPU_FMA_FLAGS, 0 },
886 { STRING_COMMA_LEN (".fma4"), PROCESSOR_UNKNOWN,
887 CPU_FMA4_FLAGS, 0 },
888 { STRING_COMMA_LEN (".xop"), PROCESSOR_UNKNOWN,
889 CPU_XOP_FLAGS, 0 },
890 { STRING_COMMA_LEN (".lwp"), PROCESSOR_UNKNOWN,
891 CPU_LWP_FLAGS, 0 },
892 { STRING_COMMA_LEN (".movbe"), PROCESSOR_UNKNOWN,
893 CPU_MOVBE_FLAGS, 0 },
894 { STRING_COMMA_LEN (".cx16"), PROCESSOR_UNKNOWN,
895 CPU_CX16_FLAGS, 0 },
896 { STRING_COMMA_LEN (".ept"), PROCESSOR_UNKNOWN,
897 CPU_EPT_FLAGS, 0 },
898 { STRING_COMMA_LEN (".lzcnt"), PROCESSOR_UNKNOWN,
899 CPU_LZCNT_FLAGS, 0 },
900 { STRING_COMMA_LEN (".hle"), PROCESSOR_UNKNOWN,
901 CPU_HLE_FLAGS, 0 },
902 { STRING_COMMA_LEN (".rtm"), PROCESSOR_UNKNOWN,
903 CPU_RTM_FLAGS, 0 },
904 { STRING_COMMA_LEN (".invpcid"), PROCESSOR_UNKNOWN,
905 CPU_INVPCID_FLAGS, 0 },
906 { STRING_COMMA_LEN (".clflush"), PROCESSOR_UNKNOWN,
907 CPU_CLFLUSH_FLAGS, 0 },
908 { STRING_COMMA_LEN (".nop"), PROCESSOR_UNKNOWN,
909 CPU_NOP_FLAGS, 0 },
910 { STRING_COMMA_LEN (".syscall"), PROCESSOR_UNKNOWN,
911 CPU_SYSCALL_FLAGS, 0 },
912 { STRING_COMMA_LEN (".rdtscp"), PROCESSOR_UNKNOWN,
913 CPU_RDTSCP_FLAGS, 0 },
914 { STRING_COMMA_LEN (".3dnow"), PROCESSOR_UNKNOWN,
915 CPU_3DNOW_FLAGS, 0 },
916 { STRING_COMMA_LEN (".3dnowa"), PROCESSOR_UNKNOWN,
917 CPU_3DNOWA_FLAGS, 0 },
918 { STRING_COMMA_LEN (".padlock"), PROCESSOR_UNKNOWN,
919 CPU_PADLOCK_FLAGS, 0 },
920 { STRING_COMMA_LEN (".pacifica"), PROCESSOR_UNKNOWN,
921 CPU_SVME_FLAGS, 1 },
922 { STRING_COMMA_LEN (".svme"), PROCESSOR_UNKNOWN,
923 CPU_SVME_FLAGS, 0 },
924 { STRING_COMMA_LEN (".sse4a"), PROCESSOR_UNKNOWN,
925 CPU_SSE4A_FLAGS, 0 },
926 { STRING_COMMA_LEN (".abm"), PROCESSOR_UNKNOWN,
927 CPU_ABM_FLAGS, 0 },
928 { STRING_COMMA_LEN (".bmi"), PROCESSOR_UNKNOWN,
929 CPU_BMI_FLAGS, 0 },
930 { STRING_COMMA_LEN (".tbm"), PROCESSOR_UNKNOWN,
931 CPU_TBM_FLAGS, 0 },
932 { STRING_COMMA_LEN (".adx"), PROCESSOR_UNKNOWN,
933 CPU_ADX_FLAGS, 0 },
934 { STRING_COMMA_LEN (".rdseed"), PROCESSOR_UNKNOWN,
935 CPU_RDSEED_FLAGS, 0 },
936 { STRING_COMMA_LEN (".prfchw"), PROCESSOR_UNKNOWN,
937 CPU_PRFCHW_FLAGS, 0 },
938 { STRING_COMMA_LEN (".smap"), PROCESSOR_UNKNOWN,
939 CPU_SMAP_FLAGS, 0 },
940 { STRING_COMMA_LEN (".mpx"), PROCESSOR_UNKNOWN,
941 CPU_MPX_FLAGS, 0 },
942 { STRING_COMMA_LEN (".sha"), PROCESSOR_UNKNOWN,
943 CPU_SHA_FLAGS, 0 },
944 { STRING_COMMA_LEN (".clflushopt"), PROCESSOR_UNKNOWN,
945 CPU_CLFLUSHOPT_FLAGS, 0 },
946 { STRING_COMMA_LEN (".prefetchwt1"), PROCESSOR_UNKNOWN,
947 CPU_PREFETCHWT1_FLAGS, 0 },
948 { STRING_COMMA_LEN (".se1"), PROCESSOR_UNKNOWN,
949 CPU_SE1_FLAGS, 0 },
950 { STRING_COMMA_LEN (".clwb"), PROCESSOR_UNKNOWN,
951 CPU_CLWB_FLAGS, 0 },
952 { STRING_COMMA_LEN (".pcommit"), PROCESSOR_UNKNOWN,
953 CPU_PCOMMIT_FLAGS, 0 },
954 { STRING_COMMA_LEN (".avx512ifma"), PROCESSOR_UNKNOWN,
955 CPU_AVX512IFMA_FLAGS, 0 },
956 { STRING_COMMA_LEN (".avx512vbmi"), PROCESSOR_UNKNOWN,
957 CPU_AVX512VBMI_FLAGS, 0 },
958 { STRING_COMMA_LEN (".clzero"), PROCESSOR_UNKNOWN,
959 CPU_CLZERO_FLAGS, 0 },
960 { STRING_COMMA_LEN (".mwaitx"), PROCESSOR_UNKNOWN,
961 CPU_MWAITX_FLAGS, 0 },
962 { STRING_COMMA_LEN (".ospke"), PROCESSOR_UNKNOWN,
963 CPU_OSPKE_FLAGS, 0 },
964 { STRING_COMMA_LEN (".rdpid"), PROCESSOR_UNKNOWN,
965 CPU_RDPID_FLAGS, 0 },
966 };
967
968 static const noarch_entry cpu_noarch[] =
969 {
970 { STRING_COMMA_LEN ("no87"), CPU_ANY_X87_FLAGS },
971 { STRING_COMMA_LEN ("nommx"), CPU_ANY_MMX_FLAGS },
972 { STRING_COMMA_LEN ("nosse"), CPU_ANY_SSE_FLAGS },
973 { STRING_COMMA_LEN ("noavx"), CPU_ANY_AVX_FLAGS },
974 };
975
976 #ifdef I386COFF
977 /* Like s_lcomm_internal in gas/read.c but the alignment string
978 is allowed to be optional. */
979
980 static symbolS *
981 pe_lcomm_internal (int needs_align, symbolS *symbolP, addressT size)
982 {
983 addressT align = 0;
984
985 SKIP_WHITESPACE ();
986
987 if (needs_align
988 && *input_line_pointer == ',')
989 {
990 align = parse_align (needs_align - 1);
991
992 if (align == (addressT) -1)
993 return NULL;
994 }
995 else
996 {
997 if (size >= 8)
998 align = 3;
999 else if (size >= 4)
1000 align = 2;
1001 else if (size >= 2)
1002 align = 1;
1003 else
1004 align = 0;
1005 }
1006
1007 bss_alloc (symbolP, size, align);
1008 return symbolP;
1009 }
1010
1011 static void
1012 pe_lcomm (int needs_align)
1013 {
1014 s_comm_internal (needs_align * 2, pe_lcomm_internal);
1015 }
1016 #endif
1017
1018 const pseudo_typeS md_pseudo_table[] =
1019 {
1020 #if !defined(OBJ_AOUT) && !defined(USE_ALIGN_PTWO)
1021 {"align", s_align_bytes, 0},
1022 #else
1023 {"align", s_align_ptwo, 0},
1024 #endif
1025 {"arch", set_cpu_arch, 0},
1026 #ifndef I386COFF
1027 {"bss", s_bss, 0},
1028 #else
1029 {"lcomm", pe_lcomm, 1},
1030 #endif
1031 {"ffloat", float_cons, 'f'},
1032 {"dfloat", float_cons, 'd'},
1033 {"tfloat", float_cons, 'x'},
1034 {"value", cons, 2},
1035 {"slong", signed_cons, 4},
1036 {"noopt", s_ignore, 0},
1037 {"optim", s_ignore, 0},
1038 {"code16gcc", set_16bit_gcc_code_flag, CODE_16BIT},
1039 {"code16", set_code_flag, CODE_16BIT},
1040 {"code32", set_code_flag, CODE_32BIT},
1041 {"code64", set_code_flag, CODE_64BIT},
1042 {"intel_syntax", set_intel_syntax, 1},
1043 {"att_syntax", set_intel_syntax, 0},
1044 {"intel_mnemonic", set_intel_mnemonic, 1},
1045 {"att_mnemonic", set_intel_mnemonic, 0},
1046 {"allow_index_reg", set_allow_index_reg, 1},
1047 {"disallow_index_reg", set_allow_index_reg, 0},
1048 {"sse_check", set_check, 0},
1049 {"operand_check", set_check, 1},
1050 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
1051 {"largecomm", handle_large_common, 0},
1052 #else
1053 {"file", (void (*) (int)) dwarf2_directive_file, 0},
1054 {"loc", dwarf2_directive_loc, 0},
1055 {"loc_mark_labels", dwarf2_directive_loc_mark_labels, 0},
1056 #endif
1057 #ifdef TE_PE
1058 {"secrel32", pe_directive_secrel, 0},
1059 #endif
1060 {0, 0, 0}
1061 };
1062
1063 /* For interface with expression (). */
1064 extern char *input_line_pointer;
1065
1066 /* Hash table for instruction mnemonic lookup. */
1067 static struct hash_control *op_hash;
1068
1069 /* Hash table for register lookup. */
1070 static struct hash_control *reg_hash;
1071 \f
1072 void
1073 i386_align_code (fragS *fragP, int count)
1074 {
1075 /* Various efficient no-op patterns for aligning code labels.
1076 Note: Don't try to assemble the instructions in the comments.
1077 0L and 0w are not legal. */
1078 static const unsigned char f32_1[] =
1079 {0x90}; /* nop */
1080 static const unsigned char f32_2[] =
1081 {0x66,0x90}; /* xchg %ax,%ax */
1082 static const unsigned char f32_3[] =
1083 {0x8d,0x76,0x00}; /* leal 0(%esi),%esi */
1084 static const unsigned char f32_4[] =
1085 {0x8d,0x74,0x26,0x00}; /* leal 0(%esi,1),%esi */
1086 static const unsigned char f32_5[] =
1087 {0x90, /* nop */
1088 0x8d,0x74,0x26,0x00}; /* leal 0(%esi,1),%esi */
1089 static const unsigned char f32_6[] =
1090 {0x8d,0xb6,0x00,0x00,0x00,0x00}; /* leal 0L(%esi),%esi */
1091 static const unsigned char f32_7[] =
1092 {0x8d,0xb4,0x26,0x00,0x00,0x00,0x00}; /* leal 0L(%esi,1),%esi */
1093 static const unsigned char f32_8[] =
1094 {0x90, /* nop */
1095 0x8d,0xb4,0x26,0x00,0x00,0x00,0x00}; /* leal 0L(%esi,1),%esi */
1096 static const unsigned char f32_9[] =
1097 {0x89,0xf6, /* movl %esi,%esi */
1098 0x8d,0xbc,0x27,0x00,0x00,0x00,0x00}; /* leal 0L(%edi,1),%edi */
1099 static const unsigned char f32_10[] =
1100 {0x8d,0x76,0x00, /* leal 0(%esi),%esi */
1101 0x8d,0xbc,0x27,0x00,0x00,0x00,0x00}; /* leal 0L(%edi,1),%edi */
1102 static const unsigned char f32_11[] =
1103 {0x8d,0x74,0x26,0x00, /* leal 0(%esi,1),%esi */
1104 0x8d,0xbc,0x27,0x00,0x00,0x00,0x00}; /* leal 0L(%edi,1),%edi */
1105 static const unsigned char f32_12[] =
1106 {0x8d,0xb6,0x00,0x00,0x00,0x00, /* leal 0L(%esi),%esi */
1107 0x8d,0xbf,0x00,0x00,0x00,0x00}; /* leal 0L(%edi),%edi */
1108 static const unsigned char f32_13[] =
1109 {0x8d,0xb6,0x00,0x00,0x00,0x00, /* leal 0L(%esi),%esi */
1110 0x8d,0xbc,0x27,0x00,0x00,0x00,0x00}; /* leal 0L(%edi,1),%edi */
1111 static const unsigned char f32_14[] =
1112 {0x8d,0xb4,0x26,0x00,0x00,0x00,0x00, /* leal 0L(%esi,1),%esi */
1113 0x8d,0xbc,0x27,0x00,0x00,0x00,0x00}; /* leal 0L(%edi,1),%edi */
1114 static const unsigned char f16_3[] =
1115 {0x8d,0x74,0x00}; /* lea 0(%esi),%esi */
1116 static const unsigned char f16_4[] =
1117 {0x8d,0xb4,0x00,0x00}; /* lea 0w(%si),%si */
1118 static const unsigned char f16_5[] =
1119 {0x90, /* nop */
1120 0x8d,0xb4,0x00,0x00}; /* lea 0w(%si),%si */
1121 static const unsigned char f16_6[] =
1122 {0x89,0xf6, /* mov %si,%si */
1123 0x8d,0xbd,0x00,0x00}; /* lea 0w(%di),%di */
1124 static const unsigned char f16_7[] =
1125 {0x8d,0x74,0x00, /* lea 0(%si),%si */
1126 0x8d,0xbd,0x00,0x00}; /* lea 0w(%di),%di */
1127 static const unsigned char f16_8[] =
1128 {0x8d,0xb4,0x00,0x00, /* lea 0w(%si),%si */
1129 0x8d,0xbd,0x00,0x00}; /* lea 0w(%di),%di */
1130 static const unsigned char jump_31[] =
1131 {0xeb,0x1d,0x90,0x90,0x90,0x90,0x90, /* jmp .+31; lotsa nops */
1132 0x90,0x90,0x90,0x90,0x90,0x90,0x90,0x90,
1133 0x90,0x90,0x90,0x90,0x90,0x90,0x90,0x90,
1134 0x90,0x90,0x90,0x90,0x90,0x90,0x90,0x90};
1135 static const unsigned char *const f32_patt[] = {
1136 f32_1, f32_2, f32_3, f32_4, f32_5, f32_6, f32_7, f32_8,
1137 f32_9, f32_10, f32_11, f32_12, f32_13, f32_14
1138 };
1139 static const unsigned char *const f16_patt[] = {
1140 f32_1, f32_2, f16_3, f16_4, f16_5, f16_6, f16_7, f16_8
1141 };
1142 /* nopl (%[re]ax) */
1143 static const unsigned char alt_3[] =
1144 {0x0f,0x1f,0x00};
1145 /* nopl 0(%[re]ax) */
1146 static const unsigned char alt_4[] =
1147 {0x0f,0x1f,0x40,0x00};
1148 /* nopl 0(%[re]ax,%[re]ax,1) */
1149 static const unsigned char alt_5[] =
1150 {0x0f,0x1f,0x44,0x00,0x00};
1151 /* nopw 0(%[re]ax,%[re]ax,1) */
1152 static const unsigned char alt_6[] =
1153 {0x66,0x0f,0x1f,0x44,0x00,0x00};
1154 /* nopl 0L(%[re]ax) */
1155 static const unsigned char alt_7[] =
1156 {0x0f,0x1f,0x80,0x00,0x00,0x00,0x00};
1157 /* nopl 0L(%[re]ax,%[re]ax,1) */
1158 static const unsigned char alt_8[] =
1159 {0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
1160 /* nopw 0L(%[re]ax,%[re]ax,1) */
1161 static const unsigned char alt_9[] =
1162 {0x66,0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
1163 /* nopw %cs:0L(%[re]ax,%[re]ax,1) */
1164 static const unsigned char alt_10[] =
1165 {0x66,0x2e,0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
1166 static const unsigned char *const alt_patt[] = {
1167 f32_1, f32_2, alt_3, alt_4, alt_5, alt_6, alt_7, alt_8,
1168 alt_9, alt_10
1169 };
1170
1171 /* Only align for at least a positive non-zero boundary. */
1172 if (count <= 0 || count > MAX_MEM_FOR_RS_ALIGN_CODE)
1173 return;
1174
1175 /* We need to decide which NOP sequence to use for 32bit and
1176 64bit. When -mtune= is used:
1177
1178 1. For PROCESSOR_I386, PROCESSOR_I486, PROCESSOR_PENTIUM and
1179 PROCESSOR_GENERIC32, f32_patt will be used.
1180 2. For the rest, alt_patt will be used.
1181
1182 When -mtune= isn't used, alt_patt will be used if
1183 cpu_arch_isa_flags has CpuNop. Otherwise, f32_patt will
1184 be used.
1185
1186 When -march= or .arch is used, we can't use anything beyond
1187 cpu_arch_isa_flags. */
1188
1189 if (flag_code == CODE_16BIT)
1190 {
1191 if (count > 8)
1192 {
1193 memcpy (fragP->fr_literal + fragP->fr_fix,
1194 jump_31, count);
1195 /* Adjust jump offset. */
1196 fragP->fr_literal[fragP->fr_fix + 1] = count - 2;
1197 }
1198 else
1199 memcpy (fragP->fr_literal + fragP->fr_fix,
1200 f16_patt[count - 1], count);
1201 }
1202 else
1203 {
1204 const unsigned char *const *patt = NULL;
1205
1206 if (fragP->tc_frag_data.isa == PROCESSOR_UNKNOWN)
1207 {
1208 /* PROCESSOR_UNKNOWN means that all ISAs may be used. */
1209 switch (cpu_arch_tune)
1210 {
1211 case PROCESSOR_UNKNOWN:
1212 /* We use cpu_arch_isa_flags to check if we SHOULD
1213 optimize with nops. */
1214 if (fragP->tc_frag_data.isa_flags.bitfield.cpunop)
1215 patt = alt_patt;
1216 else
1217 patt = f32_patt;
1218 break;
1219 case PROCESSOR_PENTIUM4:
1220 case PROCESSOR_NOCONA:
1221 case PROCESSOR_CORE:
1222 case PROCESSOR_CORE2:
1223 case PROCESSOR_COREI7:
1224 case PROCESSOR_L1OM:
1225 case PROCESSOR_K1OM:
1226 case PROCESSOR_GENERIC64:
1227 case PROCESSOR_K6:
1228 case PROCESSOR_ATHLON:
1229 case PROCESSOR_K8:
1230 case PROCESSOR_AMDFAM10:
1231 case PROCESSOR_BD:
1232 case PROCESSOR_ZNVER:
1233 case PROCESSOR_BT:
1234 patt = alt_patt;
1235 break;
1236 case PROCESSOR_I386:
1237 case PROCESSOR_I486:
1238 case PROCESSOR_PENTIUM:
1239 case PROCESSOR_PENTIUMPRO:
1240 case PROCESSOR_IAMCU:
1241 case PROCESSOR_GENERIC32:
1242 patt = f32_patt;
1243 break;
1244 }
1245 }
1246 else
1247 {
1248 switch (fragP->tc_frag_data.tune)
1249 {
1250 case PROCESSOR_UNKNOWN:
1251 /* When cpu_arch_isa is set, cpu_arch_tune shouldn't be
1252 PROCESSOR_UNKNOWN. */
1253 abort ();
1254 break;
1255
1256 case PROCESSOR_I386:
1257 case PROCESSOR_I486:
1258 case PROCESSOR_PENTIUM:
1259 case PROCESSOR_IAMCU:
1260 case PROCESSOR_K6:
1261 case PROCESSOR_ATHLON:
1262 case PROCESSOR_K8:
1263 case PROCESSOR_AMDFAM10:
1264 case PROCESSOR_BD:
1265 case PROCESSOR_ZNVER:
1266 case PROCESSOR_BT:
1267 case PROCESSOR_GENERIC32:
1268 /* We use cpu_arch_isa_flags to check if we CAN optimize
1269 with nops. */
1270 if (fragP->tc_frag_data.isa_flags.bitfield.cpunop)
1271 patt = alt_patt;
1272 else
1273 patt = f32_patt;
1274 break;
1275 case PROCESSOR_PENTIUMPRO:
1276 case PROCESSOR_PENTIUM4:
1277 case PROCESSOR_NOCONA:
1278 case PROCESSOR_CORE:
1279 case PROCESSOR_CORE2:
1280 case PROCESSOR_COREI7:
1281 case PROCESSOR_L1OM:
1282 case PROCESSOR_K1OM:
1283 if (fragP->tc_frag_data.isa_flags.bitfield.cpunop)
1284 patt = alt_patt;
1285 else
1286 patt = f32_patt;
1287 break;
1288 case PROCESSOR_GENERIC64:
1289 patt = alt_patt;
1290 break;
1291 }
1292 }
1293
1294 if (patt == f32_patt)
1295 {
1296 /* If the padding is less than 15 bytes, we use the normal
1297 ones. Otherwise, we use a jump instruction and adjust
1298 its offset. */
1299 int limit;
1300
1301 /* For 64bit, the limit is 3 bytes. */
1302 if (flag_code == CODE_64BIT
1303 && fragP->tc_frag_data.isa_flags.bitfield.cpulm)
1304 limit = 3;
1305 else
1306 limit = 15;
1307 if (count < limit)
1308 memcpy (fragP->fr_literal + fragP->fr_fix,
1309 patt[count - 1], count);
1310 else
1311 {
1312 memcpy (fragP->fr_literal + fragP->fr_fix,
1313 jump_31, count);
1314 /* Adjust jump offset. */
1315 fragP->fr_literal[fragP->fr_fix + 1] = count - 2;
1316 }
1317 }
1318 else
1319 {
1320 /* Maximum length of an instruction is 10 byte. If the
1321 padding is greater than 10 bytes and we don't use jump,
1322 we have to break it into smaller pieces. */
1323 int padding = count;
1324 while (padding > 10)
1325 {
1326 padding -= 10;
1327 memcpy (fragP->fr_literal + fragP->fr_fix + padding,
1328 patt [9], 10);
1329 }
1330
1331 if (padding)
1332 memcpy (fragP->fr_literal + fragP->fr_fix,
1333 patt [padding - 1], padding);
1334 }
1335 }
1336 fragP->fr_var = count;
1337 }
1338
1339 static INLINE int
1340 operand_type_all_zero (const union i386_operand_type *x)
1341 {
1342 switch (ARRAY_SIZE(x->array))
1343 {
1344 case 3:
1345 if (x->array[2])
1346 return 0;
1347 case 2:
1348 if (x->array[1])
1349 return 0;
1350 case 1:
1351 return !x->array[0];
1352 default:
1353 abort ();
1354 }
1355 }
1356
1357 static INLINE void
1358 operand_type_set (union i386_operand_type *x, unsigned int v)
1359 {
1360 switch (ARRAY_SIZE(x->array))
1361 {
1362 case 3:
1363 x->array[2] = v;
1364 case 2:
1365 x->array[1] = v;
1366 case 1:
1367 x->array[0] = v;
1368 break;
1369 default:
1370 abort ();
1371 }
1372 }
1373
1374 static INLINE int
1375 operand_type_equal (const union i386_operand_type *x,
1376 const union i386_operand_type *y)
1377 {
1378 switch (ARRAY_SIZE(x->array))
1379 {
1380 case 3:
1381 if (x->array[2] != y->array[2])
1382 return 0;
1383 case 2:
1384 if (x->array[1] != y->array[1])
1385 return 0;
1386 case 1:
1387 return x->array[0] == y->array[0];
1388 break;
1389 default:
1390 abort ();
1391 }
1392 }
1393
1394 static INLINE int
1395 cpu_flags_all_zero (const union i386_cpu_flags *x)
1396 {
1397 switch (ARRAY_SIZE(x->array))
1398 {
1399 case 3:
1400 if (x->array[2])
1401 return 0;
1402 case 2:
1403 if (x->array[1])
1404 return 0;
1405 case 1:
1406 return !x->array[0];
1407 default:
1408 abort ();
1409 }
1410 }
1411
1412 static INLINE int
1413 cpu_flags_equal (const union i386_cpu_flags *x,
1414 const union i386_cpu_flags *y)
1415 {
1416 switch (ARRAY_SIZE(x->array))
1417 {
1418 case 3:
1419 if (x->array[2] != y->array[2])
1420 return 0;
1421 case 2:
1422 if (x->array[1] != y->array[1])
1423 return 0;
1424 case 1:
1425 return x->array[0] == y->array[0];
1426 break;
1427 default:
1428 abort ();
1429 }
1430 }
1431
1432 static INLINE int
1433 cpu_flags_check_cpu64 (i386_cpu_flags f)
1434 {
1435 return !((flag_code == CODE_64BIT && f.bitfield.cpuno64)
1436 || (flag_code != CODE_64BIT && f.bitfield.cpu64));
1437 }
1438
1439 static INLINE i386_cpu_flags
1440 cpu_flags_and (i386_cpu_flags x, i386_cpu_flags y)
1441 {
1442 switch (ARRAY_SIZE (x.array))
1443 {
1444 case 3:
1445 x.array [2] &= y.array [2];
1446 case 2:
1447 x.array [1] &= y.array [1];
1448 case 1:
1449 x.array [0] &= y.array [0];
1450 break;
1451 default:
1452 abort ();
1453 }
1454 return x;
1455 }
1456
1457 static INLINE i386_cpu_flags
1458 cpu_flags_or (i386_cpu_flags x, i386_cpu_flags y)
1459 {
1460 switch (ARRAY_SIZE (x.array))
1461 {
1462 case 3:
1463 x.array [2] |= y.array [2];
1464 case 2:
1465 x.array [1] |= y.array [1];
1466 case 1:
1467 x.array [0] |= y.array [0];
1468 break;
1469 default:
1470 abort ();
1471 }
1472 return x;
1473 }
1474
1475 static INLINE i386_cpu_flags
1476 cpu_flags_and_not (i386_cpu_flags x, i386_cpu_flags y)
1477 {
1478 switch (ARRAY_SIZE (x.array))
1479 {
1480 case 3:
1481 x.array [2] &= ~y.array [2];
1482 case 2:
1483 x.array [1] &= ~y.array [1];
1484 case 1:
1485 x.array [0] &= ~y.array [0];
1486 break;
1487 default:
1488 abort ();
1489 }
1490 return x;
1491 }
1492
1493 static int
1494 valid_iamcu_cpu_flags (const i386_cpu_flags *flags)
1495 {
1496 if (cpu_arch_isa == PROCESSOR_IAMCU)
1497 {
1498 static const i386_cpu_flags iamcu_flags = CPU_IAMCU_COMPAT_FLAGS;
1499 i386_cpu_flags compat_flags;
1500 compat_flags = cpu_flags_and_not (*flags, iamcu_flags);
1501 return cpu_flags_all_zero (&compat_flags);
1502 }
1503 else
1504 return 1;
1505 }
1506
1507 #define CPU_FLAGS_ARCH_MATCH 0x1
1508 #define CPU_FLAGS_64BIT_MATCH 0x2
1509 #define CPU_FLAGS_AES_MATCH 0x4
1510 #define CPU_FLAGS_PCLMUL_MATCH 0x8
1511 #define CPU_FLAGS_AVX_MATCH 0x10
1512
1513 #define CPU_FLAGS_32BIT_MATCH \
1514 (CPU_FLAGS_ARCH_MATCH | CPU_FLAGS_AES_MATCH \
1515 | CPU_FLAGS_PCLMUL_MATCH | CPU_FLAGS_AVX_MATCH)
1516 #define CPU_FLAGS_PERFECT_MATCH \
1517 (CPU_FLAGS_32BIT_MATCH | CPU_FLAGS_64BIT_MATCH)
1518
1519 /* Return CPU flags match bits. */
1520
1521 static int
1522 cpu_flags_match (const insn_template *t)
1523 {
1524 i386_cpu_flags x = t->cpu_flags;
1525 int match = cpu_flags_check_cpu64 (x) ? CPU_FLAGS_64BIT_MATCH : 0;
1526
1527 x.bitfield.cpu64 = 0;
1528 x.bitfield.cpuno64 = 0;
1529
1530 if (cpu_flags_all_zero (&x))
1531 {
1532 /* This instruction is available on all archs. */
1533 match |= CPU_FLAGS_32BIT_MATCH;
1534 }
1535 else
1536 {
1537 /* This instruction is available only on some archs. */
1538 i386_cpu_flags cpu = cpu_arch_flags;
1539
1540 cpu.bitfield.cpu64 = 0;
1541 cpu.bitfield.cpuno64 = 0;
1542 cpu = cpu_flags_and (x, cpu);
1543 if (!cpu_flags_all_zero (&cpu))
1544 {
1545 if (x.bitfield.cpuavx)
1546 {
1547 /* We only need to check AES/PCLMUL/SSE2AVX with AVX. */
1548 if (cpu.bitfield.cpuavx)
1549 {
1550 /* Check SSE2AVX. */
1551 if (!t->opcode_modifier.sse2avx|| sse2avx)
1552 {
1553 match |= (CPU_FLAGS_ARCH_MATCH
1554 | CPU_FLAGS_AVX_MATCH);
1555 /* Check AES. */
1556 if (!x.bitfield.cpuaes || cpu.bitfield.cpuaes)
1557 match |= CPU_FLAGS_AES_MATCH;
1558 /* Check PCLMUL. */
1559 if (!x.bitfield.cpupclmul
1560 || cpu.bitfield.cpupclmul)
1561 match |= CPU_FLAGS_PCLMUL_MATCH;
1562 }
1563 }
1564 else
1565 match |= CPU_FLAGS_ARCH_MATCH;
1566 }
1567 else
1568 match |= CPU_FLAGS_32BIT_MATCH;
1569 }
1570 }
1571 return match;
1572 }
1573
1574 static INLINE i386_operand_type
1575 operand_type_and (i386_operand_type x, i386_operand_type y)
1576 {
1577 switch (ARRAY_SIZE (x.array))
1578 {
1579 case 3:
1580 x.array [2] &= y.array [2];
1581 case 2:
1582 x.array [1] &= y.array [1];
1583 case 1:
1584 x.array [0] &= y.array [0];
1585 break;
1586 default:
1587 abort ();
1588 }
1589 return x;
1590 }
1591
1592 static INLINE i386_operand_type
1593 operand_type_or (i386_operand_type x, i386_operand_type y)
1594 {
1595 switch (ARRAY_SIZE (x.array))
1596 {
1597 case 3:
1598 x.array [2] |= y.array [2];
1599 case 2:
1600 x.array [1] |= y.array [1];
1601 case 1:
1602 x.array [0] |= y.array [0];
1603 break;
1604 default:
1605 abort ();
1606 }
1607 return x;
1608 }
1609
1610 static INLINE i386_operand_type
1611 operand_type_xor (i386_operand_type x, i386_operand_type y)
1612 {
1613 switch (ARRAY_SIZE (x.array))
1614 {
1615 case 3:
1616 x.array [2] ^= y.array [2];
1617 case 2:
1618 x.array [1] ^= y.array [1];
1619 case 1:
1620 x.array [0] ^= y.array [0];
1621 break;
1622 default:
1623 abort ();
1624 }
1625 return x;
1626 }
1627
1628 static const i386_operand_type acc32 = OPERAND_TYPE_ACC32;
1629 static const i386_operand_type acc64 = OPERAND_TYPE_ACC64;
1630 static const i386_operand_type control = OPERAND_TYPE_CONTROL;
1631 static const i386_operand_type inoutportreg
1632 = OPERAND_TYPE_INOUTPORTREG;
1633 static const i386_operand_type reg16_inoutportreg
1634 = OPERAND_TYPE_REG16_INOUTPORTREG;
1635 static const i386_operand_type disp16 = OPERAND_TYPE_DISP16;
1636 static const i386_operand_type disp32 = OPERAND_TYPE_DISP32;
1637 static const i386_operand_type disp32s = OPERAND_TYPE_DISP32S;
1638 static const i386_operand_type disp16_32 = OPERAND_TYPE_DISP16_32;
1639 static const i386_operand_type anydisp
1640 = OPERAND_TYPE_ANYDISP;
1641 static const i386_operand_type regxmm = OPERAND_TYPE_REGXMM;
1642 static const i386_operand_type regymm = OPERAND_TYPE_REGYMM;
1643 static const i386_operand_type regzmm = OPERAND_TYPE_REGZMM;
1644 static const i386_operand_type regmask = OPERAND_TYPE_REGMASK;
1645 static const i386_operand_type imm8 = OPERAND_TYPE_IMM8;
1646 static const i386_operand_type imm8s = OPERAND_TYPE_IMM8S;
1647 static const i386_operand_type imm16 = OPERAND_TYPE_IMM16;
1648 static const i386_operand_type imm32 = OPERAND_TYPE_IMM32;
1649 static const i386_operand_type imm32s = OPERAND_TYPE_IMM32S;
1650 static const i386_operand_type imm64 = OPERAND_TYPE_IMM64;
1651 static const i386_operand_type imm16_32 = OPERAND_TYPE_IMM16_32;
1652 static const i386_operand_type imm16_32s = OPERAND_TYPE_IMM16_32S;
1653 static const i386_operand_type imm16_32_32s = OPERAND_TYPE_IMM16_32_32S;
1654 static const i386_operand_type vec_imm4 = OPERAND_TYPE_VEC_IMM4;
1655
1656 enum operand_type
1657 {
1658 reg,
1659 imm,
1660 disp,
1661 anymem
1662 };
1663
1664 static INLINE int
1665 operand_type_check (i386_operand_type t, enum operand_type c)
1666 {
1667 switch (c)
1668 {
1669 case reg:
1670 return (t.bitfield.reg8
1671 || t.bitfield.reg16
1672 || t.bitfield.reg32
1673 || t.bitfield.reg64);
1674
1675 case imm:
1676 return (t.bitfield.imm8
1677 || t.bitfield.imm8s
1678 || t.bitfield.imm16
1679 || t.bitfield.imm32
1680 || t.bitfield.imm32s
1681 || t.bitfield.imm64);
1682
1683 case disp:
1684 return (t.bitfield.disp8
1685 || t.bitfield.disp16
1686 || t.bitfield.disp32
1687 || t.bitfield.disp32s
1688 || t.bitfield.disp64);
1689
1690 case anymem:
1691 return (t.bitfield.disp8
1692 || t.bitfield.disp16
1693 || t.bitfield.disp32
1694 || t.bitfield.disp32s
1695 || t.bitfield.disp64
1696 || t.bitfield.baseindex);
1697
1698 default:
1699 abort ();
1700 }
1701
1702 return 0;
1703 }
1704
1705 /* Return 1 if there is no conflict in 8bit/16bit/32bit/64bit on
1706 operand J for instruction template T. */
1707
1708 static INLINE int
1709 match_reg_size (const insn_template *t, unsigned int j)
1710 {
1711 return !((i.types[j].bitfield.byte
1712 && !t->operand_types[j].bitfield.byte)
1713 || (i.types[j].bitfield.word
1714 && !t->operand_types[j].bitfield.word)
1715 || (i.types[j].bitfield.dword
1716 && !t->operand_types[j].bitfield.dword)
1717 || (i.types[j].bitfield.qword
1718 && !t->operand_types[j].bitfield.qword));
1719 }
1720
1721 /* Return 1 if there is no conflict in any size on operand J for
1722 instruction template T. */
1723
1724 static INLINE int
1725 match_mem_size (const insn_template *t, unsigned int j)
1726 {
1727 return (match_reg_size (t, j)
1728 && !((i.types[j].bitfield.unspecified
1729 && !i.broadcast
1730 && !t->operand_types[j].bitfield.unspecified)
1731 || (i.types[j].bitfield.fword
1732 && !t->operand_types[j].bitfield.fword)
1733 || (i.types[j].bitfield.tbyte
1734 && !t->operand_types[j].bitfield.tbyte)
1735 || (i.types[j].bitfield.xmmword
1736 && !t->operand_types[j].bitfield.xmmword)
1737 || (i.types[j].bitfield.ymmword
1738 && !t->operand_types[j].bitfield.ymmword)
1739 || (i.types[j].bitfield.zmmword
1740 && !t->operand_types[j].bitfield.zmmword)));
1741 }
1742
1743 /* Return 1 if there is no size conflict on any operands for
1744 instruction template T. */
1745
1746 static INLINE int
1747 operand_size_match (const insn_template *t)
1748 {
1749 unsigned int j;
1750 int match = 1;
1751
1752 /* Don't check jump instructions. */
1753 if (t->opcode_modifier.jump
1754 || t->opcode_modifier.jumpbyte
1755 || t->opcode_modifier.jumpdword
1756 || t->opcode_modifier.jumpintersegment)
1757 return match;
1758
1759 /* Check memory and accumulator operand size. */
1760 for (j = 0; j < i.operands; j++)
1761 {
1762 if (t->operand_types[j].bitfield.anysize)
1763 continue;
1764
1765 if (t->operand_types[j].bitfield.acc && !match_reg_size (t, j))
1766 {
1767 match = 0;
1768 break;
1769 }
1770
1771 if (i.types[j].bitfield.mem && !match_mem_size (t, j))
1772 {
1773 match = 0;
1774 break;
1775 }
1776 }
1777
1778 if (match)
1779 return match;
1780 else if (!t->opcode_modifier.d && !t->opcode_modifier.floatd)
1781 {
1782 mismatch:
1783 i.error = operand_size_mismatch;
1784 return 0;
1785 }
1786
1787 /* Check reverse. */
1788 gas_assert (i.operands == 2);
1789
1790 match = 1;
1791 for (j = 0; j < 2; j++)
1792 {
1793 if (t->operand_types[j].bitfield.acc
1794 && !match_reg_size (t, j ? 0 : 1))
1795 goto mismatch;
1796
1797 if (i.types[j].bitfield.mem
1798 && !match_mem_size (t, j ? 0 : 1))
1799 goto mismatch;
1800 }
1801
1802 return match;
1803 }
1804
1805 static INLINE int
1806 operand_type_match (i386_operand_type overlap,
1807 i386_operand_type given)
1808 {
1809 i386_operand_type temp = overlap;
1810
1811 temp.bitfield.jumpabsolute = 0;
1812 temp.bitfield.unspecified = 0;
1813 temp.bitfield.byte = 0;
1814 temp.bitfield.word = 0;
1815 temp.bitfield.dword = 0;
1816 temp.bitfield.fword = 0;
1817 temp.bitfield.qword = 0;
1818 temp.bitfield.tbyte = 0;
1819 temp.bitfield.xmmword = 0;
1820 temp.bitfield.ymmword = 0;
1821 temp.bitfield.zmmword = 0;
1822 if (operand_type_all_zero (&temp))
1823 goto mismatch;
1824
1825 if (given.bitfield.baseindex == overlap.bitfield.baseindex
1826 && given.bitfield.jumpabsolute == overlap.bitfield.jumpabsolute)
1827 return 1;
1828
1829 mismatch:
1830 i.error = operand_type_mismatch;
1831 return 0;
1832 }
1833
1834 /* If given types g0 and g1 are registers they must be of the same type
1835 unless the expected operand type register overlap is null.
1836 Note that Acc in a template matches every size of reg. */
1837
1838 static INLINE int
1839 operand_type_register_match (i386_operand_type m0,
1840 i386_operand_type g0,
1841 i386_operand_type t0,
1842 i386_operand_type m1,
1843 i386_operand_type g1,
1844 i386_operand_type t1)
1845 {
1846 if (!operand_type_check (g0, reg))
1847 return 1;
1848
1849 if (!operand_type_check (g1, reg))
1850 return 1;
1851
1852 if (g0.bitfield.reg8 == g1.bitfield.reg8
1853 && g0.bitfield.reg16 == g1.bitfield.reg16
1854 && g0.bitfield.reg32 == g1.bitfield.reg32
1855 && g0.bitfield.reg64 == g1.bitfield.reg64)
1856 return 1;
1857
1858 if (m0.bitfield.acc)
1859 {
1860 t0.bitfield.reg8 = 1;
1861 t0.bitfield.reg16 = 1;
1862 t0.bitfield.reg32 = 1;
1863 t0.bitfield.reg64 = 1;
1864 }
1865
1866 if (m1.bitfield.acc)
1867 {
1868 t1.bitfield.reg8 = 1;
1869 t1.bitfield.reg16 = 1;
1870 t1.bitfield.reg32 = 1;
1871 t1.bitfield.reg64 = 1;
1872 }
1873
1874 if (!(t0.bitfield.reg8 & t1.bitfield.reg8)
1875 && !(t0.bitfield.reg16 & t1.bitfield.reg16)
1876 && !(t0.bitfield.reg32 & t1.bitfield.reg32)
1877 && !(t0.bitfield.reg64 & t1.bitfield.reg64))
1878 return 1;
1879
1880 i.error = register_type_mismatch;
1881
1882 return 0;
1883 }
1884
1885 static INLINE unsigned int
1886 register_number (const reg_entry *r)
1887 {
1888 unsigned int nr = r->reg_num;
1889
1890 if (r->reg_flags & RegRex)
1891 nr += 8;
1892
1893 if (r->reg_flags & RegVRex)
1894 nr += 16;
1895
1896 return nr;
1897 }
1898
1899 static INLINE unsigned int
1900 mode_from_disp_size (i386_operand_type t)
1901 {
1902 if (t.bitfield.disp8 || t.bitfield.vec_disp8)
1903 return 1;
1904 else if (t.bitfield.disp16
1905 || t.bitfield.disp32
1906 || t.bitfield.disp32s)
1907 return 2;
1908 else
1909 return 0;
1910 }
1911
1912 static INLINE int
1913 fits_in_signed_byte (addressT num)
1914 {
1915 return num + 0x80 <= 0xff;
1916 }
1917
1918 static INLINE int
1919 fits_in_unsigned_byte (addressT num)
1920 {
1921 return num <= 0xff;
1922 }
1923
1924 static INLINE int
1925 fits_in_unsigned_word (addressT num)
1926 {
1927 return num <= 0xffff;
1928 }
1929
1930 static INLINE int
1931 fits_in_signed_word (addressT num)
1932 {
1933 return num + 0x8000 <= 0xffff;
1934 }
1935
1936 static INLINE int
1937 fits_in_signed_long (addressT num ATTRIBUTE_UNUSED)
1938 {
1939 #ifndef BFD64
1940 return 1;
1941 #else
1942 return num + 0x80000000 <= 0xffffffff;
1943 #endif
1944 } /* fits_in_signed_long() */
1945
1946 static INLINE int
1947 fits_in_unsigned_long (addressT num ATTRIBUTE_UNUSED)
1948 {
1949 #ifndef BFD64
1950 return 1;
1951 #else
1952 return num <= 0xffffffff;
1953 #endif
1954 } /* fits_in_unsigned_long() */
1955
1956 static INLINE int
1957 fits_in_vec_disp8 (offsetT num)
1958 {
1959 int shift = i.memshift;
1960 unsigned int mask;
1961
1962 if (shift == -1)
1963 abort ();
1964
1965 mask = (1 << shift) - 1;
1966
1967 /* Return 0 if NUM isn't properly aligned. */
1968 if ((num & mask))
1969 return 0;
1970
1971 /* Check if NUM will fit in 8bit after shift. */
1972 return fits_in_signed_byte (num >> shift);
1973 }
1974
1975 static INLINE int
1976 fits_in_imm4 (offsetT num)
1977 {
1978 return (num & 0xf) == num;
1979 }
1980
1981 static i386_operand_type
1982 smallest_imm_type (offsetT num)
1983 {
1984 i386_operand_type t;
1985
1986 operand_type_set (&t, 0);
1987 t.bitfield.imm64 = 1;
1988
1989 if (cpu_arch_tune != PROCESSOR_I486 && num == 1)
1990 {
1991 /* This code is disabled on the 486 because all the Imm1 forms
1992 in the opcode table are slower on the i486. They're the
1993 versions with the implicitly specified single-position
1994 displacement, which has another syntax if you really want to
1995 use that form. */
1996 t.bitfield.imm1 = 1;
1997 t.bitfield.imm8 = 1;
1998 t.bitfield.imm8s = 1;
1999 t.bitfield.imm16 = 1;
2000 t.bitfield.imm32 = 1;
2001 t.bitfield.imm32s = 1;
2002 }
2003 else if (fits_in_signed_byte (num))
2004 {
2005 t.bitfield.imm8 = 1;
2006 t.bitfield.imm8s = 1;
2007 t.bitfield.imm16 = 1;
2008 t.bitfield.imm32 = 1;
2009 t.bitfield.imm32s = 1;
2010 }
2011 else if (fits_in_unsigned_byte (num))
2012 {
2013 t.bitfield.imm8 = 1;
2014 t.bitfield.imm16 = 1;
2015 t.bitfield.imm32 = 1;
2016 t.bitfield.imm32s = 1;
2017 }
2018 else if (fits_in_signed_word (num) || fits_in_unsigned_word (num))
2019 {
2020 t.bitfield.imm16 = 1;
2021 t.bitfield.imm32 = 1;
2022 t.bitfield.imm32s = 1;
2023 }
2024 else if (fits_in_signed_long (num))
2025 {
2026 t.bitfield.imm32 = 1;
2027 t.bitfield.imm32s = 1;
2028 }
2029 else if (fits_in_unsigned_long (num))
2030 t.bitfield.imm32 = 1;
2031
2032 return t;
2033 }
2034
2035 static offsetT
2036 offset_in_range (offsetT val, int size)
2037 {
2038 addressT mask;
2039
2040 switch (size)
2041 {
2042 case 1: mask = ((addressT) 1 << 8) - 1; break;
2043 case 2: mask = ((addressT) 1 << 16) - 1; break;
2044 case 4: mask = ((addressT) 2 << 31) - 1; break;
2045 #ifdef BFD64
2046 case 8: mask = ((addressT) 2 << 63) - 1; break;
2047 #endif
2048 default: abort ();
2049 }
2050
2051 #ifdef BFD64
2052 /* If BFD64, sign extend val for 32bit address mode. */
2053 if (flag_code != CODE_64BIT
2054 || i.prefix[ADDR_PREFIX])
2055 if ((val & ~(((addressT) 2 << 31) - 1)) == 0)
2056 val = (val ^ ((addressT) 1 << 31)) - ((addressT) 1 << 31);
2057 #endif
2058
2059 if ((val & ~mask) != 0 && (val & ~mask) != ~mask)
2060 {
2061 char buf1[40], buf2[40];
2062
2063 sprint_value (buf1, val);
2064 sprint_value (buf2, val & mask);
2065 as_warn (_("%s shortened to %s"), buf1, buf2);
2066 }
2067 return val & mask;
2068 }
2069
2070 enum PREFIX_GROUP
2071 {
2072 PREFIX_EXIST = 0,
2073 PREFIX_LOCK,
2074 PREFIX_REP,
2075 PREFIX_OTHER
2076 };
2077
2078 /* Returns
2079 a. PREFIX_EXIST if attempting to add a prefix where one from the
2080 same class already exists.
2081 b. PREFIX_LOCK if lock prefix is added.
2082 c. PREFIX_REP if rep/repne prefix is added.
2083 d. PREFIX_OTHER if other prefix is added.
2084 */
2085
2086 static enum PREFIX_GROUP
2087 add_prefix (unsigned int prefix)
2088 {
2089 enum PREFIX_GROUP ret = PREFIX_OTHER;
2090 unsigned int q;
2091
2092 if (prefix >= REX_OPCODE && prefix < REX_OPCODE + 16
2093 && flag_code == CODE_64BIT)
2094 {
2095 if ((i.prefix[REX_PREFIX] & prefix & REX_W)
2096 || ((i.prefix[REX_PREFIX] & (REX_R | REX_X | REX_B))
2097 && (prefix & (REX_R | REX_X | REX_B))))
2098 ret = PREFIX_EXIST;
2099 q = REX_PREFIX;
2100 }
2101 else
2102 {
2103 switch (prefix)
2104 {
2105 default:
2106 abort ();
2107
2108 case CS_PREFIX_OPCODE:
2109 case DS_PREFIX_OPCODE:
2110 case ES_PREFIX_OPCODE:
2111 case FS_PREFIX_OPCODE:
2112 case GS_PREFIX_OPCODE:
2113 case SS_PREFIX_OPCODE:
2114 q = SEG_PREFIX;
2115 break;
2116
2117 case REPNE_PREFIX_OPCODE:
2118 case REPE_PREFIX_OPCODE:
2119 q = REP_PREFIX;
2120 ret = PREFIX_REP;
2121 break;
2122
2123 case LOCK_PREFIX_OPCODE:
2124 q = LOCK_PREFIX;
2125 ret = PREFIX_LOCK;
2126 break;
2127
2128 case FWAIT_OPCODE:
2129 q = WAIT_PREFIX;
2130 break;
2131
2132 case ADDR_PREFIX_OPCODE:
2133 q = ADDR_PREFIX;
2134 break;
2135
2136 case DATA_PREFIX_OPCODE:
2137 q = DATA_PREFIX;
2138 break;
2139 }
2140 if (i.prefix[q] != 0)
2141 ret = PREFIX_EXIST;
2142 }
2143
2144 if (ret)
2145 {
2146 if (!i.prefix[q])
2147 ++i.prefixes;
2148 i.prefix[q] |= prefix;
2149 }
2150 else
2151 as_bad (_("same type of prefix used twice"));
2152
2153 return ret;
2154 }
2155
2156 static void
2157 update_code_flag (int value, int check)
2158 {
2159 PRINTF_LIKE ((*as_error));
2160
2161 flag_code = (enum flag_code) value;
2162 if (flag_code == CODE_64BIT)
2163 {
2164 cpu_arch_flags.bitfield.cpu64 = 1;
2165 cpu_arch_flags.bitfield.cpuno64 = 0;
2166 }
2167 else
2168 {
2169 cpu_arch_flags.bitfield.cpu64 = 0;
2170 cpu_arch_flags.bitfield.cpuno64 = 1;
2171 }
2172 if (value == CODE_64BIT && !cpu_arch_flags.bitfield.cpulm )
2173 {
2174 if (check)
2175 as_error = as_fatal;
2176 else
2177 as_error = as_bad;
2178 (*as_error) (_("64bit mode not supported on `%s'."),
2179 cpu_arch_name ? cpu_arch_name : default_arch);
2180 }
2181 if (value == CODE_32BIT && !cpu_arch_flags.bitfield.cpui386)
2182 {
2183 if (check)
2184 as_error = as_fatal;
2185 else
2186 as_error = as_bad;
2187 (*as_error) (_("32bit mode not supported on `%s'."),
2188 cpu_arch_name ? cpu_arch_name : default_arch);
2189 }
2190 stackop_size = '\0';
2191 }
2192
2193 static void
2194 set_code_flag (int value)
2195 {
2196 update_code_flag (value, 0);
2197 }
2198
2199 static void
2200 set_16bit_gcc_code_flag (int new_code_flag)
2201 {
2202 flag_code = (enum flag_code) new_code_flag;
2203 if (flag_code != CODE_16BIT)
2204 abort ();
2205 cpu_arch_flags.bitfield.cpu64 = 0;
2206 cpu_arch_flags.bitfield.cpuno64 = 1;
2207 stackop_size = LONG_MNEM_SUFFIX;
2208 }
2209
2210 static void
2211 set_intel_syntax (int syntax_flag)
2212 {
2213 /* Find out if register prefixing is specified. */
2214 int ask_naked_reg = 0;
2215
2216 SKIP_WHITESPACE ();
2217 if (!is_end_of_line[(unsigned char) *input_line_pointer])
2218 {
2219 char *string;
2220 int e = get_symbol_name (&string);
2221
2222 if (strcmp (string, "prefix") == 0)
2223 ask_naked_reg = 1;
2224 else if (strcmp (string, "noprefix") == 0)
2225 ask_naked_reg = -1;
2226 else
2227 as_bad (_("bad argument to syntax directive."));
2228 (void) restore_line_pointer (e);
2229 }
2230 demand_empty_rest_of_line ();
2231
2232 intel_syntax = syntax_flag;
2233
2234 if (ask_naked_reg == 0)
2235 allow_naked_reg = (intel_syntax
2236 && (bfd_get_symbol_leading_char (stdoutput) != '\0'));
2237 else
2238 allow_naked_reg = (ask_naked_reg < 0);
2239
2240 expr_set_rank (O_full_ptr, syntax_flag ? 10 : 0);
2241
2242 identifier_chars['%'] = intel_syntax && allow_naked_reg ? '%' : 0;
2243 identifier_chars['$'] = intel_syntax ? '$' : 0;
2244 register_prefix = allow_naked_reg ? "" : "%";
2245 }
2246
2247 static void
2248 set_intel_mnemonic (int mnemonic_flag)
2249 {
2250 intel_mnemonic = mnemonic_flag;
2251 }
2252
2253 static void
2254 set_allow_index_reg (int flag)
2255 {
2256 allow_index_reg = flag;
2257 }
2258
2259 static void
2260 set_check (int what)
2261 {
2262 enum check_kind *kind;
2263 const char *str;
2264
2265 if (what)
2266 {
2267 kind = &operand_check;
2268 str = "operand";
2269 }
2270 else
2271 {
2272 kind = &sse_check;
2273 str = "sse";
2274 }
2275
2276 SKIP_WHITESPACE ();
2277
2278 if (!is_end_of_line[(unsigned char) *input_line_pointer])
2279 {
2280 char *string;
2281 int e = get_symbol_name (&string);
2282
2283 if (strcmp (string, "none") == 0)
2284 *kind = check_none;
2285 else if (strcmp (string, "warning") == 0)
2286 *kind = check_warning;
2287 else if (strcmp (string, "error") == 0)
2288 *kind = check_error;
2289 else
2290 as_bad (_("bad argument to %s_check directive."), str);
2291 (void) restore_line_pointer (e);
2292 }
2293 else
2294 as_bad (_("missing argument for %s_check directive"), str);
2295
2296 demand_empty_rest_of_line ();
2297 }
2298
2299 static void
2300 check_cpu_arch_compatible (const char *name ATTRIBUTE_UNUSED,
2301 i386_cpu_flags new_flag ATTRIBUTE_UNUSED)
2302 {
2303 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
2304 static const char *arch;
2305
2306 /* Intel LIOM is only supported on ELF. */
2307 if (!IS_ELF)
2308 return;
2309
2310 if (!arch)
2311 {
2312 /* Use cpu_arch_name if it is set in md_parse_option. Otherwise
2313 use default_arch. */
2314 arch = cpu_arch_name;
2315 if (!arch)
2316 arch = default_arch;
2317 }
2318
2319 /* If we are targeting Intel MCU, we must enable it. */
2320 if (get_elf_backend_data (stdoutput)->elf_machine_code != EM_IAMCU
2321 || new_flag.bitfield.cpuiamcu)
2322 return;
2323
2324 /* If we are targeting Intel L1OM, we must enable it. */
2325 if (get_elf_backend_data (stdoutput)->elf_machine_code != EM_L1OM
2326 || new_flag.bitfield.cpul1om)
2327 return;
2328
2329 /* If we are targeting Intel K1OM, we must enable it. */
2330 if (get_elf_backend_data (stdoutput)->elf_machine_code != EM_K1OM
2331 || new_flag.bitfield.cpuk1om)
2332 return;
2333
2334 as_bad (_("`%s' is not supported on `%s'"), name, arch);
2335 #endif
2336 }
2337
2338 static void
2339 set_cpu_arch (int dummy ATTRIBUTE_UNUSED)
2340 {
2341 SKIP_WHITESPACE ();
2342
2343 if (!is_end_of_line[(unsigned char) *input_line_pointer])
2344 {
2345 char *string;
2346 int e = get_symbol_name (&string);
2347 unsigned int j;
2348 i386_cpu_flags flags;
2349
2350 for (j = 0; j < ARRAY_SIZE (cpu_arch); j++)
2351 {
2352 if (strcmp (string, cpu_arch[j].name) == 0)
2353 {
2354 check_cpu_arch_compatible (string, cpu_arch[j].flags);
2355
2356 if (*string != '.')
2357 {
2358 cpu_arch_name = cpu_arch[j].name;
2359 cpu_sub_arch_name = NULL;
2360 cpu_arch_flags = cpu_arch[j].flags;
2361 if (flag_code == CODE_64BIT)
2362 {
2363 cpu_arch_flags.bitfield.cpu64 = 1;
2364 cpu_arch_flags.bitfield.cpuno64 = 0;
2365 }
2366 else
2367 {
2368 cpu_arch_flags.bitfield.cpu64 = 0;
2369 cpu_arch_flags.bitfield.cpuno64 = 1;
2370 }
2371 cpu_arch_isa = cpu_arch[j].type;
2372 cpu_arch_isa_flags = cpu_arch[j].flags;
2373 if (!cpu_arch_tune_set)
2374 {
2375 cpu_arch_tune = cpu_arch_isa;
2376 cpu_arch_tune_flags = cpu_arch_isa_flags;
2377 }
2378 break;
2379 }
2380
2381 flags = cpu_flags_or (cpu_arch_flags,
2382 cpu_arch[j].flags);
2383
2384 if (!valid_iamcu_cpu_flags (&flags))
2385 as_fatal (_("`%s' isn't valid for Intel MCU"),
2386 cpu_arch[j].name);
2387 else if (!cpu_flags_equal (&flags, &cpu_arch_flags))
2388 {
2389 if (cpu_sub_arch_name)
2390 {
2391 char *name = cpu_sub_arch_name;
2392 cpu_sub_arch_name = concat (name,
2393 cpu_arch[j].name,
2394 (const char *) NULL);
2395 free (name);
2396 }
2397 else
2398 cpu_sub_arch_name = xstrdup (cpu_arch[j].name);
2399 cpu_arch_flags = flags;
2400 cpu_arch_isa_flags = flags;
2401 }
2402 (void) restore_line_pointer (e);
2403 demand_empty_rest_of_line ();
2404 return;
2405 }
2406 }
2407
2408 if (*string == '.' && j >= ARRAY_SIZE (cpu_arch))
2409 {
2410 /* Disable an ISA entension. */
2411 for (j = 0; j < ARRAY_SIZE (cpu_noarch); j++)
2412 if (strcmp (string + 1, cpu_noarch [j].name) == 0)
2413 {
2414 flags = cpu_flags_and_not (cpu_arch_flags,
2415 cpu_noarch[j].flags);
2416 if (!cpu_flags_equal (&flags, &cpu_arch_flags))
2417 {
2418 if (cpu_sub_arch_name)
2419 {
2420 char *name = cpu_sub_arch_name;
2421 cpu_sub_arch_name = concat (name, string,
2422 (const char *) NULL);
2423 free (name);
2424 }
2425 else
2426 cpu_sub_arch_name = xstrdup (string);
2427 cpu_arch_flags = flags;
2428 cpu_arch_isa_flags = flags;
2429 }
2430 (void) restore_line_pointer (e);
2431 demand_empty_rest_of_line ();
2432 return;
2433 }
2434
2435 j = ARRAY_SIZE (cpu_arch);
2436 }
2437
2438 if (j >= ARRAY_SIZE (cpu_arch))
2439 as_bad (_("no such architecture: `%s'"), string);
2440
2441 *input_line_pointer = e;
2442 }
2443 else
2444 as_bad (_("missing cpu architecture"));
2445
2446 no_cond_jump_promotion = 0;
2447 if (*input_line_pointer == ','
2448 && !is_end_of_line[(unsigned char) input_line_pointer[1]])
2449 {
2450 char *string;
2451 char e;
2452
2453 ++input_line_pointer;
2454 e = get_symbol_name (&string);
2455
2456 if (strcmp (string, "nojumps") == 0)
2457 no_cond_jump_promotion = 1;
2458 else if (strcmp (string, "jumps") == 0)
2459 ;
2460 else
2461 as_bad (_("no such architecture modifier: `%s'"), string);
2462
2463 (void) restore_line_pointer (e);
2464 }
2465
2466 demand_empty_rest_of_line ();
2467 }
2468
2469 enum bfd_architecture
2470 i386_arch (void)
2471 {
2472 if (cpu_arch_isa == PROCESSOR_L1OM)
2473 {
2474 if (OUTPUT_FLAVOR != bfd_target_elf_flavour
2475 || flag_code != CODE_64BIT)
2476 as_fatal (_("Intel L1OM is 64bit ELF only"));
2477 return bfd_arch_l1om;
2478 }
2479 else if (cpu_arch_isa == PROCESSOR_K1OM)
2480 {
2481 if (OUTPUT_FLAVOR != bfd_target_elf_flavour
2482 || flag_code != CODE_64BIT)
2483 as_fatal (_("Intel K1OM is 64bit ELF only"));
2484 return bfd_arch_k1om;
2485 }
2486 else if (cpu_arch_isa == PROCESSOR_IAMCU)
2487 {
2488 if (OUTPUT_FLAVOR != bfd_target_elf_flavour
2489 || flag_code == CODE_64BIT)
2490 as_fatal (_("Intel MCU is 32bit ELF only"));
2491 return bfd_arch_iamcu;
2492 }
2493 else
2494 return bfd_arch_i386;
2495 }
2496
2497 unsigned long
2498 i386_mach (void)
2499 {
2500 if (!strncmp (default_arch, "x86_64", 6))
2501 {
2502 if (cpu_arch_isa == PROCESSOR_L1OM)
2503 {
2504 if (OUTPUT_FLAVOR != bfd_target_elf_flavour
2505 || default_arch[6] != '\0')
2506 as_fatal (_("Intel L1OM is 64bit ELF only"));
2507 return bfd_mach_l1om;
2508 }
2509 else if (cpu_arch_isa == PROCESSOR_K1OM)
2510 {
2511 if (OUTPUT_FLAVOR != bfd_target_elf_flavour
2512 || default_arch[6] != '\0')
2513 as_fatal (_("Intel K1OM is 64bit ELF only"));
2514 return bfd_mach_k1om;
2515 }
2516 else if (default_arch[6] == '\0')
2517 return bfd_mach_x86_64;
2518 else
2519 return bfd_mach_x64_32;
2520 }
2521 else if (!strcmp (default_arch, "i386")
2522 || !strcmp (default_arch, "iamcu"))
2523 {
2524 if (cpu_arch_isa == PROCESSOR_IAMCU)
2525 {
2526 if (OUTPUT_FLAVOR != bfd_target_elf_flavour)
2527 as_fatal (_("Intel MCU is 32bit ELF only"));
2528 return bfd_mach_i386_iamcu;
2529 }
2530 else
2531 return bfd_mach_i386_i386;
2532 }
2533 else
2534 as_fatal (_("unknown architecture"));
2535 }
2536 \f
2537 void
2538 md_begin (void)
2539 {
2540 const char *hash_err;
2541
2542 /* Initialize op_hash hash table. */
2543 op_hash = hash_new ();
2544
2545 {
2546 const insn_template *optab;
2547 templates *core_optab;
2548
2549 /* Setup for loop. */
2550 optab = i386_optab;
2551 core_optab = XNEW (templates);
2552 core_optab->start = optab;
2553
2554 while (1)
2555 {
2556 ++optab;
2557 if (optab->name == NULL
2558 || strcmp (optab->name, (optab - 1)->name) != 0)
2559 {
2560 /* different name --> ship out current template list;
2561 add to hash table; & begin anew. */
2562 core_optab->end = optab;
2563 hash_err = hash_insert (op_hash,
2564 (optab - 1)->name,
2565 (void *) core_optab);
2566 if (hash_err)
2567 {
2568 as_fatal (_("can't hash %s: %s"),
2569 (optab - 1)->name,
2570 hash_err);
2571 }
2572 if (optab->name == NULL)
2573 break;
2574 core_optab = XNEW (templates);
2575 core_optab->start = optab;
2576 }
2577 }
2578 }
2579
2580 /* Initialize reg_hash hash table. */
2581 reg_hash = hash_new ();
2582 {
2583 const reg_entry *regtab;
2584 unsigned int regtab_size = i386_regtab_size;
2585
2586 for (regtab = i386_regtab; regtab_size--; regtab++)
2587 {
2588 hash_err = hash_insert (reg_hash, regtab->reg_name, (void *) regtab);
2589 if (hash_err)
2590 as_fatal (_("can't hash %s: %s"),
2591 regtab->reg_name,
2592 hash_err);
2593 }
2594 }
2595
2596 /* Fill in lexical tables: mnemonic_chars, operand_chars. */
2597 {
2598 int c;
2599 char *p;
2600
2601 for (c = 0; c < 256; c++)
2602 {
2603 if (ISDIGIT (c))
2604 {
2605 digit_chars[c] = c;
2606 mnemonic_chars[c] = c;
2607 register_chars[c] = c;
2608 operand_chars[c] = c;
2609 }
2610 else if (ISLOWER (c))
2611 {
2612 mnemonic_chars[c] = c;
2613 register_chars[c] = c;
2614 operand_chars[c] = c;
2615 }
2616 else if (ISUPPER (c))
2617 {
2618 mnemonic_chars[c] = TOLOWER (c);
2619 register_chars[c] = mnemonic_chars[c];
2620 operand_chars[c] = c;
2621 }
2622 else if (c == '{' || c == '}')
2623 operand_chars[c] = c;
2624
2625 if (ISALPHA (c) || ISDIGIT (c))
2626 identifier_chars[c] = c;
2627 else if (c >= 128)
2628 {
2629 identifier_chars[c] = c;
2630 operand_chars[c] = c;
2631 }
2632 }
2633
2634 #ifdef LEX_AT
2635 identifier_chars['@'] = '@';
2636 #endif
2637 #ifdef LEX_QM
2638 identifier_chars['?'] = '?';
2639 operand_chars['?'] = '?';
2640 #endif
2641 digit_chars['-'] = '-';
2642 mnemonic_chars['_'] = '_';
2643 mnemonic_chars['-'] = '-';
2644 mnemonic_chars['.'] = '.';
2645 identifier_chars['_'] = '_';
2646 identifier_chars['.'] = '.';
2647
2648 for (p = operand_special_chars; *p != '\0'; p++)
2649 operand_chars[(unsigned char) *p] = *p;
2650 }
2651
2652 if (flag_code == CODE_64BIT)
2653 {
2654 #if defined (OBJ_COFF) && defined (TE_PE)
2655 x86_dwarf2_return_column = (OUTPUT_FLAVOR == bfd_target_coff_flavour
2656 ? 32 : 16);
2657 #else
2658 x86_dwarf2_return_column = 16;
2659 #endif
2660 x86_cie_data_alignment = -8;
2661 }
2662 else
2663 {
2664 x86_dwarf2_return_column = 8;
2665 x86_cie_data_alignment = -4;
2666 }
2667 }
2668
2669 void
2670 i386_print_statistics (FILE *file)
2671 {
2672 hash_print_statistics (file, "i386 opcode", op_hash);
2673 hash_print_statistics (file, "i386 register", reg_hash);
2674 }
2675 \f
2676 #ifdef DEBUG386
2677
2678 /* Debugging routines for md_assemble. */
2679 static void pte (insn_template *);
2680 static void pt (i386_operand_type);
2681 static void pe (expressionS *);
2682 static void ps (symbolS *);
2683
2684 static void
2685 pi (char *line, i386_insn *x)
2686 {
2687 unsigned int j;
2688
2689 fprintf (stdout, "%s: template ", line);
2690 pte (&x->tm);
2691 fprintf (stdout, " address: base %s index %s scale %x\n",
2692 x->base_reg ? x->base_reg->reg_name : "none",
2693 x->index_reg ? x->index_reg->reg_name : "none",
2694 x->log2_scale_factor);
2695 fprintf (stdout, " modrm: mode %x reg %x reg/mem %x\n",
2696 x->rm.mode, x->rm.reg, x->rm.regmem);
2697 fprintf (stdout, " sib: base %x index %x scale %x\n",
2698 x->sib.base, x->sib.index, x->sib.scale);
2699 fprintf (stdout, " rex: 64bit %x extX %x extY %x extZ %x\n",
2700 (x->rex & REX_W) != 0,
2701 (x->rex & REX_R) != 0,
2702 (x->rex & REX_X) != 0,
2703 (x->rex & REX_B) != 0);
2704 for (j = 0; j < x->operands; j++)
2705 {
2706 fprintf (stdout, " #%d: ", j + 1);
2707 pt (x->types[j]);
2708 fprintf (stdout, "\n");
2709 if (x->types[j].bitfield.reg8
2710 || x->types[j].bitfield.reg16
2711 || x->types[j].bitfield.reg32
2712 || x->types[j].bitfield.reg64
2713 || x->types[j].bitfield.regmmx
2714 || x->types[j].bitfield.regxmm
2715 || x->types[j].bitfield.regymm
2716 || x->types[j].bitfield.regzmm
2717 || x->types[j].bitfield.sreg2
2718 || x->types[j].bitfield.sreg3
2719 || x->types[j].bitfield.control
2720 || x->types[j].bitfield.debug
2721 || x->types[j].bitfield.test)
2722 fprintf (stdout, "%s\n", x->op[j].regs->reg_name);
2723 if (operand_type_check (x->types[j], imm))
2724 pe (x->op[j].imms);
2725 if (operand_type_check (x->types[j], disp))
2726 pe (x->op[j].disps);
2727 }
2728 }
2729
2730 static void
2731 pte (insn_template *t)
2732 {
2733 unsigned int j;
2734 fprintf (stdout, " %d operands ", t->operands);
2735 fprintf (stdout, "opcode %x ", t->base_opcode);
2736 if (t->extension_opcode != None)
2737 fprintf (stdout, "ext %x ", t->extension_opcode);
2738 if (t->opcode_modifier.d)
2739 fprintf (stdout, "D");
2740 if (t->opcode_modifier.w)
2741 fprintf (stdout, "W");
2742 fprintf (stdout, "\n");
2743 for (j = 0; j < t->operands; j++)
2744 {
2745 fprintf (stdout, " #%d type ", j + 1);
2746 pt (t->operand_types[j]);
2747 fprintf (stdout, "\n");
2748 }
2749 }
2750
2751 static void
2752 pe (expressionS *e)
2753 {
2754 fprintf (stdout, " operation %d\n", e->X_op);
2755 fprintf (stdout, " add_number %ld (%lx)\n",
2756 (long) e->X_add_number, (long) e->X_add_number);
2757 if (e->X_add_symbol)
2758 {
2759 fprintf (stdout, " add_symbol ");
2760 ps (e->X_add_symbol);
2761 fprintf (stdout, "\n");
2762 }
2763 if (e->X_op_symbol)
2764 {
2765 fprintf (stdout, " op_symbol ");
2766 ps (e->X_op_symbol);
2767 fprintf (stdout, "\n");
2768 }
2769 }
2770
2771 static void
2772 ps (symbolS *s)
2773 {
2774 fprintf (stdout, "%s type %s%s",
2775 S_GET_NAME (s),
2776 S_IS_EXTERNAL (s) ? "EXTERNAL " : "",
2777 segment_name (S_GET_SEGMENT (s)));
2778 }
2779
2780 static struct type_name
2781 {
2782 i386_operand_type mask;
2783 const char *name;
2784 }
2785 const type_names[] =
2786 {
2787 { OPERAND_TYPE_REG8, "r8" },
2788 { OPERAND_TYPE_REG16, "r16" },
2789 { OPERAND_TYPE_REG32, "r32" },
2790 { OPERAND_TYPE_REG64, "r64" },
2791 { OPERAND_TYPE_IMM8, "i8" },
2792 { OPERAND_TYPE_IMM8, "i8s" },
2793 { OPERAND_TYPE_IMM16, "i16" },
2794 { OPERAND_TYPE_IMM32, "i32" },
2795 { OPERAND_TYPE_IMM32S, "i32s" },
2796 { OPERAND_TYPE_IMM64, "i64" },
2797 { OPERAND_TYPE_IMM1, "i1" },
2798 { OPERAND_TYPE_BASEINDEX, "BaseIndex" },
2799 { OPERAND_TYPE_DISP8, "d8" },
2800 { OPERAND_TYPE_DISP16, "d16" },
2801 { OPERAND_TYPE_DISP32, "d32" },
2802 { OPERAND_TYPE_DISP32S, "d32s" },
2803 { OPERAND_TYPE_DISP64, "d64" },
2804 { OPERAND_TYPE_VEC_DISP8, "Vector d8" },
2805 { OPERAND_TYPE_INOUTPORTREG, "InOutPortReg" },
2806 { OPERAND_TYPE_SHIFTCOUNT, "ShiftCount" },
2807 { OPERAND_TYPE_CONTROL, "control reg" },
2808 { OPERAND_TYPE_TEST, "test reg" },
2809 { OPERAND_TYPE_DEBUG, "debug reg" },
2810 { OPERAND_TYPE_FLOATREG, "FReg" },
2811 { OPERAND_TYPE_FLOATACC, "FAcc" },
2812 { OPERAND_TYPE_SREG2, "SReg2" },
2813 { OPERAND_TYPE_SREG3, "SReg3" },
2814 { OPERAND_TYPE_ACC, "Acc" },
2815 { OPERAND_TYPE_JUMPABSOLUTE, "Jump Absolute" },
2816 { OPERAND_TYPE_REGMMX, "rMMX" },
2817 { OPERAND_TYPE_REGXMM, "rXMM" },
2818 { OPERAND_TYPE_REGYMM, "rYMM" },
2819 { OPERAND_TYPE_REGZMM, "rZMM" },
2820 { OPERAND_TYPE_REGMASK, "Mask reg" },
2821 { OPERAND_TYPE_ESSEG, "es" },
2822 };
2823
2824 static void
2825 pt (i386_operand_type t)
2826 {
2827 unsigned int j;
2828 i386_operand_type a;
2829
2830 for (j = 0; j < ARRAY_SIZE (type_names); j++)
2831 {
2832 a = operand_type_and (t, type_names[j].mask);
2833 if (!operand_type_all_zero (&a))
2834 fprintf (stdout, "%s, ", type_names[j].name);
2835 }
2836 fflush (stdout);
2837 }
2838
2839 #endif /* DEBUG386 */
2840 \f
2841 static bfd_reloc_code_real_type
2842 reloc (unsigned int size,
2843 int pcrel,
2844 int sign,
2845 bfd_reloc_code_real_type other)
2846 {
2847 if (other != NO_RELOC)
2848 {
2849 reloc_howto_type *rel;
2850
2851 if (size == 8)
2852 switch (other)
2853 {
2854 case BFD_RELOC_X86_64_GOT32:
2855 return BFD_RELOC_X86_64_GOT64;
2856 break;
2857 case BFD_RELOC_X86_64_GOTPLT64:
2858 return BFD_RELOC_X86_64_GOTPLT64;
2859 break;
2860 case BFD_RELOC_X86_64_PLTOFF64:
2861 return BFD_RELOC_X86_64_PLTOFF64;
2862 break;
2863 case BFD_RELOC_X86_64_GOTPC32:
2864 other = BFD_RELOC_X86_64_GOTPC64;
2865 break;
2866 case BFD_RELOC_X86_64_GOTPCREL:
2867 other = BFD_RELOC_X86_64_GOTPCREL64;
2868 break;
2869 case BFD_RELOC_X86_64_TPOFF32:
2870 other = BFD_RELOC_X86_64_TPOFF64;
2871 break;
2872 case BFD_RELOC_X86_64_DTPOFF32:
2873 other = BFD_RELOC_X86_64_DTPOFF64;
2874 break;
2875 default:
2876 break;
2877 }
2878
2879 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
2880 if (other == BFD_RELOC_SIZE32)
2881 {
2882 if (size == 8)
2883 other = BFD_RELOC_SIZE64;
2884 if (pcrel)
2885 {
2886 as_bad (_("there are no pc-relative size relocations"));
2887 return NO_RELOC;
2888 }
2889 }
2890 #endif
2891
2892 /* Sign-checking 4-byte relocations in 16-/32-bit code is pointless. */
2893 if (size == 4 && (flag_code != CODE_64BIT || disallow_64bit_reloc))
2894 sign = -1;
2895
2896 rel = bfd_reloc_type_lookup (stdoutput, other);
2897 if (!rel)
2898 as_bad (_("unknown relocation (%u)"), other);
2899 else if (size != bfd_get_reloc_size (rel))
2900 as_bad (_("%u-byte relocation cannot be applied to %u-byte field"),
2901 bfd_get_reloc_size (rel),
2902 size);
2903 else if (pcrel && !rel->pc_relative)
2904 as_bad (_("non-pc-relative relocation for pc-relative field"));
2905 else if ((rel->complain_on_overflow == complain_overflow_signed
2906 && !sign)
2907 || (rel->complain_on_overflow == complain_overflow_unsigned
2908 && sign > 0))
2909 as_bad (_("relocated field and relocation type differ in signedness"));
2910 else
2911 return other;
2912 return NO_RELOC;
2913 }
2914
2915 if (pcrel)
2916 {
2917 if (!sign)
2918 as_bad (_("there are no unsigned pc-relative relocations"));
2919 switch (size)
2920 {
2921 case 1: return BFD_RELOC_8_PCREL;
2922 case 2: return BFD_RELOC_16_PCREL;
2923 case 4: return BFD_RELOC_32_PCREL;
2924 case 8: return BFD_RELOC_64_PCREL;
2925 }
2926 as_bad (_("cannot do %u byte pc-relative relocation"), size);
2927 }
2928 else
2929 {
2930 if (sign > 0)
2931 switch (size)
2932 {
2933 case 4: return BFD_RELOC_X86_64_32S;
2934 }
2935 else
2936 switch (size)
2937 {
2938 case 1: return BFD_RELOC_8;
2939 case 2: return BFD_RELOC_16;
2940 case 4: return BFD_RELOC_32;
2941 case 8: return BFD_RELOC_64;
2942 }
2943 as_bad (_("cannot do %s %u byte relocation"),
2944 sign > 0 ? "signed" : "unsigned", size);
2945 }
2946
2947 return NO_RELOC;
2948 }
2949
2950 /* Here we decide which fixups can be adjusted to make them relative to
2951 the beginning of the section instead of the symbol. Basically we need
2952 to make sure that the dynamic relocations are done correctly, so in
2953 some cases we force the original symbol to be used. */
2954
2955 int
2956 tc_i386_fix_adjustable (fixS *fixP ATTRIBUTE_UNUSED)
2957 {
2958 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
2959 if (!IS_ELF)
2960 return 1;
2961
2962 /* Don't adjust pc-relative references to merge sections in 64-bit
2963 mode. */
2964 if (use_rela_relocations
2965 && (S_GET_SEGMENT (fixP->fx_addsy)->flags & SEC_MERGE) != 0
2966 && fixP->fx_pcrel)
2967 return 0;
2968
2969 /* The x86_64 GOTPCREL are represented as 32bit PCrel relocations
2970 and changed later by validate_fix. */
2971 if (GOT_symbol && fixP->fx_subsy == GOT_symbol
2972 && fixP->fx_r_type == BFD_RELOC_32_PCREL)
2973 return 0;
2974
2975 /* Adjust_reloc_syms doesn't know about the GOT. Need to keep symbol
2976 for size relocations. */
2977 if (fixP->fx_r_type == BFD_RELOC_SIZE32
2978 || fixP->fx_r_type == BFD_RELOC_SIZE64
2979 || fixP->fx_r_type == BFD_RELOC_386_GOTOFF
2980 || fixP->fx_r_type == BFD_RELOC_386_PLT32
2981 || fixP->fx_r_type == BFD_RELOC_386_GOT32
2982 || fixP->fx_r_type == BFD_RELOC_386_GOT32X
2983 || fixP->fx_r_type == BFD_RELOC_386_TLS_GD
2984 || fixP->fx_r_type == BFD_RELOC_386_TLS_LDM
2985 || fixP->fx_r_type == BFD_RELOC_386_TLS_LDO_32
2986 || fixP->fx_r_type == BFD_RELOC_386_TLS_IE_32
2987 || fixP->fx_r_type == BFD_RELOC_386_TLS_IE
2988 || fixP->fx_r_type == BFD_RELOC_386_TLS_GOTIE
2989 || fixP->fx_r_type == BFD_RELOC_386_TLS_LE_32
2990 || fixP->fx_r_type == BFD_RELOC_386_TLS_LE
2991 || fixP->fx_r_type == BFD_RELOC_386_TLS_GOTDESC
2992 || fixP->fx_r_type == BFD_RELOC_386_TLS_DESC_CALL
2993 || fixP->fx_r_type == BFD_RELOC_X86_64_PLT32
2994 || fixP->fx_r_type == BFD_RELOC_X86_64_GOT32
2995 || fixP->fx_r_type == BFD_RELOC_X86_64_GOTPCREL
2996 || fixP->fx_r_type == BFD_RELOC_X86_64_GOTPCRELX
2997 || fixP->fx_r_type == BFD_RELOC_X86_64_REX_GOTPCRELX
2998 || fixP->fx_r_type == BFD_RELOC_X86_64_TLSGD
2999 || fixP->fx_r_type == BFD_RELOC_X86_64_TLSLD
3000 || fixP->fx_r_type == BFD_RELOC_X86_64_DTPOFF32
3001 || fixP->fx_r_type == BFD_RELOC_X86_64_DTPOFF64
3002 || fixP->fx_r_type == BFD_RELOC_X86_64_GOTTPOFF
3003 || fixP->fx_r_type == BFD_RELOC_X86_64_TPOFF32
3004 || fixP->fx_r_type == BFD_RELOC_X86_64_TPOFF64
3005 || fixP->fx_r_type == BFD_RELOC_X86_64_GOTOFF64
3006 || fixP->fx_r_type == BFD_RELOC_X86_64_GOTPC32_TLSDESC
3007 || fixP->fx_r_type == BFD_RELOC_X86_64_TLSDESC_CALL
3008 || fixP->fx_r_type == BFD_RELOC_VTABLE_INHERIT
3009 || fixP->fx_r_type == BFD_RELOC_VTABLE_ENTRY)
3010 return 0;
3011 #endif
3012 return 1;
3013 }
3014
3015 static int
3016 intel_float_operand (const char *mnemonic)
3017 {
3018 /* Note that the value returned is meaningful only for opcodes with (memory)
3019 operands, hence the code here is free to improperly handle opcodes that
3020 have no operands (for better performance and smaller code). */
3021
3022 if (mnemonic[0] != 'f')
3023 return 0; /* non-math */
3024
3025 switch (mnemonic[1])
3026 {
3027 /* fclex, fdecstp, fdisi, femms, feni, fincstp, finit, fsetpm, and
3028 the fs segment override prefix not currently handled because no
3029 call path can make opcodes without operands get here */
3030 case 'i':
3031 return 2 /* integer op */;
3032 case 'l':
3033 if (mnemonic[2] == 'd' && (mnemonic[3] == 'c' || mnemonic[3] == 'e'))
3034 return 3; /* fldcw/fldenv */
3035 break;
3036 case 'n':
3037 if (mnemonic[2] != 'o' /* fnop */)
3038 return 3; /* non-waiting control op */
3039 break;
3040 case 'r':
3041 if (mnemonic[2] == 's')
3042 return 3; /* frstor/frstpm */
3043 break;
3044 case 's':
3045 if (mnemonic[2] == 'a')
3046 return 3; /* fsave */
3047 if (mnemonic[2] == 't')
3048 {
3049 switch (mnemonic[3])
3050 {
3051 case 'c': /* fstcw */
3052 case 'd': /* fstdw */
3053 case 'e': /* fstenv */
3054 case 's': /* fsts[gw] */
3055 return 3;
3056 }
3057 }
3058 break;
3059 case 'x':
3060 if (mnemonic[2] == 'r' || mnemonic[2] == 's')
3061 return 0; /* fxsave/fxrstor are not really math ops */
3062 break;
3063 }
3064
3065 return 1;
3066 }
3067
3068 /* Build the VEX prefix. */
3069
3070 static void
3071 build_vex_prefix (const insn_template *t)
3072 {
3073 unsigned int register_specifier;
3074 unsigned int implied_prefix;
3075 unsigned int vector_length;
3076
3077 /* Check register specifier. */
3078 if (i.vex.register_specifier)
3079 {
3080 register_specifier =
3081 ~register_number (i.vex.register_specifier) & 0xf;
3082 gas_assert ((i.vex.register_specifier->reg_flags & RegVRex) == 0);
3083 }
3084 else
3085 register_specifier = 0xf;
3086
3087 /* Use 2-byte VEX prefix by swappping destination and source
3088 operand. */
3089 if (!i.swap_operand
3090 && i.operands == i.reg_operands
3091 && i.tm.opcode_modifier.vexopcode == VEX0F
3092 && i.tm.opcode_modifier.s
3093 && i.rex == REX_B)
3094 {
3095 unsigned int xchg = i.operands - 1;
3096 union i386_op temp_op;
3097 i386_operand_type temp_type;
3098
3099 temp_type = i.types[xchg];
3100 i.types[xchg] = i.types[0];
3101 i.types[0] = temp_type;
3102 temp_op = i.op[xchg];
3103 i.op[xchg] = i.op[0];
3104 i.op[0] = temp_op;
3105
3106 gas_assert (i.rm.mode == 3);
3107
3108 i.rex = REX_R;
3109 xchg = i.rm.regmem;
3110 i.rm.regmem = i.rm.reg;
3111 i.rm.reg = xchg;
3112
3113 /* Use the next insn. */
3114 i.tm = t[1];
3115 }
3116
3117 if (i.tm.opcode_modifier.vex == VEXScalar)
3118 vector_length = avxscalar;
3119 else
3120 vector_length = i.tm.opcode_modifier.vex == VEX256 ? 1 : 0;
3121
3122 switch ((i.tm.base_opcode >> 8) & 0xff)
3123 {
3124 case 0:
3125 implied_prefix = 0;
3126 break;
3127 case DATA_PREFIX_OPCODE:
3128 implied_prefix = 1;
3129 break;
3130 case REPE_PREFIX_OPCODE:
3131 implied_prefix = 2;
3132 break;
3133 case REPNE_PREFIX_OPCODE:
3134 implied_prefix = 3;
3135 break;
3136 default:
3137 abort ();
3138 }
3139
3140 /* Use 2-byte VEX prefix if possible. */
3141 if (i.tm.opcode_modifier.vexopcode == VEX0F
3142 && i.tm.opcode_modifier.vexw != VEXW1
3143 && (i.rex & (REX_W | REX_X | REX_B)) == 0)
3144 {
3145 /* 2-byte VEX prefix. */
3146 unsigned int r;
3147
3148 i.vex.length = 2;
3149 i.vex.bytes[0] = 0xc5;
3150
3151 /* Check the REX.R bit. */
3152 r = (i.rex & REX_R) ? 0 : 1;
3153 i.vex.bytes[1] = (r << 7
3154 | register_specifier << 3
3155 | vector_length << 2
3156 | implied_prefix);
3157 }
3158 else
3159 {
3160 /* 3-byte VEX prefix. */
3161 unsigned int m, w;
3162
3163 i.vex.length = 3;
3164
3165 switch (i.tm.opcode_modifier.vexopcode)
3166 {
3167 case VEX0F:
3168 m = 0x1;
3169 i.vex.bytes[0] = 0xc4;
3170 break;
3171 case VEX0F38:
3172 m = 0x2;
3173 i.vex.bytes[0] = 0xc4;
3174 break;
3175 case VEX0F3A:
3176 m = 0x3;
3177 i.vex.bytes[0] = 0xc4;
3178 break;
3179 case XOP08:
3180 m = 0x8;
3181 i.vex.bytes[0] = 0x8f;
3182 break;
3183 case XOP09:
3184 m = 0x9;
3185 i.vex.bytes[0] = 0x8f;
3186 break;
3187 case XOP0A:
3188 m = 0xa;
3189 i.vex.bytes[0] = 0x8f;
3190 break;
3191 default:
3192 abort ();
3193 }
3194
3195 /* The high 3 bits of the second VEX byte are 1's compliment
3196 of RXB bits from REX. */
3197 i.vex.bytes[1] = (~i.rex & 0x7) << 5 | m;
3198
3199 /* Check the REX.W bit. */
3200 w = (i.rex & REX_W) ? 1 : 0;
3201 if (i.tm.opcode_modifier.vexw == VEXW1)
3202 w = 1;
3203
3204 i.vex.bytes[2] = (w << 7
3205 | register_specifier << 3
3206 | vector_length << 2
3207 | implied_prefix);
3208 }
3209 }
3210
3211 /* Build the EVEX prefix. */
3212
3213 static void
3214 build_evex_prefix (void)
3215 {
3216 unsigned int register_specifier;
3217 unsigned int implied_prefix;
3218 unsigned int m, w;
3219 rex_byte vrex_used = 0;
3220
3221 /* Check register specifier. */
3222 if (i.vex.register_specifier)
3223 {
3224 gas_assert ((i.vrex & REX_X) == 0);
3225
3226 register_specifier = i.vex.register_specifier->reg_num;
3227 if ((i.vex.register_specifier->reg_flags & RegRex))
3228 register_specifier += 8;
3229 /* The upper 16 registers are encoded in the fourth byte of the
3230 EVEX prefix. */
3231 if (!(i.vex.register_specifier->reg_flags & RegVRex))
3232 i.vex.bytes[3] = 0x8;
3233 register_specifier = ~register_specifier & 0xf;
3234 }
3235 else
3236 {
3237 register_specifier = 0xf;
3238
3239 /* Encode upper 16 vector index register in the fourth byte of
3240 the EVEX prefix. */
3241 if (!(i.vrex & REX_X))
3242 i.vex.bytes[3] = 0x8;
3243 else
3244 vrex_used |= REX_X;
3245 }
3246
3247 switch ((i.tm.base_opcode >> 8) & 0xff)
3248 {
3249 case 0:
3250 implied_prefix = 0;
3251 break;
3252 case DATA_PREFIX_OPCODE:
3253 implied_prefix = 1;
3254 break;
3255 case REPE_PREFIX_OPCODE:
3256 implied_prefix = 2;
3257 break;
3258 case REPNE_PREFIX_OPCODE:
3259 implied_prefix = 3;
3260 break;
3261 default:
3262 abort ();
3263 }
3264
3265 /* 4 byte EVEX prefix. */
3266 i.vex.length = 4;
3267 i.vex.bytes[0] = 0x62;
3268
3269 /* mmmm bits. */
3270 switch (i.tm.opcode_modifier.vexopcode)
3271 {
3272 case VEX0F:
3273 m = 1;
3274 break;
3275 case VEX0F38:
3276 m = 2;
3277 break;
3278 case VEX0F3A:
3279 m = 3;
3280 break;
3281 default:
3282 abort ();
3283 break;
3284 }
3285
3286 /* The high 3 bits of the second EVEX byte are 1's compliment of RXB
3287 bits from REX. */
3288 i.vex.bytes[1] = (~i.rex & 0x7) << 5 | m;
3289
3290 /* The fifth bit of the second EVEX byte is 1's compliment of the
3291 REX_R bit in VREX. */
3292 if (!(i.vrex & REX_R))
3293 i.vex.bytes[1] |= 0x10;
3294 else
3295 vrex_used |= REX_R;
3296
3297 if ((i.reg_operands + i.imm_operands) == i.operands)
3298 {
3299 /* When all operands are registers, the REX_X bit in REX is not
3300 used. We reuse it to encode the upper 16 registers, which is
3301 indicated by the REX_B bit in VREX. The REX_X bit is encoded
3302 as 1's compliment. */
3303 if ((i.vrex & REX_B))
3304 {
3305 vrex_used |= REX_B;
3306 i.vex.bytes[1] &= ~0x40;
3307 }
3308 }
3309
3310 /* EVEX instructions shouldn't need the REX prefix. */
3311 i.vrex &= ~vrex_used;
3312 gas_assert (i.vrex == 0);
3313
3314 /* Check the REX.W bit. */
3315 w = (i.rex & REX_W) ? 1 : 0;
3316 if (i.tm.opcode_modifier.vexw)
3317 {
3318 if (i.tm.opcode_modifier.vexw == VEXW1)
3319 w = 1;
3320 }
3321 /* If w is not set it means we are dealing with WIG instruction. */
3322 else if (!w)
3323 {
3324 if (evexwig == evexw1)
3325 w = 1;
3326 }
3327
3328 /* Encode the U bit. */
3329 implied_prefix |= 0x4;
3330
3331 /* The third byte of the EVEX prefix. */
3332 i.vex.bytes[2] = (w << 7 | register_specifier << 3 | implied_prefix);
3333
3334 /* The fourth byte of the EVEX prefix. */
3335 /* The zeroing-masking bit. */
3336 if (i.mask && i.mask->zeroing)
3337 i.vex.bytes[3] |= 0x80;
3338
3339 /* Don't always set the broadcast bit if there is no RC. */
3340 if (!i.rounding)
3341 {
3342 /* Encode the vector length. */
3343 unsigned int vec_length;
3344
3345 switch (i.tm.opcode_modifier.evex)
3346 {
3347 case EVEXLIG: /* LL' is ignored */
3348 vec_length = evexlig << 5;
3349 break;
3350 case EVEX128:
3351 vec_length = 0 << 5;
3352 break;
3353 case EVEX256:
3354 vec_length = 1 << 5;
3355 break;
3356 case EVEX512:
3357 vec_length = 2 << 5;
3358 break;
3359 default:
3360 abort ();
3361 break;
3362 }
3363 i.vex.bytes[3] |= vec_length;
3364 /* Encode the broadcast bit. */
3365 if (i.broadcast)
3366 i.vex.bytes[3] |= 0x10;
3367 }
3368 else
3369 {
3370 if (i.rounding->type != saeonly)
3371 i.vex.bytes[3] |= 0x10 | (i.rounding->type << 5);
3372 else
3373 i.vex.bytes[3] |= 0x10 | (evexrcig << 5);
3374 }
3375
3376 if (i.mask && i.mask->mask)
3377 i.vex.bytes[3] |= i.mask->mask->reg_num;
3378 }
3379
3380 static void
3381 process_immext (void)
3382 {
3383 expressionS *exp;
3384
3385 if ((i.tm.cpu_flags.bitfield.cpusse3 || i.tm.cpu_flags.bitfield.cpusvme)
3386 && i.operands > 0)
3387 {
3388 /* MONITOR/MWAIT as well as SVME instructions have fixed operands
3389 with an opcode suffix which is coded in the same place as an
3390 8-bit immediate field would be.
3391 Here we check those operands and remove them afterwards. */
3392 unsigned int x;
3393
3394 for (x = 0; x < i.operands; x++)
3395 if (register_number (i.op[x].regs) != x)
3396 as_bad (_("can't use register '%s%s' as operand %d in '%s'."),
3397 register_prefix, i.op[x].regs->reg_name, x + 1,
3398 i.tm.name);
3399
3400 i.operands = 0;
3401 }
3402
3403 if (i.tm.cpu_flags.bitfield.cpumwaitx && i.operands > 0)
3404 {
3405 /* MONITORX/MWAITX instructions have fixed operands with an opcode
3406 suffix which is coded in the same place as an 8-bit immediate
3407 field would be.
3408 Here we check those operands and remove them afterwards. */
3409 unsigned int x;
3410
3411 if (i.operands != 3)
3412 abort();
3413
3414 for (x = 0; x < 2; x++)
3415 if (register_number (i.op[x].regs) != x)
3416 goto bad_register_operand;
3417
3418 /* Check for third operand for mwaitx/monitorx insn. */
3419 if (register_number (i.op[x].regs)
3420 != (x + (i.tm.extension_opcode == 0xfb)))
3421 {
3422 bad_register_operand:
3423 as_bad (_("can't use register '%s%s' as operand %d in '%s'."),
3424 register_prefix, i.op[x].regs->reg_name, x+1,
3425 i.tm.name);
3426 }
3427
3428 i.operands = 0;
3429 }
3430
3431 /* These AMD 3DNow! and SSE2 instructions have an opcode suffix
3432 which is coded in the same place as an 8-bit immediate field
3433 would be. Here we fake an 8-bit immediate operand from the
3434 opcode suffix stored in tm.extension_opcode.
3435
3436 AVX instructions also use this encoding, for some of
3437 3 argument instructions. */
3438
3439 gas_assert (i.imm_operands <= 1
3440 && (i.operands <= 2
3441 || ((i.tm.opcode_modifier.vex
3442 || i.tm.opcode_modifier.evex)
3443 && i.operands <= 4)));
3444
3445 exp = &im_expressions[i.imm_operands++];
3446 i.op[i.operands].imms = exp;
3447 i.types[i.operands] = imm8;
3448 i.operands++;
3449 exp->X_op = O_constant;
3450 exp->X_add_number = i.tm.extension_opcode;
3451 i.tm.extension_opcode = None;
3452 }
3453
3454
3455 static int
3456 check_hle (void)
3457 {
3458 switch (i.tm.opcode_modifier.hleprefixok)
3459 {
3460 default:
3461 abort ();
3462 case HLEPrefixNone:
3463 as_bad (_("invalid instruction `%s' after `%s'"),
3464 i.tm.name, i.hle_prefix);
3465 return 0;
3466 case HLEPrefixLock:
3467 if (i.prefix[LOCK_PREFIX])
3468 return 1;
3469 as_bad (_("missing `lock' with `%s'"), i.hle_prefix);
3470 return 0;
3471 case HLEPrefixAny:
3472 return 1;
3473 case HLEPrefixRelease:
3474 if (i.prefix[HLE_PREFIX] != XRELEASE_PREFIX_OPCODE)
3475 {
3476 as_bad (_("instruction `%s' after `xacquire' not allowed"),
3477 i.tm.name);
3478 return 0;
3479 }
3480 if (i.mem_operands == 0
3481 || !operand_type_check (i.types[i.operands - 1], anymem))
3482 {
3483 as_bad (_("memory destination needed for instruction `%s'"
3484 " after `xrelease'"), i.tm.name);
3485 return 0;
3486 }
3487 return 1;
3488 }
3489 }
3490
3491 /* This is the guts of the machine-dependent assembler. LINE points to a
3492 machine dependent instruction. This function is supposed to emit
3493 the frags/bytes it assembles to. */
3494
3495 void
3496 md_assemble (char *line)
3497 {
3498 unsigned int j;
3499 char mnemonic[MAX_MNEM_SIZE];
3500 const insn_template *t;
3501
3502 /* Initialize globals. */
3503 memset (&i, '\0', sizeof (i));
3504 for (j = 0; j < MAX_OPERANDS; j++)
3505 i.reloc[j] = NO_RELOC;
3506 memset (disp_expressions, '\0', sizeof (disp_expressions));
3507 memset (im_expressions, '\0', sizeof (im_expressions));
3508 save_stack_p = save_stack;
3509
3510 /* First parse an instruction mnemonic & call i386_operand for the operands.
3511 We assume that the scrubber has arranged it so that line[0] is the valid
3512 start of a (possibly prefixed) mnemonic. */
3513
3514 line = parse_insn (line, mnemonic);
3515 if (line == NULL)
3516 return;
3517
3518 line = parse_operands (line, mnemonic);
3519 this_operand = -1;
3520 if (line == NULL)
3521 return;
3522
3523 /* Now we've parsed the mnemonic into a set of templates, and have the
3524 operands at hand. */
3525
3526 /* All intel opcodes have reversed operands except for "bound" and
3527 "enter". We also don't reverse intersegment "jmp" and "call"
3528 instructions with 2 immediate operands so that the immediate segment
3529 precedes the offset, as it does when in AT&T mode. */
3530 if (intel_syntax
3531 && i.operands > 1
3532 && (strcmp (mnemonic, "bound") != 0)
3533 && (strcmp (mnemonic, "invlpga") != 0)
3534 && !(operand_type_check (i.types[0], imm)
3535 && operand_type_check (i.types[1], imm)))
3536 swap_operands ();
3537
3538 /* The order of the immediates should be reversed
3539 for 2 immediates extrq and insertq instructions */
3540 if (i.imm_operands == 2
3541 && (strcmp (mnemonic, "extrq") == 0
3542 || strcmp (mnemonic, "insertq") == 0))
3543 swap_2_operands (0, 1);
3544
3545 if (i.imm_operands)
3546 optimize_imm ();
3547
3548 /* Don't optimize displacement for movabs since it only takes 64bit
3549 displacement. */
3550 if (i.disp_operands
3551 && i.disp_encoding != disp_encoding_32bit
3552 && (flag_code != CODE_64BIT
3553 || strcmp (mnemonic, "movabs") != 0))
3554 optimize_disp ();
3555
3556 /* Next, we find a template that matches the given insn,
3557 making sure the overlap of the given operands types is consistent
3558 with the template operand types. */
3559
3560 if (!(t = match_template ()))
3561 return;
3562
3563 if (sse_check != check_none
3564 && !i.tm.opcode_modifier.noavx
3565 && (i.tm.cpu_flags.bitfield.cpusse
3566 || i.tm.cpu_flags.bitfield.cpusse2
3567 || i.tm.cpu_flags.bitfield.cpusse3
3568 || i.tm.cpu_flags.bitfield.cpussse3
3569 || i.tm.cpu_flags.bitfield.cpusse4_1
3570 || i.tm.cpu_flags.bitfield.cpusse4_2))
3571 {
3572 (sse_check == check_warning
3573 ? as_warn
3574 : as_bad) (_("SSE instruction `%s' is used"), i.tm.name);
3575 }
3576
3577 /* Zap movzx and movsx suffix. The suffix has been set from
3578 "word ptr" or "byte ptr" on the source operand in Intel syntax
3579 or extracted from mnemonic in AT&T syntax. But we'll use
3580 the destination register to choose the suffix for encoding. */
3581 if ((i.tm.base_opcode & ~9) == 0x0fb6)
3582 {
3583 /* In Intel syntax, there must be a suffix. In AT&T syntax, if
3584 there is no suffix, the default will be byte extension. */
3585 if (i.reg_operands != 2
3586 && !i.suffix
3587 && intel_syntax)
3588 as_bad (_("ambiguous operand size for `%s'"), i.tm.name);
3589
3590 i.suffix = 0;
3591 }
3592
3593 if (i.tm.opcode_modifier.fwait)
3594 if (!add_prefix (FWAIT_OPCODE))
3595 return;
3596
3597 /* Check if REP prefix is OK. */
3598 if (i.rep_prefix && !i.tm.opcode_modifier.repprefixok)
3599 {
3600 as_bad (_("invalid instruction `%s' after `%s'"),
3601 i.tm.name, i.rep_prefix);
3602 return;
3603 }
3604
3605 /* Check for lock without a lockable instruction. Destination operand
3606 must be memory unless it is xchg (0x86). */
3607 if (i.prefix[LOCK_PREFIX]
3608 && (!i.tm.opcode_modifier.islockable
3609 || i.mem_operands == 0
3610 || (i.tm.base_opcode != 0x86
3611 && !operand_type_check (i.types[i.operands - 1], anymem))))
3612 {
3613 as_bad (_("expecting lockable instruction after `lock'"));
3614 return;
3615 }
3616
3617 /* Check if HLE prefix is OK. */
3618 if (i.hle_prefix && !check_hle ())
3619 return;
3620
3621 /* Check BND prefix. */
3622 if (i.bnd_prefix && !i.tm.opcode_modifier.bndprefixok)
3623 as_bad (_("expecting valid branch instruction after `bnd'"));
3624
3625 if (i.tm.cpu_flags.bitfield.cpumpx
3626 && flag_code == CODE_64BIT
3627 && i.prefix[ADDR_PREFIX])
3628 as_bad (_("32-bit address isn't allowed in 64-bit MPX instructions."));
3629
3630 /* Insert BND prefix. */
3631 if (add_bnd_prefix
3632 && i.tm.opcode_modifier.bndprefixok
3633 && !i.prefix[BND_PREFIX])
3634 add_prefix (BND_PREFIX_OPCODE);
3635
3636 /* Check string instruction segment overrides. */
3637 if (i.tm.opcode_modifier.isstring && i.mem_operands != 0)
3638 {
3639 if (!check_string ())
3640 return;
3641 i.disp_operands = 0;
3642 }
3643
3644 if (!process_suffix ())
3645 return;
3646
3647 /* Update operand types. */
3648 for (j = 0; j < i.operands; j++)
3649 i.types[j] = operand_type_and (i.types[j], i.tm.operand_types[j]);
3650
3651 /* Make still unresolved immediate matches conform to size of immediate
3652 given in i.suffix. */
3653 if (!finalize_imm ())
3654 return;
3655
3656 if (i.types[0].bitfield.imm1)
3657 i.imm_operands = 0; /* kludge for shift insns. */
3658
3659 /* We only need to check those implicit registers for instructions
3660 with 3 operands or less. */
3661 if (i.operands <= 3)
3662 for (j = 0; j < i.operands; j++)
3663 if (i.types[j].bitfield.inoutportreg
3664 || i.types[j].bitfield.shiftcount
3665 || i.types[j].bitfield.acc
3666 || i.types[j].bitfield.floatacc)
3667 i.reg_operands--;
3668
3669 /* ImmExt should be processed after SSE2AVX. */
3670 if (!i.tm.opcode_modifier.sse2avx
3671 && i.tm.opcode_modifier.immext)
3672 process_immext ();
3673
3674 /* For insns with operands there are more diddles to do to the opcode. */
3675 if (i.operands)
3676 {
3677 if (!process_operands ())
3678 return;
3679 }
3680 else if (!quiet_warnings && i.tm.opcode_modifier.ugh)
3681 {
3682 /* UnixWare fsub no args is alias for fsubp, fadd -> faddp, etc. */
3683 as_warn (_("translating to `%sp'"), i.tm.name);
3684 }
3685
3686 if (i.tm.opcode_modifier.vex || i.tm.opcode_modifier.evex)
3687 {
3688 if (flag_code == CODE_16BIT)
3689 {
3690 as_bad (_("instruction `%s' isn't supported in 16-bit mode."),
3691 i.tm.name);
3692 return;
3693 }
3694
3695 if (i.tm.opcode_modifier.vex)
3696 build_vex_prefix (t);
3697 else
3698 build_evex_prefix ();
3699 }
3700
3701 /* Handle conversion of 'int $3' --> special int3 insn. XOP or FMA4
3702 instructions may define INT_OPCODE as well, so avoid this corner
3703 case for those instructions that use MODRM. */
3704 if (i.tm.base_opcode == INT_OPCODE
3705 && !i.tm.opcode_modifier.modrm
3706 && i.op[0].imms->X_add_number == 3)
3707 {
3708 i.tm.base_opcode = INT3_OPCODE;
3709 i.imm_operands = 0;
3710 }
3711
3712 if ((i.tm.opcode_modifier.jump
3713 || i.tm.opcode_modifier.jumpbyte
3714 || i.tm.opcode_modifier.jumpdword)
3715 && i.op[0].disps->X_op == O_constant)
3716 {
3717 /* Convert "jmp constant" (and "call constant") to a jump (call) to
3718 the absolute address given by the constant. Since ix86 jumps and
3719 calls are pc relative, we need to generate a reloc. */
3720 i.op[0].disps->X_add_symbol = &abs_symbol;
3721 i.op[0].disps->X_op = O_symbol;
3722 }
3723
3724 if (i.tm.opcode_modifier.rex64)
3725 i.rex |= REX_W;
3726
3727 /* For 8 bit registers we need an empty rex prefix. Also if the
3728 instruction already has a prefix, we need to convert old
3729 registers to new ones. */
3730
3731 if ((i.types[0].bitfield.reg8
3732 && (i.op[0].regs->reg_flags & RegRex64) != 0)
3733 || (i.types[1].bitfield.reg8
3734 && (i.op[1].regs->reg_flags & RegRex64) != 0)
3735 || ((i.types[0].bitfield.reg8
3736 || i.types[1].bitfield.reg8)
3737 && i.rex != 0))
3738 {
3739 int x;
3740
3741 i.rex |= REX_OPCODE;
3742 for (x = 0; x < 2; x++)
3743 {
3744 /* Look for 8 bit operand that uses old registers. */
3745 if (i.types[x].bitfield.reg8
3746 && (i.op[x].regs->reg_flags & RegRex64) == 0)
3747 {
3748 /* In case it is "hi" register, give up. */
3749 if (i.op[x].regs->reg_num > 3)
3750 as_bad (_("can't encode register '%s%s' in an "
3751 "instruction requiring REX prefix."),
3752 register_prefix, i.op[x].regs->reg_name);
3753
3754 /* Otherwise it is equivalent to the extended register.
3755 Since the encoding doesn't change this is merely
3756 cosmetic cleanup for debug output. */
3757
3758 i.op[x].regs = i.op[x].regs + 8;
3759 }
3760 }
3761 }
3762
3763 if (i.rex != 0)
3764 add_prefix (REX_OPCODE | i.rex);
3765
3766 /* We are ready to output the insn. */
3767 output_insn ();
3768 }
3769
3770 static char *
3771 parse_insn (char *line, char *mnemonic)
3772 {
3773 char *l = line;
3774 char *token_start = l;
3775 char *mnem_p;
3776 int supported;
3777 const insn_template *t;
3778 char *dot_p = NULL;
3779
3780 while (1)
3781 {
3782 mnem_p = mnemonic;
3783 while ((*mnem_p = mnemonic_chars[(unsigned char) *l]) != 0)
3784 {
3785 if (*mnem_p == '.')
3786 dot_p = mnem_p;
3787 mnem_p++;
3788 if (mnem_p >= mnemonic + MAX_MNEM_SIZE)
3789 {
3790 as_bad (_("no such instruction: `%s'"), token_start);
3791 return NULL;
3792 }
3793 l++;
3794 }
3795 if (!is_space_char (*l)
3796 && *l != END_OF_INSN
3797 && (intel_syntax
3798 || (*l != PREFIX_SEPARATOR
3799 && *l != ',')))
3800 {
3801 as_bad (_("invalid character %s in mnemonic"),
3802 output_invalid (*l));
3803 return NULL;
3804 }
3805 if (token_start == l)
3806 {
3807 if (!intel_syntax && *l == PREFIX_SEPARATOR)
3808 as_bad (_("expecting prefix; got nothing"));
3809 else
3810 as_bad (_("expecting mnemonic; got nothing"));
3811 return NULL;
3812 }
3813
3814 /* Look up instruction (or prefix) via hash table. */
3815 current_templates = (const templates *) hash_find (op_hash, mnemonic);
3816
3817 if (*l != END_OF_INSN
3818 && (!is_space_char (*l) || l[1] != END_OF_INSN)
3819 && current_templates
3820 && current_templates->start->opcode_modifier.isprefix)
3821 {
3822 if (!cpu_flags_check_cpu64 (current_templates->start->cpu_flags))
3823 {
3824 as_bad ((flag_code != CODE_64BIT
3825 ? _("`%s' is only supported in 64-bit mode")
3826 : _("`%s' is not supported in 64-bit mode")),
3827 current_templates->start->name);
3828 return NULL;
3829 }
3830 /* If we are in 16-bit mode, do not allow addr16 or data16.
3831 Similarly, in 32-bit mode, do not allow addr32 or data32. */
3832 if ((current_templates->start->opcode_modifier.size16
3833 || current_templates->start->opcode_modifier.size32)
3834 && flag_code != CODE_64BIT
3835 && (current_templates->start->opcode_modifier.size32
3836 ^ (flag_code == CODE_16BIT)))
3837 {
3838 as_bad (_("redundant %s prefix"),
3839 current_templates->start->name);
3840 return NULL;
3841 }
3842 /* Add prefix, checking for repeated prefixes. */
3843 switch (add_prefix (current_templates->start->base_opcode))
3844 {
3845 case PREFIX_EXIST:
3846 return NULL;
3847 case PREFIX_REP:
3848 if (current_templates->start->cpu_flags.bitfield.cpuhle)
3849 i.hle_prefix = current_templates->start->name;
3850 else if (current_templates->start->cpu_flags.bitfield.cpumpx)
3851 i.bnd_prefix = current_templates->start->name;
3852 else
3853 i.rep_prefix = current_templates->start->name;
3854 break;
3855 default:
3856 break;
3857 }
3858 /* Skip past PREFIX_SEPARATOR and reset token_start. */
3859 token_start = ++l;
3860 }
3861 else
3862 break;
3863 }
3864
3865 if (!current_templates)
3866 {
3867 /* Check if we should swap operand or force 32bit displacement in
3868 encoding. */
3869 if (mnem_p - 2 == dot_p && dot_p[1] == 's')
3870 i.swap_operand = 1;
3871 else if (mnem_p - 3 == dot_p
3872 && dot_p[1] == 'd'
3873 && dot_p[2] == '8')
3874 i.disp_encoding = disp_encoding_8bit;
3875 else if (mnem_p - 4 == dot_p
3876 && dot_p[1] == 'd'
3877 && dot_p[2] == '3'
3878 && dot_p[3] == '2')
3879 i.disp_encoding = disp_encoding_32bit;
3880 else
3881 goto check_suffix;
3882 mnem_p = dot_p;
3883 *dot_p = '\0';
3884 current_templates = (const templates *) hash_find (op_hash, mnemonic);
3885 }
3886
3887 if (!current_templates)
3888 {
3889 check_suffix:
3890 /* See if we can get a match by trimming off a suffix. */
3891 switch (mnem_p[-1])
3892 {
3893 case WORD_MNEM_SUFFIX:
3894 if (intel_syntax && (intel_float_operand (mnemonic) & 2))
3895 i.suffix = SHORT_MNEM_SUFFIX;
3896 else
3897 case BYTE_MNEM_SUFFIX:
3898 case QWORD_MNEM_SUFFIX:
3899 i.suffix = mnem_p[-1];
3900 mnem_p[-1] = '\0';
3901 current_templates = (const templates *) hash_find (op_hash,
3902 mnemonic);
3903 break;
3904 case SHORT_MNEM_SUFFIX:
3905 case LONG_MNEM_SUFFIX:
3906 if (!intel_syntax)
3907 {
3908 i.suffix = mnem_p[-1];
3909 mnem_p[-1] = '\0';
3910 current_templates = (const templates *) hash_find (op_hash,
3911 mnemonic);
3912 }
3913 break;
3914
3915 /* Intel Syntax. */
3916 case 'd':
3917 if (intel_syntax)
3918 {
3919 if (intel_float_operand (mnemonic) == 1)
3920 i.suffix = SHORT_MNEM_SUFFIX;
3921 else
3922 i.suffix = LONG_MNEM_SUFFIX;
3923 mnem_p[-1] = '\0';
3924 current_templates = (const templates *) hash_find (op_hash,
3925 mnemonic);
3926 }
3927 break;
3928 }
3929 if (!current_templates)
3930 {
3931 as_bad (_("no such instruction: `%s'"), token_start);
3932 return NULL;
3933 }
3934 }
3935
3936 if (current_templates->start->opcode_modifier.jump
3937 || current_templates->start->opcode_modifier.jumpbyte)
3938 {
3939 /* Check for a branch hint. We allow ",pt" and ",pn" for
3940 predict taken and predict not taken respectively.
3941 I'm not sure that branch hints actually do anything on loop
3942 and jcxz insns (JumpByte) for current Pentium4 chips. They
3943 may work in the future and it doesn't hurt to accept them
3944 now. */
3945 if (l[0] == ',' && l[1] == 'p')
3946 {
3947 if (l[2] == 't')
3948 {
3949 if (!add_prefix (DS_PREFIX_OPCODE))
3950 return NULL;
3951 l += 3;
3952 }
3953 else if (l[2] == 'n')
3954 {
3955 if (!add_prefix (CS_PREFIX_OPCODE))
3956 return NULL;
3957 l += 3;
3958 }
3959 }
3960 }
3961 /* Any other comma loses. */
3962 if (*l == ',')
3963 {
3964 as_bad (_("invalid character %s in mnemonic"),
3965 output_invalid (*l));
3966 return NULL;
3967 }
3968
3969 /* Check if instruction is supported on specified architecture. */
3970 supported = 0;
3971 for (t = current_templates->start; t < current_templates->end; ++t)
3972 {
3973 supported |= cpu_flags_match (t);
3974 if (supported == CPU_FLAGS_PERFECT_MATCH)
3975 goto skip;
3976 }
3977
3978 if (!(supported & CPU_FLAGS_64BIT_MATCH))
3979 {
3980 as_bad (flag_code == CODE_64BIT
3981 ? _("`%s' is not supported in 64-bit mode")
3982 : _("`%s' is only supported in 64-bit mode"),
3983 current_templates->start->name);
3984 return NULL;
3985 }
3986 if (supported != CPU_FLAGS_PERFECT_MATCH)
3987 {
3988 as_bad (_("`%s' is not supported on `%s%s'"),
3989 current_templates->start->name,
3990 cpu_arch_name ? cpu_arch_name : default_arch,
3991 cpu_sub_arch_name ? cpu_sub_arch_name : "");
3992 return NULL;
3993 }
3994
3995 skip:
3996 if (!cpu_arch_flags.bitfield.cpui386
3997 && (flag_code != CODE_16BIT))
3998 {
3999 as_warn (_("use .code16 to ensure correct addressing mode"));
4000 }
4001
4002 return l;
4003 }
4004
4005 static char *
4006 parse_operands (char *l, const char *mnemonic)
4007 {
4008 char *token_start;
4009
4010 /* 1 if operand is pending after ','. */
4011 unsigned int expecting_operand = 0;
4012
4013 /* Non-zero if operand parens not balanced. */
4014 unsigned int paren_not_balanced;
4015
4016 while (*l != END_OF_INSN)
4017 {
4018 /* Skip optional white space before operand. */
4019 if (is_space_char (*l))
4020 ++l;
4021 if (!is_operand_char (*l) && *l != END_OF_INSN && *l != '"')
4022 {
4023 as_bad (_("invalid character %s before operand %d"),
4024 output_invalid (*l),
4025 i.operands + 1);
4026 return NULL;
4027 }
4028 token_start = l; /* After white space. */
4029 paren_not_balanced = 0;
4030 while (paren_not_balanced || *l != ',')
4031 {
4032 if (*l == END_OF_INSN)
4033 {
4034 if (paren_not_balanced)
4035 {
4036 if (!intel_syntax)
4037 as_bad (_("unbalanced parenthesis in operand %d."),
4038 i.operands + 1);
4039 else
4040 as_bad (_("unbalanced brackets in operand %d."),
4041 i.operands + 1);
4042 return NULL;
4043 }
4044 else
4045 break; /* we are done */
4046 }
4047 else if (!is_operand_char (*l) && !is_space_char (*l) && *l != '"')
4048 {
4049 as_bad (_("invalid character %s in operand %d"),
4050 output_invalid (*l),
4051 i.operands + 1);
4052 return NULL;
4053 }
4054 if (!intel_syntax)
4055 {
4056 if (*l == '(')
4057 ++paren_not_balanced;
4058 if (*l == ')')
4059 --paren_not_balanced;
4060 }
4061 else
4062 {
4063 if (*l == '[')
4064 ++paren_not_balanced;
4065 if (*l == ']')
4066 --paren_not_balanced;
4067 }
4068 l++;
4069 }
4070 if (l != token_start)
4071 { /* Yes, we've read in another operand. */
4072 unsigned int operand_ok;
4073 this_operand = i.operands++;
4074 i.types[this_operand].bitfield.unspecified = 1;
4075 if (i.operands > MAX_OPERANDS)
4076 {
4077 as_bad (_("spurious operands; (%d operands/instruction max)"),
4078 MAX_OPERANDS);
4079 return NULL;
4080 }
4081 /* Now parse operand adding info to 'i' as we go along. */
4082 END_STRING_AND_SAVE (l);
4083
4084 if (intel_syntax)
4085 operand_ok =
4086 i386_intel_operand (token_start,
4087 intel_float_operand (mnemonic));
4088 else
4089 operand_ok = i386_att_operand (token_start);
4090
4091 RESTORE_END_STRING (l);
4092 if (!operand_ok)
4093 return NULL;
4094 }
4095 else
4096 {
4097 if (expecting_operand)
4098 {
4099 expecting_operand_after_comma:
4100 as_bad (_("expecting operand after ','; got nothing"));
4101 return NULL;
4102 }
4103 if (*l == ',')
4104 {
4105 as_bad (_("expecting operand before ','; got nothing"));
4106 return NULL;
4107 }
4108 }
4109
4110 /* Now *l must be either ',' or END_OF_INSN. */
4111 if (*l == ',')
4112 {
4113 if (*++l == END_OF_INSN)
4114 {
4115 /* Just skip it, if it's \n complain. */
4116 goto expecting_operand_after_comma;
4117 }
4118 expecting_operand = 1;
4119 }
4120 }
4121 return l;
4122 }
4123
4124 static void
4125 swap_2_operands (int xchg1, int xchg2)
4126 {
4127 union i386_op temp_op;
4128 i386_operand_type temp_type;
4129 enum bfd_reloc_code_real temp_reloc;
4130
4131 temp_type = i.types[xchg2];
4132 i.types[xchg2] = i.types[xchg1];
4133 i.types[xchg1] = temp_type;
4134 temp_op = i.op[xchg2];
4135 i.op[xchg2] = i.op[xchg1];
4136 i.op[xchg1] = temp_op;
4137 temp_reloc = i.reloc[xchg2];
4138 i.reloc[xchg2] = i.reloc[xchg1];
4139 i.reloc[xchg1] = temp_reloc;
4140
4141 if (i.mask)
4142 {
4143 if (i.mask->operand == xchg1)
4144 i.mask->operand = xchg2;
4145 else if (i.mask->operand == xchg2)
4146 i.mask->operand = xchg1;
4147 }
4148 if (i.broadcast)
4149 {
4150 if (i.broadcast->operand == xchg1)
4151 i.broadcast->operand = xchg2;
4152 else if (i.broadcast->operand == xchg2)
4153 i.broadcast->operand = xchg1;
4154 }
4155 if (i.rounding)
4156 {
4157 if (i.rounding->operand == xchg1)
4158 i.rounding->operand = xchg2;
4159 else if (i.rounding->operand == xchg2)
4160 i.rounding->operand = xchg1;
4161 }
4162 }
4163
4164 static void
4165 swap_operands (void)
4166 {
4167 switch (i.operands)
4168 {
4169 case 5:
4170 case 4:
4171 swap_2_operands (1, i.operands - 2);
4172 case 3:
4173 case 2:
4174 swap_2_operands (0, i.operands - 1);
4175 break;
4176 default:
4177 abort ();
4178 }
4179
4180 if (i.mem_operands == 2)
4181 {
4182 const seg_entry *temp_seg;
4183 temp_seg = i.seg[0];
4184 i.seg[0] = i.seg[1];
4185 i.seg[1] = temp_seg;
4186 }
4187 }
4188
4189 /* Try to ensure constant immediates are represented in the smallest
4190 opcode possible. */
4191 static void
4192 optimize_imm (void)
4193 {
4194 char guess_suffix = 0;
4195 int op;
4196
4197 if (i.suffix)
4198 guess_suffix = i.suffix;
4199 else if (i.reg_operands)
4200 {
4201 /* Figure out a suffix from the last register operand specified.
4202 We can't do this properly yet, ie. excluding InOutPortReg,
4203 but the following works for instructions with immediates.
4204 In any case, we can't set i.suffix yet. */
4205 for (op = i.operands; --op >= 0;)
4206 if (i.types[op].bitfield.reg8)
4207 {
4208 guess_suffix = BYTE_MNEM_SUFFIX;
4209 break;
4210 }
4211 else if (i.types[op].bitfield.reg16)
4212 {
4213 guess_suffix = WORD_MNEM_SUFFIX;
4214 break;
4215 }
4216 else if (i.types[op].bitfield.reg32)
4217 {
4218 guess_suffix = LONG_MNEM_SUFFIX;
4219 break;
4220 }
4221 else if (i.types[op].bitfield.reg64)
4222 {
4223 guess_suffix = QWORD_MNEM_SUFFIX;
4224 break;
4225 }
4226 }
4227 else if ((flag_code == CODE_16BIT) ^ (i.prefix[DATA_PREFIX] != 0))
4228 guess_suffix = WORD_MNEM_SUFFIX;
4229
4230 for (op = i.operands; --op >= 0;)
4231 if (operand_type_check (i.types[op], imm))
4232 {
4233 switch (i.op[op].imms->X_op)
4234 {
4235 case O_constant:
4236 /* If a suffix is given, this operand may be shortened. */
4237 switch (guess_suffix)
4238 {
4239 case LONG_MNEM_SUFFIX:
4240 i.types[op].bitfield.imm32 = 1;
4241 i.types[op].bitfield.imm64 = 1;
4242 break;
4243 case WORD_MNEM_SUFFIX:
4244 i.types[op].bitfield.imm16 = 1;
4245 i.types[op].bitfield.imm32 = 1;
4246 i.types[op].bitfield.imm32s = 1;
4247 i.types[op].bitfield.imm64 = 1;
4248 break;
4249 case BYTE_MNEM_SUFFIX:
4250 i.types[op].bitfield.imm8 = 1;
4251 i.types[op].bitfield.imm8s = 1;
4252 i.types[op].bitfield.imm16 = 1;
4253 i.types[op].bitfield.imm32 = 1;
4254 i.types[op].bitfield.imm32s = 1;
4255 i.types[op].bitfield.imm64 = 1;
4256 break;
4257 }
4258
4259 /* If this operand is at most 16 bits, convert it
4260 to a signed 16 bit number before trying to see
4261 whether it will fit in an even smaller size.
4262 This allows a 16-bit operand such as $0xffe0 to
4263 be recognised as within Imm8S range. */
4264 if ((i.types[op].bitfield.imm16)
4265 && (i.op[op].imms->X_add_number & ~(offsetT) 0xffff) == 0)
4266 {
4267 i.op[op].imms->X_add_number =
4268 (((i.op[op].imms->X_add_number & 0xffff) ^ 0x8000) - 0x8000);
4269 }
4270 #ifdef BFD64
4271 /* Store 32-bit immediate in 64-bit for 64-bit BFD. */
4272 if ((i.types[op].bitfield.imm32)
4273 && ((i.op[op].imms->X_add_number & ~(((offsetT) 2 << 31) - 1))
4274 == 0))
4275 {
4276 i.op[op].imms->X_add_number = ((i.op[op].imms->X_add_number
4277 ^ ((offsetT) 1 << 31))
4278 - ((offsetT) 1 << 31));
4279 }
4280 #endif
4281 i.types[op]
4282 = operand_type_or (i.types[op],
4283 smallest_imm_type (i.op[op].imms->X_add_number));
4284
4285 /* We must avoid matching of Imm32 templates when 64bit
4286 only immediate is available. */
4287 if (guess_suffix == QWORD_MNEM_SUFFIX)
4288 i.types[op].bitfield.imm32 = 0;
4289 break;
4290
4291 case O_absent:
4292 case O_register:
4293 abort ();
4294
4295 /* Symbols and expressions. */
4296 default:
4297 /* Convert symbolic operand to proper sizes for matching, but don't
4298 prevent matching a set of insns that only supports sizes other
4299 than those matching the insn suffix. */
4300 {
4301 i386_operand_type mask, allowed;
4302 const insn_template *t;
4303
4304 operand_type_set (&mask, 0);
4305 operand_type_set (&allowed, 0);
4306
4307 for (t = current_templates->start;
4308 t < current_templates->end;
4309 ++t)
4310 allowed = operand_type_or (allowed,
4311 t->operand_types[op]);
4312 switch (guess_suffix)
4313 {
4314 case QWORD_MNEM_SUFFIX:
4315 mask.bitfield.imm64 = 1;
4316 mask.bitfield.imm32s = 1;
4317 break;
4318 case LONG_MNEM_SUFFIX:
4319 mask.bitfield.imm32 = 1;
4320 break;
4321 case WORD_MNEM_SUFFIX:
4322 mask.bitfield.imm16 = 1;
4323 break;
4324 case BYTE_MNEM_SUFFIX:
4325 mask.bitfield.imm8 = 1;
4326 break;
4327 default:
4328 break;
4329 }
4330 allowed = operand_type_and (mask, allowed);
4331 if (!operand_type_all_zero (&allowed))
4332 i.types[op] = operand_type_and (i.types[op], mask);
4333 }
4334 break;
4335 }
4336 }
4337 }
4338
4339 /* Try to use the smallest displacement type too. */
4340 static void
4341 optimize_disp (void)
4342 {
4343 int op;
4344
4345 for (op = i.operands; --op >= 0;)
4346 if (operand_type_check (i.types[op], disp))
4347 {
4348 if (i.op[op].disps->X_op == O_constant)
4349 {
4350 offsetT op_disp = i.op[op].disps->X_add_number;
4351
4352 if (i.types[op].bitfield.disp16
4353 && (op_disp & ~(offsetT) 0xffff) == 0)
4354 {
4355 /* If this operand is at most 16 bits, convert
4356 to a signed 16 bit number and don't use 64bit
4357 displacement. */
4358 op_disp = (((op_disp & 0xffff) ^ 0x8000) - 0x8000);
4359 i.types[op].bitfield.disp64 = 0;
4360 }
4361 #ifdef BFD64
4362 /* Optimize 64-bit displacement to 32-bit for 64-bit BFD. */
4363 if (i.types[op].bitfield.disp32
4364 && (op_disp & ~(((offsetT) 2 << 31) - 1)) == 0)
4365 {
4366 /* If this operand is at most 32 bits, convert
4367 to a signed 32 bit number and don't use 64bit
4368 displacement. */
4369 op_disp &= (((offsetT) 2 << 31) - 1);
4370 op_disp = (op_disp ^ ((offsetT) 1 << 31)) - ((addressT) 1 << 31);
4371 i.types[op].bitfield.disp64 = 0;
4372 }
4373 #endif
4374 if (!op_disp && i.types[op].bitfield.baseindex)
4375 {
4376 i.types[op].bitfield.disp8 = 0;
4377 i.types[op].bitfield.disp16 = 0;
4378 i.types[op].bitfield.disp32 = 0;
4379 i.types[op].bitfield.disp32s = 0;
4380 i.types[op].bitfield.disp64 = 0;
4381 i.op[op].disps = 0;
4382 i.disp_operands--;
4383 }
4384 else if (flag_code == CODE_64BIT)
4385 {
4386 if (fits_in_signed_long (op_disp))
4387 {
4388 i.types[op].bitfield.disp64 = 0;
4389 i.types[op].bitfield.disp32s = 1;
4390 }
4391 if (i.prefix[ADDR_PREFIX]
4392 && fits_in_unsigned_long (op_disp))
4393 i.types[op].bitfield.disp32 = 1;
4394 }
4395 if ((i.types[op].bitfield.disp32
4396 || i.types[op].bitfield.disp32s
4397 || i.types[op].bitfield.disp16)
4398 && fits_in_signed_byte (op_disp))
4399 i.types[op].bitfield.disp8 = 1;
4400 }
4401 else if (i.reloc[op] == BFD_RELOC_386_TLS_DESC_CALL
4402 || i.reloc[op] == BFD_RELOC_X86_64_TLSDESC_CALL)
4403 {
4404 fix_new_exp (frag_now, frag_more (0) - frag_now->fr_literal, 0,
4405 i.op[op].disps, 0, i.reloc[op]);
4406 i.types[op].bitfield.disp8 = 0;
4407 i.types[op].bitfield.disp16 = 0;
4408 i.types[op].bitfield.disp32 = 0;
4409 i.types[op].bitfield.disp32s = 0;
4410 i.types[op].bitfield.disp64 = 0;
4411 }
4412 else
4413 /* We only support 64bit displacement on constants. */
4414 i.types[op].bitfield.disp64 = 0;
4415 }
4416 }
4417
4418 /* Check if operands are valid for the instruction. */
4419
4420 static int
4421 check_VecOperands (const insn_template *t)
4422 {
4423 unsigned int op;
4424
4425 /* Without VSIB byte, we can't have a vector register for index. */
4426 if (!t->opcode_modifier.vecsib
4427 && i.index_reg
4428 && (i.index_reg->reg_type.bitfield.regxmm
4429 || i.index_reg->reg_type.bitfield.regymm
4430 || i.index_reg->reg_type.bitfield.regzmm))
4431 {
4432 i.error = unsupported_vector_index_register;
4433 return 1;
4434 }
4435
4436 /* Check if default mask is allowed. */
4437 if (t->opcode_modifier.nodefmask
4438 && (!i.mask || i.mask->mask->reg_num == 0))
4439 {
4440 i.error = no_default_mask;
4441 return 1;
4442 }
4443
4444 /* For VSIB byte, we need a vector register for index, and all vector
4445 registers must be distinct. */
4446 if (t->opcode_modifier.vecsib)
4447 {
4448 if (!i.index_reg
4449 || !((t->opcode_modifier.vecsib == VecSIB128
4450 && i.index_reg->reg_type.bitfield.regxmm)
4451 || (t->opcode_modifier.vecsib == VecSIB256
4452 && i.index_reg->reg_type.bitfield.regymm)
4453 || (t->opcode_modifier.vecsib == VecSIB512
4454 && i.index_reg->reg_type.bitfield.regzmm)))
4455 {
4456 i.error = invalid_vsib_address;
4457 return 1;
4458 }
4459
4460 gas_assert (i.reg_operands == 2 || i.mask);
4461 if (i.reg_operands == 2 && !i.mask)
4462 {
4463 gas_assert (i.types[0].bitfield.regxmm
4464 || i.types[0].bitfield.regymm);
4465 gas_assert (i.types[2].bitfield.regxmm
4466 || i.types[2].bitfield.regymm);
4467 if (operand_check == check_none)
4468 return 0;
4469 if (register_number (i.op[0].regs)
4470 != register_number (i.index_reg)
4471 && register_number (i.op[2].regs)
4472 != register_number (i.index_reg)
4473 && register_number (i.op[0].regs)
4474 != register_number (i.op[2].regs))
4475 return 0;
4476 if (operand_check == check_error)
4477 {
4478 i.error = invalid_vector_register_set;
4479 return 1;
4480 }
4481 as_warn (_("mask, index, and destination registers should be distinct"));
4482 }
4483 else if (i.reg_operands == 1 && i.mask)
4484 {
4485 if ((i.types[1].bitfield.regymm
4486 || i.types[1].bitfield.regzmm)
4487 && (register_number (i.op[1].regs)
4488 == register_number (i.index_reg)))
4489 {
4490 if (operand_check == check_error)
4491 {
4492 i.error = invalid_vector_register_set;
4493 return 1;
4494 }
4495 if (operand_check != check_none)
4496 as_warn (_("index and destination registers should be distinct"));
4497 }
4498 }
4499 }
4500
4501 /* Check if broadcast is supported by the instruction and is applied
4502 to the memory operand. */
4503 if (i.broadcast)
4504 {
4505 int broadcasted_opnd_size;
4506
4507 /* Check if specified broadcast is supported in this instruction,
4508 and it's applied to memory operand of DWORD or QWORD type,
4509 depending on VecESize. */
4510 if (i.broadcast->type != t->opcode_modifier.broadcast
4511 || !i.types[i.broadcast->operand].bitfield.mem
4512 || (t->opcode_modifier.vecesize == 0
4513 && !i.types[i.broadcast->operand].bitfield.dword
4514 && !i.types[i.broadcast->operand].bitfield.unspecified)
4515 || (t->opcode_modifier.vecesize == 1
4516 && !i.types[i.broadcast->operand].bitfield.qword
4517 && !i.types[i.broadcast->operand].bitfield.unspecified))
4518 goto bad_broadcast;
4519
4520 broadcasted_opnd_size = t->opcode_modifier.vecesize ? 64 : 32;
4521 if (i.broadcast->type == BROADCAST_1TO16)
4522 broadcasted_opnd_size <<= 4; /* Broadcast 1to16. */
4523 else if (i.broadcast->type == BROADCAST_1TO8)
4524 broadcasted_opnd_size <<= 3; /* Broadcast 1to8. */
4525 else if (i.broadcast->type == BROADCAST_1TO4)
4526 broadcasted_opnd_size <<= 2; /* Broadcast 1to4. */
4527 else if (i.broadcast->type == BROADCAST_1TO2)
4528 broadcasted_opnd_size <<= 1; /* Broadcast 1to2. */
4529 else
4530 goto bad_broadcast;
4531
4532 if ((broadcasted_opnd_size == 256
4533 && !t->operand_types[i.broadcast->operand].bitfield.ymmword)
4534 || (broadcasted_opnd_size == 512
4535 && !t->operand_types[i.broadcast->operand].bitfield.zmmword))
4536 {
4537 bad_broadcast:
4538 i.error = unsupported_broadcast;
4539 return 1;
4540 }
4541 }
4542 /* If broadcast is supported in this instruction, we need to check if
4543 operand of one-element size isn't specified without broadcast. */
4544 else if (t->opcode_modifier.broadcast && i.mem_operands)
4545 {
4546 /* Find memory operand. */
4547 for (op = 0; op < i.operands; op++)
4548 if (operand_type_check (i.types[op], anymem))
4549 break;
4550 gas_assert (op < i.operands);
4551 /* Check size of the memory operand. */
4552 if ((t->opcode_modifier.vecesize == 0
4553 && i.types[op].bitfield.dword)
4554 || (t->opcode_modifier.vecesize == 1
4555 && i.types[op].bitfield.qword))
4556 {
4557 i.error = broadcast_needed;
4558 return 1;
4559 }
4560 }
4561
4562 /* Check if requested masking is supported. */
4563 if (i.mask
4564 && (!t->opcode_modifier.masking
4565 || (i.mask->zeroing
4566 && t->opcode_modifier.masking == MERGING_MASKING)))
4567 {
4568 i.error = unsupported_masking;
4569 return 1;
4570 }
4571
4572 /* Check if masking is applied to dest operand. */
4573 if (i.mask && (i.mask->operand != (int) (i.operands - 1)))
4574 {
4575 i.error = mask_not_on_destination;
4576 return 1;
4577 }
4578
4579 /* Check RC/SAE. */
4580 if (i.rounding)
4581 {
4582 if ((i.rounding->type != saeonly
4583 && !t->opcode_modifier.staticrounding)
4584 || (i.rounding->type == saeonly
4585 && (t->opcode_modifier.staticrounding
4586 || !t->opcode_modifier.sae)))
4587 {
4588 i.error = unsupported_rc_sae;
4589 return 1;
4590 }
4591 /* If the instruction has several immediate operands and one of
4592 them is rounding, the rounding operand should be the last
4593 immediate operand. */
4594 if (i.imm_operands > 1
4595 && i.rounding->operand != (int) (i.imm_operands - 1))
4596 {
4597 i.error = rc_sae_operand_not_last_imm;
4598 return 1;
4599 }
4600 }
4601
4602 /* Check vector Disp8 operand. */
4603 if (t->opcode_modifier.disp8memshift)
4604 {
4605 if (i.broadcast)
4606 i.memshift = t->opcode_modifier.vecesize ? 3 : 2;
4607 else
4608 i.memshift = t->opcode_modifier.disp8memshift;
4609
4610 for (op = 0; op < i.operands; op++)
4611 if (operand_type_check (i.types[op], disp)
4612 && i.op[op].disps->X_op == O_constant)
4613 {
4614 offsetT value = i.op[op].disps->X_add_number;
4615 int vec_disp8_ok
4616 = (i.disp_encoding != disp_encoding_32bit
4617 && fits_in_vec_disp8 (value));
4618 if (t->operand_types [op].bitfield.vec_disp8)
4619 {
4620 if (vec_disp8_ok)
4621 i.types[op].bitfield.vec_disp8 = 1;
4622 else
4623 {
4624 /* Vector insn can only have Vec_Disp8/Disp32 in
4625 32/64bit modes, and Vec_Disp8/Disp16 in 16bit
4626 mode. */
4627 i.types[op].bitfield.disp8 = 0;
4628 if (flag_code != CODE_16BIT)
4629 i.types[op].bitfield.disp16 = 0;
4630 }
4631 }
4632 else if (flag_code != CODE_16BIT)
4633 {
4634 /* One form of this instruction supports vector Disp8.
4635 Try vector Disp8 if we need to use Disp32. */
4636 if (vec_disp8_ok && !fits_in_signed_byte (value))
4637 {
4638 i.error = try_vector_disp8;
4639 return 1;
4640 }
4641 }
4642 }
4643 }
4644 else
4645 i.memshift = -1;
4646
4647 return 0;
4648 }
4649
4650 /* Check if operands are valid for the instruction. Update VEX
4651 operand types. */
4652
4653 static int
4654 VEX_check_operands (const insn_template *t)
4655 {
4656 /* VREX is only valid with EVEX prefix. */
4657 if (i.need_vrex && !t->opcode_modifier.evex)
4658 {
4659 i.error = invalid_register_operand;
4660 return 1;
4661 }
4662
4663 if (!t->opcode_modifier.vex)
4664 return 0;
4665
4666 /* Only check VEX_Imm4, which must be the first operand. */
4667 if (t->operand_types[0].bitfield.vec_imm4)
4668 {
4669 if (i.op[0].imms->X_op != O_constant
4670 || !fits_in_imm4 (i.op[0].imms->X_add_number))
4671 {
4672 i.error = bad_imm4;
4673 return 1;
4674 }
4675
4676 /* Turn off Imm8 so that update_imm won't complain. */
4677 i.types[0] = vec_imm4;
4678 }
4679
4680 return 0;
4681 }
4682
4683 static const insn_template *
4684 match_template (void)
4685 {
4686 /* Points to template once we've found it. */
4687 const insn_template *t;
4688 i386_operand_type overlap0, overlap1, overlap2, overlap3;
4689 i386_operand_type overlap4;
4690 unsigned int found_reverse_match;
4691 i386_opcode_modifier suffix_check;
4692 i386_operand_type operand_types [MAX_OPERANDS];
4693 int addr_prefix_disp;
4694 unsigned int j;
4695 unsigned int found_cpu_match;
4696 unsigned int check_register;
4697 enum i386_error specific_error = 0;
4698
4699 #if MAX_OPERANDS != 5
4700 # error "MAX_OPERANDS must be 5."
4701 #endif
4702
4703 found_reverse_match = 0;
4704 addr_prefix_disp = -1;
4705
4706 memset (&suffix_check, 0, sizeof (suffix_check));
4707 if (i.suffix == BYTE_MNEM_SUFFIX)
4708 suffix_check.no_bsuf = 1;
4709 else if (i.suffix == WORD_MNEM_SUFFIX)
4710 suffix_check.no_wsuf = 1;
4711 else if (i.suffix == SHORT_MNEM_SUFFIX)
4712 suffix_check.no_ssuf = 1;
4713 else if (i.suffix == LONG_MNEM_SUFFIX)
4714 suffix_check.no_lsuf = 1;
4715 else if (i.suffix == QWORD_MNEM_SUFFIX)
4716 suffix_check.no_qsuf = 1;
4717 else if (i.suffix == LONG_DOUBLE_MNEM_SUFFIX)
4718 suffix_check.no_ldsuf = 1;
4719
4720 /* Must have right number of operands. */
4721 i.error = number_of_operands_mismatch;
4722
4723 for (t = current_templates->start; t < current_templates->end; t++)
4724 {
4725 addr_prefix_disp = -1;
4726
4727 if (i.operands != t->operands)
4728 continue;
4729
4730 /* Check processor support. */
4731 i.error = unsupported;
4732 found_cpu_match = (cpu_flags_match (t)
4733 == CPU_FLAGS_PERFECT_MATCH);
4734 if (!found_cpu_match)
4735 continue;
4736
4737 /* Check old gcc support. */
4738 i.error = old_gcc_only;
4739 if (!old_gcc && t->opcode_modifier.oldgcc)
4740 continue;
4741
4742 /* Check AT&T mnemonic. */
4743 i.error = unsupported_with_intel_mnemonic;
4744 if (intel_mnemonic && t->opcode_modifier.attmnemonic)
4745 continue;
4746
4747 /* Check AT&T/Intel syntax. */
4748 i.error = unsupported_syntax;
4749 if ((intel_syntax && t->opcode_modifier.attsyntax)
4750 || (!intel_syntax && t->opcode_modifier.intelsyntax))
4751 continue;
4752
4753 /* Check the suffix, except for some instructions in intel mode. */
4754 i.error = invalid_instruction_suffix;
4755 if ((!intel_syntax || !t->opcode_modifier.ignoresize)
4756 && ((t->opcode_modifier.no_bsuf && suffix_check.no_bsuf)
4757 || (t->opcode_modifier.no_wsuf && suffix_check.no_wsuf)
4758 || (t->opcode_modifier.no_lsuf && suffix_check.no_lsuf)
4759 || (t->opcode_modifier.no_ssuf && suffix_check.no_ssuf)
4760 || (t->opcode_modifier.no_qsuf && suffix_check.no_qsuf)
4761 || (t->opcode_modifier.no_ldsuf && suffix_check.no_ldsuf)))
4762 continue;
4763
4764 if (!operand_size_match (t))
4765 continue;
4766
4767 for (j = 0; j < MAX_OPERANDS; j++)
4768 operand_types[j] = t->operand_types[j];
4769
4770 /* In general, don't allow 64-bit operands in 32-bit mode. */
4771 if (i.suffix == QWORD_MNEM_SUFFIX
4772 && flag_code != CODE_64BIT
4773 && (intel_syntax
4774 ? (!t->opcode_modifier.ignoresize
4775 && !intel_float_operand (t->name))
4776 : intel_float_operand (t->name) != 2)
4777 && ((!operand_types[0].bitfield.regmmx
4778 && !operand_types[0].bitfield.regxmm
4779 && !operand_types[0].bitfield.regymm
4780 && !operand_types[0].bitfield.regzmm)
4781 || (!operand_types[t->operands > 1].bitfield.regmmx
4782 && operand_types[t->operands > 1].bitfield.regxmm
4783 && operand_types[t->operands > 1].bitfield.regymm
4784 && operand_types[t->operands > 1].bitfield.regzmm))
4785 && (t->base_opcode != 0x0fc7
4786 || t->extension_opcode != 1 /* cmpxchg8b */))
4787 continue;
4788
4789 /* In general, don't allow 32-bit operands on pre-386. */
4790 else if (i.suffix == LONG_MNEM_SUFFIX
4791 && !cpu_arch_flags.bitfield.cpui386
4792 && (intel_syntax
4793 ? (!t->opcode_modifier.ignoresize
4794 && !intel_float_operand (t->name))
4795 : intel_float_operand (t->name) != 2)
4796 && ((!operand_types[0].bitfield.regmmx
4797 && !operand_types[0].bitfield.regxmm)
4798 || (!operand_types[t->operands > 1].bitfield.regmmx
4799 && operand_types[t->operands > 1].bitfield.regxmm)))
4800 continue;
4801
4802 /* Do not verify operands when there are none. */
4803 else
4804 {
4805 if (!t->operands)
4806 /* We've found a match; break out of loop. */
4807 break;
4808 }
4809
4810 /* Address size prefix will turn Disp64/Disp32/Disp16 operand
4811 into Disp32/Disp16/Disp32 operand. */
4812 if (i.prefix[ADDR_PREFIX] != 0)
4813 {
4814 /* There should be only one Disp operand. */
4815 switch (flag_code)
4816 {
4817 case CODE_16BIT:
4818 for (j = 0; j < MAX_OPERANDS; j++)
4819 {
4820 if (operand_types[j].bitfield.disp16)
4821 {
4822 addr_prefix_disp = j;
4823 operand_types[j].bitfield.disp32 = 1;
4824 operand_types[j].bitfield.disp16 = 0;
4825 break;
4826 }
4827 }
4828 break;
4829 case CODE_32BIT:
4830 for (j = 0; j < MAX_OPERANDS; j++)
4831 {
4832 if (operand_types[j].bitfield.disp32)
4833 {
4834 addr_prefix_disp = j;
4835 operand_types[j].bitfield.disp32 = 0;
4836 operand_types[j].bitfield.disp16 = 1;
4837 break;
4838 }
4839 }
4840 break;
4841 case CODE_64BIT:
4842 for (j = 0; j < MAX_OPERANDS; j++)
4843 {
4844 if (operand_types[j].bitfield.disp64)
4845 {
4846 addr_prefix_disp = j;
4847 operand_types[j].bitfield.disp64 = 0;
4848 operand_types[j].bitfield.disp32 = 1;
4849 break;
4850 }
4851 }
4852 break;
4853 }
4854 }
4855
4856 /* Force 0x8b encoding for "mov foo@GOT, %eax". */
4857 if (i.reloc[0] == BFD_RELOC_386_GOT32 && t->base_opcode == 0xa0)
4858 continue;
4859
4860 /* We check register size if needed. */
4861 check_register = t->opcode_modifier.checkregsize;
4862 overlap0 = operand_type_and (i.types[0], operand_types[0]);
4863 switch (t->operands)
4864 {
4865 case 1:
4866 if (!operand_type_match (overlap0, i.types[0]))
4867 continue;
4868 break;
4869 case 2:
4870 /* xchg %eax, %eax is a special case. It is an aliase for nop
4871 only in 32bit mode and we can use opcode 0x90. In 64bit
4872 mode, we can't use 0x90 for xchg %eax, %eax since it should
4873 zero-extend %eax to %rax. */
4874 if (flag_code == CODE_64BIT
4875 && t->base_opcode == 0x90
4876 && operand_type_equal (&i.types [0], &acc32)
4877 && operand_type_equal (&i.types [1], &acc32))
4878 continue;
4879 if (i.swap_operand)
4880 {
4881 /* If we swap operand in encoding, we either match
4882 the next one or reverse direction of operands. */
4883 if (t->opcode_modifier.s)
4884 continue;
4885 else if (t->opcode_modifier.d)
4886 goto check_reverse;
4887 }
4888
4889 case 3:
4890 /* If we swap operand in encoding, we match the next one. */
4891 if (i.swap_operand && t->opcode_modifier.s)
4892 continue;
4893 case 4:
4894 case 5:
4895 overlap1 = operand_type_and (i.types[1], operand_types[1]);
4896 if (!operand_type_match (overlap0, i.types[0])
4897 || !operand_type_match (overlap1, i.types[1])
4898 || (check_register
4899 && !operand_type_register_match (overlap0, i.types[0],
4900 operand_types[0],
4901 overlap1, i.types[1],
4902 operand_types[1])))
4903 {
4904 /* Check if other direction is valid ... */
4905 if (!t->opcode_modifier.d && !t->opcode_modifier.floatd)
4906 continue;
4907
4908 check_reverse:
4909 /* Try reversing direction of operands. */
4910 overlap0 = operand_type_and (i.types[0], operand_types[1]);
4911 overlap1 = operand_type_and (i.types[1], operand_types[0]);
4912 if (!operand_type_match (overlap0, i.types[0])
4913 || !operand_type_match (overlap1, i.types[1])
4914 || (check_register
4915 && !operand_type_register_match (overlap0,
4916 i.types[0],
4917 operand_types[1],
4918 overlap1,
4919 i.types[1],
4920 operand_types[0])))
4921 {
4922 /* Does not match either direction. */
4923 continue;
4924 }
4925 /* found_reverse_match holds which of D or FloatDR
4926 we've found. */
4927 if (t->opcode_modifier.d)
4928 found_reverse_match = Opcode_D;
4929 else if (t->opcode_modifier.floatd)
4930 found_reverse_match = Opcode_FloatD;
4931 else
4932 found_reverse_match = 0;
4933 if (t->opcode_modifier.floatr)
4934 found_reverse_match |= Opcode_FloatR;
4935 }
4936 else
4937 {
4938 /* Found a forward 2 operand match here. */
4939 switch (t->operands)
4940 {
4941 case 5:
4942 overlap4 = operand_type_and (i.types[4],
4943 operand_types[4]);
4944 case 4:
4945 overlap3 = operand_type_and (i.types[3],
4946 operand_types[3]);
4947 case 3:
4948 overlap2 = operand_type_and (i.types[2],
4949 operand_types[2]);
4950 break;
4951 }
4952
4953 switch (t->operands)
4954 {
4955 case 5:
4956 if (!operand_type_match (overlap4, i.types[4])
4957 || !operand_type_register_match (overlap3,
4958 i.types[3],
4959 operand_types[3],
4960 overlap4,
4961 i.types[4],
4962 operand_types[4]))
4963 continue;
4964 case 4:
4965 if (!operand_type_match (overlap3, i.types[3])
4966 || (check_register
4967 && !operand_type_register_match (overlap2,
4968 i.types[2],
4969 operand_types[2],
4970 overlap3,
4971 i.types[3],
4972 operand_types[3])))
4973 continue;
4974 case 3:
4975 /* Here we make use of the fact that there are no
4976 reverse match 3 operand instructions, and all 3
4977 operand instructions only need to be checked for
4978 register consistency between operands 2 and 3. */
4979 if (!operand_type_match (overlap2, i.types[2])
4980 || (check_register
4981 && !operand_type_register_match (overlap1,
4982 i.types[1],
4983 operand_types[1],
4984 overlap2,
4985 i.types[2],
4986 operand_types[2])))
4987 continue;
4988 break;
4989 }
4990 }
4991 /* Found either forward/reverse 2, 3 or 4 operand match here:
4992 slip through to break. */
4993 }
4994 if (!found_cpu_match)
4995 {
4996 found_reverse_match = 0;
4997 continue;
4998 }
4999
5000 /* Check if vector and VEX operands are valid. */
5001 if (check_VecOperands (t) || VEX_check_operands (t))
5002 {
5003 specific_error = i.error;
5004 continue;
5005 }
5006
5007 /* We've found a match; break out of loop. */
5008 break;
5009 }
5010
5011 if (t == current_templates->end)
5012 {
5013 /* We found no match. */
5014 const char *err_msg;
5015 switch (specific_error ? specific_error : i.error)
5016 {
5017 default:
5018 abort ();
5019 case operand_size_mismatch:
5020 err_msg = _("operand size mismatch");
5021 break;
5022 case operand_type_mismatch:
5023 err_msg = _("operand type mismatch");
5024 break;
5025 case register_type_mismatch:
5026 err_msg = _("register type mismatch");
5027 break;
5028 case number_of_operands_mismatch:
5029 err_msg = _("number of operands mismatch");
5030 break;
5031 case invalid_instruction_suffix:
5032 err_msg = _("invalid instruction suffix");
5033 break;
5034 case bad_imm4:
5035 err_msg = _("constant doesn't fit in 4 bits");
5036 break;
5037 case old_gcc_only:
5038 err_msg = _("only supported with old gcc");
5039 break;
5040 case unsupported_with_intel_mnemonic:
5041 err_msg = _("unsupported with Intel mnemonic");
5042 break;
5043 case unsupported_syntax:
5044 err_msg = _("unsupported syntax");
5045 break;
5046 case unsupported:
5047 as_bad (_("unsupported instruction `%s'"),
5048 current_templates->start->name);
5049 return NULL;
5050 case invalid_vsib_address:
5051 err_msg = _("invalid VSIB address");
5052 break;
5053 case invalid_vector_register_set:
5054 err_msg = _("mask, index, and destination registers must be distinct");
5055 break;
5056 case unsupported_vector_index_register:
5057 err_msg = _("unsupported vector index register");
5058 break;
5059 case unsupported_broadcast:
5060 err_msg = _("unsupported broadcast");
5061 break;
5062 case broadcast_not_on_src_operand:
5063 err_msg = _("broadcast not on source memory operand");
5064 break;
5065 case broadcast_needed:
5066 err_msg = _("broadcast is needed for operand of such type");
5067 break;
5068 case unsupported_masking:
5069 err_msg = _("unsupported masking");
5070 break;
5071 case mask_not_on_destination:
5072 err_msg = _("mask not on destination operand");
5073 break;
5074 case no_default_mask:
5075 err_msg = _("default mask isn't allowed");
5076 break;
5077 case unsupported_rc_sae:
5078 err_msg = _("unsupported static rounding/sae");
5079 break;
5080 case rc_sae_operand_not_last_imm:
5081 if (intel_syntax)
5082 err_msg = _("RC/SAE operand must precede immediate operands");
5083 else
5084 err_msg = _("RC/SAE operand must follow immediate operands");
5085 break;
5086 case invalid_register_operand:
5087 err_msg = _("invalid register operand");
5088 break;
5089 }
5090 as_bad (_("%s for `%s'"), err_msg,
5091 current_templates->start->name);
5092 return NULL;
5093 }
5094
5095 if (!quiet_warnings)
5096 {
5097 if (!intel_syntax
5098 && (i.types[0].bitfield.jumpabsolute
5099 != operand_types[0].bitfield.jumpabsolute))
5100 {
5101 as_warn (_("indirect %s without `*'"), t->name);
5102 }
5103
5104 if (t->opcode_modifier.isprefix
5105 && t->opcode_modifier.ignoresize)
5106 {
5107 /* Warn them that a data or address size prefix doesn't
5108 affect assembly of the next line of code. */
5109 as_warn (_("stand-alone `%s' prefix"), t->name);
5110 }
5111 }
5112
5113 /* Copy the template we found. */
5114 i.tm = *t;
5115
5116 if (addr_prefix_disp != -1)
5117 i.tm.operand_types[addr_prefix_disp]
5118 = operand_types[addr_prefix_disp];
5119
5120 if (found_reverse_match)
5121 {
5122 /* If we found a reverse match we must alter the opcode
5123 direction bit. found_reverse_match holds bits to change
5124 (different for int & float insns). */
5125
5126 i.tm.base_opcode ^= found_reverse_match;
5127
5128 i.tm.operand_types[0] = operand_types[1];
5129 i.tm.operand_types[1] = operand_types[0];
5130 }
5131
5132 return t;
5133 }
5134
5135 static int
5136 check_string (void)
5137 {
5138 int mem_op = operand_type_check (i.types[0], anymem) ? 0 : 1;
5139 if (i.tm.operand_types[mem_op].bitfield.esseg)
5140 {
5141 if (i.seg[0] != NULL && i.seg[0] != &es)
5142 {
5143 as_bad (_("`%s' operand %d must use `%ses' segment"),
5144 i.tm.name,
5145 mem_op + 1,
5146 register_prefix);
5147 return 0;
5148 }
5149 /* There's only ever one segment override allowed per instruction.
5150 This instruction possibly has a legal segment override on the
5151 second operand, so copy the segment to where non-string
5152 instructions store it, allowing common code. */
5153 i.seg[0] = i.seg[1];
5154 }
5155 else if (i.tm.operand_types[mem_op + 1].bitfield.esseg)
5156 {
5157 if (i.seg[1] != NULL && i.seg[1] != &es)
5158 {
5159 as_bad (_("`%s' operand %d must use `%ses' segment"),
5160 i.tm.name,
5161 mem_op + 2,
5162 register_prefix);
5163 return 0;
5164 }
5165 }
5166 return 1;
5167 }
5168
5169 static int
5170 process_suffix (void)
5171 {
5172 /* If matched instruction specifies an explicit instruction mnemonic
5173 suffix, use it. */
5174 if (i.tm.opcode_modifier.size16)
5175 i.suffix = WORD_MNEM_SUFFIX;
5176 else if (i.tm.opcode_modifier.size32)
5177 i.suffix = LONG_MNEM_SUFFIX;
5178 else if (i.tm.opcode_modifier.size64)
5179 i.suffix = QWORD_MNEM_SUFFIX;
5180 else if (i.reg_operands)
5181 {
5182 /* If there's no instruction mnemonic suffix we try to invent one
5183 based on register operands. */
5184 if (!i.suffix)
5185 {
5186 /* We take i.suffix from the last register operand specified,
5187 Destination register type is more significant than source
5188 register type. crc32 in SSE4.2 prefers source register
5189 type. */
5190 if (i.tm.base_opcode == 0xf20f38f1)
5191 {
5192 if (i.types[0].bitfield.reg16)
5193 i.suffix = WORD_MNEM_SUFFIX;
5194 else if (i.types[0].bitfield.reg32)
5195 i.suffix = LONG_MNEM_SUFFIX;
5196 else if (i.types[0].bitfield.reg64)
5197 i.suffix = QWORD_MNEM_SUFFIX;
5198 }
5199 else if (i.tm.base_opcode == 0xf20f38f0)
5200 {
5201 if (i.types[0].bitfield.reg8)
5202 i.suffix = BYTE_MNEM_SUFFIX;
5203 }
5204
5205 if (!i.suffix)
5206 {
5207 int op;
5208
5209 if (i.tm.base_opcode == 0xf20f38f1
5210 || i.tm.base_opcode == 0xf20f38f0)
5211 {
5212 /* We have to know the operand size for crc32. */
5213 as_bad (_("ambiguous memory operand size for `%s`"),
5214 i.tm.name);
5215 return 0;
5216 }
5217
5218 for (op = i.operands; --op >= 0;)
5219 if (!i.tm.operand_types[op].bitfield.inoutportreg)
5220 {
5221 if (i.types[op].bitfield.reg8)
5222 {
5223 i.suffix = BYTE_MNEM_SUFFIX;
5224 break;
5225 }
5226 else if (i.types[op].bitfield.reg16)
5227 {
5228 i.suffix = WORD_MNEM_SUFFIX;
5229 break;
5230 }
5231 else if (i.types[op].bitfield.reg32)
5232 {
5233 i.suffix = LONG_MNEM_SUFFIX;
5234 break;
5235 }
5236 else if (i.types[op].bitfield.reg64)
5237 {
5238 i.suffix = QWORD_MNEM_SUFFIX;
5239 break;
5240 }
5241 }
5242 }
5243 }
5244 else if (i.suffix == BYTE_MNEM_SUFFIX)
5245 {
5246 if (intel_syntax
5247 && i.tm.opcode_modifier.ignoresize
5248 && i.tm.opcode_modifier.no_bsuf)
5249 i.suffix = 0;
5250 else if (!check_byte_reg ())
5251 return 0;
5252 }
5253 else if (i.suffix == LONG_MNEM_SUFFIX)
5254 {
5255 if (intel_syntax
5256 && i.tm.opcode_modifier.ignoresize
5257 && i.tm.opcode_modifier.no_lsuf)
5258 i.suffix = 0;
5259 else if (!check_long_reg ())
5260 return 0;
5261 }
5262 else if (i.suffix == QWORD_MNEM_SUFFIX)
5263 {
5264 if (intel_syntax
5265 && i.tm.opcode_modifier.ignoresize
5266 && i.tm.opcode_modifier.no_qsuf)
5267 i.suffix = 0;
5268 else if (!check_qword_reg ())
5269 return 0;
5270 }
5271 else if (i.suffix == WORD_MNEM_SUFFIX)
5272 {
5273 if (intel_syntax
5274 && i.tm.opcode_modifier.ignoresize
5275 && i.tm.opcode_modifier.no_wsuf)
5276 i.suffix = 0;
5277 else if (!check_word_reg ())
5278 return 0;
5279 }
5280 else if (i.suffix == XMMWORD_MNEM_SUFFIX
5281 || i.suffix == YMMWORD_MNEM_SUFFIX
5282 || i.suffix == ZMMWORD_MNEM_SUFFIX)
5283 {
5284 /* Skip if the instruction has x/y/z suffix. match_template
5285 should check if it is a valid suffix. */
5286 }
5287 else if (intel_syntax && i.tm.opcode_modifier.ignoresize)
5288 /* Do nothing if the instruction is going to ignore the prefix. */
5289 ;
5290 else
5291 abort ();
5292 }
5293 else if (i.tm.opcode_modifier.defaultsize
5294 && !i.suffix
5295 /* exclude fldenv/frstor/fsave/fstenv */
5296 && i.tm.opcode_modifier.no_ssuf)
5297 {
5298 i.suffix = stackop_size;
5299 }
5300 else if (intel_syntax
5301 && !i.suffix
5302 && (i.tm.operand_types[0].bitfield.jumpabsolute
5303 || i.tm.opcode_modifier.jumpbyte
5304 || i.tm.opcode_modifier.jumpintersegment
5305 || (i.tm.base_opcode == 0x0f01 /* [ls][gi]dt */
5306 && i.tm.extension_opcode <= 3)))
5307 {
5308 switch (flag_code)
5309 {
5310 case CODE_64BIT:
5311 if (!i.tm.opcode_modifier.no_qsuf)
5312 {
5313 i.suffix = QWORD_MNEM_SUFFIX;
5314 break;
5315 }
5316 case CODE_32BIT:
5317 if (!i.tm.opcode_modifier.no_lsuf)
5318 i.suffix = LONG_MNEM_SUFFIX;
5319 break;
5320 case CODE_16BIT:
5321 if (!i.tm.opcode_modifier.no_wsuf)
5322 i.suffix = WORD_MNEM_SUFFIX;
5323 break;
5324 }
5325 }
5326
5327 if (!i.suffix)
5328 {
5329 if (!intel_syntax)
5330 {
5331 if (i.tm.opcode_modifier.w)
5332 {
5333 as_bad (_("no instruction mnemonic suffix given and "
5334 "no register operands; can't size instruction"));
5335 return 0;
5336 }
5337 }
5338 else
5339 {
5340 unsigned int suffixes;
5341
5342 suffixes = !i.tm.opcode_modifier.no_bsuf;
5343 if (!i.tm.opcode_modifier.no_wsuf)
5344 suffixes |= 1 << 1;
5345 if (!i.tm.opcode_modifier.no_lsuf)
5346 suffixes |= 1 << 2;
5347 if (!i.tm.opcode_modifier.no_ldsuf)
5348 suffixes |= 1 << 3;
5349 if (!i.tm.opcode_modifier.no_ssuf)
5350 suffixes |= 1 << 4;
5351 if (!i.tm.opcode_modifier.no_qsuf)
5352 suffixes |= 1 << 5;
5353
5354 /* There are more than suffix matches. */
5355 if (i.tm.opcode_modifier.w
5356 || ((suffixes & (suffixes - 1))
5357 && !i.tm.opcode_modifier.defaultsize
5358 && !i.tm.opcode_modifier.ignoresize))
5359 {
5360 as_bad (_("ambiguous operand size for `%s'"), i.tm.name);
5361 return 0;
5362 }
5363 }
5364 }
5365
5366 /* Change the opcode based on the operand size given by i.suffix;
5367 We don't need to change things for byte insns. */
5368
5369 if (i.suffix
5370 && i.suffix != BYTE_MNEM_SUFFIX
5371 && i.suffix != XMMWORD_MNEM_SUFFIX
5372 && i.suffix != YMMWORD_MNEM_SUFFIX
5373 && i.suffix != ZMMWORD_MNEM_SUFFIX)
5374 {
5375 /* It's not a byte, select word/dword operation. */
5376 if (i.tm.opcode_modifier.w)
5377 {
5378 if (i.tm.opcode_modifier.shortform)
5379 i.tm.base_opcode |= 8;
5380 else
5381 i.tm.base_opcode |= 1;
5382 }
5383
5384 /* Now select between word & dword operations via the operand
5385 size prefix, except for instructions that will ignore this
5386 prefix anyway. */
5387 if (i.tm.opcode_modifier.addrprefixop0)
5388 {
5389 /* The address size override prefix changes the size of the
5390 first operand. */
5391 if ((flag_code == CODE_32BIT
5392 && i.op->regs[0].reg_type.bitfield.reg16)
5393 || (flag_code != CODE_32BIT
5394 && i.op->regs[0].reg_type.bitfield.reg32))
5395 if (!add_prefix (ADDR_PREFIX_OPCODE))
5396 return 0;
5397 }
5398 else if (i.suffix != QWORD_MNEM_SUFFIX
5399 && i.suffix != LONG_DOUBLE_MNEM_SUFFIX
5400 && !i.tm.opcode_modifier.ignoresize
5401 && !i.tm.opcode_modifier.floatmf
5402 && ((i.suffix == LONG_MNEM_SUFFIX) == (flag_code == CODE_16BIT)
5403 || (flag_code == CODE_64BIT
5404 && i.tm.opcode_modifier.jumpbyte)))
5405 {
5406 unsigned int prefix = DATA_PREFIX_OPCODE;
5407
5408 if (i.tm.opcode_modifier.jumpbyte) /* jcxz, loop */
5409 prefix = ADDR_PREFIX_OPCODE;
5410
5411 if (!add_prefix (prefix))
5412 return 0;
5413 }
5414
5415 /* Set mode64 for an operand. */
5416 if (i.suffix == QWORD_MNEM_SUFFIX
5417 && flag_code == CODE_64BIT
5418 && !i.tm.opcode_modifier.norex64)
5419 {
5420 /* Special case for xchg %rax,%rax. It is NOP and doesn't
5421 need rex64. cmpxchg8b is also a special case. */
5422 if (! (i.operands == 2
5423 && i.tm.base_opcode == 0x90
5424 && i.tm.extension_opcode == None
5425 && operand_type_equal (&i.types [0], &acc64)
5426 && operand_type_equal (&i.types [1], &acc64))
5427 && ! (i.operands == 1
5428 && i.tm.base_opcode == 0xfc7
5429 && i.tm.extension_opcode == 1
5430 && !operand_type_check (i.types [0], reg)
5431 && operand_type_check (i.types [0], anymem)))
5432 i.rex |= REX_W;
5433 }
5434
5435 /* Size floating point instruction. */
5436 if (i.suffix == LONG_MNEM_SUFFIX)
5437 if (i.tm.opcode_modifier.floatmf)
5438 i.tm.base_opcode ^= 4;
5439 }
5440
5441 return 1;
5442 }
5443
5444 static int
5445 check_byte_reg (void)
5446 {
5447 int op;
5448
5449 for (op = i.operands; --op >= 0;)
5450 {
5451 /* If this is an eight bit register, it's OK. If it's the 16 or
5452 32 bit version of an eight bit register, we will just use the
5453 low portion, and that's OK too. */
5454 if (i.types[op].bitfield.reg8)
5455 continue;
5456
5457 /* I/O port address operands are OK too. */
5458 if (i.tm.operand_types[op].bitfield.inoutportreg)
5459 continue;
5460
5461 /* crc32 doesn't generate this warning. */
5462 if (i.tm.base_opcode == 0xf20f38f0)
5463 continue;
5464
5465 if ((i.types[op].bitfield.reg16
5466 || i.types[op].bitfield.reg32
5467 || i.types[op].bitfield.reg64)
5468 && i.op[op].regs->reg_num < 4
5469 /* Prohibit these changes in 64bit mode, since the lowering
5470 would be more complicated. */
5471 && flag_code != CODE_64BIT)
5472 {
5473 #if REGISTER_WARNINGS
5474 if (!quiet_warnings)
5475 as_warn (_("using `%s%s' instead of `%s%s' due to `%c' suffix"),
5476 register_prefix,
5477 (i.op[op].regs + (i.types[op].bitfield.reg16
5478 ? REGNAM_AL - REGNAM_AX
5479 : REGNAM_AL - REGNAM_EAX))->reg_name,
5480 register_prefix,
5481 i.op[op].regs->reg_name,
5482 i.suffix);
5483 #endif
5484 continue;
5485 }
5486 /* Any other register is bad. */
5487 if (i.types[op].bitfield.reg16
5488 || i.types[op].bitfield.reg32
5489 || i.types[op].bitfield.reg64
5490 || i.types[op].bitfield.regmmx
5491 || i.types[op].bitfield.regxmm
5492 || i.types[op].bitfield.regymm
5493 || i.types[op].bitfield.regzmm
5494 || i.types[op].bitfield.sreg2
5495 || i.types[op].bitfield.sreg3
5496 || i.types[op].bitfield.control
5497 || i.types[op].bitfield.debug
5498 || i.types[op].bitfield.test
5499 || i.types[op].bitfield.floatreg
5500 || i.types[op].bitfield.floatacc)
5501 {
5502 as_bad (_("`%s%s' not allowed with `%s%c'"),
5503 register_prefix,
5504 i.op[op].regs->reg_name,
5505 i.tm.name,
5506 i.suffix);
5507 return 0;
5508 }
5509 }
5510 return 1;
5511 }
5512
5513 static int
5514 check_long_reg (void)
5515 {
5516 int op;
5517
5518 for (op = i.operands; --op >= 0;)
5519 /* Reject eight bit registers, except where the template requires
5520 them. (eg. movzb) */
5521 if (i.types[op].bitfield.reg8
5522 && (i.tm.operand_types[op].bitfield.reg16
5523 || i.tm.operand_types[op].bitfield.reg32
5524 || i.tm.operand_types[op].bitfield.acc))
5525 {
5526 as_bad (_("`%s%s' not allowed with `%s%c'"),
5527 register_prefix,
5528 i.op[op].regs->reg_name,
5529 i.tm.name,
5530 i.suffix);
5531 return 0;
5532 }
5533 /* Warn if the e prefix on a general reg is missing. */
5534 else if ((!quiet_warnings || flag_code == CODE_64BIT)
5535 && i.types[op].bitfield.reg16
5536 && (i.tm.operand_types[op].bitfield.reg32
5537 || i.tm.operand_types[op].bitfield.acc))
5538 {
5539 /* Prohibit these changes in the 64bit mode, since the
5540 lowering is more complicated. */
5541 if (flag_code == CODE_64BIT)
5542 {
5543 as_bad (_("incorrect register `%s%s' used with `%c' suffix"),
5544 register_prefix, i.op[op].regs->reg_name,
5545 i.suffix);
5546 return 0;
5547 }
5548 #if REGISTER_WARNINGS
5549 as_warn (_("using `%s%s' instead of `%s%s' due to `%c' suffix"),
5550 register_prefix,
5551 (i.op[op].regs + REGNAM_EAX - REGNAM_AX)->reg_name,
5552 register_prefix, i.op[op].regs->reg_name, i.suffix);
5553 #endif
5554 }
5555 /* Warn if the r prefix on a general reg is present. */
5556 else if (i.types[op].bitfield.reg64
5557 && (i.tm.operand_types[op].bitfield.reg32
5558 || i.tm.operand_types[op].bitfield.acc))
5559 {
5560 if (intel_syntax
5561 && i.tm.opcode_modifier.toqword
5562 && !i.types[0].bitfield.regxmm)
5563 {
5564 /* Convert to QWORD. We want REX byte. */
5565 i.suffix = QWORD_MNEM_SUFFIX;
5566 }
5567 else
5568 {
5569 as_bad (_("incorrect register `%s%s' used with `%c' suffix"),
5570 register_prefix, i.op[op].regs->reg_name,
5571 i.suffix);
5572 return 0;
5573 }
5574 }
5575 return 1;
5576 }
5577
5578 static int
5579 check_qword_reg (void)
5580 {
5581 int op;
5582
5583 for (op = i.operands; --op >= 0; )
5584 /* Reject eight bit registers, except where the template requires
5585 them. (eg. movzb) */
5586 if (i.types[op].bitfield.reg8
5587 && (i.tm.operand_types[op].bitfield.reg16
5588 || i.tm.operand_types[op].bitfield.reg32
5589 || i.tm.operand_types[op].bitfield.acc))
5590 {
5591 as_bad (_("`%s%s' not allowed with `%s%c'"),
5592 register_prefix,
5593 i.op[op].regs->reg_name,
5594 i.tm.name,
5595 i.suffix);
5596 return 0;
5597 }
5598 /* Warn if the r prefix on a general reg is missing. */
5599 else if ((i.types[op].bitfield.reg16
5600 || i.types[op].bitfield.reg32)
5601 && (i.tm.operand_types[op].bitfield.reg32
5602 || i.tm.operand_types[op].bitfield.acc))
5603 {
5604 /* Prohibit these changes in the 64bit mode, since the
5605 lowering is more complicated. */
5606 if (intel_syntax
5607 && i.tm.opcode_modifier.todword
5608 && !i.types[0].bitfield.regxmm)
5609 {
5610 /* Convert to DWORD. We don't want REX byte. */
5611 i.suffix = LONG_MNEM_SUFFIX;
5612 }
5613 else
5614 {
5615 as_bad (_("incorrect register `%s%s' used with `%c' suffix"),
5616 register_prefix, i.op[op].regs->reg_name,
5617 i.suffix);
5618 return 0;
5619 }
5620 }
5621 return 1;
5622 }
5623
5624 static int
5625 check_word_reg (void)
5626 {
5627 int op;
5628 for (op = i.operands; --op >= 0;)
5629 /* Reject eight bit registers, except where the template requires
5630 them. (eg. movzb) */
5631 if (i.types[op].bitfield.reg8
5632 && (i.tm.operand_types[op].bitfield.reg16
5633 || i.tm.operand_types[op].bitfield.reg32
5634 || i.tm.operand_types[op].bitfield.acc))
5635 {
5636 as_bad (_("`%s%s' not allowed with `%s%c'"),
5637 register_prefix,
5638 i.op[op].regs->reg_name,
5639 i.tm.name,
5640 i.suffix);
5641 return 0;
5642 }
5643 /* Warn if the e or r prefix on a general reg is present. */
5644 else if ((!quiet_warnings || flag_code == CODE_64BIT)
5645 && (i.types[op].bitfield.reg32
5646 || i.types[op].bitfield.reg64)
5647 && (i.tm.operand_types[op].bitfield.reg16
5648 || i.tm.operand_types[op].bitfield.acc))
5649 {
5650 /* Prohibit these changes in the 64bit mode, since the
5651 lowering is more complicated. */
5652 if (flag_code == CODE_64BIT)
5653 {
5654 as_bad (_("incorrect register `%s%s' used with `%c' suffix"),
5655 register_prefix, i.op[op].regs->reg_name,
5656 i.suffix);
5657 return 0;
5658 }
5659 #if REGISTER_WARNINGS
5660 as_warn (_("using `%s%s' instead of `%s%s' due to `%c' suffix"),
5661 register_prefix,
5662 (i.op[op].regs + REGNAM_AX - REGNAM_EAX)->reg_name,
5663 register_prefix, i.op[op].regs->reg_name, i.suffix);
5664 #endif
5665 }
5666 return 1;
5667 }
5668
5669 static int
5670 update_imm (unsigned int j)
5671 {
5672 i386_operand_type overlap = i.types[j];
5673 if ((overlap.bitfield.imm8
5674 || overlap.bitfield.imm8s
5675 || overlap.bitfield.imm16
5676 || overlap.bitfield.imm32
5677 || overlap.bitfield.imm32s
5678 || overlap.bitfield.imm64)
5679 && !operand_type_equal (&overlap, &imm8)
5680 && !operand_type_equal (&overlap, &imm8s)
5681 && !operand_type_equal (&overlap, &imm16)
5682 && !operand_type_equal (&overlap, &imm32)
5683 && !operand_type_equal (&overlap, &imm32s)
5684 && !operand_type_equal (&overlap, &imm64))
5685 {
5686 if (i.suffix)
5687 {
5688 i386_operand_type temp;
5689
5690 operand_type_set (&temp, 0);
5691 if (i.suffix == BYTE_MNEM_SUFFIX)
5692 {
5693 temp.bitfield.imm8 = overlap.bitfield.imm8;
5694 temp.bitfield.imm8s = overlap.bitfield.imm8s;
5695 }
5696 else if (i.suffix == WORD_MNEM_SUFFIX)
5697 temp.bitfield.imm16 = overlap.bitfield.imm16;
5698 else if (i.suffix == QWORD_MNEM_SUFFIX)
5699 {
5700 temp.bitfield.imm64 = overlap.bitfield.imm64;
5701 temp.bitfield.imm32s = overlap.bitfield.imm32s;
5702 }
5703 else
5704 temp.bitfield.imm32 = overlap.bitfield.imm32;
5705 overlap = temp;
5706 }
5707 else if (operand_type_equal (&overlap, &imm16_32_32s)
5708 || operand_type_equal (&overlap, &imm16_32)
5709 || operand_type_equal (&overlap, &imm16_32s))
5710 {
5711 if ((flag_code == CODE_16BIT) ^ (i.prefix[DATA_PREFIX] != 0))
5712 overlap = imm16;
5713 else
5714 overlap = imm32s;
5715 }
5716 if (!operand_type_equal (&overlap, &imm8)
5717 && !operand_type_equal (&overlap, &imm8s)
5718 && !operand_type_equal (&overlap, &imm16)
5719 && !operand_type_equal (&overlap, &imm32)
5720 && !operand_type_equal (&overlap, &imm32s)
5721 && !operand_type_equal (&overlap, &imm64))
5722 {
5723 as_bad (_("no instruction mnemonic suffix given; "
5724 "can't determine immediate size"));
5725 return 0;
5726 }
5727 }
5728 i.types[j] = overlap;
5729
5730 return 1;
5731 }
5732
5733 static int
5734 finalize_imm (void)
5735 {
5736 unsigned int j, n;
5737
5738 /* Update the first 2 immediate operands. */
5739 n = i.operands > 2 ? 2 : i.operands;
5740 if (n)
5741 {
5742 for (j = 0; j < n; j++)
5743 if (update_imm (j) == 0)
5744 return 0;
5745
5746 /* The 3rd operand can't be immediate operand. */
5747 gas_assert (operand_type_check (i.types[2], imm) == 0);
5748 }
5749
5750 return 1;
5751 }
5752
5753 static int
5754 bad_implicit_operand (int xmm)
5755 {
5756 const char *ireg = xmm ? "xmm0" : "ymm0";
5757
5758 if (intel_syntax)
5759 as_bad (_("the last operand of `%s' must be `%s%s'"),
5760 i.tm.name, register_prefix, ireg);
5761 else
5762 as_bad (_("the first operand of `%s' must be `%s%s'"),
5763 i.tm.name, register_prefix, ireg);
5764 return 0;
5765 }
5766
5767 static int
5768 process_operands (void)
5769 {
5770 /* Default segment register this instruction will use for memory
5771 accesses. 0 means unknown. This is only for optimizing out
5772 unnecessary segment overrides. */
5773 const seg_entry *default_seg = 0;
5774
5775 if (i.tm.opcode_modifier.sse2avx && i.tm.opcode_modifier.vexvvvv)
5776 {
5777 unsigned int dupl = i.operands;
5778 unsigned int dest = dupl - 1;
5779 unsigned int j;
5780
5781 /* The destination must be an xmm register. */
5782 gas_assert (i.reg_operands
5783 && MAX_OPERANDS > dupl
5784 && operand_type_equal (&i.types[dest], &regxmm));
5785
5786 if (i.tm.opcode_modifier.firstxmm0)
5787 {
5788 /* The first operand is implicit and must be xmm0. */
5789 gas_assert (operand_type_equal (&i.types[0], &regxmm));
5790 if (register_number (i.op[0].regs) != 0)
5791 return bad_implicit_operand (1);
5792
5793 if (i.tm.opcode_modifier.vexsources == VEX3SOURCES)
5794 {
5795 /* Keep xmm0 for instructions with VEX prefix and 3
5796 sources. */
5797 goto duplicate;
5798 }
5799 else
5800 {
5801 /* We remove the first xmm0 and keep the number of
5802 operands unchanged, which in fact duplicates the
5803 destination. */
5804 for (j = 1; j < i.operands; j++)
5805 {
5806 i.op[j - 1] = i.op[j];
5807 i.types[j - 1] = i.types[j];
5808 i.tm.operand_types[j - 1] = i.tm.operand_types[j];
5809 }
5810 }
5811 }
5812 else if (i.tm.opcode_modifier.implicit1stxmm0)
5813 {
5814 gas_assert ((MAX_OPERANDS - 1) > dupl
5815 && (i.tm.opcode_modifier.vexsources
5816 == VEX3SOURCES));
5817
5818 /* Add the implicit xmm0 for instructions with VEX prefix
5819 and 3 sources. */
5820 for (j = i.operands; j > 0; j--)
5821 {
5822 i.op[j] = i.op[j - 1];
5823 i.types[j] = i.types[j - 1];
5824 i.tm.operand_types[j] = i.tm.operand_types[j - 1];
5825 }
5826 i.op[0].regs
5827 = (const reg_entry *) hash_find (reg_hash, "xmm0");
5828 i.types[0] = regxmm;
5829 i.tm.operand_types[0] = regxmm;
5830
5831 i.operands += 2;
5832 i.reg_operands += 2;
5833 i.tm.operands += 2;
5834
5835 dupl++;
5836 dest++;
5837 i.op[dupl] = i.op[dest];
5838 i.types[dupl] = i.types[dest];
5839 i.tm.operand_types[dupl] = i.tm.operand_types[dest];
5840 }
5841 else
5842 {
5843 duplicate:
5844 i.operands++;
5845 i.reg_operands++;
5846 i.tm.operands++;
5847
5848 i.op[dupl] = i.op[dest];
5849 i.types[dupl] = i.types[dest];
5850 i.tm.operand_types[dupl] = i.tm.operand_types[dest];
5851 }
5852
5853 if (i.tm.opcode_modifier.immext)
5854 process_immext ();
5855 }
5856 else if (i.tm.opcode_modifier.firstxmm0)
5857 {
5858 unsigned int j;
5859
5860 /* The first operand is implicit and must be xmm0/ymm0/zmm0. */
5861 gas_assert (i.reg_operands
5862 && (operand_type_equal (&i.types[0], &regxmm)
5863 || operand_type_equal (&i.types[0], &regymm)
5864 || operand_type_equal (&i.types[0], &regzmm)));
5865 if (register_number (i.op[0].regs) != 0)
5866 return bad_implicit_operand (i.types[0].bitfield.regxmm);
5867
5868 for (j = 1; j < i.operands; j++)
5869 {
5870 i.op[j - 1] = i.op[j];
5871 i.types[j - 1] = i.types[j];
5872
5873 /* We need to adjust fields in i.tm since they are used by
5874 build_modrm_byte. */
5875 i.tm.operand_types [j - 1] = i.tm.operand_types [j];
5876 }
5877
5878 i.operands--;
5879 i.reg_operands--;
5880 i.tm.operands--;
5881 }
5882 else if (i.tm.opcode_modifier.regkludge)
5883 {
5884 /* The imul $imm, %reg instruction is converted into
5885 imul $imm, %reg, %reg, and the clr %reg instruction
5886 is converted into xor %reg, %reg. */
5887
5888 unsigned int first_reg_op;
5889
5890 if (operand_type_check (i.types[0], reg))
5891 first_reg_op = 0;
5892 else
5893 first_reg_op = 1;
5894 /* Pretend we saw the extra register operand. */
5895 gas_assert (i.reg_operands == 1
5896 && i.op[first_reg_op + 1].regs == 0);
5897 i.op[first_reg_op + 1].regs = i.op[first_reg_op].regs;
5898 i.types[first_reg_op + 1] = i.types[first_reg_op];
5899 i.operands++;
5900 i.reg_operands++;
5901 }
5902
5903 if (i.tm.opcode_modifier.shortform)
5904 {
5905 if (i.types[0].bitfield.sreg2
5906 || i.types[0].bitfield.sreg3)
5907 {
5908 if (i.tm.base_opcode == POP_SEG_SHORT
5909 && i.op[0].regs->reg_num == 1)
5910 {
5911 as_bad (_("you can't `pop %scs'"), register_prefix);
5912 return 0;
5913 }
5914 i.tm.base_opcode |= (i.op[0].regs->reg_num << 3);
5915 if ((i.op[0].regs->reg_flags & RegRex) != 0)
5916 i.rex |= REX_B;
5917 }
5918 else
5919 {
5920 /* The register or float register operand is in operand
5921 0 or 1. */
5922 unsigned int op;
5923
5924 if (i.types[0].bitfield.floatreg
5925 || operand_type_check (i.types[0], reg))
5926 op = 0;
5927 else
5928 op = 1;
5929 /* Register goes in low 3 bits of opcode. */
5930 i.tm.base_opcode |= i.op[op].regs->reg_num;
5931 if ((i.op[op].regs->reg_flags & RegRex) != 0)
5932 i.rex |= REX_B;
5933 if (!quiet_warnings && i.tm.opcode_modifier.ugh)
5934 {
5935 /* Warn about some common errors, but press on regardless.
5936 The first case can be generated by gcc (<= 2.8.1). */
5937 if (i.operands == 2)
5938 {
5939 /* Reversed arguments on faddp, fsubp, etc. */
5940 as_warn (_("translating to `%s %s%s,%s%s'"), i.tm.name,
5941 register_prefix, i.op[!intel_syntax].regs->reg_name,
5942 register_prefix, i.op[intel_syntax].regs->reg_name);
5943 }
5944 else
5945 {
5946 /* Extraneous `l' suffix on fp insn. */
5947 as_warn (_("translating to `%s %s%s'"), i.tm.name,
5948 register_prefix, i.op[0].regs->reg_name);
5949 }
5950 }
5951 }
5952 }
5953 else if (i.tm.opcode_modifier.modrm)
5954 {
5955 /* The opcode is completed (modulo i.tm.extension_opcode which
5956 must be put into the modrm byte). Now, we make the modrm and
5957 index base bytes based on all the info we've collected. */
5958
5959 default_seg = build_modrm_byte ();
5960 }
5961 else if ((i.tm.base_opcode & ~0x3) == MOV_AX_DISP32)
5962 {
5963 default_seg = &ds;
5964 }
5965 else if (i.tm.opcode_modifier.isstring)
5966 {
5967 /* For the string instructions that allow a segment override
5968 on one of their operands, the default segment is ds. */
5969 default_seg = &ds;
5970 }
5971
5972 if (i.tm.base_opcode == 0x8d /* lea */
5973 && i.seg[0]
5974 && !quiet_warnings)
5975 as_warn (_("segment override on `%s' is ineffectual"), i.tm.name);
5976
5977 /* If a segment was explicitly specified, and the specified segment
5978 is not the default, use an opcode prefix to select it. If we
5979 never figured out what the default segment is, then default_seg
5980 will be zero at this point, and the specified segment prefix will
5981 always be used. */
5982 if ((i.seg[0]) && (i.seg[0] != default_seg))
5983 {
5984 if (!add_prefix (i.seg[0]->seg_prefix))
5985 return 0;
5986 }
5987 return 1;
5988 }
5989
5990 static const seg_entry *
5991 build_modrm_byte (void)
5992 {
5993 const seg_entry *default_seg = 0;
5994 unsigned int source, dest;
5995 int vex_3_sources;
5996
5997 /* The first operand of instructions with VEX prefix and 3 sources
5998 must be VEX_Imm4. */
5999 vex_3_sources = i.tm.opcode_modifier.vexsources == VEX3SOURCES;
6000 if (vex_3_sources)
6001 {
6002 unsigned int nds, reg_slot;
6003 expressionS *exp;
6004
6005 if (i.tm.opcode_modifier.veximmext
6006 && i.tm.opcode_modifier.immext)
6007 {
6008 dest = i.operands - 2;
6009 gas_assert (dest == 3);
6010 }
6011 else
6012 dest = i.operands - 1;
6013 nds = dest - 1;
6014
6015 /* There are 2 kinds of instructions:
6016 1. 5 operands: 4 register operands or 3 register operands
6017 plus 1 memory operand plus one Vec_Imm4 operand, VexXDS, and
6018 VexW0 or VexW1. The destination must be either XMM, YMM or
6019 ZMM register.
6020 2. 4 operands: 4 register operands or 3 register operands
6021 plus 1 memory operand, VexXDS, and VexImmExt */
6022 gas_assert ((i.reg_operands == 4
6023 || (i.reg_operands == 3 && i.mem_operands == 1))
6024 && i.tm.opcode_modifier.vexvvvv == VEXXDS
6025 && (i.tm.opcode_modifier.veximmext
6026 || (i.imm_operands == 1
6027 && i.types[0].bitfield.vec_imm4
6028 && (i.tm.opcode_modifier.vexw == VEXW0
6029 || i.tm.opcode_modifier.vexw == VEXW1)
6030 && (operand_type_equal (&i.tm.operand_types[dest], &regxmm)
6031 || operand_type_equal (&i.tm.operand_types[dest], &regymm)
6032 || operand_type_equal (&i.tm.operand_types[dest], &regzmm)))));
6033
6034 if (i.imm_operands == 0)
6035 {
6036 /* When there is no immediate operand, generate an 8bit
6037 immediate operand to encode the first operand. */
6038 exp = &im_expressions[i.imm_operands++];
6039 i.op[i.operands].imms = exp;
6040 i.types[i.operands] = imm8;
6041 i.operands++;
6042 /* If VexW1 is set, the first operand is the source and
6043 the second operand is encoded in the immediate operand. */
6044 if (i.tm.opcode_modifier.vexw == VEXW1)
6045 {
6046 source = 0;
6047 reg_slot = 1;
6048 }
6049 else
6050 {
6051 source = 1;
6052 reg_slot = 0;
6053 }
6054
6055 /* FMA swaps REG and NDS. */
6056 if (i.tm.cpu_flags.bitfield.cpufma)
6057 {
6058 unsigned int tmp;
6059 tmp = reg_slot;
6060 reg_slot = nds;
6061 nds = tmp;
6062 }
6063
6064 gas_assert (operand_type_equal (&i.tm.operand_types[reg_slot],
6065 &regxmm)
6066 || operand_type_equal (&i.tm.operand_types[reg_slot],
6067 &regymm)
6068 || operand_type_equal (&i.tm.operand_types[reg_slot],
6069 &regzmm));
6070 exp->X_op = O_constant;
6071 exp->X_add_number = register_number (i.op[reg_slot].regs) << 4;
6072 gas_assert ((i.op[reg_slot].regs->reg_flags & RegVRex) == 0);
6073 }
6074 else
6075 {
6076 unsigned int imm_slot;
6077
6078 if (i.tm.opcode_modifier.vexw == VEXW0)
6079 {
6080 /* If VexW0 is set, the third operand is the source and
6081 the second operand is encoded in the immediate
6082 operand. */
6083 source = 2;
6084 reg_slot = 1;
6085 }
6086 else
6087 {
6088 /* VexW1 is set, the second operand is the source and
6089 the third operand is encoded in the immediate
6090 operand. */
6091 source = 1;
6092 reg_slot = 2;
6093 }
6094
6095 if (i.tm.opcode_modifier.immext)
6096 {
6097 /* When ImmExt is set, the immdiate byte is the last
6098 operand. */
6099 imm_slot = i.operands - 1;
6100 source--;
6101 reg_slot--;
6102 }
6103 else
6104 {
6105 imm_slot = 0;
6106
6107 /* Turn on Imm8 so that output_imm will generate it. */
6108 i.types[imm_slot].bitfield.imm8 = 1;
6109 }
6110
6111 gas_assert (operand_type_equal (&i.tm.operand_types[reg_slot],
6112 &regxmm)
6113 || operand_type_equal (&i.tm.operand_types[reg_slot],
6114 &regymm)
6115 || operand_type_equal (&i.tm.operand_types[reg_slot],
6116 &regzmm));
6117 i.op[imm_slot].imms->X_add_number
6118 |= register_number (i.op[reg_slot].regs) << 4;
6119 gas_assert ((i.op[reg_slot].regs->reg_flags & RegVRex) == 0);
6120 }
6121
6122 gas_assert (operand_type_equal (&i.tm.operand_types[nds], &regxmm)
6123 || operand_type_equal (&i.tm.operand_types[nds],
6124 &regymm)
6125 || operand_type_equal (&i.tm.operand_types[nds],
6126 &regzmm));
6127 i.vex.register_specifier = i.op[nds].regs;
6128 }
6129 else
6130 source = dest = 0;
6131
6132 /* i.reg_operands MUST be the number of real register operands;
6133 implicit registers do not count. If there are 3 register
6134 operands, it must be a instruction with VexNDS. For a
6135 instruction with VexNDD, the destination register is encoded
6136 in VEX prefix. If there are 4 register operands, it must be
6137 a instruction with VEX prefix and 3 sources. */
6138 if (i.mem_operands == 0
6139 && ((i.reg_operands == 2
6140 && i.tm.opcode_modifier.vexvvvv <= VEXXDS)
6141 || (i.reg_operands == 3
6142 && i.tm.opcode_modifier.vexvvvv == VEXXDS)
6143 || (i.reg_operands == 4 && vex_3_sources)))
6144 {
6145 switch (i.operands)
6146 {
6147 case 2:
6148 source = 0;
6149 break;
6150 case 3:
6151 /* When there are 3 operands, one of them may be immediate,
6152 which may be the first or the last operand. Otherwise,
6153 the first operand must be shift count register (cl) or it
6154 is an instruction with VexNDS. */
6155 gas_assert (i.imm_operands == 1
6156 || (i.imm_operands == 0
6157 && (i.tm.opcode_modifier.vexvvvv == VEXXDS
6158 || i.types[0].bitfield.shiftcount)));
6159 if (operand_type_check (i.types[0], imm)
6160 || i.types[0].bitfield.shiftcount)
6161 source = 1;
6162 else
6163 source = 0;
6164 break;
6165 case 4:
6166 /* When there are 4 operands, the first two must be 8bit
6167 immediate operands. The source operand will be the 3rd
6168 one.
6169
6170 For instructions with VexNDS, if the first operand
6171 an imm8, the source operand is the 2nd one. If the last
6172 operand is imm8, the source operand is the first one. */
6173 gas_assert ((i.imm_operands == 2
6174 && i.types[0].bitfield.imm8
6175 && i.types[1].bitfield.imm8)
6176 || (i.tm.opcode_modifier.vexvvvv == VEXXDS
6177 && i.imm_operands == 1
6178 && (i.types[0].bitfield.imm8
6179 || i.types[i.operands - 1].bitfield.imm8
6180 || i.rounding)));
6181 if (i.imm_operands == 2)
6182 source = 2;
6183 else
6184 {
6185 if (i.types[0].bitfield.imm8)
6186 source = 1;
6187 else
6188 source = 0;
6189 }
6190 break;
6191 case 5:
6192 if (i.tm.opcode_modifier.evex)
6193 {
6194 /* For EVEX instructions, when there are 5 operands, the
6195 first one must be immediate operand. If the second one
6196 is immediate operand, the source operand is the 3th
6197 one. If the last one is immediate operand, the source
6198 operand is the 2nd one. */
6199 gas_assert (i.imm_operands == 2
6200 && i.tm.opcode_modifier.sae
6201 && operand_type_check (i.types[0], imm));
6202 if (operand_type_check (i.types[1], imm))
6203 source = 2;
6204 else if (operand_type_check (i.types[4], imm))
6205 source = 1;
6206 else
6207 abort ();
6208 }
6209 break;
6210 default:
6211 abort ();
6212 }
6213
6214 if (!vex_3_sources)
6215 {
6216 dest = source + 1;
6217
6218 /* RC/SAE operand could be between DEST and SRC. That happens
6219 when one operand is GPR and the other one is XMM/YMM/ZMM
6220 register. */
6221 if (i.rounding && i.rounding->operand == (int) dest)
6222 dest++;
6223
6224 if (i.tm.opcode_modifier.vexvvvv == VEXXDS)
6225 {
6226 /* For instructions with VexNDS, the register-only source
6227 operand must be 32/64bit integer, XMM, YMM or ZMM
6228 register. It is encoded in VEX prefix. We need to
6229 clear RegMem bit before calling operand_type_equal. */
6230
6231 i386_operand_type op;
6232 unsigned int vvvv;
6233
6234 /* Check register-only source operand when two source
6235 operands are swapped. */
6236 if (!i.tm.operand_types[source].bitfield.baseindex
6237 && i.tm.operand_types[dest].bitfield.baseindex)
6238 {
6239 vvvv = source;
6240 source = dest;
6241 }
6242 else
6243 vvvv = dest;
6244
6245 op = i.tm.operand_types[vvvv];
6246 op.bitfield.regmem = 0;
6247 if ((dest + 1) >= i.operands
6248 || (!op.bitfield.reg32
6249 && op.bitfield.reg64
6250 && !operand_type_equal (&op, &regxmm)
6251 && !operand_type_equal (&op, &regymm)
6252 && !operand_type_equal (&op, &regzmm)
6253 && !operand_type_equal (&op, &regmask)))
6254 abort ();
6255 i.vex.register_specifier = i.op[vvvv].regs;
6256 dest++;
6257 }
6258 }
6259
6260 i.rm.mode = 3;
6261 /* One of the register operands will be encoded in the i.tm.reg
6262 field, the other in the combined i.tm.mode and i.tm.regmem
6263 fields. If no form of this instruction supports a memory
6264 destination operand, then we assume the source operand may
6265 sometimes be a memory operand and so we need to store the
6266 destination in the i.rm.reg field. */
6267 if (!i.tm.operand_types[dest].bitfield.regmem
6268 && operand_type_check (i.tm.operand_types[dest], anymem) == 0)
6269 {
6270 i.rm.reg = i.op[dest].regs->reg_num;
6271 i.rm.regmem = i.op[source].regs->reg_num;
6272 if ((i.op[dest].regs->reg_flags & RegRex) != 0)
6273 i.rex |= REX_R;
6274 if ((i.op[dest].regs->reg_flags & RegVRex) != 0)
6275 i.vrex |= REX_R;
6276 if ((i.op[source].regs->reg_flags & RegRex) != 0)
6277 i.rex |= REX_B;
6278 if ((i.op[source].regs->reg_flags & RegVRex) != 0)
6279 i.vrex |= REX_B;
6280 }
6281 else
6282 {
6283 i.rm.reg = i.op[source].regs->reg_num;
6284 i.rm.regmem = i.op[dest].regs->reg_num;
6285 if ((i.op[dest].regs->reg_flags & RegRex) != 0)
6286 i.rex |= REX_B;
6287 if ((i.op[dest].regs->reg_flags & RegVRex) != 0)
6288 i.vrex |= REX_B;
6289 if ((i.op[source].regs->reg_flags & RegRex) != 0)
6290 i.rex |= REX_R;
6291 if ((i.op[source].regs->reg_flags & RegVRex) != 0)
6292 i.vrex |= REX_R;
6293 }
6294 if (flag_code != CODE_64BIT && (i.rex & (REX_R | REX_B)))
6295 {
6296 if (!i.types[0].bitfield.control
6297 && !i.types[1].bitfield.control)
6298 abort ();
6299 i.rex &= ~(REX_R | REX_B);
6300 add_prefix (LOCK_PREFIX_OPCODE);
6301 }
6302 }
6303 else
6304 { /* If it's not 2 reg operands... */
6305 unsigned int mem;
6306
6307 if (i.mem_operands)
6308 {
6309 unsigned int fake_zero_displacement = 0;
6310 unsigned int op;
6311
6312 for (op = 0; op < i.operands; op++)
6313 if (operand_type_check (i.types[op], anymem))
6314 break;
6315 gas_assert (op < i.operands);
6316
6317 if (i.tm.opcode_modifier.vecsib)
6318 {
6319 if (i.index_reg->reg_num == RegEiz
6320 || i.index_reg->reg_num == RegRiz)
6321 abort ();
6322
6323 i.rm.regmem = ESCAPE_TO_TWO_BYTE_ADDRESSING;
6324 if (!i.base_reg)
6325 {
6326 i.sib.base = NO_BASE_REGISTER;
6327 i.sib.scale = i.log2_scale_factor;
6328 /* No Vec_Disp8 if there is no base. */
6329 i.types[op].bitfield.vec_disp8 = 0;
6330 i.types[op].bitfield.disp8 = 0;
6331 i.types[op].bitfield.disp16 = 0;
6332 i.types[op].bitfield.disp64 = 0;
6333 if (flag_code != CODE_64BIT)
6334 {
6335 /* Must be 32 bit */
6336 i.types[op].bitfield.disp32 = 1;
6337 i.types[op].bitfield.disp32s = 0;
6338 }
6339 else
6340 {
6341 i.types[op].bitfield.disp32 = 0;
6342 i.types[op].bitfield.disp32s = 1;
6343 }
6344 }
6345 i.sib.index = i.index_reg->reg_num;
6346 if ((i.index_reg->reg_flags & RegRex) != 0)
6347 i.rex |= REX_X;
6348 if ((i.index_reg->reg_flags & RegVRex) != 0)
6349 i.vrex |= REX_X;
6350 }
6351
6352 default_seg = &ds;
6353
6354 if (i.base_reg == 0)
6355 {
6356 i.rm.mode = 0;
6357 if (!i.disp_operands)
6358 {
6359 fake_zero_displacement = 1;
6360 /* Instructions with VSIB byte need 32bit displacement
6361 if there is no base register. */
6362 if (i.tm.opcode_modifier.vecsib)
6363 i.types[op].bitfield.disp32 = 1;
6364 }
6365 if (i.index_reg == 0)
6366 {
6367 gas_assert (!i.tm.opcode_modifier.vecsib);
6368 /* Operand is just <disp> */
6369 if (flag_code == CODE_64BIT)
6370 {
6371 /* 64bit mode overwrites the 32bit absolute
6372 addressing by RIP relative addressing and
6373 absolute addressing is encoded by one of the
6374 redundant SIB forms. */
6375 i.rm.regmem = ESCAPE_TO_TWO_BYTE_ADDRESSING;
6376 i.sib.base = NO_BASE_REGISTER;
6377 i.sib.index = NO_INDEX_REGISTER;
6378 i.types[op] = ((i.prefix[ADDR_PREFIX] == 0)
6379 ? disp32s : disp32);
6380 }
6381 else if ((flag_code == CODE_16BIT)
6382 ^ (i.prefix[ADDR_PREFIX] != 0))
6383 {
6384 i.rm.regmem = NO_BASE_REGISTER_16;
6385 i.types[op] = disp16;
6386 }
6387 else
6388 {
6389 i.rm.regmem = NO_BASE_REGISTER;
6390 i.types[op] = disp32;
6391 }
6392 }
6393 else if (!i.tm.opcode_modifier.vecsib)
6394 {
6395 /* !i.base_reg && i.index_reg */
6396 if (i.index_reg->reg_num == RegEiz
6397 || i.index_reg->reg_num == RegRiz)
6398 i.sib.index = NO_INDEX_REGISTER;
6399 else
6400 i.sib.index = i.index_reg->reg_num;
6401 i.sib.base = NO_BASE_REGISTER;
6402 i.sib.scale = i.log2_scale_factor;
6403 i.rm.regmem = ESCAPE_TO_TWO_BYTE_ADDRESSING;
6404 /* No Vec_Disp8 if there is no base. */
6405 i.types[op].bitfield.vec_disp8 = 0;
6406 i.types[op].bitfield.disp8 = 0;
6407 i.types[op].bitfield.disp16 = 0;
6408 i.types[op].bitfield.disp64 = 0;
6409 if (flag_code != CODE_64BIT)
6410 {
6411 /* Must be 32 bit */
6412 i.types[op].bitfield.disp32 = 1;
6413 i.types[op].bitfield.disp32s = 0;
6414 }
6415 else
6416 {
6417 i.types[op].bitfield.disp32 = 0;
6418 i.types[op].bitfield.disp32s = 1;
6419 }
6420 if ((i.index_reg->reg_flags & RegRex) != 0)
6421 i.rex |= REX_X;
6422 }
6423 }
6424 /* RIP addressing for 64bit mode. */
6425 else if (i.base_reg->reg_num == RegRip ||
6426 i.base_reg->reg_num == RegEip)
6427 {
6428 gas_assert (!i.tm.opcode_modifier.vecsib);
6429 i.rm.regmem = NO_BASE_REGISTER;
6430 i.types[op].bitfield.disp8 = 0;
6431 i.types[op].bitfield.disp16 = 0;
6432 i.types[op].bitfield.disp32 = 0;
6433 i.types[op].bitfield.disp32s = 1;
6434 i.types[op].bitfield.disp64 = 0;
6435 i.types[op].bitfield.vec_disp8 = 0;
6436 i.flags[op] |= Operand_PCrel;
6437 if (! i.disp_operands)
6438 fake_zero_displacement = 1;
6439 }
6440 else if (i.base_reg->reg_type.bitfield.reg16)
6441 {
6442 gas_assert (!i.tm.opcode_modifier.vecsib);
6443 switch (i.base_reg->reg_num)
6444 {
6445 case 3: /* (%bx) */
6446 if (i.index_reg == 0)
6447 i.rm.regmem = 7;
6448 else /* (%bx,%si) -> 0, or (%bx,%di) -> 1 */
6449 i.rm.regmem = i.index_reg->reg_num - 6;
6450 break;
6451 case 5: /* (%bp) */
6452 default_seg = &ss;
6453 if (i.index_reg == 0)
6454 {
6455 i.rm.regmem = 6;
6456 if (operand_type_check (i.types[op], disp) == 0)
6457 {
6458 /* fake (%bp) into 0(%bp) */
6459 if (i.tm.operand_types[op].bitfield.vec_disp8)
6460 i.types[op].bitfield.vec_disp8 = 1;
6461 else
6462 i.types[op].bitfield.disp8 = 1;
6463 fake_zero_displacement = 1;
6464 }
6465 }
6466 else /* (%bp,%si) -> 2, or (%bp,%di) -> 3 */
6467 i.rm.regmem = i.index_reg->reg_num - 6 + 2;
6468 break;
6469 default: /* (%si) -> 4 or (%di) -> 5 */
6470 i.rm.regmem = i.base_reg->reg_num - 6 + 4;
6471 }
6472 i.rm.mode = mode_from_disp_size (i.types[op]);
6473 }
6474 else /* i.base_reg and 32/64 bit mode */
6475 {
6476 if (flag_code == CODE_64BIT
6477 && operand_type_check (i.types[op], disp))
6478 {
6479 i386_operand_type temp;
6480 operand_type_set (&temp, 0);
6481 temp.bitfield.disp8 = i.types[op].bitfield.disp8;
6482 temp.bitfield.vec_disp8
6483 = i.types[op].bitfield.vec_disp8;
6484 i.types[op] = temp;
6485 if (i.prefix[ADDR_PREFIX] == 0)
6486 i.types[op].bitfield.disp32s = 1;
6487 else
6488 i.types[op].bitfield.disp32 = 1;
6489 }
6490
6491 if (!i.tm.opcode_modifier.vecsib)
6492 i.rm.regmem = i.base_reg->reg_num;
6493 if ((i.base_reg->reg_flags & RegRex) != 0)
6494 i.rex |= REX_B;
6495 i.sib.base = i.base_reg->reg_num;
6496 /* x86-64 ignores REX prefix bit here to avoid decoder
6497 complications. */
6498 if (!(i.base_reg->reg_flags & RegRex)
6499 && (i.base_reg->reg_num == EBP_REG_NUM
6500 || i.base_reg->reg_num == ESP_REG_NUM))
6501 default_seg = &ss;
6502 if (i.base_reg->reg_num == 5 && i.disp_operands == 0)
6503 {
6504 fake_zero_displacement = 1;
6505 if (i.tm.operand_types [op].bitfield.vec_disp8)
6506 i.types[op].bitfield.vec_disp8 = 1;
6507 else
6508 i.types[op].bitfield.disp8 = 1;
6509 }
6510 i.sib.scale = i.log2_scale_factor;
6511 if (i.index_reg == 0)
6512 {
6513 gas_assert (!i.tm.opcode_modifier.vecsib);
6514 /* <disp>(%esp) becomes two byte modrm with no index
6515 register. We've already stored the code for esp
6516 in i.rm.regmem ie. ESCAPE_TO_TWO_BYTE_ADDRESSING.
6517 Any base register besides %esp will not use the
6518 extra modrm byte. */
6519 i.sib.index = NO_INDEX_REGISTER;
6520 }
6521 else if (!i.tm.opcode_modifier.vecsib)
6522 {
6523 if (i.index_reg->reg_num == RegEiz
6524 || i.index_reg->reg_num == RegRiz)
6525 i.sib.index = NO_INDEX_REGISTER;
6526 else
6527 i.sib.index = i.index_reg->reg_num;
6528 i.rm.regmem = ESCAPE_TO_TWO_BYTE_ADDRESSING;
6529 if ((i.index_reg->reg_flags & RegRex) != 0)
6530 i.rex |= REX_X;
6531 }
6532
6533 if (i.disp_operands
6534 && (i.reloc[op] == BFD_RELOC_386_TLS_DESC_CALL
6535 || i.reloc[op] == BFD_RELOC_X86_64_TLSDESC_CALL))
6536 i.rm.mode = 0;
6537 else
6538 {
6539 if (!fake_zero_displacement
6540 && !i.disp_operands
6541 && i.disp_encoding)
6542 {
6543 fake_zero_displacement = 1;
6544 if (i.disp_encoding == disp_encoding_8bit)
6545 i.types[op].bitfield.disp8 = 1;
6546 else
6547 i.types[op].bitfield.disp32 = 1;
6548 }
6549 i.rm.mode = mode_from_disp_size (i.types[op]);
6550 }
6551 }
6552
6553 if (fake_zero_displacement)
6554 {
6555 /* Fakes a zero displacement assuming that i.types[op]
6556 holds the correct displacement size. */
6557 expressionS *exp;
6558
6559 gas_assert (i.op[op].disps == 0);
6560 exp = &disp_expressions[i.disp_operands++];
6561 i.op[op].disps = exp;
6562 exp->X_op = O_constant;
6563 exp->X_add_number = 0;
6564 exp->X_add_symbol = (symbolS *) 0;
6565 exp->X_op_symbol = (symbolS *) 0;
6566 }
6567
6568 mem = op;
6569 }
6570 else
6571 mem = ~0;
6572
6573 if (i.tm.opcode_modifier.vexsources == XOP2SOURCES)
6574 {
6575 if (operand_type_check (i.types[0], imm))
6576 i.vex.register_specifier = NULL;
6577 else
6578 {
6579 /* VEX.vvvv encodes one of the sources when the first
6580 operand is not an immediate. */
6581 if (i.tm.opcode_modifier.vexw == VEXW0)
6582 i.vex.register_specifier = i.op[0].regs;
6583 else
6584 i.vex.register_specifier = i.op[1].regs;
6585 }
6586
6587 /* Destination is a XMM register encoded in the ModRM.reg
6588 and VEX.R bit. */
6589 i.rm.reg = i.op[2].regs->reg_num;
6590 if ((i.op[2].regs->reg_flags & RegRex) != 0)
6591 i.rex |= REX_R;
6592
6593 /* ModRM.rm and VEX.B encodes the other source. */
6594 if (!i.mem_operands)
6595 {
6596 i.rm.mode = 3;
6597
6598 if (i.tm.opcode_modifier.vexw == VEXW0)
6599 i.rm.regmem = i.op[1].regs->reg_num;
6600 else
6601 i.rm.regmem = i.op[0].regs->reg_num;
6602
6603 if ((i.op[1].regs->reg_flags & RegRex) != 0)
6604 i.rex |= REX_B;
6605 }
6606 }
6607 else if (i.tm.opcode_modifier.vexvvvv == VEXLWP)
6608 {
6609 i.vex.register_specifier = i.op[2].regs;
6610 if (!i.mem_operands)
6611 {
6612 i.rm.mode = 3;
6613 i.rm.regmem = i.op[1].regs->reg_num;
6614 if ((i.op[1].regs->reg_flags & RegRex) != 0)
6615 i.rex |= REX_B;
6616 }
6617 }
6618 /* Fill in i.rm.reg or i.rm.regmem field with register operand
6619 (if any) based on i.tm.extension_opcode. Again, we must be
6620 careful to make sure that segment/control/debug/test/MMX
6621 registers are coded into the i.rm.reg field. */
6622 else if (i.reg_operands)
6623 {
6624 unsigned int op;
6625 unsigned int vex_reg = ~0;
6626
6627 for (op = 0; op < i.operands; op++)
6628 if (i.types[op].bitfield.reg8
6629 || i.types[op].bitfield.reg16
6630 || i.types[op].bitfield.reg32
6631 || i.types[op].bitfield.reg64
6632 || i.types[op].bitfield.regmmx
6633 || i.types[op].bitfield.regxmm
6634 || i.types[op].bitfield.regymm
6635 || i.types[op].bitfield.regbnd
6636 || i.types[op].bitfield.regzmm
6637 || i.types[op].bitfield.regmask
6638 || i.types[op].bitfield.sreg2
6639 || i.types[op].bitfield.sreg3
6640 || i.types[op].bitfield.control
6641 || i.types[op].bitfield.debug
6642 || i.types[op].bitfield.test)
6643 break;
6644
6645 if (vex_3_sources)
6646 op = dest;
6647 else if (i.tm.opcode_modifier.vexvvvv == VEXXDS)
6648 {
6649 /* For instructions with VexNDS, the register-only
6650 source operand is encoded in VEX prefix. */
6651 gas_assert (mem != (unsigned int) ~0);
6652
6653 if (op > mem)
6654 {
6655 vex_reg = op++;
6656 gas_assert (op < i.operands);
6657 }
6658 else
6659 {
6660 /* Check register-only source operand when two source
6661 operands are swapped. */
6662 if (!i.tm.operand_types[op].bitfield.baseindex
6663 && i.tm.operand_types[op + 1].bitfield.baseindex)
6664 {
6665 vex_reg = op;
6666 op += 2;
6667 gas_assert (mem == (vex_reg + 1)
6668 && op < i.operands);
6669 }
6670 else
6671 {
6672 vex_reg = op + 1;
6673 gas_assert (vex_reg < i.operands);
6674 }
6675 }
6676 }
6677 else if (i.tm.opcode_modifier.vexvvvv == VEXNDD)
6678 {
6679 /* For instructions with VexNDD, the register destination
6680 is encoded in VEX prefix. */
6681 if (i.mem_operands == 0)
6682 {
6683 /* There is no memory operand. */
6684 gas_assert ((op + 2) == i.operands);
6685 vex_reg = op + 1;
6686 }
6687 else
6688 {
6689 /* There are only 2 operands. */
6690 gas_assert (op < 2 && i.operands == 2);
6691 vex_reg = 1;
6692 }
6693 }
6694 else
6695 gas_assert (op < i.operands);
6696
6697 if (vex_reg != (unsigned int) ~0)
6698 {
6699 i386_operand_type *type = &i.tm.operand_types[vex_reg];
6700
6701 if (type->bitfield.reg32 != 1
6702 && type->bitfield.reg64 != 1
6703 && !operand_type_equal (type, &regxmm)
6704 && !operand_type_equal (type, &regymm)
6705 && !operand_type_equal (type, &regzmm)
6706 && !operand_type_equal (type, &regmask))
6707 abort ();
6708
6709 i.vex.register_specifier = i.op[vex_reg].regs;
6710 }
6711
6712 /* Don't set OP operand twice. */
6713 if (vex_reg != op)
6714 {
6715 /* If there is an extension opcode to put here, the
6716 register number must be put into the regmem field. */
6717 if (i.tm.extension_opcode != None)
6718 {
6719 i.rm.regmem = i.op[op].regs->reg_num;
6720 if ((i.op[op].regs->reg_flags & RegRex) != 0)
6721 i.rex |= REX_B;
6722 if ((i.op[op].regs->reg_flags & RegVRex) != 0)
6723 i.vrex |= REX_B;
6724 }
6725 else
6726 {
6727 i.rm.reg = i.op[op].regs->reg_num;
6728 if ((i.op[op].regs->reg_flags & RegRex) != 0)
6729 i.rex |= REX_R;
6730 if ((i.op[op].regs->reg_flags & RegVRex) != 0)
6731 i.vrex |= REX_R;
6732 }
6733 }
6734
6735 /* Now, if no memory operand has set i.rm.mode = 0, 1, 2 we
6736 must set it to 3 to indicate this is a register operand
6737 in the regmem field. */
6738 if (!i.mem_operands)
6739 i.rm.mode = 3;
6740 }
6741
6742 /* Fill in i.rm.reg field with extension opcode (if any). */
6743 if (i.tm.extension_opcode != None)
6744 i.rm.reg = i.tm.extension_opcode;
6745 }
6746 return default_seg;
6747 }
6748
6749 static void
6750 output_branch (void)
6751 {
6752 char *p;
6753 int size;
6754 int code16;
6755 int prefix;
6756 relax_substateT subtype;
6757 symbolS *sym;
6758 offsetT off;
6759
6760 code16 = flag_code == CODE_16BIT ? CODE16 : 0;
6761 size = i.disp_encoding == disp_encoding_32bit ? BIG : SMALL;
6762
6763 prefix = 0;
6764 if (i.prefix[DATA_PREFIX] != 0)
6765 {
6766 prefix = 1;
6767 i.prefixes -= 1;
6768 code16 ^= CODE16;
6769 }
6770 /* Pentium4 branch hints. */
6771 if (i.prefix[SEG_PREFIX] == CS_PREFIX_OPCODE /* not taken */
6772 || i.prefix[SEG_PREFIX] == DS_PREFIX_OPCODE /* taken */)
6773 {
6774 prefix++;
6775 i.prefixes--;
6776 }
6777 if (i.prefix[REX_PREFIX] != 0)
6778 {
6779 prefix++;
6780 i.prefixes--;
6781 }
6782
6783 /* BND prefixed jump. */
6784 if (i.prefix[BND_PREFIX] != 0)
6785 {
6786 FRAG_APPEND_1_CHAR (i.prefix[BND_PREFIX]);
6787 i.prefixes -= 1;
6788 }
6789
6790 if (i.prefixes != 0 && !intel_syntax)
6791 as_warn (_("skipping prefixes on this instruction"));
6792
6793 /* It's always a symbol; End frag & setup for relax.
6794 Make sure there is enough room in this frag for the largest
6795 instruction we may generate in md_convert_frag. This is 2
6796 bytes for the opcode and room for the prefix and largest
6797 displacement. */
6798 frag_grow (prefix + 2 + 4);
6799 /* Prefix and 1 opcode byte go in fr_fix. */
6800 p = frag_more (prefix + 1);
6801 if (i.prefix[DATA_PREFIX] != 0)
6802 *p++ = DATA_PREFIX_OPCODE;
6803 if (i.prefix[SEG_PREFIX] == CS_PREFIX_OPCODE
6804 || i.prefix[SEG_PREFIX] == DS_PREFIX_OPCODE)
6805 *p++ = i.prefix[SEG_PREFIX];
6806 if (i.prefix[REX_PREFIX] != 0)
6807 *p++ = i.prefix[REX_PREFIX];
6808 *p = i.tm.base_opcode;
6809
6810 if ((unsigned char) *p == JUMP_PC_RELATIVE)
6811 subtype = ENCODE_RELAX_STATE (UNCOND_JUMP, size);
6812 else if (cpu_arch_flags.bitfield.cpui386)
6813 subtype = ENCODE_RELAX_STATE (COND_JUMP, size);
6814 else
6815 subtype = ENCODE_RELAX_STATE (COND_JUMP86, size);
6816 subtype |= code16;
6817
6818 sym = i.op[0].disps->X_add_symbol;
6819 off = i.op[0].disps->X_add_number;
6820
6821 if (i.op[0].disps->X_op != O_constant
6822 && i.op[0].disps->X_op != O_symbol)
6823 {
6824 /* Handle complex expressions. */
6825 sym = make_expr_symbol (i.op[0].disps);
6826 off = 0;
6827 }
6828
6829 /* 1 possible extra opcode + 4 byte displacement go in var part.
6830 Pass reloc in fr_var. */
6831 frag_var (rs_machine_dependent, 5, i.reloc[0], subtype, sym, off, p);
6832 }
6833
6834 static void
6835 output_jump (void)
6836 {
6837 char *p;
6838 int size;
6839 fixS *fixP;
6840
6841 if (i.tm.opcode_modifier.jumpbyte)
6842 {
6843 /* This is a loop or jecxz type instruction. */
6844 size = 1;
6845 if (i.prefix[ADDR_PREFIX] != 0)
6846 {
6847 FRAG_APPEND_1_CHAR (ADDR_PREFIX_OPCODE);
6848 i.prefixes -= 1;
6849 }
6850 /* Pentium4 branch hints. */
6851 if (i.prefix[SEG_PREFIX] == CS_PREFIX_OPCODE /* not taken */
6852 || i.prefix[SEG_PREFIX] == DS_PREFIX_OPCODE /* taken */)
6853 {
6854 FRAG_APPEND_1_CHAR (i.prefix[SEG_PREFIX]);
6855 i.prefixes--;
6856 }
6857 }
6858 else
6859 {
6860 int code16;
6861
6862 code16 = 0;
6863 if (flag_code == CODE_16BIT)
6864 code16 = CODE16;
6865
6866 if (i.prefix[DATA_PREFIX] != 0)
6867 {
6868 FRAG_APPEND_1_CHAR (DATA_PREFIX_OPCODE);
6869 i.prefixes -= 1;
6870 code16 ^= CODE16;
6871 }
6872
6873 size = 4;
6874 if (code16)
6875 size = 2;
6876 }
6877
6878 if (i.prefix[REX_PREFIX] != 0)
6879 {
6880 FRAG_APPEND_1_CHAR (i.prefix[REX_PREFIX]);
6881 i.prefixes -= 1;
6882 }
6883
6884 /* BND prefixed jump. */
6885 if (i.prefix[BND_PREFIX] != 0)
6886 {
6887 FRAG_APPEND_1_CHAR (i.prefix[BND_PREFIX]);
6888 i.prefixes -= 1;
6889 }
6890
6891 if (i.prefixes != 0 && !intel_syntax)
6892 as_warn (_("skipping prefixes on this instruction"));
6893
6894 p = frag_more (i.tm.opcode_length + size);
6895 switch (i.tm.opcode_length)
6896 {
6897 case 2:
6898 *p++ = i.tm.base_opcode >> 8;
6899 case 1:
6900 *p++ = i.tm.base_opcode;
6901 break;
6902 default:
6903 abort ();
6904 }
6905
6906 fixP = fix_new_exp (frag_now, p - frag_now->fr_literal, size,
6907 i.op[0].disps, 1, reloc (size, 1, 1, i.reloc[0]));
6908
6909 /* All jumps handled here are signed, but don't use a signed limit
6910 check for 32 and 16 bit jumps as we want to allow wrap around at
6911 4G and 64k respectively. */
6912 if (size == 1)
6913 fixP->fx_signed = 1;
6914 }
6915
6916 static void
6917 output_interseg_jump (void)
6918 {
6919 char *p;
6920 int size;
6921 int prefix;
6922 int code16;
6923
6924 code16 = 0;
6925 if (flag_code == CODE_16BIT)
6926 code16 = CODE16;
6927
6928 prefix = 0;
6929 if (i.prefix[DATA_PREFIX] != 0)
6930 {
6931 prefix = 1;
6932 i.prefixes -= 1;
6933 code16 ^= CODE16;
6934 }
6935 if (i.prefix[REX_PREFIX] != 0)
6936 {
6937 prefix++;
6938 i.prefixes -= 1;
6939 }
6940
6941 size = 4;
6942 if (code16)
6943 size = 2;
6944
6945 if (i.prefixes != 0 && !intel_syntax)
6946 as_warn (_("skipping prefixes on this instruction"));
6947
6948 /* 1 opcode; 2 segment; offset */
6949 p = frag_more (prefix + 1 + 2 + size);
6950
6951 if (i.prefix[DATA_PREFIX] != 0)
6952 *p++ = DATA_PREFIX_OPCODE;
6953
6954 if (i.prefix[REX_PREFIX] != 0)
6955 *p++ = i.prefix[REX_PREFIX];
6956
6957 *p++ = i.tm.base_opcode;
6958 if (i.op[1].imms->X_op == O_constant)
6959 {
6960 offsetT n = i.op[1].imms->X_add_number;
6961
6962 if (size == 2
6963 && !fits_in_unsigned_word (n)
6964 && !fits_in_signed_word (n))
6965 {
6966 as_bad (_("16-bit jump out of range"));
6967 return;
6968 }
6969 md_number_to_chars (p, n, size);
6970 }
6971 else
6972 fix_new_exp (frag_now, p - frag_now->fr_literal, size,
6973 i.op[1].imms, 0, reloc (size, 0, 0, i.reloc[1]));
6974 if (i.op[0].imms->X_op != O_constant)
6975 as_bad (_("can't handle non absolute segment in `%s'"),
6976 i.tm.name);
6977 md_number_to_chars (p + size, (valueT) i.op[0].imms->X_add_number, 2);
6978 }
6979
6980 static void
6981 output_insn (void)
6982 {
6983 fragS *insn_start_frag;
6984 offsetT insn_start_off;
6985
6986 /* Tie dwarf2 debug info to the address at the start of the insn.
6987 We can't do this after the insn has been output as the current
6988 frag may have been closed off. eg. by frag_var. */
6989 dwarf2_emit_insn (0);
6990
6991 insn_start_frag = frag_now;
6992 insn_start_off = frag_now_fix ();
6993
6994 /* Output jumps. */
6995 if (i.tm.opcode_modifier.jump)
6996 output_branch ();
6997 else if (i.tm.opcode_modifier.jumpbyte
6998 || i.tm.opcode_modifier.jumpdword)
6999 output_jump ();
7000 else if (i.tm.opcode_modifier.jumpintersegment)
7001 output_interseg_jump ();
7002 else
7003 {
7004 /* Output normal instructions here. */
7005 char *p;
7006 unsigned char *q;
7007 unsigned int j;
7008 unsigned int prefix;
7009
7010 if (avoid_fence
7011 && i.tm.base_opcode == 0xfae
7012 && i.operands == 1
7013 && i.imm_operands == 1
7014 && (i.op[0].imms->X_add_number == 0xe8
7015 || i.op[0].imms->X_add_number == 0xf0
7016 || i.op[0].imms->X_add_number == 0xf8))
7017 {
7018 /* Encode lfence, mfence, and sfence as
7019 f0 83 04 24 00 lock addl $0x0, (%{re}sp). */
7020 offsetT val = 0x240483f0ULL;
7021 p = frag_more (5);
7022 md_number_to_chars (p, val, 5);
7023 return;
7024 }
7025
7026 /* Some processors fail on LOCK prefix. This options makes
7027 assembler ignore LOCK prefix and serves as a workaround. */
7028 if (omit_lock_prefix)
7029 {
7030 if (i.tm.base_opcode == LOCK_PREFIX_OPCODE)
7031 return;
7032 i.prefix[LOCK_PREFIX] = 0;
7033 }
7034
7035 /* Since the VEX/EVEX prefix contains the implicit prefix, we
7036 don't need the explicit prefix. */
7037 if (!i.tm.opcode_modifier.vex && !i.tm.opcode_modifier.evex)
7038 {
7039 switch (i.tm.opcode_length)
7040 {
7041 case 3:
7042 if (i.tm.base_opcode & 0xff000000)
7043 {
7044 prefix = (i.tm.base_opcode >> 24) & 0xff;
7045 goto check_prefix;
7046 }
7047 break;
7048 case 2:
7049 if ((i.tm.base_opcode & 0xff0000) != 0)
7050 {
7051 prefix = (i.tm.base_opcode >> 16) & 0xff;
7052 if (i.tm.cpu_flags.bitfield.cpupadlock)
7053 {
7054 check_prefix:
7055 if (prefix != REPE_PREFIX_OPCODE
7056 || (i.prefix[REP_PREFIX]
7057 != REPE_PREFIX_OPCODE))
7058 add_prefix (prefix);
7059 }
7060 else
7061 add_prefix (prefix);
7062 }
7063 break;
7064 case 1:
7065 break;
7066 default:
7067 abort ();
7068 }
7069
7070 #if defined (OBJ_MAYBE_ELF) || defined (OBJ_ELF)
7071 /* For x32, add a dummy REX_OPCODE prefix for mov/add with
7072 R_X86_64_GOTTPOFF relocation so that linker can safely
7073 perform IE->LE optimization. */
7074 if (x86_elf_abi == X86_64_X32_ABI
7075 && i.operands == 2
7076 && i.reloc[0] == BFD_RELOC_X86_64_GOTTPOFF
7077 && i.prefix[REX_PREFIX] == 0)
7078 add_prefix (REX_OPCODE);
7079 #endif
7080
7081 /* The prefix bytes. */
7082 for (j = ARRAY_SIZE (i.prefix), q = i.prefix; j > 0; j--, q++)
7083 if (*q)
7084 FRAG_APPEND_1_CHAR (*q);
7085 }
7086 else
7087 {
7088 for (j = 0, q = i.prefix; j < ARRAY_SIZE (i.prefix); j++, q++)
7089 if (*q)
7090 switch (j)
7091 {
7092 case REX_PREFIX:
7093 /* REX byte is encoded in VEX prefix. */
7094 break;
7095 case SEG_PREFIX:
7096 case ADDR_PREFIX:
7097 FRAG_APPEND_1_CHAR (*q);
7098 break;
7099 default:
7100 /* There should be no other prefixes for instructions
7101 with VEX prefix. */
7102 abort ();
7103 }
7104
7105 /* For EVEX instructions i.vrex should become 0 after
7106 build_evex_prefix. For VEX instructions upper 16 registers
7107 aren't available, so VREX should be 0. */
7108 if (i.vrex)
7109 abort ();
7110 /* Now the VEX prefix. */
7111 p = frag_more (i.vex.length);
7112 for (j = 0; j < i.vex.length; j++)
7113 p[j] = i.vex.bytes[j];
7114 }
7115
7116 /* Now the opcode; be careful about word order here! */
7117 if (i.tm.opcode_length == 1)
7118 {
7119 FRAG_APPEND_1_CHAR (i.tm.base_opcode);
7120 }
7121 else
7122 {
7123 switch (i.tm.opcode_length)
7124 {
7125 case 4:
7126 p = frag_more (4);
7127 *p++ = (i.tm.base_opcode >> 24) & 0xff;
7128 *p++ = (i.tm.base_opcode >> 16) & 0xff;
7129 break;
7130 case 3:
7131 p = frag_more (3);
7132 *p++ = (i.tm.base_opcode >> 16) & 0xff;
7133 break;
7134 case 2:
7135 p = frag_more (2);
7136 break;
7137 default:
7138 abort ();
7139 break;
7140 }
7141
7142 /* Put out high byte first: can't use md_number_to_chars! */
7143 *p++ = (i.tm.base_opcode >> 8) & 0xff;
7144 *p = i.tm.base_opcode & 0xff;
7145 }
7146
7147 /* Now the modrm byte and sib byte (if present). */
7148 if (i.tm.opcode_modifier.modrm)
7149 {
7150 FRAG_APPEND_1_CHAR ((i.rm.regmem << 0
7151 | i.rm.reg << 3
7152 | i.rm.mode << 6));
7153 /* If i.rm.regmem == ESP (4)
7154 && i.rm.mode != (Register mode)
7155 && not 16 bit
7156 ==> need second modrm byte. */
7157 if (i.rm.regmem == ESCAPE_TO_TWO_BYTE_ADDRESSING
7158 && i.rm.mode != 3
7159 && !(i.base_reg && i.base_reg->reg_type.bitfield.reg16))
7160 FRAG_APPEND_1_CHAR ((i.sib.base << 0
7161 | i.sib.index << 3
7162 | i.sib.scale << 6));
7163 }
7164
7165 if (i.disp_operands)
7166 output_disp (insn_start_frag, insn_start_off);
7167
7168 if (i.imm_operands)
7169 output_imm (insn_start_frag, insn_start_off);
7170 }
7171
7172 #ifdef DEBUG386
7173 if (flag_debug)
7174 {
7175 pi ("" /*line*/, &i);
7176 }
7177 #endif /* DEBUG386 */
7178 }
7179
7180 /* Return the size of the displacement operand N. */
7181
7182 static int
7183 disp_size (unsigned int n)
7184 {
7185 int size = 4;
7186
7187 /* Vec_Disp8 has to be 8bit. */
7188 if (i.types[n].bitfield.vec_disp8)
7189 size = 1;
7190 else if (i.types[n].bitfield.disp64)
7191 size = 8;
7192 else if (i.types[n].bitfield.disp8)
7193 size = 1;
7194 else if (i.types[n].bitfield.disp16)
7195 size = 2;
7196 return size;
7197 }
7198
7199 /* Return the size of the immediate operand N. */
7200
7201 static int
7202 imm_size (unsigned int n)
7203 {
7204 int size = 4;
7205 if (i.types[n].bitfield.imm64)
7206 size = 8;
7207 else if (i.types[n].bitfield.imm8 || i.types[n].bitfield.imm8s)
7208 size = 1;
7209 else if (i.types[n].bitfield.imm16)
7210 size = 2;
7211 return size;
7212 }
7213
7214 static void
7215 output_disp (fragS *insn_start_frag, offsetT insn_start_off)
7216 {
7217 char *p;
7218 unsigned int n;
7219
7220 for (n = 0; n < i.operands; n++)
7221 {
7222 if (i.types[n].bitfield.vec_disp8
7223 || operand_type_check (i.types[n], disp))
7224 {
7225 if (i.op[n].disps->X_op == O_constant)
7226 {
7227 int size = disp_size (n);
7228 offsetT val = i.op[n].disps->X_add_number;
7229
7230 if (i.types[n].bitfield.vec_disp8)
7231 val >>= i.memshift;
7232 val = offset_in_range (val, size);
7233 p = frag_more (size);
7234 md_number_to_chars (p, val, size);
7235 }
7236 else
7237 {
7238 enum bfd_reloc_code_real reloc_type;
7239 int size = disp_size (n);
7240 int sign = i.types[n].bitfield.disp32s;
7241 int pcrel = (i.flags[n] & Operand_PCrel) != 0;
7242 fixS *fixP;
7243
7244 /* We can't have 8 bit displacement here. */
7245 gas_assert (!i.types[n].bitfield.disp8);
7246
7247 /* The PC relative address is computed relative
7248 to the instruction boundary, so in case immediate
7249 fields follows, we need to adjust the value. */
7250 if (pcrel && i.imm_operands)
7251 {
7252 unsigned int n1;
7253 int sz = 0;
7254
7255 for (n1 = 0; n1 < i.operands; n1++)
7256 if (operand_type_check (i.types[n1], imm))
7257 {
7258 /* Only one immediate is allowed for PC
7259 relative address. */
7260 gas_assert (sz == 0);
7261 sz = imm_size (n1);
7262 i.op[n].disps->X_add_number -= sz;
7263 }
7264 /* We should find the immediate. */
7265 gas_assert (sz != 0);
7266 }
7267
7268 p = frag_more (size);
7269 reloc_type = reloc (size, pcrel, sign, i.reloc[n]);
7270 if (GOT_symbol
7271 && GOT_symbol == i.op[n].disps->X_add_symbol
7272 && (((reloc_type == BFD_RELOC_32
7273 || reloc_type == BFD_RELOC_X86_64_32S
7274 || (reloc_type == BFD_RELOC_64
7275 && object_64bit))
7276 && (i.op[n].disps->X_op == O_symbol
7277 || (i.op[n].disps->X_op == O_add
7278 && ((symbol_get_value_expression
7279 (i.op[n].disps->X_op_symbol)->X_op)
7280 == O_subtract))))
7281 || reloc_type == BFD_RELOC_32_PCREL))
7282 {
7283 offsetT add;
7284
7285 if (insn_start_frag == frag_now)
7286 add = (p - frag_now->fr_literal) - insn_start_off;
7287 else
7288 {
7289 fragS *fr;
7290
7291 add = insn_start_frag->fr_fix - insn_start_off;
7292 for (fr = insn_start_frag->fr_next;
7293 fr && fr != frag_now; fr = fr->fr_next)
7294 add += fr->fr_fix;
7295 add += p - frag_now->fr_literal;
7296 }
7297
7298 if (!object_64bit)
7299 {
7300 reloc_type = BFD_RELOC_386_GOTPC;
7301 i.op[n].imms->X_add_number += add;
7302 }
7303 else if (reloc_type == BFD_RELOC_64)
7304 reloc_type = BFD_RELOC_X86_64_GOTPC64;
7305 else
7306 /* Don't do the adjustment for x86-64, as there
7307 the pcrel addressing is relative to the _next_
7308 insn, and that is taken care of in other code. */
7309 reloc_type = BFD_RELOC_X86_64_GOTPC32;
7310 }
7311 fixP = fix_new_exp (frag_now, p - frag_now->fr_literal,
7312 size, i.op[n].disps, pcrel,
7313 reloc_type);
7314 /* Check for "call/jmp *mem", "mov mem, %reg",
7315 "test %reg, mem" and "binop mem, %reg" where binop
7316 is one of adc, add, and, cmp, or, sbb, sub, xor
7317 instructions. Always generate R_386_GOT32X for
7318 "sym*GOT" operand in 32-bit mode. */
7319 if ((generate_relax_relocations
7320 || (!object_64bit
7321 && i.rm.mode == 0
7322 && i.rm.regmem == 5))
7323 && (i.rm.mode == 2
7324 || (i.rm.mode == 0 && i.rm.regmem == 5))
7325 && ((i.operands == 1
7326 && i.tm.base_opcode == 0xff
7327 && (i.rm.reg == 2 || i.rm.reg == 4))
7328 || (i.operands == 2
7329 && (i.tm.base_opcode == 0x8b
7330 || i.tm.base_opcode == 0x85
7331 || (i.tm.base_opcode & 0xc7) == 0x03))))
7332 {
7333 if (object_64bit)
7334 {
7335 fixP->fx_tcbit = i.rex != 0;
7336 if (i.base_reg
7337 && (i.base_reg->reg_num == RegRip
7338 || i.base_reg->reg_num == RegEip))
7339 fixP->fx_tcbit2 = 1;
7340 }
7341 else
7342 fixP->fx_tcbit2 = 1;
7343 }
7344 }
7345 }
7346 }
7347 }
7348
7349 static void
7350 output_imm (fragS *insn_start_frag, offsetT insn_start_off)
7351 {
7352 char *p;
7353 unsigned int n;
7354
7355 for (n = 0; n < i.operands; n++)
7356 {
7357 /* Skip SAE/RC Imm operand in EVEX. They are already handled. */
7358 if (i.rounding && (int) n == i.rounding->operand)
7359 continue;
7360
7361 if (operand_type_check (i.types[n], imm))
7362 {
7363 if (i.op[n].imms->X_op == O_constant)
7364 {
7365 int size = imm_size (n);
7366 offsetT val;
7367
7368 val = offset_in_range (i.op[n].imms->X_add_number,
7369 size);
7370 p = frag_more (size);
7371 md_number_to_chars (p, val, size);
7372 }
7373 else
7374 {
7375 /* Not absolute_section.
7376 Need a 32-bit fixup (don't support 8bit
7377 non-absolute imms). Try to support other
7378 sizes ... */
7379 enum bfd_reloc_code_real reloc_type;
7380 int size = imm_size (n);
7381 int sign;
7382
7383 if (i.types[n].bitfield.imm32s
7384 && (i.suffix == QWORD_MNEM_SUFFIX
7385 || (!i.suffix && i.tm.opcode_modifier.no_lsuf)))
7386 sign = 1;
7387 else
7388 sign = 0;
7389
7390 p = frag_more (size);
7391 reloc_type = reloc (size, 0, sign, i.reloc[n]);
7392
7393 /* This is tough to explain. We end up with this one if we
7394 * have operands that look like
7395 * "_GLOBAL_OFFSET_TABLE_+[.-.L284]". The goal here is to
7396 * obtain the absolute address of the GOT, and it is strongly
7397 * preferable from a performance point of view to avoid using
7398 * a runtime relocation for this. The actual sequence of
7399 * instructions often look something like:
7400 *
7401 * call .L66
7402 * .L66:
7403 * popl %ebx
7404 * addl $_GLOBAL_OFFSET_TABLE_+[.-.L66],%ebx
7405 *
7406 * The call and pop essentially return the absolute address
7407 * of the label .L66 and store it in %ebx. The linker itself
7408 * will ultimately change the first operand of the addl so
7409 * that %ebx points to the GOT, but to keep things simple, the
7410 * .o file must have this operand set so that it generates not
7411 * the absolute address of .L66, but the absolute address of
7412 * itself. This allows the linker itself simply treat a GOTPC
7413 * relocation as asking for a pcrel offset to the GOT to be
7414 * added in, and the addend of the relocation is stored in the
7415 * operand field for the instruction itself.
7416 *
7417 * Our job here is to fix the operand so that it would add
7418 * the correct offset so that %ebx would point to itself. The
7419 * thing that is tricky is that .-.L66 will point to the
7420 * beginning of the instruction, so we need to further modify
7421 * the operand so that it will point to itself. There are
7422 * other cases where you have something like:
7423 *
7424 * .long $_GLOBAL_OFFSET_TABLE_+[.-.L66]
7425 *
7426 * and here no correction would be required. Internally in
7427 * the assembler we treat operands of this form as not being
7428 * pcrel since the '.' is explicitly mentioned, and I wonder
7429 * whether it would simplify matters to do it this way. Who
7430 * knows. In earlier versions of the PIC patches, the
7431 * pcrel_adjust field was used to store the correction, but
7432 * since the expression is not pcrel, I felt it would be
7433 * confusing to do it this way. */
7434
7435 if ((reloc_type == BFD_RELOC_32
7436 || reloc_type == BFD_RELOC_X86_64_32S
7437 || reloc_type == BFD_RELOC_64)
7438 && GOT_symbol
7439 && GOT_symbol == i.op[n].imms->X_add_symbol
7440 && (i.op[n].imms->X_op == O_symbol
7441 || (i.op[n].imms->X_op == O_add
7442 && ((symbol_get_value_expression
7443 (i.op[n].imms->X_op_symbol)->X_op)
7444 == O_subtract))))
7445 {
7446 offsetT add;
7447
7448 if (insn_start_frag == frag_now)
7449 add = (p - frag_now->fr_literal) - insn_start_off;
7450 else
7451 {
7452 fragS *fr;
7453
7454 add = insn_start_frag->fr_fix - insn_start_off;
7455 for (fr = insn_start_frag->fr_next;
7456 fr && fr != frag_now; fr = fr->fr_next)
7457 add += fr->fr_fix;
7458 add += p - frag_now->fr_literal;
7459 }
7460
7461 if (!object_64bit)
7462 reloc_type = BFD_RELOC_386_GOTPC;
7463 else if (size == 4)
7464 reloc_type = BFD_RELOC_X86_64_GOTPC32;
7465 else if (size == 8)
7466 reloc_type = BFD_RELOC_X86_64_GOTPC64;
7467 i.op[n].imms->X_add_number += add;
7468 }
7469 fix_new_exp (frag_now, p - frag_now->fr_literal, size,
7470 i.op[n].imms, 0, reloc_type);
7471 }
7472 }
7473 }
7474 }
7475 \f
7476 /* x86_cons_fix_new is called via the expression parsing code when a
7477 reloc is needed. We use this hook to get the correct .got reloc. */
7478 static int cons_sign = -1;
7479
7480 void
7481 x86_cons_fix_new (fragS *frag, unsigned int off, unsigned int len,
7482 expressionS *exp, bfd_reloc_code_real_type r)
7483 {
7484 r = reloc (len, 0, cons_sign, r);
7485
7486 #ifdef TE_PE
7487 if (exp->X_op == O_secrel)
7488 {
7489 exp->X_op = O_symbol;
7490 r = BFD_RELOC_32_SECREL;
7491 }
7492 #endif
7493
7494 fix_new_exp (frag, off, len, exp, 0, r);
7495 }
7496
7497 /* Export the ABI address size for use by TC_ADDRESS_BYTES for the
7498 purpose of the `.dc.a' internal pseudo-op. */
7499
7500 int
7501 x86_address_bytes (void)
7502 {
7503 if ((stdoutput->arch_info->mach & bfd_mach_x64_32))
7504 return 4;
7505 return stdoutput->arch_info->bits_per_address / 8;
7506 }
7507
7508 #if !(defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) || defined (OBJ_MACH_O)) \
7509 || defined (LEX_AT)
7510 # define lex_got(reloc, adjust, types) NULL
7511 #else
7512 /* Parse operands of the form
7513 <symbol>@GOTOFF+<nnn>
7514 and similar .plt or .got references.
7515
7516 If we find one, set up the correct relocation in RELOC and copy the
7517 input string, minus the `@GOTOFF' into a malloc'd buffer for
7518 parsing by the calling routine. Return this buffer, and if ADJUST
7519 is non-null set it to the length of the string we removed from the
7520 input line. Otherwise return NULL. */
7521 static char *
7522 lex_got (enum bfd_reloc_code_real *rel,
7523 int *adjust,
7524 i386_operand_type *types)
7525 {
7526 /* Some of the relocations depend on the size of what field is to
7527 be relocated. But in our callers i386_immediate and i386_displacement
7528 we don't yet know the operand size (this will be set by insn
7529 matching). Hence we record the word32 relocation here,
7530 and adjust the reloc according to the real size in reloc(). */
7531 static const struct {
7532 const char *str;
7533 int len;
7534 const enum bfd_reloc_code_real rel[2];
7535 const i386_operand_type types64;
7536 } gotrel[] = {
7537 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
7538 { STRING_COMMA_LEN ("SIZE"), { BFD_RELOC_SIZE32,
7539 BFD_RELOC_SIZE32 },
7540 OPERAND_TYPE_IMM32_64 },
7541 #endif
7542 { STRING_COMMA_LEN ("PLTOFF"), { _dummy_first_bfd_reloc_code_real,
7543 BFD_RELOC_X86_64_PLTOFF64 },
7544 OPERAND_TYPE_IMM64 },
7545 { STRING_COMMA_LEN ("PLT"), { BFD_RELOC_386_PLT32,
7546 BFD_RELOC_X86_64_PLT32 },
7547 OPERAND_TYPE_IMM32_32S_DISP32 },
7548 { STRING_COMMA_LEN ("GOTPLT"), { _dummy_first_bfd_reloc_code_real,
7549 BFD_RELOC_X86_64_GOTPLT64 },
7550 OPERAND_TYPE_IMM64_DISP64 },
7551 { STRING_COMMA_LEN ("GOTOFF"), { BFD_RELOC_386_GOTOFF,
7552 BFD_RELOC_X86_64_GOTOFF64 },
7553 OPERAND_TYPE_IMM64_DISP64 },
7554 { STRING_COMMA_LEN ("GOTPCREL"), { _dummy_first_bfd_reloc_code_real,
7555 BFD_RELOC_X86_64_GOTPCREL },
7556 OPERAND_TYPE_IMM32_32S_DISP32 },
7557 { STRING_COMMA_LEN ("TLSGD"), { BFD_RELOC_386_TLS_GD,
7558 BFD_RELOC_X86_64_TLSGD },
7559 OPERAND_TYPE_IMM32_32S_DISP32 },
7560 { STRING_COMMA_LEN ("TLSLDM"), { BFD_RELOC_386_TLS_LDM,
7561 _dummy_first_bfd_reloc_code_real },
7562 OPERAND_TYPE_NONE },
7563 { STRING_COMMA_LEN ("TLSLD"), { _dummy_first_bfd_reloc_code_real,
7564 BFD_RELOC_X86_64_TLSLD },
7565 OPERAND_TYPE_IMM32_32S_DISP32 },
7566 { STRING_COMMA_LEN ("GOTTPOFF"), { BFD_RELOC_386_TLS_IE_32,
7567 BFD_RELOC_X86_64_GOTTPOFF },
7568 OPERAND_TYPE_IMM32_32S_DISP32 },
7569 { STRING_COMMA_LEN ("TPOFF"), { BFD_RELOC_386_TLS_LE_32,
7570 BFD_RELOC_X86_64_TPOFF32 },
7571 OPERAND_TYPE_IMM32_32S_64_DISP32_64 },
7572 { STRING_COMMA_LEN ("NTPOFF"), { BFD_RELOC_386_TLS_LE,
7573 _dummy_first_bfd_reloc_code_real },
7574 OPERAND_TYPE_NONE },
7575 { STRING_COMMA_LEN ("DTPOFF"), { BFD_RELOC_386_TLS_LDO_32,
7576 BFD_RELOC_X86_64_DTPOFF32 },
7577 OPERAND_TYPE_IMM32_32S_64_DISP32_64 },
7578 { STRING_COMMA_LEN ("GOTNTPOFF"),{ BFD_RELOC_386_TLS_GOTIE,
7579 _dummy_first_bfd_reloc_code_real },
7580 OPERAND_TYPE_NONE },
7581 { STRING_COMMA_LEN ("INDNTPOFF"),{ BFD_RELOC_386_TLS_IE,
7582 _dummy_first_bfd_reloc_code_real },
7583 OPERAND_TYPE_NONE },
7584 { STRING_COMMA_LEN ("GOT"), { BFD_RELOC_386_GOT32,
7585 BFD_RELOC_X86_64_GOT32 },
7586 OPERAND_TYPE_IMM32_32S_64_DISP32 },
7587 { STRING_COMMA_LEN ("TLSDESC"), { BFD_RELOC_386_TLS_GOTDESC,
7588 BFD_RELOC_X86_64_GOTPC32_TLSDESC },
7589 OPERAND_TYPE_IMM32_32S_DISP32 },
7590 { STRING_COMMA_LEN ("TLSCALL"), { BFD_RELOC_386_TLS_DESC_CALL,
7591 BFD_RELOC_X86_64_TLSDESC_CALL },
7592 OPERAND_TYPE_IMM32_32S_DISP32 },
7593 };
7594 char *cp;
7595 unsigned int j;
7596
7597 #if defined (OBJ_MAYBE_ELF)
7598 if (!IS_ELF)
7599 return NULL;
7600 #endif
7601
7602 for (cp = input_line_pointer; *cp != '@'; cp++)
7603 if (is_end_of_line[(unsigned char) *cp] || *cp == ',')
7604 return NULL;
7605
7606 for (j = 0; j < ARRAY_SIZE (gotrel); j++)
7607 {
7608 int len = gotrel[j].len;
7609 if (strncasecmp (cp + 1, gotrel[j].str, len) == 0)
7610 {
7611 if (gotrel[j].rel[object_64bit] != 0)
7612 {
7613 int first, second;
7614 char *tmpbuf, *past_reloc;
7615
7616 *rel = gotrel[j].rel[object_64bit];
7617
7618 if (types)
7619 {
7620 if (flag_code != CODE_64BIT)
7621 {
7622 types->bitfield.imm32 = 1;
7623 types->bitfield.disp32 = 1;
7624 }
7625 else
7626 *types = gotrel[j].types64;
7627 }
7628
7629 if (j != 0 && GOT_symbol == NULL)
7630 GOT_symbol = symbol_find_or_make (GLOBAL_OFFSET_TABLE_NAME);
7631
7632 /* The length of the first part of our input line. */
7633 first = cp - input_line_pointer;
7634
7635 /* The second part goes from after the reloc token until
7636 (and including) an end_of_line char or comma. */
7637 past_reloc = cp + 1 + len;
7638 cp = past_reloc;
7639 while (!is_end_of_line[(unsigned char) *cp] && *cp != ',')
7640 ++cp;
7641 second = cp + 1 - past_reloc;
7642
7643 /* Allocate and copy string. The trailing NUL shouldn't
7644 be necessary, but be safe. */
7645 tmpbuf = XNEWVEC (char, first + second + 2);
7646 memcpy (tmpbuf, input_line_pointer, first);
7647 if (second != 0 && *past_reloc != ' ')
7648 /* Replace the relocation token with ' ', so that
7649 errors like foo@GOTOFF1 will be detected. */
7650 tmpbuf[first++] = ' ';
7651 else
7652 /* Increment length by 1 if the relocation token is
7653 removed. */
7654 len++;
7655 if (adjust)
7656 *adjust = len;
7657 memcpy (tmpbuf + first, past_reloc, second);
7658 tmpbuf[first + second] = '\0';
7659 return tmpbuf;
7660 }
7661
7662 as_bad (_("@%s reloc is not supported with %d-bit output format"),
7663 gotrel[j].str, 1 << (5 + object_64bit));
7664 return NULL;
7665 }
7666 }
7667
7668 /* Might be a symbol version string. Don't as_bad here. */
7669 return NULL;
7670 }
7671 #endif
7672
7673 #ifdef TE_PE
7674 #ifdef lex_got
7675 #undef lex_got
7676 #endif
7677 /* Parse operands of the form
7678 <symbol>@SECREL32+<nnn>
7679
7680 If we find one, set up the correct relocation in RELOC and copy the
7681 input string, minus the `@SECREL32' into a malloc'd buffer for
7682 parsing by the calling routine. Return this buffer, and if ADJUST
7683 is non-null set it to the length of the string we removed from the
7684 input line. Otherwise return NULL.
7685
7686 This function is copied from the ELF version above adjusted for PE targets. */
7687
7688 static char *
7689 lex_got (enum bfd_reloc_code_real *rel ATTRIBUTE_UNUSED,
7690 int *adjust ATTRIBUTE_UNUSED,
7691 i386_operand_type *types)
7692 {
7693 static const struct
7694 {
7695 const char *str;
7696 int len;
7697 const enum bfd_reloc_code_real rel[2];
7698 const i386_operand_type types64;
7699 }
7700 gotrel[] =
7701 {
7702 { STRING_COMMA_LEN ("SECREL32"), { BFD_RELOC_32_SECREL,
7703 BFD_RELOC_32_SECREL },
7704 OPERAND_TYPE_IMM32_32S_64_DISP32_64 },
7705 };
7706
7707 char *cp;
7708 unsigned j;
7709
7710 for (cp = input_line_pointer; *cp != '@'; cp++)
7711 if (is_end_of_line[(unsigned char) *cp] || *cp == ',')
7712 return NULL;
7713
7714 for (j = 0; j < ARRAY_SIZE (gotrel); j++)
7715 {
7716 int len = gotrel[j].len;
7717
7718 if (strncasecmp (cp + 1, gotrel[j].str, len) == 0)
7719 {
7720 if (gotrel[j].rel[object_64bit] != 0)
7721 {
7722 int first, second;
7723 char *tmpbuf, *past_reloc;
7724
7725 *rel = gotrel[j].rel[object_64bit];
7726 if (adjust)
7727 *adjust = len;
7728
7729 if (types)
7730 {
7731 if (flag_code != CODE_64BIT)
7732 {
7733 types->bitfield.imm32 = 1;
7734 types->bitfield.disp32 = 1;
7735 }
7736 else
7737 *types = gotrel[j].types64;
7738 }
7739
7740 /* The length of the first part of our input line. */
7741 first = cp - input_line_pointer;
7742
7743 /* The second part goes from after the reloc token until
7744 (and including) an end_of_line char or comma. */
7745 past_reloc = cp + 1 + len;
7746 cp = past_reloc;
7747 while (!is_end_of_line[(unsigned char) *cp] && *cp != ',')
7748 ++cp;
7749 second = cp + 1 - past_reloc;
7750
7751 /* Allocate and copy string. The trailing NUL shouldn't
7752 be necessary, but be safe. */
7753 tmpbuf = XNEWVEC (char, first + second + 2);
7754 memcpy (tmpbuf, input_line_pointer, first);
7755 if (second != 0 && *past_reloc != ' ')
7756 /* Replace the relocation token with ' ', so that
7757 errors like foo@SECLREL321 will be detected. */
7758 tmpbuf[first++] = ' ';
7759 memcpy (tmpbuf + first, past_reloc, second);
7760 tmpbuf[first + second] = '\0';
7761 return tmpbuf;
7762 }
7763
7764 as_bad (_("@%s reloc is not supported with %d-bit output format"),
7765 gotrel[j].str, 1 << (5 + object_64bit));
7766 return NULL;
7767 }
7768 }
7769
7770 /* Might be a symbol version string. Don't as_bad here. */
7771 return NULL;
7772 }
7773
7774 #endif /* TE_PE */
7775
7776 bfd_reloc_code_real_type
7777 x86_cons (expressionS *exp, int size)
7778 {
7779 bfd_reloc_code_real_type got_reloc = NO_RELOC;
7780
7781 intel_syntax = -intel_syntax;
7782
7783 exp->X_md = 0;
7784 if (size == 4 || (object_64bit && size == 8))
7785 {
7786 /* Handle @GOTOFF and the like in an expression. */
7787 char *save;
7788 char *gotfree_input_line;
7789 int adjust = 0;
7790
7791 save = input_line_pointer;
7792 gotfree_input_line = lex_got (&got_reloc, &adjust, NULL);
7793 if (gotfree_input_line)
7794 input_line_pointer = gotfree_input_line;
7795
7796 expression (exp);
7797
7798 if (gotfree_input_line)
7799 {
7800 /* expression () has merrily parsed up to the end of line,
7801 or a comma - in the wrong buffer. Transfer how far
7802 input_line_pointer has moved to the right buffer. */
7803 input_line_pointer = (save
7804 + (input_line_pointer - gotfree_input_line)
7805 + adjust);
7806 free (gotfree_input_line);
7807 if (exp->X_op == O_constant
7808 || exp->X_op == O_absent
7809 || exp->X_op == O_illegal
7810 || exp->X_op == O_register
7811 || exp->X_op == O_big)
7812 {
7813 char c = *input_line_pointer;
7814 *input_line_pointer = 0;
7815 as_bad (_("missing or invalid expression `%s'"), save);
7816 *input_line_pointer = c;
7817 }
7818 }
7819 }
7820 else
7821 expression (exp);
7822
7823 intel_syntax = -intel_syntax;
7824
7825 if (intel_syntax)
7826 i386_intel_simplify (exp);
7827
7828 return got_reloc;
7829 }
7830
7831 static void
7832 signed_cons (int size)
7833 {
7834 if (flag_code == CODE_64BIT)
7835 cons_sign = 1;
7836 cons (size);
7837 cons_sign = -1;
7838 }
7839
7840 #ifdef TE_PE
7841 static void
7842 pe_directive_secrel (int dummy ATTRIBUTE_UNUSED)
7843 {
7844 expressionS exp;
7845
7846 do
7847 {
7848 expression (&exp);
7849 if (exp.X_op == O_symbol)
7850 exp.X_op = O_secrel;
7851
7852 emit_expr (&exp, 4);
7853 }
7854 while (*input_line_pointer++ == ',');
7855
7856 input_line_pointer--;
7857 demand_empty_rest_of_line ();
7858 }
7859 #endif
7860
7861 /* Handle Vector operations. */
7862
7863 static char *
7864 check_VecOperations (char *op_string, char *op_end)
7865 {
7866 const reg_entry *mask;
7867 const char *saved;
7868 char *end_op;
7869
7870 while (*op_string
7871 && (op_end == NULL || op_string < op_end))
7872 {
7873 saved = op_string;
7874 if (*op_string == '{')
7875 {
7876 op_string++;
7877
7878 /* Check broadcasts. */
7879 if (strncmp (op_string, "1to", 3) == 0)
7880 {
7881 int bcst_type;
7882
7883 if (i.broadcast)
7884 goto duplicated_vec_op;
7885
7886 op_string += 3;
7887 if (*op_string == '8')
7888 bcst_type = BROADCAST_1TO8;
7889 else if (*op_string == '4')
7890 bcst_type = BROADCAST_1TO4;
7891 else if (*op_string == '2')
7892 bcst_type = BROADCAST_1TO2;
7893 else if (*op_string == '1'
7894 && *(op_string+1) == '6')
7895 {
7896 bcst_type = BROADCAST_1TO16;
7897 op_string++;
7898 }
7899 else
7900 {
7901 as_bad (_("Unsupported broadcast: `%s'"), saved);
7902 return NULL;
7903 }
7904 op_string++;
7905
7906 broadcast_op.type = bcst_type;
7907 broadcast_op.operand = this_operand;
7908 i.broadcast = &broadcast_op;
7909 }
7910 /* Check masking operation. */
7911 else if ((mask = parse_register (op_string, &end_op)) != NULL)
7912 {
7913 /* k0 can't be used for write mask. */
7914 if (mask->reg_num == 0)
7915 {
7916 as_bad (_("`%s' can't be used for write mask"),
7917 op_string);
7918 return NULL;
7919 }
7920
7921 if (!i.mask)
7922 {
7923 mask_op.mask = mask;
7924 mask_op.zeroing = 0;
7925 mask_op.operand = this_operand;
7926 i.mask = &mask_op;
7927 }
7928 else
7929 {
7930 if (i.mask->mask)
7931 goto duplicated_vec_op;
7932
7933 i.mask->mask = mask;
7934
7935 /* Only "{z}" is allowed here. No need to check
7936 zeroing mask explicitly. */
7937 if (i.mask->operand != this_operand)
7938 {
7939 as_bad (_("invalid write mask `%s'"), saved);
7940 return NULL;
7941 }
7942 }
7943
7944 op_string = end_op;
7945 }
7946 /* Check zeroing-flag for masking operation. */
7947 else if (*op_string == 'z')
7948 {
7949 if (!i.mask)
7950 {
7951 mask_op.mask = NULL;
7952 mask_op.zeroing = 1;
7953 mask_op.operand = this_operand;
7954 i.mask = &mask_op;
7955 }
7956 else
7957 {
7958 if (i.mask->zeroing)
7959 {
7960 duplicated_vec_op:
7961 as_bad (_("duplicated `%s'"), saved);
7962 return NULL;
7963 }
7964
7965 i.mask->zeroing = 1;
7966
7967 /* Only "{%k}" is allowed here. No need to check mask
7968 register explicitly. */
7969 if (i.mask->operand != this_operand)
7970 {
7971 as_bad (_("invalid zeroing-masking `%s'"),
7972 saved);
7973 return NULL;
7974 }
7975 }
7976
7977 op_string++;
7978 }
7979 else
7980 goto unknown_vec_op;
7981
7982 if (*op_string != '}')
7983 {
7984 as_bad (_("missing `}' in `%s'"), saved);
7985 return NULL;
7986 }
7987 op_string++;
7988 continue;
7989 }
7990 unknown_vec_op:
7991 /* We don't know this one. */
7992 as_bad (_("unknown vector operation: `%s'"), saved);
7993 return NULL;
7994 }
7995
7996 return op_string;
7997 }
7998
7999 static int
8000 i386_immediate (char *imm_start)
8001 {
8002 char *save_input_line_pointer;
8003 char *gotfree_input_line;
8004 segT exp_seg = 0;
8005 expressionS *exp;
8006 i386_operand_type types;
8007
8008 operand_type_set (&types, ~0);
8009
8010 if (i.imm_operands == MAX_IMMEDIATE_OPERANDS)
8011 {
8012 as_bad (_("at most %d immediate operands are allowed"),
8013 MAX_IMMEDIATE_OPERANDS);
8014 return 0;
8015 }
8016
8017 exp = &im_expressions[i.imm_operands++];
8018 i.op[this_operand].imms = exp;
8019
8020 if (is_space_char (*imm_start))
8021 ++imm_start;
8022
8023 save_input_line_pointer = input_line_pointer;
8024 input_line_pointer = imm_start;
8025
8026 gotfree_input_line = lex_got (&i.reloc[this_operand], NULL, &types);
8027 if (gotfree_input_line)
8028 input_line_pointer = gotfree_input_line;
8029
8030 exp_seg = expression (exp);
8031
8032 SKIP_WHITESPACE ();
8033
8034 /* Handle vector operations. */
8035 if (*input_line_pointer == '{')
8036 {
8037 input_line_pointer = check_VecOperations (input_line_pointer,
8038 NULL);
8039 if (input_line_pointer == NULL)
8040 return 0;
8041 }
8042
8043 if (*input_line_pointer)
8044 as_bad (_("junk `%s' after expression"), input_line_pointer);
8045
8046 input_line_pointer = save_input_line_pointer;
8047 if (gotfree_input_line)
8048 {
8049 free (gotfree_input_line);
8050
8051 if (exp->X_op == O_constant || exp->X_op == O_register)
8052 exp->X_op = O_illegal;
8053 }
8054
8055 return i386_finalize_immediate (exp_seg, exp, types, imm_start);
8056 }
8057
8058 static int
8059 i386_finalize_immediate (segT exp_seg ATTRIBUTE_UNUSED, expressionS *exp,
8060 i386_operand_type types, const char *imm_start)
8061 {
8062 if (exp->X_op == O_absent || exp->X_op == O_illegal || exp->X_op == O_big)
8063 {
8064 if (imm_start)
8065 as_bad (_("missing or invalid immediate expression `%s'"),
8066 imm_start);
8067 return 0;
8068 }
8069 else if (exp->X_op == O_constant)
8070 {
8071 /* Size it properly later. */
8072 i.types[this_operand].bitfield.imm64 = 1;
8073 /* If not 64bit, sign extend val. */
8074 if (flag_code != CODE_64BIT
8075 && (exp->X_add_number & ~(((addressT) 2 << 31) - 1)) == 0)
8076 exp->X_add_number
8077 = (exp->X_add_number ^ ((addressT) 1 << 31)) - ((addressT) 1 << 31);
8078 }
8079 #if (defined (OBJ_AOUT) || defined (OBJ_MAYBE_AOUT))
8080 else if (OUTPUT_FLAVOR == bfd_target_aout_flavour
8081 && exp_seg != absolute_section
8082 && exp_seg != text_section
8083 && exp_seg != data_section
8084 && exp_seg != bss_section
8085 && exp_seg != undefined_section
8086 && !bfd_is_com_section (exp_seg))
8087 {
8088 as_bad (_("unimplemented segment %s in operand"), exp_seg->name);
8089 return 0;
8090 }
8091 #endif
8092 else if (!intel_syntax && exp_seg == reg_section)
8093 {
8094 if (imm_start)
8095 as_bad (_("illegal immediate register operand %s"), imm_start);
8096 return 0;
8097 }
8098 else
8099 {
8100 /* This is an address. The size of the address will be
8101 determined later, depending on destination register,
8102 suffix, or the default for the section. */
8103 i.types[this_operand].bitfield.imm8 = 1;
8104 i.types[this_operand].bitfield.imm16 = 1;
8105 i.types[this_operand].bitfield.imm32 = 1;
8106 i.types[this_operand].bitfield.imm32s = 1;
8107 i.types[this_operand].bitfield.imm64 = 1;
8108 i.types[this_operand] = operand_type_and (i.types[this_operand],
8109 types);
8110 }
8111
8112 return 1;
8113 }
8114
8115 static char *
8116 i386_scale (char *scale)
8117 {
8118 offsetT val;
8119 char *save = input_line_pointer;
8120
8121 input_line_pointer = scale;
8122 val = get_absolute_expression ();
8123
8124 switch (val)
8125 {
8126 case 1:
8127 i.log2_scale_factor = 0;
8128 break;
8129 case 2:
8130 i.log2_scale_factor = 1;
8131 break;
8132 case 4:
8133 i.log2_scale_factor = 2;
8134 break;
8135 case 8:
8136 i.log2_scale_factor = 3;
8137 break;
8138 default:
8139 {
8140 char sep = *input_line_pointer;
8141
8142 *input_line_pointer = '\0';
8143 as_bad (_("expecting scale factor of 1, 2, 4, or 8: got `%s'"),
8144 scale);
8145 *input_line_pointer = sep;
8146 input_line_pointer = save;
8147 return NULL;
8148 }
8149 }
8150 if (i.log2_scale_factor != 0 && i.index_reg == 0)
8151 {
8152 as_warn (_("scale factor of %d without an index register"),
8153 1 << i.log2_scale_factor);
8154 i.log2_scale_factor = 0;
8155 }
8156 scale = input_line_pointer;
8157 input_line_pointer = save;
8158 return scale;
8159 }
8160
8161 static int
8162 i386_displacement (char *disp_start, char *disp_end)
8163 {
8164 expressionS *exp;
8165 segT exp_seg = 0;
8166 char *save_input_line_pointer;
8167 char *gotfree_input_line;
8168 int override;
8169 i386_operand_type bigdisp, types = anydisp;
8170 int ret;
8171
8172 if (i.disp_operands == MAX_MEMORY_OPERANDS)
8173 {
8174 as_bad (_("at most %d displacement operands are allowed"),
8175 MAX_MEMORY_OPERANDS);
8176 return 0;
8177 }
8178
8179 operand_type_set (&bigdisp, 0);
8180 if ((i.types[this_operand].bitfield.jumpabsolute)
8181 || (!current_templates->start->opcode_modifier.jump
8182 && !current_templates->start->opcode_modifier.jumpdword))
8183 {
8184 bigdisp.bitfield.disp32 = 1;
8185 override = (i.prefix[ADDR_PREFIX] != 0);
8186 if (flag_code == CODE_64BIT)
8187 {
8188 if (!override)
8189 {
8190 bigdisp.bitfield.disp32s = 1;
8191 bigdisp.bitfield.disp64 = 1;
8192 }
8193 }
8194 else if ((flag_code == CODE_16BIT) ^ override)
8195 {
8196 bigdisp.bitfield.disp32 = 0;
8197 bigdisp.bitfield.disp16 = 1;
8198 }
8199 }
8200 else
8201 {
8202 /* For PC-relative branches, the width of the displacement
8203 is dependent upon data size, not address size. */
8204 override = (i.prefix[DATA_PREFIX] != 0);
8205 if (flag_code == CODE_64BIT)
8206 {
8207 if (override || i.suffix == WORD_MNEM_SUFFIX)
8208 bigdisp.bitfield.disp16 = 1;
8209 else
8210 {
8211 bigdisp.bitfield.disp32 = 1;
8212 bigdisp.bitfield.disp32s = 1;
8213 }
8214 }
8215 else
8216 {
8217 if (!override)
8218 override = (i.suffix == (flag_code != CODE_16BIT
8219 ? WORD_MNEM_SUFFIX
8220 : LONG_MNEM_SUFFIX));
8221 bigdisp.bitfield.disp32 = 1;
8222 if ((flag_code == CODE_16BIT) ^ override)
8223 {
8224 bigdisp.bitfield.disp32 = 0;
8225 bigdisp.bitfield.disp16 = 1;
8226 }
8227 }
8228 }
8229 i.types[this_operand] = operand_type_or (i.types[this_operand],
8230 bigdisp);
8231
8232 exp = &disp_expressions[i.disp_operands];
8233 i.op[this_operand].disps = exp;
8234 i.disp_operands++;
8235 save_input_line_pointer = input_line_pointer;
8236 input_line_pointer = disp_start;
8237 END_STRING_AND_SAVE (disp_end);
8238
8239 #ifndef GCC_ASM_O_HACK
8240 #define GCC_ASM_O_HACK 0
8241 #endif
8242 #if GCC_ASM_O_HACK
8243 END_STRING_AND_SAVE (disp_end + 1);
8244 if (i.types[this_operand].bitfield.baseIndex
8245 && displacement_string_end[-1] == '+')
8246 {
8247 /* This hack is to avoid a warning when using the "o"
8248 constraint within gcc asm statements.
8249 For instance:
8250
8251 #define _set_tssldt_desc(n,addr,limit,type) \
8252 __asm__ __volatile__ ( \
8253 "movw %w2,%0\n\t" \
8254 "movw %w1,2+%0\n\t" \
8255 "rorl $16,%1\n\t" \
8256 "movb %b1,4+%0\n\t" \
8257 "movb %4,5+%0\n\t" \
8258 "movb $0,6+%0\n\t" \
8259 "movb %h1,7+%0\n\t" \
8260 "rorl $16,%1" \
8261 : "=o"(*(n)) : "q" (addr), "ri"(limit), "i"(type))
8262
8263 This works great except that the output assembler ends
8264 up looking a bit weird if it turns out that there is
8265 no offset. You end up producing code that looks like:
8266
8267 #APP
8268 movw $235,(%eax)
8269 movw %dx,2+(%eax)
8270 rorl $16,%edx
8271 movb %dl,4+(%eax)
8272 movb $137,5+(%eax)
8273 movb $0,6+(%eax)
8274 movb %dh,7+(%eax)
8275 rorl $16,%edx
8276 #NO_APP
8277
8278 So here we provide the missing zero. */
8279
8280 *displacement_string_end = '0';
8281 }
8282 #endif
8283 gotfree_input_line = lex_got (&i.reloc[this_operand], NULL, &types);
8284 if (gotfree_input_line)
8285 input_line_pointer = gotfree_input_line;
8286
8287 exp_seg = expression (exp);
8288
8289 SKIP_WHITESPACE ();
8290 if (*input_line_pointer)
8291 as_bad (_("junk `%s' after expression"), input_line_pointer);
8292 #if GCC_ASM_O_HACK
8293 RESTORE_END_STRING (disp_end + 1);
8294 #endif
8295 input_line_pointer = save_input_line_pointer;
8296 if (gotfree_input_line)
8297 {
8298 free (gotfree_input_line);
8299
8300 if (exp->X_op == O_constant || exp->X_op == O_register)
8301 exp->X_op = O_illegal;
8302 }
8303
8304 ret = i386_finalize_displacement (exp_seg, exp, types, disp_start);
8305
8306 RESTORE_END_STRING (disp_end);
8307
8308 return ret;
8309 }
8310
8311 static int
8312 i386_finalize_displacement (segT exp_seg ATTRIBUTE_UNUSED, expressionS *exp,
8313 i386_operand_type types, const char *disp_start)
8314 {
8315 i386_operand_type bigdisp;
8316 int ret = 1;
8317
8318 /* We do this to make sure that the section symbol is in
8319 the symbol table. We will ultimately change the relocation
8320 to be relative to the beginning of the section. */
8321 if (i.reloc[this_operand] == BFD_RELOC_386_GOTOFF
8322 || i.reloc[this_operand] == BFD_RELOC_X86_64_GOTPCREL
8323 || i.reloc[this_operand] == BFD_RELOC_X86_64_GOTOFF64)
8324 {
8325 if (exp->X_op != O_symbol)
8326 goto inv_disp;
8327
8328 if (S_IS_LOCAL (exp->X_add_symbol)
8329 && S_GET_SEGMENT (exp->X_add_symbol) != undefined_section
8330 && S_GET_SEGMENT (exp->X_add_symbol) != expr_section)
8331 section_symbol (S_GET_SEGMENT (exp->X_add_symbol));
8332 exp->X_op = O_subtract;
8333 exp->X_op_symbol = GOT_symbol;
8334 if (i.reloc[this_operand] == BFD_RELOC_X86_64_GOTPCREL)
8335 i.reloc[this_operand] = BFD_RELOC_32_PCREL;
8336 else if (i.reloc[this_operand] == BFD_RELOC_X86_64_GOTOFF64)
8337 i.reloc[this_operand] = BFD_RELOC_64;
8338 else
8339 i.reloc[this_operand] = BFD_RELOC_32;
8340 }
8341
8342 else if (exp->X_op == O_absent
8343 || exp->X_op == O_illegal
8344 || exp->X_op == O_big)
8345 {
8346 inv_disp:
8347 as_bad (_("missing or invalid displacement expression `%s'"),
8348 disp_start);
8349 ret = 0;
8350 }
8351
8352 else if (flag_code == CODE_64BIT
8353 && !i.prefix[ADDR_PREFIX]
8354 && exp->X_op == O_constant)
8355 {
8356 /* Since displacement is signed extended to 64bit, don't allow
8357 disp32 and turn off disp32s if they are out of range. */
8358 i.types[this_operand].bitfield.disp32 = 0;
8359 if (!fits_in_signed_long (exp->X_add_number))
8360 {
8361 i.types[this_operand].bitfield.disp32s = 0;
8362 if (i.types[this_operand].bitfield.baseindex)
8363 {
8364 as_bad (_("0x%lx out range of signed 32bit displacement"),
8365 (long) exp->X_add_number);
8366 ret = 0;
8367 }
8368 }
8369 }
8370
8371 #if (defined (OBJ_AOUT) || defined (OBJ_MAYBE_AOUT))
8372 else if (exp->X_op != O_constant
8373 && OUTPUT_FLAVOR == bfd_target_aout_flavour
8374 && exp_seg != absolute_section
8375 && exp_seg != text_section
8376 && exp_seg != data_section
8377 && exp_seg != bss_section
8378 && exp_seg != undefined_section
8379 && !bfd_is_com_section (exp_seg))
8380 {
8381 as_bad (_("unimplemented segment %s in operand"), exp_seg->name);
8382 ret = 0;
8383 }
8384 #endif
8385
8386 /* Check if this is a displacement only operand. */
8387 bigdisp = i.types[this_operand];
8388 bigdisp.bitfield.disp8 = 0;
8389 bigdisp.bitfield.disp16 = 0;
8390 bigdisp.bitfield.disp32 = 0;
8391 bigdisp.bitfield.disp32s = 0;
8392 bigdisp.bitfield.disp64 = 0;
8393 if (operand_type_all_zero (&bigdisp))
8394 i.types[this_operand] = operand_type_and (i.types[this_operand],
8395 types);
8396
8397 return ret;
8398 }
8399
8400 /* Make sure the memory operand we've been dealt is valid.
8401 Return 1 on success, 0 on a failure. */
8402
8403 static int
8404 i386_index_check (const char *operand_string)
8405 {
8406 const char *kind = "base/index";
8407 enum flag_code addr_mode;
8408
8409 if (i.prefix[ADDR_PREFIX])
8410 addr_mode = flag_code == CODE_32BIT ? CODE_16BIT : CODE_32BIT;
8411 else
8412 {
8413 addr_mode = flag_code;
8414
8415 #if INFER_ADDR_PREFIX
8416 if (i.mem_operands == 0)
8417 {
8418 /* Infer address prefix from the first memory operand. */
8419 const reg_entry *addr_reg = i.base_reg;
8420
8421 if (addr_reg == NULL)
8422 addr_reg = i.index_reg;
8423
8424 if (addr_reg)
8425 {
8426 if (addr_reg->reg_num == RegEip
8427 || addr_reg->reg_num == RegEiz
8428 || addr_reg->reg_type.bitfield.reg32)
8429 addr_mode = CODE_32BIT;
8430 else if (flag_code != CODE_64BIT
8431 && addr_reg->reg_type.bitfield.reg16)
8432 addr_mode = CODE_16BIT;
8433
8434 if (addr_mode != flag_code)
8435 {
8436 i.prefix[ADDR_PREFIX] = ADDR_PREFIX_OPCODE;
8437 i.prefixes += 1;
8438 /* Change the size of any displacement too. At most one
8439 of Disp16 or Disp32 is set.
8440 FIXME. There doesn't seem to be any real need for
8441 separate Disp16 and Disp32 flags. The same goes for
8442 Imm16 and Imm32. Removing them would probably clean
8443 up the code quite a lot. */
8444 if (flag_code != CODE_64BIT
8445 && (i.types[this_operand].bitfield.disp16
8446 || i.types[this_operand].bitfield.disp32))
8447 i.types[this_operand]
8448 = operand_type_xor (i.types[this_operand], disp16_32);
8449 }
8450 }
8451 }
8452 #endif
8453 }
8454
8455 if (current_templates->start->opcode_modifier.isstring
8456 && !current_templates->start->opcode_modifier.immext
8457 && (current_templates->end[-1].opcode_modifier.isstring
8458 || i.mem_operands))
8459 {
8460 /* Memory operands of string insns are special in that they only allow
8461 a single register (rDI, rSI, or rBX) as their memory address. */
8462 const reg_entry *expected_reg;
8463 static const char *di_si[][2] =
8464 {
8465 { "esi", "edi" },
8466 { "si", "di" },
8467 { "rsi", "rdi" }
8468 };
8469 static const char *bx[] = { "ebx", "bx", "rbx" };
8470
8471 kind = "string address";
8472
8473 if (current_templates->start->opcode_modifier.w)
8474 {
8475 i386_operand_type type = current_templates->end[-1].operand_types[0];
8476
8477 if (!type.bitfield.baseindex
8478 || ((!i.mem_operands != !intel_syntax)
8479 && current_templates->end[-1].operand_types[1]
8480 .bitfield.baseindex))
8481 type = current_templates->end[-1].operand_types[1];
8482 expected_reg = hash_find (reg_hash,
8483 di_si[addr_mode][type.bitfield.esseg]);
8484
8485 }
8486 else
8487 expected_reg = hash_find (reg_hash, bx[addr_mode]);
8488
8489 if (i.base_reg != expected_reg
8490 || i.index_reg
8491 || operand_type_check (i.types[this_operand], disp))
8492 {
8493 /* The second memory operand must have the same size as
8494 the first one. */
8495 if (i.mem_operands
8496 && i.base_reg
8497 && !((addr_mode == CODE_64BIT
8498 && i.base_reg->reg_type.bitfield.reg64)
8499 || (addr_mode == CODE_32BIT
8500 ? i.base_reg->reg_type.bitfield.reg32
8501 : i.base_reg->reg_type.bitfield.reg16)))
8502 goto bad_address;
8503
8504 as_warn (_("`%s' is not valid here (expected `%c%s%s%c')"),
8505 operand_string,
8506 intel_syntax ? '[' : '(',
8507 register_prefix,
8508 expected_reg->reg_name,
8509 intel_syntax ? ']' : ')');
8510 return 1;
8511 }
8512 else
8513 return 1;
8514
8515 bad_address:
8516 as_bad (_("`%s' is not a valid %s expression"),
8517 operand_string, kind);
8518 return 0;
8519 }
8520 else
8521 {
8522 if (addr_mode != CODE_16BIT)
8523 {
8524 /* 32-bit/64-bit checks. */
8525 if ((i.base_reg
8526 && (addr_mode == CODE_64BIT
8527 ? !i.base_reg->reg_type.bitfield.reg64
8528 : !i.base_reg->reg_type.bitfield.reg32)
8529 && (i.index_reg
8530 || (i.base_reg->reg_num
8531 != (addr_mode == CODE_64BIT ? RegRip : RegEip))))
8532 || (i.index_reg
8533 && !i.index_reg->reg_type.bitfield.regxmm
8534 && !i.index_reg->reg_type.bitfield.regymm
8535 && !i.index_reg->reg_type.bitfield.regzmm
8536 && ((addr_mode == CODE_64BIT
8537 ? !(i.index_reg->reg_type.bitfield.reg64
8538 || i.index_reg->reg_num == RegRiz)
8539 : !(i.index_reg->reg_type.bitfield.reg32
8540 || i.index_reg->reg_num == RegEiz))
8541 || !i.index_reg->reg_type.bitfield.baseindex)))
8542 goto bad_address;
8543 }
8544 else
8545 {
8546 /* 16-bit checks. */
8547 if ((i.base_reg
8548 && (!i.base_reg->reg_type.bitfield.reg16
8549 || !i.base_reg->reg_type.bitfield.baseindex))
8550 || (i.index_reg
8551 && (!i.index_reg->reg_type.bitfield.reg16
8552 || !i.index_reg->reg_type.bitfield.baseindex
8553 || !(i.base_reg
8554 && i.base_reg->reg_num < 6
8555 && i.index_reg->reg_num >= 6
8556 && i.log2_scale_factor == 0))))
8557 goto bad_address;
8558 }
8559 }
8560 return 1;
8561 }
8562
8563 /* Handle vector immediates. */
8564
8565 static int
8566 RC_SAE_immediate (const char *imm_start)
8567 {
8568 unsigned int match_found, j;
8569 const char *pstr = imm_start;
8570 expressionS *exp;
8571
8572 if (*pstr != '{')
8573 return 0;
8574
8575 pstr++;
8576 match_found = 0;
8577 for (j = 0; j < ARRAY_SIZE (RC_NamesTable); j++)
8578 {
8579 if (!strncmp (pstr, RC_NamesTable[j].name, RC_NamesTable[j].len))
8580 {
8581 if (!i.rounding)
8582 {
8583 rc_op.type = RC_NamesTable[j].type;
8584 rc_op.operand = this_operand;
8585 i.rounding = &rc_op;
8586 }
8587 else
8588 {
8589 as_bad (_("duplicated `%s'"), imm_start);
8590 return 0;
8591 }
8592 pstr += RC_NamesTable[j].len;
8593 match_found = 1;
8594 break;
8595 }
8596 }
8597 if (!match_found)
8598 return 0;
8599
8600 if (*pstr++ != '}')
8601 {
8602 as_bad (_("Missing '}': '%s'"), imm_start);
8603 return 0;
8604 }
8605 /* RC/SAE immediate string should contain nothing more. */;
8606 if (*pstr != 0)
8607 {
8608 as_bad (_("Junk after '}': '%s'"), imm_start);
8609 return 0;
8610 }
8611
8612 exp = &im_expressions[i.imm_operands++];
8613 i.op[this_operand].imms = exp;
8614
8615 exp->X_op = O_constant;
8616 exp->X_add_number = 0;
8617 exp->X_add_symbol = (symbolS *) 0;
8618 exp->X_op_symbol = (symbolS *) 0;
8619
8620 i.types[this_operand].bitfield.imm8 = 1;
8621 return 1;
8622 }
8623
8624 /* Parse OPERAND_STRING into the i386_insn structure I. Returns zero
8625 on error. */
8626
8627 static int
8628 i386_att_operand (char *operand_string)
8629 {
8630 const reg_entry *r;
8631 char *end_op;
8632 char *op_string = operand_string;
8633
8634 if (is_space_char (*op_string))
8635 ++op_string;
8636
8637 /* We check for an absolute prefix (differentiating,
8638 for example, 'jmp pc_relative_label' from 'jmp *absolute_label'. */
8639 if (*op_string == ABSOLUTE_PREFIX)
8640 {
8641 ++op_string;
8642 if (is_space_char (*op_string))
8643 ++op_string;
8644 i.types[this_operand].bitfield.jumpabsolute = 1;
8645 }
8646
8647 /* Check if operand is a register. */
8648 if ((r = parse_register (op_string, &end_op)) != NULL)
8649 {
8650 i386_operand_type temp;
8651
8652 /* Check for a segment override by searching for ':' after a
8653 segment register. */
8654 op_string = end_op;
8655 if (is_space_char (*op_string))
8656 ++op_string;
8657 if (*op_string == ':'
8658 && (r->reg_type.bitfield.sreg2
8659 || r->reg_type.bitfield.sreg3))
8660 {
8661 switch (r->reg_num)
8662 {
8663 case 0:
8664 i.seg[i.mem_operands] = &es;
8665 break;
8666 case 1:
8667 i.seg[i.mem_operands] = &cs;
8668 break;
8669 case 2:
8670 i.seg[i.mem_operands] = &ss;
8671 break;
8672 case 3:
8673 i.seg[i.mem_operands] = &ds;
8674 break;
8675 case 4:
8676 i.seg[i.mem_operands] = &fs;
8677 break;
8678 case 5:
8679 i.seg[i.mem_operands] = &gs;
8680 break;
8681 }
8682
8683 /* Skip the ':' and whitespace. */
8684 ++op_string;
8685 if (is_space_char (*op_string))
8686 ++op_string;
8687
8688 if (!is_digit_char (*op_string)
8689 && !is_identifier_char (*op_string)
8690 && *op_string != '('
8691 && *op_string != ABSOLUTE_PREFIX)
8692 {
8693 as_bad (_("bad memory operand `%s'"), op_string);
8694 return 0;
8695 }
8696 /* Handle case of %es:*foo. */
8697 if (*op_string == ABSOLUTE_PREFIX)
8698 {
8699 ++op_string;
8700 if (is_space_char (*op_string))
8701 ++op_string;
8702 i.types[this_operand].bitfield.jumpabsolute = 1;
8703 }
8704 goto do_memory_reference;
8705 }
8706
8707 /* Handle vector operations. */
8708 if (*op_string == '{')
8709 {
8710 op_string = check_VecOperations (op_string, NULL);
8711 if (op_string == NULL)
8712 return 0;
8713 }
8714
8715 if (*op_string)
8716 {
8717 as_bad (_("junk `%s' after register"), op_string);
8718 return 0;
8719 }
8720 temp = r->reg_type;
8721 temp.bitfield.baseindex = 0;
8722 i.types[this_operand] = operand_type_or (i.types[this_operand],
8723 temp);
8724 i.types[this_operand].bitfield.unspecified = 0;
8725 i.op[this_operand].regs = r;
8726 i.reg_operands++;
8727 }
8728 else if (*op_string == REGISTER_PREFIX)
8729 {
8730 as_bad (_("bad register name `%s'"), op_string);
8731 return 0;
8732 }
8733 else if (*op_string == IMMEDIATE_PREFIX)
8734 {
8735 ++op_string;
8736 if (i.types[this_operand].bitfield.jumpabsolute)
8737 {
8738 as_bad (_("immediate operand illegal with absolute jump"));
8739 return 0;
8740 }
8741 if (!i386_immediate (op_string))
8742 return 0;
8743 }
8744 else if (RC_SAE_immediate (operand_string))
8745 {
8746 /* If it is a RC or SAE immediate, do nothing. */
8747 ;
8748 }
8749 else if (is_digit_char (*op_string)
8750 || is_identifier_char (*op_string)
8751 || *op_string == '"'
8752 || *op_string == '(')
8753 {
8754 /* This is a memory reference of some sort. */
8755 char *base_string;
8756
8757 /* Start and end of displacement string expression (if found). */
8758 char *displacement_string_start;
8759 char *displacement_string_end;
8760 char *vop_start;
8761
8762 do_memory_reference:
8763 if ((i.mem_operands == 1
8764 && !current_templates->start->opcode_modifier.isstring)
8765 || i.mem_operands == 2)
8766 {
8767 as_bad (_("too many memory references for `%s'"),
8768 current_templates->start->name);
8769 return 0;
8770 }
8771
8772 /* Check for base index form. We detect the base index form by
8773 looking for an ')' at the end of the operand, searching
8774 for the '(' matching it, and finding a REGISTER_PREFIX or ','
8775 after the '('. */
8776 base_string = op_string + strlen (op_string);
8777
8778 /* Handle vector operations. */
8779 vop_start = strchr (op_string, '{');
8780 if (vop_start && vop_start < base_string)
8781 {
8782 if (check_VecOperations (vop_start, base_string) == NULL)
8783 return 0;
8784 base_string = vop_start;
8785 }
8786
8787 --base_string;
8788 if (is_space_char (*base_string))
8789 --base_string;
8790
8791 /* If we only have a displacement, set-up for it to be parsed later. */
8792 displacement_string_start = op_string;
8793 displacement_string_end = base_string + 1;
8794
8795 if (*base_string == ')')
8796 {
8797 char *temp_string;
8798 unsigned int parens_balanced = 1;
8799 /* We've already checked that the number of left & right ()'s are
8800 equal, so this loop will not be infinite. */
8801 do
8802 {
8803 base_string--;
8804 if (*base_string == ')')
8805 parens_balanced++;
8806 if (*base_string == '(')
8807 parens_balanced--;
8808 }
8809 while (parens_balanced);
8810
8811 temp_string = base_string;
8812
8813 /* Skip past '(' and whitespace. */
8814 ++base_string;
8815 if (is_space_char (*base_string))
8816 ++base_string;
8817
8818 if (*base_string == ','
8819 || ((i.base_reg = parse_register (base_string, &end_op))
8820 != NULL))
8821 {
8822 displacement_string_end = temp_string;
8823
8824 i.types[this_operand].bitfield.baseindex = 1;
8825
8826 if (i.base_reg)
8827 {
8828 base_string = end_op;
8829 if (is_space_char (*base_string))
8830 ++base_string;
8831 }
8832
8833 /* There may be an index reg or scale factor here. */
8834 if (*base_string == ',')
8835 {
8836 ++base_string;
8837 if (is_space_char (*base_string))
8838 ++base_string;
8839
8840 if ((i.index_reg = parse_register (base_string, &end_op))
8841 != NULL)
8842 {
8843 base_string = end_op;
8844 if (is_space_char (*base_string))
8845 ++base_string;
8846 if (*base_string == ',')
8847 {
8848 ++base_string;
8849 if (is_space_char (*base_string))
8850 ++base_string;
8851 }
8852 else if (*base_string != ')')
8853 {
8854 as_bad (_("expecting `,' or `)' "
8855 "after index register in `%s'"),
8856 operand_string);
8857 return 0;
8858 }
8859 }
8860 else if (*base_string == REGISTER_PREFIX)
8861 {
8862 end_op = strchr (base_string, ',');
8863 if (end_op)
8864 *end_op = '\0';
8865 as_bad (_("bad register name `%s'"), base_string);
8866 return 0;
8867 }
8868
8869 /* Check for scale factor. */
8870 if (*base_string != ')')
8871 {
8872 char *end_scale = i386_scale (base_string);
8873
8874 if (!end_scale)
8875 return 0;
8876
8877 base_string = end_scale;
8878 if (is_space_char (*base_string))
8879 ++base_string;
8880 if (*base_string != ')')
8881 {
8882 as_bad (_("expecting `)' "
8883 "after scale factor in `%s'"),
8884 operand_string);
8885 return 0;
8886 }
8887 }
8888 else if (!i.index_reg)
8889 {
8890 as_bad (_("expecting index register or scale factor "
8891 "after `,'; got '%c'"),
8892 *base_string);
8893 return 0;
8894 }
8895 }
8896 else if (*base_string != ')')
8897 {
8898 as_bad (_("expecting `,' or `)' "
8899 "after base register in `%s'"),
8900 operand_string);
8901 return 0;
8902 }
8903 }
8904 else if (*base_string == REGISTER_PREFIX)
8905 {
8906 end_op = strchr (base_string, ',');
8907 if (end_op)
8908 *end_op = '\0';
8909 as_bad (_("bad register name `%s'"), base_string);
8910 return 0;
8911 }
8912 }
8913
8914 /* If there's an expression beginning the operand, parse it,
8915 assuming displacement_string_start and
8916 displacement_string_end are meaningful. */
8917 if (displacement_string_start != displacement_string_end)
8918 {
8919 if (!i386_displacement (displacement_string_start,
8920 displacement_string_end))
8921 return 0;
8922 }
8923
8924 /* Special case for (%dx) while doing input/output op. */
8925 if (i.base_reg
8926 && operand_type_equal (&i.base_reg->reg_type,
8927 &reg16_inoutportreg)
8928 && i.index_reg == 0
8929 && i.log2_scale_factor == 0
8930 && i.seg[i.mem_operands] == 0
8931 && !operand_type_check (i.types[this_operand], disp))
8932 {
8933 i.types[this_operand] = inoutportreg;
8934 return 1;
8935 }
8936
8937 if (i386_index_check (operand_string) == 0)
8938 return 0;
8939 i.types[this_operand].bitfield.mem = 1;
8940 i.mem_operands++;
8941 }
8942 else
8943 {
8944 /* It's not a memory operand; argh! */
8945 as_bad (_("invalid char %s beginning operand %d `%s'"),
8946 output_invalid (*op_string),
8947 this_operand + 1,
8948 op_string);
8949 return 0;
8950 }
8951 return 1; /* Normal return. */
8952 }
8953 \f
8954 /* Calculate the maximum variable size (i.e., excluding fr_fix)
8955 that an rs_machine_dependent frag may reach. */
8956
8957 unsigned int
8958 i386_frag_max_var (fragS *frag)
8959 {
8960 /* The only relaxable frags are for jumps.
8961 Unconditional jumps can grow by 4 bytes and others by 5 bytes. */
8962 gas_assert (frag->fr_type == rs_machine_dependent);
8963 return TYPE_FROM_RELAX_STATE (frag->fr_subtype) == UNCOND_JUMP ? 4 : 5;
8964 }
8965
8966 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
8967 static int
8968 elf_symbol_resolved_in_segment_p (symbolS *fr_symbol, offsetT fr_var)
8969 {
8970 /* STT_GNU_IFUNC symbol must go through PLT. */
8971 if ((symbol_get_bfdsym (fr_symbol)->flags
8972 & BSF_GNU_INDIRECT_FUNCTION) != 0)
8973 return 0;
8974
8975 if (!S_IS_EXTERNAL (fr_symbol))
8976 /* Symbol may be weak or local. */
8977 return !S_IS_WEAK (fr_symbol);
8978
8979 /* Global symbols with non-default visibility can't be preempted. */
8980 if (ELF_ST_VISIBILITY (S_GET_OTHER (fr_symbol)) != STV_DEFAULT)
8981 return 1;
8982
8983 if (fr_var != NO_RELOC)
8984 switch ((enum bfd_reloc_code_real) fr_var)
8985 {
8986 case BFD_RELOC_386_PLT32:
8987 case BFD_RELOC_X86_64_PLT32:
8988 /* Symbol with PLT relocatin may be preempted. */
8989 return 0;
8990 default:
8991 abort ();
8992 }
8993
8994 /* Global symbols with default visibility in a shared library may be
8995 preempted by another definition. */
8996 return !shared;
8997 }
8998 #endif
8999
9000 /* md_estimate_size_before_relax()
9001
9002 Called just before relax() for rs_machine_dependent frags. The x86
9003 assembler uses these frags to handle variable size jump
9004 instructions.
9005
9006 Any symbol that is now undefined will not become defined.
9007 Return the correct fr_subtype in the frag.
9008 Return the initial "guess for variable size of frag" to caller.
9009 The guess is actually the growth beyond the fixed part. Whatever
9010 we do to grow the fixed or variable part contributes to our
9011 returned value. */
9012
9013 int
9014 md_estimate_size_before_relax (fragS *fragP, segT segment)
9015 {
9016 /* We've already got fragP->fr_subtype right; all we have to do is
9017 check for un-relaxable symbols. On an ELF system, we can't relax
9018 an externally visible symbol, because it may be overridden by a
9019 shared library. */
9020 if (S_GET_SEGMENT (fragP->fr_symbol) != segment
9021 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
9022 || (IS_ELF
9023 && !elf_symbol_resolved_in_segment_p (fragP->fr_symbol,
9024 fragP->fr_var))
9025 #endif
9026 #if defined (OBJ_COFF) && defined (TE_PE)
9027 || (OUTPUT_FLAVOR == bfd_target_coff_flavour
9028 && S_IS_WEAK (fragP->fr_symbol))
9029 #endif
9030 )
9031 {
9032 /* Symbol is undefined in this segment, or we need to keep a
9033 reloc so that weak symbols can be overridden. */
9034 int size = (fragP->fr_subtype & CODE16) ? 2 : 4;
9035 enum bfd_reloc_code_real reloc_type;
9036 unsigned char *opcode;
9037 int old_fr_fix;
9038
9039 if (fragP->fr_var != NO_RELOC)
9040 reloc_type = (enum bfd_reloc_code_real) fragP->fr_var;
9041 else if (size == 2)
9042 reloc_type = BFD_RELOC_16_PCREL;
9043 else
9044 reloc_type = BFD_RELOC_32_PCREL;
9045
9046 old_fr_fix = fragP->fr_fix;
9047 opcode = (unsigned char *) fragP->fr_opcode;
9048
9049 switch (TYPE_FROM_RELAX_STATE (fragP->fr_subtype))
9050 {
9051 case UNCOND_JUMP:
9052 /* Make jmp (0xeb) a (d)word displacement jump. */
9053 opcode[0] = 0xe9;
9054 fragP->fr_fix += size;
9055 fix_new (fragP, old_fr_fix, size,
9056 fragP->fr_symbol,
9057 fragP->fr_offset, 1,
9058 reloc_type);
9059 break;
9060
9061 case COND_JUMP86:
9062 if (size == 2
9063 && (!no_cond_jump_promotion || fragP->fr_var != NO_RELOC))
9064 {
9065 /* Negate the condition, and branch past an
9066 unconditional jump. */
9067 opcode[0] ^= 1;
9068 opcode[1] = 3;
9069 /* Insert an unconditional jump. */
9070 opcode[2] = 0xe9;
9071 /* We added two extra opcode bytes, and have a two byte
9072 offset. */
9073 fragP->fr_fix += 2 + 2;
9074 fix_new (fragP, old_fr_fix + 2, 2,
9075 fragP->fr_symbol,
9076 fragP->fr_offset, 1,
9077 reloc_type);
9078 break;
9079 }
9080 /* Fall through. */
9081
9082 case COND_JUMP:
9083 if (no_cond_jump_promotion && fragP->fr_var == NO_RELOC)
9084 {
9085 fixS *fixP;
9086
9087 fragP->fr_fix += 1;
9088 fixP = fix_new (fragP, old_fr_fix, 1,
9089 fragP->fr_symbol,
9090 fragP->fr_offset, 1,
9091 BFD_RELOC_8_PCREL);
9092 fixP->fx_signed = 1;
9093 break;
9094 }
9095
9096 /* This changes the byte-displacement jump 0x7N
9097 to the (d)word-displacement jump 0x0f,0x8N. */
9098 opcode[1] = opcode[0] + 0x10;
9099 opcode[0] = TWO_BYTE_OPCODE_ESCAPE;
9100 /* We've added an opcode byte. */
9101 fragP->fr_fix += 1 + size;
9102 fix_new (fragP, old_fr_fix + 1, size,
9103 fragP->fr_symbol,
9104 fragP->fr_offset, 1,
9105 reloc_type);
9106 break;
9107
9108 default:
9109 BAD_CASE (fragP->fr_subtype);
9110 break;
9111 }
9112 frag_wane (fragP);
9113 return fragP->fr_fix - old_fr_fix;
9114 }
9115
9116 /* Guess size depending on current relax state. Initially the relax
9117 state will correspond to a short jump and we return 1, because
9118 the variable part of the frag (the branch offset) is one byte
9119 long. However, we can relax a section more than once and in that
9120 case we must either set fr_subtype back to the unrelaxed state,
9121 or return the value for the appropriate branch. */
9122 return md_relax_table[fragP->fr_subtype].rlx_length;
9123 }
9124
9125 /* Called after relax() is finished.
9126
9127 In: Address of frag.
9128 fr_type == rs_machine_dependent.
9129 fr_subtype is what the address relaxed to.
9130
9131 Out: Any fixSs and constants are set up.
9132 Caller will turn frag into a ".space 0". */
9133
9134 void
9135 md_convert_frag (bfd *abfd ATTRIBUTE_UNUSED, segT sec ATTRIBUTE_UNUSED,
9136 fragS *fragP)
9137 {
9138 unsigned char *opcode;
9139 unsigned char *where_to_put_displacement = NULL;
9140 offsetT target_address;
9141 offsetT opcode_address;
9142 unsigned int extension = 0;
9143 offsetT displacement_from_opcode_start;
9144
9145 opcode = (unsigned char *) fragP->fr_opcode;
9146
9147 /* Address we want to reach in file space. */
9148 target_address = S_GET_VALUE (fragP->fr_symbol) + fragP->fr_offset;
9149
9150 /* Address opcode resides at in file space. */
9151 opcode_address = fragP->fr_address + fragP->fr_fix;
9152
9153 /* Displacement from opcode start to fill into instruction. */
9154 displacement_from_opcode_start = target_address - opcode_address;
9155
9156 if ((fragP->fr_subtype & BIG) == 0)
9157 {
9158 /* Don't have to change opcode. */
9159 extension = 1; /* 1 opcode + 1 displacement */
9160 where_to_put_displacement = &opcode[1];
9161 }
9162 else
9163 {
9164 if (no_cond_jump_promotion
9165 && TYPE_FROM_RELAX_STATE (fragP->fr_subtype) != UNCOND_JUMP)
9166 as_warn_where (fragP->fr_file, fragP->fr_line,
9167 _("long jump required"));
9168
9169 switch (fragP->fr_subtype)
9170 {
9171 case ENCODE_RELAX_STATE (UNCOND_JUMP, BIG):
9172 extension = 4; /* 1 opcode + 4 displacement */
9173 opcode[0] = 0xe9;
9174 where_to_put_displacement = &opcode[1];
9175 break;
9176
9177 case ENCODE_RELAX_STATE (UNCOND_JUMP, BIG16):
9178 extension = 2; /* 1 opcode + 2 displacement */
9179 opcode[0] = 0xe9;
9180 where_to_put_displacement = &opcode[1];
9181 break;
9182
9183 case ENCODE_RELAX_STATE (COND_JUMP, BIG):
9184 case ENCODE_RELAX_STATE (COND_JUMP86, BIG):
9185 extension = 5; /* 2 opcode + 4 displacement */
9186 opcode[1] = opcode[0] + 0x10;
9187 opcode[0] = TWO_BYTE_OPCODE_ESCAPE;
9188 where_to_put_displacement = &opcode[2];
9189 break;
9190
9191 case ENCODE_RELAX_STATE (COND_JUMP, BIG16):
9192 extension = 3; /* 2 opcode + 2 displacement */
9193 opcode[1] = opcode[0] + 0x10;
9194 opcode[0] = TWO_BYTE_OPCODE_ESCAPE;
9195 where_to_put_displacement = &opcode[2];
9196 break;
9197
9198 case ENCODE_RELAX_STATE (COND_JUMP86, BIG16):
9199 extension = 4;
9200 opcode[0] ^= 1;
9201 opcode[1] = 3;
9202 opcode[2] = 0xe9;
9203 where_to_put_displacement = &opcode[3];
9204 break;
9205
9206 default:
9207 BAD_CASE (fragP->fr_subtype);
9208 break;
9209 }
9210 }
9211
9212 /* If size if less then four we are sure that the operand fits,
9213 but if it's 4, then it could be that the displacement is larger
9214 then -/+ 2GB. */
9215 if (DISP_SIZE_FROM_RELAX_STATE (fragP->fr_subtype) == 4
9216 && object_64bit
9217 && ((addressT) (displacement_from_opcode_start - extension
9218 + ((addressT) 1 << 31))
9219 > (((addressT) 2 << 31) - 1)))
9220 {
9221 as_bad_where (fragP->fr_file, fragP->fr_line,
9222 _("jump target out of range"));
9223 /* Make us emit 0. */
9224 displacement_from_opcode_start = extension;
9225 }
9226 /* Now put displacement after opcode. */
9227 md_number_to_chars ((char *) where_to_put_displacement,
9228 (valueT) (displacement_from_opcode_start - extension),
9229 DISP_SIZE_FROM_RELAX_STATE (fragP->fr_subtype));
9230 fragP->fr_fix += extension;
9231 }
9232 \f
9233 /* Apply a fixup (fixP) to segment data, once it has been determined
9234 by our caller that we have all the info we need to fix it up.
9235
9236 Parameter valP is the pointer to the value of the bits.
9237
9238 On the 386, immediates, displacements, and data pointers are all in
9239 the same (little-endian) format, so we don't need to care about which
9240 we are handling. */
9241
9242 void
9243 md_apply_fix (fixS *fixP, valueT *valP, segT seg ATTRIBUTE_UNUSED)
9244 {
9245 char *p = fixP->fx_where + fixP->fx_frag->fr_literal;
9246 valueT value = *valP;
9247
9248 #if !defined (TE_Mach)
9249 if (fixP->fx_pcrel)
9250 {
9251 switch (fixP->fx_r_type)
9252 {
9253 default:
9254 break;
9255
9256 case BFD_RELOC_64:
9257 fixP->fx_r_type = BFD_RELOC_64_PCREL;
9258 break;
9259 case BFD_RELOC_32:
9260 case BFD_RELOC_X86_64_32S:
9261 fixP->fx_r_type = BFD_RELOC_32_PCREL;
9262 break;
9263 case BFD_RELOC_16:
9264 fixP->fx_r_type = BFD_RELOC_16_PCREL;
9265 break;
9266 case BFD_RELOC_8:
9267 fixP->fx_r_type = BFD_RELOC_8_PCREL;
9268 break;
9269 }
9270 }
9271
9272 if (fixP->fx_addsy != NULL
9273 && (fixP->fx_r_type == BFD_RELOC_32_PCREL
9274 || fixP->fx_r_type == BFD_RELOC_64_PCREL
9275 || fixP->fx_r_type == BFD_RELOC_16_PCREL
9276 || fixP->fx_r_type == BFD_RELOC_8_PCREL)
9277 && !use_rela_relocations)
9278 {
9279 /* This is a hack. There should be a better way to handle this.
9280 This covers for the fact that bfd_install_relocation will
9281 subtract the current location (for partial_inplace, PC relative
9282 relocations); see more below. */
9283 #ifndef OBJ_AOUT
9284 if (IS_ELF
9285 #ifdef TE_PE
9286 || OUTPUT_FLAVOR == bfd_target_coff_flavour
9287 #endif
9288 )
9289 value += fixP->fx_where + fixP->fx_frag->fr_address;
9290 #endif
9291 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
9292 if (IS_ELF)
9293 {
9294 segT sym_seg = S_GET_SEGMENT (fixP->fx_addsy);
9295
9296 if ((sym_seg == seg
9297 || (symbol_section_p (fixP->fx_addsy)
9298 && sym_seg != absolute_section))
9299 && !generic_force_reloc (fixP))
9300 {
9301 /* Yes, we add the values in twice. This is because
9302 bfd_install_relocation subtracts them out again. I think
9303 bfd_install_relocation is broken, but I don't dare change
9304 it. FIXME. */
9305 value += fixP->fx_where + fixP->fx_frag->fr_address;
9306 }
9307 }
9308 #endif
9309 #if defined (OBJ_COFF) && defined (TE_PE)
9310 /* For some reason, the PE format does not store a
9311 section address offset for a PC relative symbol. */
9312 if (S_GET_SEGMENT (fixP->fx_addsy) != seg
9313 || S_IS_WEAK (fixP->fx_addsy))
9314 value += md_pcrel_from (fixP);
9315 #endif
9316 }
9317 #if defined (OBJ_COFF) && defined (TE_PE)
9318 if (fixP->fx_addsy != NULL
9319 && S_IS_WEAK (fixP->fx_addsy)
9320 /* PR 16858: Do not modify weak function references. */
9321 && ! fixP->fx_pcrel)
9322 {
9323 #if !defined (TE_PEP)
9324 /* For x86 PE weak function symbols are neither PC-relative
9325 nor do they set S_IS_FUNCTION. So the only reliable way
9326 to detect them is to check the flags of their containing
9327 section. */
9328 if (S_GET_SEGMENT (fixP->fx_addsy) != NULL
9329 && S_GET_SEGMENT (fixP->fx_addsy)->flags & SEC_CODE)
9330 ;
9331 else
9332 #endif
9333 value -= S_GET_VALUE (fixP->fx_addsy);
9334 }
9335 #endif
9336
9337 /* Fix a few things - the dynamic linker expects certain values here,
9338 and we must not disappoint it. */
9339 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
9340 if (IS_ELF && fixP->fx_addsy)
9341 switch (fixP->fx_r_type)
9342 {
9343 case BFD_RELOC_386_PLT32:
9344 case BFD_RELOC_X86_64_PLT32:
9345 /* Make the jump instruction point to the address of the operand. At
9346 runtime we merely add the offset to the actual PLT entry. */
9347 value = -4;
9348 break;
9349
9350 case BFD_RELOC_386_TLS_GD:
9351 case BFD_RELOC_386_TLS_LDM:
9352 case BFD_RELOC_386_TLS_IE_32:
9353 case BFD_RELOC_386_TLS_IE:
9354 case BFD_RELOC_386_TLS_GOTIE:
9355 case BFD_RELOC_386_TLS_GOTDESC:
9356 case BFD_RELOC_X86_64_TLSGD:
9357 case BFD_RELOC_X86_64_TLSLD:
9358 case BFD_RELOC_X86_64_GOTTPOFF:
9359 case BFD_RELOC_X86_64_GOTPC32_TLSDESC:
9360 value = 0; /* Fully resolved at runtime. No addend. */
9361 /* Fallthrough */
9362 case BFD_RELOC_386_TLS_LE:
9363 case BFD_RELOC_386_TLS_LDO_32:
9364 case BFD_RELOC_386_TLS_LE_32:
9365 case BFD_RELOC_X86_64_DTPOFF32:
9366 case BFD_RELOC_X86_64_DTPOFF64:
9367 case BFD_RELOC_X86_64_TPOFF32:
9368 case BFD_RELOC_X86_64_TPOFF64:
9369 S_SET_THREAD_LOCAL (fixP->fx_addsy);
9370 break;
9371
9372 case BFD_RELOC_386_TLS_DESC_CALL:
9373 case BFD_RELOC_X86_64_TLSDESC_CALL:
9374 value = 0; /* Fully resolved at runtime. No addend. */
9375 S_SET_THREAD_LOCAL (fixP->fx_addsy);
9376 fixP->fx_done = 0;
9377 return;
9378
9379 case BFD_RELOC_VTABLE_INHERIT:
9380 case BFD_RELOC_VTABLE_ENTRY:
9381 fixP->fx_done = 0;
9382 return;
9383
9384 default:
9385 break;
9386 }
9387 #endif /* defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) */
9388 *valP = value;
9389 #endif /* !defined (TE_Mach) */
9390
9391 /* Are we finished with this relocation now? */
9392 if (fixP->fx_addsy == NULL)
9393 fixP->fx_done = 1;
9394 #if defined (OBJ_COFF) && defined (TE_PE)
9395 else if (fixP->fx_addsy != NULL && S_IS_WEAK (fixP->fx_addsy))
9396 {
9397 fixP->fx_done = 0;
9398 /* Remember value for tc_gen_reloc. */
9399 fixP->fx_addnumber = value;
9400 /* Clear out the frag for now. */
9401 value = 0;
9402 }
9403 #endif
9404 else if (use_rela_relocations)
9405 {
9406 fixP->fx_no_overflow = 1;
9407 /* Remember value for tc_gen_reloc. */
9408 fixP->fx_addnumber = value;
9409 value = 0;
9410 }
9411
9412 md_number_to_chars (p, value, fixP->fx_size);
9413 }
9414 \f
9415 const char *
9416 md_atof (int type, char *litP, int *sizeP)
9417 {
9418 /* This outputs the LITTLENUMs in REVERSE order;
9419 in accord with the bigendian 386. */
9420 return ieee_md_atof (type, litP, sizeP, FALSE);
9421 }
9422 \f
9423 static char output_invalid_buf[sizeof (unsigned char) * 2 + 6];
9424
9425 static char *
9426 output_invalid (int c)
9427 {
9428 if (ISPRINT (c))
9429 snprintf (output_invalid_buf, sizeof (output_invalid_buf),
9430 "'%c'", c);
9431 else
9432 snprintf (output_invalid_buf, sizeof (output_invalid_buf),
9433 "(0x%x)", (unsigned char) c);
9434 return output_invalid_buf;
9435 }
9436
9437 /* REG_STRING starts *before* REGISTER_PREFIX. */
9438
9439 static const reg_entry *
9440 parse_real_register (char *reg_string, char **end_op)
9441 {
9442 char *s = reg_string;
9443 char *p;
9444 char reg_name_given[MAX_REG_NAME_SIZE + 1];
9445 const reg_entry *r;
9446
9447 /* Skip possible REGISTER_PREFIX and possible whitespace. */
9448 if (*s == REGISTER_PREFIX)
9449 ++s;
9450
9451 if (is_space_char (*s))
9452 ++s;
9453
9454 p = reg_name_given;
9455 while ((*p++ = register_chars[(unsigned char) *s]) != '\0')
9456 {
9457 if (p >= reg_name_given + MAX_REG_NAME_SIZE)
9458 return (const reg_entry *) NULL;
9459 s++;
9460 }
9461
9462 /* For naked regs, make sure that we are not dealing with an identifier.
9463 This prevents confusing an identifier like `eax_var' with register
9464 `eax'. */
9465 if (allow_naked_reg && identifier_chars[(unsigned char) *s])
9466 return (const reg_entry *) NULL;
9467
9468 *end_op = s;
9469
9470 r = (const reg_entry *) hash_find (reg_hash, reg_name_given);
9471
9472 /* Handle floating point regs, allowing spaces in the (i) part. */
9473 if (r == i386_regtab /* %st is first entry of table */)
9474 {
9475 if (is_space_char (*s))
9476 ++s;
9477 if (*s == '(')
9478 {
9479 ++s;
9480 if (is_space_char (*s))
9481 ++s;
9482 if (*s >= '0' && *s <= '7')
9483 {
9484 int fpr = *s - '0';
9485 ++s;
9486 if (is_space_char (*s))
9487 ++s;
9488 if (*s == ')')
9489 {
9490 *end_op = s + 1;
9491 r = (const reg_entry *) hash_find (reg_hash, "st(0)");
9492 know (r);
9493 return r + fpr;
9494 }
9495 }
9496 /* We have "%st(" then garbage. */
9497 return (const reg_entry *) NULL;
9498 }
9499 }
9500
9501 if (r == NULL || allow_pseudo_reg)
9502 return r;
9503
9504 if (operand_type_all_zero (&r->reg_type))
9505 return (const reg_entry *) NULL;
9506
9507 if ((r->reg_type.bitfield.reg32
9508 || r->reg_type.bitfield.sreg3
9509 || r->reg_type.bitfield.control
9510 || r->reg_type.bitfield.debug
9511 || r->reg_type.bitfield.test)
9512 && !cpu_arch_flags.bitfield.cpui386)
9513 return (const reg_entry *) NULL;
9514
9515 if (r->reg_type.bitfield.floatreg
9516 && !cpu_arch_flags.bitfield.cpu8087
9517 && !cpu_arch_flags.bitfield.cpu287
9518 && !cpu_arch_flags.bitfield.cpu387)
9519 return (const reg_entry *) NULL;
9520
9521 if (r->reg_type.bitfield.regmmx && !cpu_arch_flags.bitfield.cpummx)
9522 return (const reg_entry *) NULL;
9523
9524 if (r->reg_type.bitfield.regxmm && !cpu_arch_flags.bitfield.cpusse)
9525 return (const reg_entry *) NULL;
9526
9527 if (r->reg_type.bitfield.regymm && !cpu_arch_flags.bitfield.cpuavx)
9528 return (const reg_entry *) NULL;
9529
9530 if ((r->reg_type.bitfield.regzmm || r->reg_type.bitfield.regmask)
9531 && !cpu_arch_flags.bitfield.cpuavx512f)
9532 return (const reg_entry *) NULL;
9533
9534 /* Don't allow fake index register unless allow_index_reg isn't 0. */
9535 if (!allow_index_reg
9536 && (r->reg_num == RegEiz || r->reg_num == RegRiz))
9537 return (const reg_entry *) NULL;
9538
9539 /* Upper 16 vector register is only available with VREX in 64bit
9540 mode. */
9541 if ((r->reg_flags & RegVRex))
9542 {
9543 if (!cpu_arch_flags.bitfield.cpuvrex
9544 || flag_code != CODE_64BIT)
9545 return (const reg_entry *) NULL;
9546
9547 i.need_vrex = 1;
9548 }
9549
9550 if (((r->reg_flags & (RegRex64 | RegRex))
9551 || r->reg_type.bitfield.reg64)
9552 && (!cpu_arch_flags.bitfield.cpulm
9553 || !operand_type_equal (&r->reg_type, &control))
9554 && flag_code != CODE_64BIT)
9555 return (const reg_entry *) NULL;
9556
9557 if (r->reg_type.bitfield.sreg3 && r->reg_num == RegFlat && !intel_syntax)
9558 return (const reg_entry *) NULL;
9559
9560 return r;
9561 }
9562
9563 /* REG_STRING starts *before* REGISTER_PREFIX. */
9564
9565 static const reg_entry *
9566 parse_register (char *reg_string, char **end_op)
9567 {
9568 const reg_entry *r;
9569
9570 if (*reg_string == REGISTER_PREFIX || allow_naked_reg)
9571 r = parse_real_register (reg_string, end_op);
9572 else
9573 r = NULL;
9574 if (!r)
9575 {
9576 char *save = input_line_pointer;
9577 char c;
9578 symbolS *symbolP;
9579
9580 input_line_pointer = reg_string;
9581 c = get_symbol_name (&reg_string);
9582 symbolP = symbol_find (reg_string);
9583 if (symbolP && S_GET_SEGMENT (symbolP) == reg_section)
9584 {
9585 const expressionS *e = symbol_get_value_expression (symbolP);
9586
9587 know (e->X_op == O_register);
9588 know (e->X_add_number >= 0
9589 && (valueT) e->X_add_number < i386_regtab_size);
9590 r = i386_regtab + e->X_add_number;
9591 if ((r->reg_flags & RegVRex))
9592 i.need_vrex = 1;
9593 *end_op = input_line_pointer;
9594 }
9595 *input_line_pointer = c;
9596 input_line_pointer = save;
9597 }
9598 return r;
9599 }
9600
9601 int
9602 i386_parse_name (char *name, expressionS *e, char *nextcharP)
9603 {
9604 const reg_entry *r;
9605 char *end = input_line_pointer;
9606
9607 *end = *nextcharP;
9608 r = parse_register (name, &input_line_pointer);
9609 if (r && end <= input_line_pointer)
9610 {
9611 *nextcharP = *input_line_pointer;
9612 *input_line_pointer = 0;
9613 e->X_op = O_register;
9614 e->X_add_number = r - i386_regtab;
9615 return 1;
9616 }
9617 input_line_pointer = end;
9618 *end = 0;
9619 return intel_syntax ? i386_intel_parse_name (name, e) : 0;
9620 }
9621
9622 void
9623 md_operand (expressionS *e)
9624 {
9625 char *end;
9626 const reg_entry *r;
9627
9628 switch (*input_line_pointer)
9629 {
9630 case REGISTER_PREFIX:
9631 r = parse_real_register (input_line_pointer, &end);
9632 if (r)
9633 {
9634 e->X_op = O_register;
9635 e->X_add_number = r - i386_regtab;
9636 input_line_pointer = end;
9637 }
9638 break;
9639
9640 case '[':
9641 gas_assert (intel_syntax);
9642 end = input_line_pointer++;
9643 expression (e);
9644 if (*input_line_pointer == ']')
9645 {
9646 ++input_line_pointer;
9647 e->X_op_symbol = make_expr_symbol (e);
9648 e->X_add_symbol = NULL;
9649 e->X_add_number = 0;
9650 e->X_op = O_index;
9651 }
9652 else
9653 {
9654 e->X_op = O_absent;
9655 input_line_pointer = end;
9656 }
9657 break;
9658 }
9659 }
9660
9661 \f
9662 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
9663 const char *md_shortopts = "kVQ:sqn";
9664 #else
9665 const char *md_shortopts = "qn";
9666 #endif
9667
9668 #define OPTION_32 (OPTION_MD_BASE + 0)
9669 #define OPTION_64 (OPTION_MD_BASE + 1)
9670 #define OPTION_DIVIDE (OPTION_MD_BASE + 2)
9671 #define OPTION_MARCH (OPTION_MD_BASE + 3)
9672 #define OPTION_MTUNE (OPTION_MD_BASE + 4)
9673 #define OPTION_MMNEMONIC (OPTION_MD_BASE + 5)
9674 #define OPTION_MSYNTAX (OPTION_MD_BASE + 6)
9675 #define OPTION_MINDEX_REG (OPTION_MD_BASE + 7)
9676 #define OPTION_MNAKED_REG (OPTION_MD_BASE + 8)
9677 #define OPTION_MOLD_GCC (OPTION_MD_BASE + 9)
9678 #define OPTION_MSSE2AVX (OPTION_MD_BASE + 10)
9679 #define OPTION_MSSE_CHECK (OPTION_MD_BASE + 11)
9680 #define OPTION_MOPERAND_CHECK (OPTION_MD_BASE + 12)
9681 #define OPTION_MAVXSCALAR (OPTION_MD_BASE + 13)
9682 #define OPTION_X32 (OPTION_MD_BASE + 14)
9683 #define OPTION_MADD_BND_PREFIX (OPTION_MD_BASE + 15)
9684 #define OPTION_MEVEXLIG (OPTION_MD_BASE + 16)
9685 #define OPTION_MEVEXWIG (OPTION_MD_BASE + 17)
9686 #define OPTION_MBIG_OBJ (OPTION_MD_BASE + 18)
9687 #define OPTION_MOMIT_LOCK_PREFIX (OPTION_MD_BASE + 19)
9688 #define OPTION_MEVEXRCIG (OPTION_MD_BASE + 20)
9689 #define OPTION_MSHARED (OPTION_MD_BASE + 21)
9690 #define OPTION_MAMD64 (OPTION_MD_BASE + 22)
9691 #define OPTION_MINTEL64 (OPTION_MD_BASE + 23)
9692 #define OPTION_MFENCE_AS_LOCK_ADD (OPTION_MD_BASE + 24)
9693 #define OPTION_MRELAX_RELOCATIONS (OPTION_MD_BASE + 25)
9694
9695 struct option md_longopts[] =
9696 {
9697 {"32", no_argument, NULL, OPTION_32},
9698 #if (defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) \
9699 || defined (TE_PE) || defined (TE_PEP) || defined (OBJ_MACH_O))
9700 {"64", no_argument, NULL, OPTION_64},
9701 #endif
9702 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
9703 {"x32", no_argument, NULL, OPTION_X32},
9704 {"mshared", no_argument, NULL, OPTION_MSHARED},
9705 #endif
9706 {"divide", no_argument, NULL, OPTION_DIVIDE},
9707 {"march", required_argument, NULL, OPTION_MARCH},
9708 {"mtune", required_argument, NULL, OPTION_MTUNE},
9709 {"mmnemonic", required_argument, NULL, OPTION_MMNEMONIC},
9710 {"msyntax", required_argument, NULL, OPTION_MSYNTAX},
9711 {"mindex-reg", no_argument, NULL, OPTION_MINDEX_REG},
9712 {"mnaked-reg", no_argument, NULL, OPTION_MNAKED_REG},
9713 {"mold-gcc", no_argument, NULL, OPTION_MOLD_GCC},
9714 {"msse2avx", no_argument, NULL, OPTION_MSSE2AVX},
9715 {"msse-check", required_argument, NULL, OPTION_MSSE_CHECK},
9716 {"moperand-check", required_argument, NULL, OPTION_MOPERAND_CHECK},
9717 {"mavxscalar", required_argument, NULL, OPTION_MAVXSCALAR},
9718 {"madd-bnd-prefix", no_argument, NULL, OPTION_MADD_BND_PREFIX},
9719 {"mevexlig", required_argument, NULL, OPTION_MEVEXLIG},
9720 {"mevexwig", required_argument, NULL, OPTION_MEVEXWIG},
9721 # if defined (TE_PE) || defined (TE_PEP)
9722 {"mbig-obj", no_argument, NULL, OPTION_MBIG_OBJ},
9723 #endif
9724 {"momit-lock-prefix", required_argument, NULL, OPTION_MOMIT_LOCK_PREFIX},
9725 {"mfence-as-lock-add", required_argument, NULL, OPTION_MFENCE_AS_LOCK_ADD},
9726 {"mrelax-relocations", required_argument, NULL, OPTION_MRELAX_RELOCATIONS},
9727 {"mevexrcig", required_argument, NULL, OPTION_MEVEXRCIG},
9728 {"mamd64", no_argument, NULL, OPTION_MAMD64},
9729 {"mintel64", no_argument, NULL, OPTION_MINTEL64},
9730 {NULL, no_argument, NULL, 0}
9731 };
9732 size_t md_longopts_size = sizeof (md_longopts);
9733
9734 int
9735 md_parse_option (int c, const char *arg)
9736 {
9737 unsigned int j;
9738 char *arch, *next, *saved;
9739
9740 switch (c)
9741 {
9742 case 'n':
9743 optimize_align_code = 0;
9744 break;
9745
9746 case 'q':
9747 quiet_warnings = 1;
9748 break;
9749
9750 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
9751 /* -Qy, -Qn: SVR4 arguments controlling whether a .comment section
9752 should be emitted or not. FIXME: Not implemented. */
9753 case 'Q':
9754 break;
9755
9756 /* -V: SVR4 argument to print version ID. */
9757 case 'V':
9758 print_version_id ();
9759 break;
9760
9761 /* -k: Ignore for FreeBSD compatibility. */
9762 case 'k':
9763 break;
9764
9765 case 's':
9766 /* -s: On i386 Solaris, this tells the native assembler to use
9767 .stab instead of .stab.excl. We always use .stab anyhow. */
9768 break;
9769
9770 case OPTION_MSHARED:
9771 shared = 1;
9772 break;
9773 #endif
9774 #if (defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) \
9775 || defined (TE_PE) || defined (TE_PEP) || defined (OBJ_MACH_O))
9776 case OPTION_64:
9777 {
9778 const char **list, **l;
9779
9780 list = bfd_target_list ();
9781 for (l = list; *l != NULL; l++)
9782 if (CONST_STRNEQ (*l, "elf64-x86-64")
9783 || strcmp (*l, "coff-x86-64") == 0
9784 || strcmp (*l, "pe-x86-64") == 0
9785 || strcmp (*l, "pei-x86-64") == 0
9786 || strcmp (*l, "mach-o-x86-64") == 0)
9787 {
9788 default_arch = "x86_64";
9789 break;
9790 }
9791 if (*l == NULL)
9792 as_fatal (_("no compiled in support for x86_64"));
9793 free (list);
9794 }
9795 break;
9796 #endif
9797
9798 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
9799 case OPTION_X32:
9800 if (IS_ELF)
9801 {
9802 const char **list, **l;
9803
9804 list = bfd_target_list ();
9805 for (l = list; *l != NULL; l++)
9806 if (CONST_STRNEQ (*l, "elf32-x86-64"))
9807 {
9808 default_arch = "x86_64:32";
9809 break;
9810 }
9811 if (*l == NULL)
9812 as_fatal (_("no compiled in support for 32bit x86_64"));
9813 free (list);
9814 }
9815 else
9816 as_fatal (_("32bit x86_64 is only supported for ELF"));
9817 break;
9818 #endif
9819
9820 case OPTION_32:
9821 default_arch = "i386";
9822 break;
9823
9824 case OPTION_DIVIDE:
9825 #ifdef SVR4_COMMENT_CHARS
9826 {
9827 char *n, *t;
9828 const char *s;
9829
9830 n = XNEWVEC (char, strlen (i386_comment_chars) + 1);
9831 t = n;
9832 for (s = i386_comment_chars; *s != '\0'; s++)
9833 if (*s != '/')
9834 *t++ = *s;
9835 *t = '\0';
9836 i386_comment_chars = n;
9837 }
9838 #endif
9839 break;
9840
9841 case OPTION_MARCH:
9842 saved = xstrdup (arg);
9843 arch = saved;
9844 /* Allow -march=+nosse. */
9845 if (*arch == '+')
9846 arch++;
9847 do
9848 {
9849 if (*arch == '.')
9850 as_fatal (_("invalid -march= option: `%s'"), arg);
9851 next = strchr (arch, '+');
9852 if (next)
9853 *next++ = '\0';
9854 for (j = 0; j < ARRAY_SIZE (cpu_arch); j++)
9855 {
9856 if (strcmp (arch, cpu_arch [j].name) == 0)
9857 {
9858 /* Processor. */
9859 if (! cpu_arch[j].flags.bitfield.cpui386)
9860 continue;
9861
9862 cpu_arch_name = cpu_arch[j].name;
9863 cpu_sub_arch_name = NULL;
9864 cpu_arch_flags = cpu_arch[j].flags;
9865 cpu_arch_isa = cpu_arch[j].type;
9866 cpu_arch_isa_flags = cpu_arch[j].flags;
9867 if (!cpu_arch_tune_set)
9868 {
9869 cpu_arch_tune = cpu_arch_isa;
9870 cpu_arch_tune_flags = cpu_arch_isa_flags;
9871 }
9872 break;
9873 }
9874 else if (*cpu_arch [j].name == '.'
9875 && strcmp (arch, cpu_arch [j].name + 1) == 0)
9876 {
9877 /* ISA entension. */
9878 i386_cpu_flags flags;
9879
9880 flags = cpu_flags_or (cpu_arch_flags,
9881 cpu_arch[j].flags);
9882
9883 if (!valid_iamcu_cpu_flags (&flags))
9884 as_fatal (_("`%s' isn't valid for Intel MCU"), arch);
9885 else if (!cpu_flags_equal (&flags, &cpu_arch_flags))
9886 {
9887 if (cpu_sub_arch_name)
9888 {
9889 char *name = cpu_sub_arch_name;
9890 cpu_sub_arch_name = concat (name,
9891 cpu_arch[j].name,
9892 (const char *) NULL);
9893 free (name);
9894 }
9895 else
9896 cpu_sub_arch_name = xstrdup (cpu_arch[j].name);
9897 cpu_arch_flags = flags;
9898 cpu_arch_isa_flags = flags;
9899 }
9900 break;
9901 }
9902 }
9903
9904 if (j >= ARRAY_SIZE (cpu_arch))
9905 {
9906 /* Disable an ISA entension. */
9907 for (j = 0; j < ARRAY_SIZE (cpu_noarch); j++)
9908 if (strcmp (arch, cpu_noarch [j].name) == 0)
9909 {
9910 i386_cpu_flags flags;
9911
9912 flags = cpu_flags_and_not (cpu_arch_flags,
9913 cpu_noarch[j].flags);
9914 if (!cpu_flags_equal (&flags, &cpu_arch_flags))
9915 {
9916 if (cpu_sub_arch_name)
9917 {
9918 char *name = cpu_sub_arch_name;
9919 cpu_sub_arch_name = concat (arch,
9920 (const char *) NULL);
9921 free (name);
9922 }
9923 else
9924 cpu_sub_arch_name = xstrdup (arch);
9925 cpu_arch_flags = flags;
9926 cpu_arch_isa_flags = flags;
9927 }
9928 break;
9929 }
9930
9931 if (j >= ARRAY_SIZE (cpu_noarch))
9932 j = ARRAY_SIZE (cpu_arch);
9933 }
9934
9935 if (j >= ARRAY_SIZE (cpu_arch))
9936 as_fatal (_("invalid -march= option: `%s'"), arg);
9937
9938 arch = next;
9939 }
9940 while (next != NULL);
9941 free (saved);
9942 break;
9943
9944 case OPTION_MTUNE:
9945 if (*arg == '.')
9946 as_fatal (_("invalid -mtune= option: `%s'"), arg);
9947 for (j = 0; j < ARRAY_SIZE (cpu_arch); j++)
9948 {
9949 if (strcmp (arg, cpu_arch [j].name) == 0)
9950 {
9951 cpu_arch_tune_set = 1;
9952 cpu_arch_tune = cpu_arch [j].type;
9953 cpu_arch_tune_flags = cpu_arch[j].flags;
9954 break;
9955 }
9956 }
9957 if (j >= ARRAY_SIZE (cpu_arch))
9958 as_fatal (_("invalid -mtune= option: `%s'"), arg);
9959 break;
9960
9961 case OPTION_MMNEMONIC:
9962 if (strcasecmp (arg, "att") == 0)
9963 intel_mnemonic = 0;
9964 else if (strcasecmp (arg, "intel") == 0)
9965 intel_mnemonic = 1;
9966 else
9967 as_fatal (_("invalid -mmnemonic= option: `%s'"), arg);
9968 break;
9969
9970 case OPTION_MSYNTAX:
9971 if (strcasecmp (arg, "att") == 0)
9972 intel_syntax = 0;
9973 else if (strcasecmp (arg, "intel") == 0)
9974 intel_syntax = 1;
9975 else
9976 as_fatal (_("invalid -msyntax= option: `%s'"), arg);
9977 break;
9978
9979 case OPTION_MINDEX_REG:
9980 allow_index_reg = 1;
9981 break;
9982
9983 case OPTION_MNAKED_REG:
9984 allow_naked_reg = 1;
9985 break;
9986
9987 case OPTION_MOLD_GCC:
9988 old_gcc = 1;
9989 break;
9990
9991 case OPTION_MSSE2AVX:
9992 sse2avx = 1;
9993 break;
9994
9995 case OPTION_MSSE_CHECK:
9996 if (strcasecmp (arg, "error") == 0)
9997 sse_check = check_error;
9998 else if (strcasecmp (arg, "warning") == 0)
9999 sse_check = check_warning;
10000 else if (strcasecmp (arg, "none") == 0)
10001 sse_check = check_none;
10002 else
10003 as_fatal (_("invalid -msse-check= option: `%s'"), arg);
10004 break;
10005
10006 case OPTION_MOPERAND_CHECK:
10007 if (strcasecmp (arg, "error") == 0)
10008 operand_check = check_error;
10009 else if (strcasecmp (arg, "warning") == 0)
10010 operand_check = check_warning;
10011 else if (strcasecmp (arg, "none") == 0)
10012 operand_check = check_none;
10013 else
10014 as_fatal (_("invalid -moperand-check= option: `%s'"), arg);
10015 break;
10016
10017 case OPTION_MAVXSCALAR:
10018 if (strcasecmp (arg, "128") == 0)
10019 avxscalar = vex128;
10020 else if (strcasecmp (arg, "256") == 0)
10021 avxscalar = vex256;
10022 else
10023 as_fatal (_("invalid -mavxscalar= option: `%s'"), arg);
10024 break;
10025
10026 case OPTION_MADD_BND_PREFIX:
10027 add_bnd_prefix = 1;
10028 break;
10029
10030 case OPTION_MEVEXLIG:
10031 if (strcmp (arg, "128") == 0)
10032 evexlig = evexl128;
10033 else if (strcmp (arg, "256") == 0)
10034 evexlig = evexl256;
10035 else if (strcmp (arg, "512") == 0)
10036 evexlig = evexl512;
10037 else
10038 as_fatal (_("invalid -mevexlig= option: `%s'"), arg);
10039 break;
10040
10041 case OPTION_MEVEXRCIG:
10042 if (strcmp (arg, "rne") == 0)
10043 evexrcig = rne;
10044 else if (strcmp (arg, "rd") == 0)
10045 evexrcig = rd;
10046 else if (strcmp (arg, "ru") == 0)
10047 evexrcig = ru;
10048 else if (strcmp (arg, "rz") == 0)
10049 evexrcig = rz;
10050 else
10051 as_fatal (_("invalid -mevexrcig= option: `%s'"), arg);
10052 break;
10053
10054 case OPTION_MEVEXWIG:
10055 if (strcmp (arg, "0") == 0)
10056 evexwig = evexw0;
10057 else if (strcmp (arg, "1") == 0)
10058 evexwig = evexw1;
10059 else
10060 as_fatal (_("invalid -mevexwig= option: `%s'"), arg);
10061 break;
10062
10063 # if defined (TE_PE) || defined (TE_PEP)
10064 case OPTION_MBIG_OBJ:
10065 use_big_obj = 1;
10066 break;
10067 #endif
10068
10069 case OPTION_MOMIT_LOCK_PREFIX:
10070 if (strcasecmp (arg, "yes") == 0)
10071 omit_lock_prefix = 1;
10072 else if (strcasecmp (arg, "no") == 0)
10073 omit_lock_prefix = 0;
10074 else
10075 as_fatal (_("invalid -momit-lock-prefix= option: `%s'"), arg);
10076 break;
10077
10078 case OPTION_MFENCE_AS_LOCK_ADD:
10079 if (strcasecmp (arg, "yes") == 0)
10080 avoid_fence = 1;
10081 else if (strcasecmp (arg, "no") == 0)
10082 avoid_fence = 0;
10083 else
10084 as_fatal (_("invalid -mfence-as-lock-add= option: `%s'"), arg);
10085 break;
10086
10087 case OPTION_MRELAX_RELOCATIONS:
10088 if (strcasecmp (arg, "yes") == 0)
10089 generate_relax_relocations = 1;
10090 else if (strcasecmp (arg, "no") == 0)
10091 generate_relax_relocations = 0;
10092 else
10093 as_fatal (_("invalid -mrelax-relocations= option: `%s'"), arg);
10094 break;
10095
10096 case OPTION_MAMD64:
10097 cpu_arch_flags.bitfield.cpuamd64 = 1;
10098 cpu_arch_flags.bitfield.cpuintel64 = 0;
10099 cpu_arch_isa_flags.bitfield.cpuamd64 = 1;
10100 cpu_arch_isa_flags.bitfield.cpuintel64 = 0;
10101 break;
10102
10103 case OPTION_MINTEL64:
10104 cpu_arch_flags.bitfield.cpuamd64 = 0;
10105 cpu_arch_flags.bitfield.cpuintel64 = 1;
10106 cpu_arch_isa_flags.bitfield.cpuamd64 = 0;
10107 cpu_arch_isa_flags.bitfield.cpuintel64 = 1;
10108 break;
10109
10110 default:
10111 return 0;
10112 }
10113 return 1;
10114 }
10115
10116 #define MESSAGE_TEMPLATE \
10117 " "
10118
10119 static char *
10120 output_message (FILE *stream, char *p, char *message, char *start,
10121 int *left_p, const char *name, int len)
10122 {
10123 int size = sizeof (MESSAGE_TEMPLATE);
10124 int left = *left_p;
10125
10126 /* Reserve 2 spaces for ", " or ",\0" */
10127 left -= len + 2;
10128
10129 /* Check if there is any room. */
10130 if (left >= 0)
10131 {
10132 if (p != start)
10133 {
10134 *p++ = ',';
10135 *p++ = ' ';
10136 }
10137 p = mempcpy (p, name, len);
10138 }
10139 else
10140 {
10141 /* Output the current message now and start a new one. */
10142 *p++ = ',';
10143 *p = '\0';
10144 fprintf (stream, "%s\n", message);
10145 p = start;
10146 left = size - (start - message) - len - 2;
10147
10148 gas_assert (left >= 0);
10149
10150 p = mempcpy (p, name, len);
10151 }
10152
10153 *left_p = left;
10154 return p;
10155 }
10156
10157 static void
10158 show_arch (FILE *stream, int ext, int check)
10159 {
10160 static char message[] = MESSAGE_TEMPLATE;
10161 char *start = message + 27;
10162 char *p;
10163 int size = sizeof (MESSAGE_TEMPLATE);
10164 int left;
10165 const char *name;
10166 int len;
10167 unsigned int j;
10168
10169 p = start;
10170 left = size - (start - message);
10171 for (j = 0; j < ARRAY_SIZE (cpu_arch); j++)
10172 {
10173 /* Should it be skipped? */
10174 if (cpu_arch [j].skip)
10175 continue;
10176
10177 name = cpu_arch [j].name;
10178 len = cpu_arch [j].len;
10179 if (*name == '.')
10180 {
10181 /* It is an extension. Skip if we aren't asked to show it. */
10182 if (ext)
10183 {
10184 name++;
10185 len--;
10186 }
10187 else
10188 continue;
10189 }
10190 else if (ext)
10191 {
10192 /* It is an processor. Skip if we show only extension. */
10193 continue;
10194 }
10195 else if (check && ! cpu_arch[j].flags.bitfield.cpui386)
10196 {
10197 /* It is an impossible processor - skip. */
10198 continue;
10199 }
10200
10201 p = output_message (stream, p, message, start, &left, name, len);
10202 }
10203
10204 /* Display disabled extensions. */
10205 if (ext)
10206 for (j = 0; j < ARRAY_SIZE (cpu_noarch); j++)
10207 {
10208 name = cpu_noarch [j].name;
10209 len = cpu_noarch [j].len;
10210 p = output_message (stream, p, message, start, &left, name,
10211 len);
10212 }
10213
10214 *p = '\0';
10215 fprintf (stream, "%s\n", message);
10216 }
10217
10218 void
10219 md_show_usage (FILE *stream)
10220 {
10221 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
10222 fprintf (stream, _("\
10223 -Q ignored\n\
10224 -V print assembler version number\n\
10225 -k ignored\n"));
10226 #endif
10227 fprintf (stream, _("\
10228 -n Do not optimize code alignment\n\
10229 -q quieten some warnings\n"));
10230 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
10231 fprintf (stream, _("\
10232 -s ignored\n"));
10233 #endif
10234 #if (defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) \
10235 || defined (TE_PE) || defined (TE_PEP))
10236 fprintf (stream, _("\
10237 --32/--64/--x32 generate 32bit/64bit/x32 code\n"));
10238 #endif
10239 #ifdef SVR4_COMMENT_CHARS
10240 fprintf (stream, _("\
10241 --divide do not treat `/' as a comment character\n"));
10242 #else
10243 fprintf (stream, _("\
10244 --divide ignored\n"));
10245 #endif
10246 fprintf (stream, _("\
10247 -march=CPU[,+EXTENSION...]\n\
10248 generate code for CPU and EXTENSION, CPU is one of:\n"));
10249 show_arch (stream, 0, 1);
10250 fprintf (stream, _("\
10251 EXTENSION is combination of:\n"));
10252 show_arch (stream, 1, 0);
10253 fprintf (stream, _("\
10254 -mtune=CPU optimize for CPU, CPU is one of:\n"));
10255 show_arch (stream, 0, 0);
10256 fprintf (stream, _("\
10257 -msse2avx encode SSE instructions with VEX prefix\n"));
10258 fprintf (stream, _("\
10259 -msse-check=[none|error|warning]\n\
10260 check SSE instructions\n"));
10261 fprintf (stream, _("\
10262 -moperand-check=[none|error|warning]\n\
10263 check operand combinations for validity\n"));
10264 fprintf (stream, _("\
10265 -mavxscalar=[128|256] encode scalar AVX instructions with specific vector\n\
10266 length\n"));
10267 fprintf (stream, _("\
10268 -mevexlig=[128|256|512] encode scalar EVEX instructions with specific vector\n\
10269 length\n"));
10270 fprintf (stream, _("\
10271 -mevexwig=[0|1] encode EVEX instructions with specific EVEX.W value\n\
10272 for EVEX.W bit ignored instructions\n"));
10273 fprintf (stream, _("\
10274 -mevexrcig=[rne|rd|ru|rz]\n\
10275 encode EVEX instructions with specific EVEX.RC value\n\
10276 for SAE-only ignored instructions\n"));
10277 fprintf (stream, _("\
10278 -mmnemonic=[att|intel] use AT&T/Intel mnemonic\n"));
10279 fprintf (stream, _("\
10280 -msyntax=[att|intel] use AT&T/Intel syntax\n"));
10281 fprintf (stream, _("\
10282 -mindex-reg support pseudo index registers\n"));
10283 fprintf (stream, _("\
10284 -mnaked-reg don't require `%%' prefix for registers\n"));
10285 fprintf (stream, _("\
10286 -mold-gcc support old (<= 2.8.1) versions of gcc\n"));
10287 fprintf (stream, _("\
10288 -madd-bnd-prefix add BND prefix for all valid branches\n"));
10289 fprintf (stream, _("\
10290 -mshared disable branch optimization for shared code\n"));
10291 # if defined (TE_PE) || defined (TE_PEP)
10292 fprintf (stream, _("\
10293 -mbig-obj generate big object files\n"));
10294 #endif
10295 fprintf (stream, _("\
10296 -momit-lock-prefix=[no|yes]\n\
10297 strip all lock prefixes\n"));
10298 fprintf (stream, _("\
10299 -mfence-as-lock-add=[no|yes]\n\
10300 encode lfence, mfence and sfence as\n\
10301 lock addl $0x0, (%%{re}sp)\n"));
10302 fprintf (stream, _("\
10303 -mrelax-relocations=[no|yes]\n\
10304 generate relax relocations\n"));
10305 fprintf (stream, _("\
10306 -mamd64 accept only AMD64 ISA\n"));
10307 fprintf (stream, _("\
10308 -mintel64 accept only Intel64 ISA\n"));
10309 }
10310
10311 #if ((defined (OBJ_MAYBE_COFF) && defined (OBJ_MAYBE_AOUT)) \
10312 || defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) \
10313 || defined (TE_PE) || defined (TE_PEP) || defined (OBJ_MACH_O))
10314
10315 /* Pick the target format to use. */
10316
10317 const char *
10318 i386_target_format (void)
10319 {
10320 if (!strncmp (default_arch, "x86_64", 6))
10321 {
10322 update_code_flag (CODE_64BIT, 1);
10323 if (default_arch[6] == '\0')
10324 x86_elf_abi = X86_64_ABI;
10325 else
10326 x86_elf_abi = X86_64_X32_ABI;
10327 }
10328 else if (!strcmp (default_arch, "i386"))
10329 update_code_flag (CODE_32BIT, 1);
10330 else if (!strcmp (default_arch, "iamcu"))
10331 {
10332 update_code_flag (CODE_32BIT, 1);
10333 if (cpu_arch_isa == PROCESSOR_UNKNOWN)
10334 {
10335 static const i386_cpu_flags iamcu_flags = CPU_IAMCU_FLAGS;
10336 cpu_arch_name = "iamcu";
10337 cpu_sub_arch_name = NULL;
10338 cpu_arch_flags = iamcu_flags;
10339 cpu_arch_isa = PROCESSOR_IAMCU;
10340 cpu_arch_isa_flags = iamcu_flags;
10341 if (!cpu_arch_tune_set)
10342 {
10343 cpu_arch_tune = cpu_arch_isa;
10344 cpu_arch_tune_flags = cpu_arch_isa_flags;
10345 }
10346 }
10347 else
10348 as_fatal (_("Intel MCU doesn't support `%s' architecture"),
10349 cpu_arch_name);
10350 }
10351 else
10352 as_fatal (_("unknown architecture"));
10353
10354 if (cpu_flags_all_zero (&cpu_arch_isa_flags))
10355 cpu_arch_isa_flags = cpu_arch[flag_code == CODE_64BIT].flags;
10356 if (cpu_flags_all_zero (&cpu_arch_tune_flags))
10357 cpu_arch_tune_flags = cpu_arch[flag_code == CODE_64BIT].flags;
10358
10359 switch (OUTPUT_FLAVOR)
10360 {
10361 #if defined (OBJ_MAYBE_AOUT) || defined (OBJ_AOUT)
10362 case bfd_target_aout_flavour:
10363 return AOUT_TARGET_FORMAT;
10364 #endif
10365 #if defined (OBJ_MAYBE_COFF) || defined (OBJ_COFF)
10366 # if defined (TE_PE) || defined (TE_PEP)
10367 case bfd_target_coff_flavour:
10368 if (flag_code == CODE_64BIT)
10369 return use_big_obj ? "pe-bigobj-x86-64" : "pe-x86-64";
10370 else
10371 return "pe-i386";
10372 # elif defined (TE_GO32)
10373 case bfd_target_coff_flavour:
10374 return "coff-go32";
10375 # else
10376 case bfd_target_coff_flavour:
10377 return "coff-i386";
10378 # endif
10379 #endif
10380 #if defined (OBJ_MAYBE_ELF) || defined (OBJ_ELF)
10381 case bfd_target_elf_flavour:
10382 {
10383 const char *format;
10384
10385 switch (x86_elf_abi)
10386 {
10387 default:
10388 format = ELF_TARGET_FORMAT;
10389 break;
10390 case X86_64_ABI:
10391 use_rela_relocations = 1;
10392 object_64bit = 1;
10393 format = ELF_TARGET_FORMAT64;
10394 break;
10395 case X86_64_X32_ABI:
10396 use_rela_relocations = 1;
10397 object_64bit = 1;
10398 disallow_64bit_reloc = 1;
10399 format = ELF_TARGET_FORMAT32;
10400 break;
10401 }
10402 if (cpu_arch_isa == PROCESSOR_L1OM)
10403 {
10404 if (x86_elf_abi != X86_64_ABI)
10405 as_fatal (_("Intel L1OM is 64bit only"));
10406 return ELF_TARGET_L1OM_FORMAT;
10407 }
10408 else if (cpu_arch_isa == PROCESSOR_K1OM)
10409 {
10410 if (x86_elf_abi != X86_64_ABI)
10411 as_fatal (_("Intel K1OM is 64bit only"));
10412 return ELF_TARGET_K1OM_FORMAT;
10413 }
10414 else if (cpu_arch_isa == PROCESSOR_IAMCU)
10415 {
10416 if (x86_elf_abi != I386_ABI)
10417 as_fatal (_("Intel MCU is 32bit only"));
10418 return ELF_TARGET_IAMCU_FORMAT;
10419 }
10420 else
10421 return format;
10422 }
10423 #endif
10424 #if defined (OBJ_MACH_O)
10425 case bfd_target_mach_o_flavour:
10426 if (flag_code == CODE_64BIT)
10427 {
10428 use_rela_relocations = 1;
10429 object_64bit = 1;
10430 return "mach-o-x86-64";
10431 }
10432 else
10433 return "mach-o-i386";
10434 #endif
10435 default:
10436 abort ();
10437 return NULL;
10438 }
10439 }
10440
10441 #endif /* OBJ_MAYBE_ more than one */
10442 \f
10443 symbolS *
10444 md_undefined_symbol (char *name)
10445 {
10446 if (name[0] == GLOBAL_OFFSET_TABLE_NAME[0]
10447 && name[1] == GLOBAL_OFFSET_TABLE_NAME[1]
10448 && name[2] == GLOBAL_OFFSET_TABLE_NAME[2]
10449 && strcmp (name, GLOBAL_OFFSET_TABLE_NAME) == 0)
10450 {
10451 if (!GOT_symbol)
10452 {
10453 if (symbol_find (name))
10454 as_bad (_("GOT already in symbol table"));
10455 GOT_symbol = symbol_new (name, undefined_section,
10456 (valueT) 0, &zero_address_frag);
10457 };
10458 return GOT_symbol;
10459 }
10460 return 0;
10461 }
10462
10463 /* Round up a section size to the appropriate boundary. */
10464
10465 valueT
10466 md_section_align (segT segment ATTRIBUTE_UNUSED, valueT size)
10467 {
10468 #if (defined (OBJ_AOUT) || defined (OBJ_MAYBE_AOUT))
10469 if (OUTPUT_FLAVOR == bfd_target_aout_flavour)
10470 {
10471 /* For a.out, force the section size to be aligned. If we don't do
10472 this, BFD will align it for us, but it will not write out the
10473 final bytes of the section. This may be a bug in BFD, but it is
10474 easier to fix it here since that is how the other a.out targets
10475 work. */
10476 int align;
10477
10478 align = bfd_get_section_alignment (stdoutput, segment);
10479 size = ((size + (1 << align) - 1) & (-((valueT) 1 << align)));
10480 }
10481 #endif
10482
10483 return size;
10484 }
10485
10486 /* On the i386, PC-relative offsets are relative to the start of the
10487 next instruction. That is, the address of the offset, plus its
10488 size, since the offset is always the last part of the insn. */
10489
10490 long
10491 md_pcrel_from (fixS *fixP)
10492 {
10493 return fixP->fx_size + fixP->fx_where + fixP->fx_frag->fr_address;
10494 }
10495
10496 #ifndef I386COFF
10497
10498 static void
10499 s_bss (int ignore ATTRIBUTE_UNUSED)
10500 {
10501 int temp;
10502
10503 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
10504 if (IS_ELF)
10505 obj_elf_section_change_hook ();
10506 #endif
10507 temp = get_absolute_expression ();
10508 subseg_set (bss_section, (subsegT) temp);
10509 demand_empty_rest_of_line ();
10510 }
10511
10512 #endif
10513
10514 void
10515 i386_validate_fix (fixS *fixp)
10516 {
10517 if (fixp->fx_subsy)
10518 {
10519 if (fixp->fx_subsy == GOT_symbol)
10520 {
10521 if (fixp->fx_r_type == BFD_RELOC_32_PCREL)
10522 {
10523 if (!object_64bit)
10524 abort ();
10525 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
10526 if (fixp->fx_tcbit2)
10527 fixp->fx_r_type = (fixp->fx_tcbit
10528 ? BFD_RELOC_X86_64_REX_GOTPCRELX
10529 : BFD_RELOC_X86_64_GOTPCRELX);
10530 else
10531 #endif
10532 fixp->fx_r_type = BFD_RELOC_X86_64_GOTPCREL;
10533 }
10534 else
10535 {
10536 if (!object_64bit)
10537 fixp->fx_r_type = BFD_RELOC_386_GOTOFF;
10538 else
10539 fixp->fx_r_type = BFD_RELOC_X86_64_GOTOFF64;
10540 }
10541 fixp->fx_subsy = 0;
10542 }
10543 }
10544 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
10545 else if (!object_64bit)
10546 {
10547 if (fixp->fx_r_type == BFD_RELOC_386_GOT32
10548 && fixp->fx_tcbit2)
10549 fixp->fx_r_type = BFD_RELOC_386_GOT32X;
10550 }
10551 #endif
10552 }
10553
10554 arelent *
10555 tc_gen_reloc (asection *section ATTRIBUTE_UNUSED, fixS *fixp)
10556 {
10557 arelent *rel;
10558 bfd_reloc_code_real_type code;
10559
10560 switch (fixp->fx_r_type)
10561 {
10562 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
10563 case BFD_RELOC_SIZE32:
10564 case BFD_RELOC_SIZE64:
10565 if (S_IS_DEFINED (fixp->fx_addsy)
10566 && !S_IS_EXTERNAL (fixp->fx_addsy))
10567 {
10568 /* Resolve size relocation against local symbol to size of
10569 the symbol plus addend. */
10570 valueT value = S_GET_SIZE (fixp->fx_addsy) + fixp->fx_offset;
10571 if (fixp->fx_r_type == BFD_RELOC_SIZE32
10572 && !fits_in_unsigned_long (value))
10573 as_bad_where (fixp->fx_file, fixp->fx_line,
10574 _("symbol size computation overflow"));
10575 fixp->fx_addsy = NULL;
10576 fixp->fx_subsy = NULL;
10577 md_apply_fix (fixp, (valueT *) &value, NULL);
10578 return NULL;
10579 }
10580 #endif
10581
10582 case BFD_RELOC_X86_64_PLT32:
10583 case BFD_RELOC_X86_64_GOT32:
10584 case BFD_RELOC_X86_64_GOTPCREL:
10585 case BFD_RELOC_X86_64_GOTPCRELX:
10586 case BFD_RELOC_X86_64_REX_GOTPCRELX:
10587 case BFD_RELOC_386_PLT32:
10588 case BFD_RELOC_386_GOT32:
10589 case BFD_RELOC_386_GOT32X:
10590 case BFD_RELOC_386_GOTOFF:
10591 case BFD_RELOC_386_GOTPC:
10592 case BFD_RELOC_386_TLS_GD:
10593 case BFD_RELOC_386_TLS_LDM:
10594 case BFD_RELOC_386_TLS_LDO_32:
10595 case BFD_RELOC_386_TLS_IE_32:
10596 case BFD_RELOC_386_TLS_IE:
10597 case BFD_RELOC_386_TLS_GOTIE:
10598 case BFD_RELOC_386_TLS_LE_32:
10599 case BFD_RELOC_386_TLS_LE:
10600 case BFD_RELOC_386_TLS_GOTDESC:
10601 case BFD_RELOC_386_TLS_DESC_CALL:
10602 case BFD_RELOC_X86_64_TLSGD:
10603 case BFD_RELOC_X86_64_TLSLD:
10604 case BFD_RELOC_X86_64_DTPOFF32:
10605 case BFD_RELOC_X86_64_DTPOFF64:
10606 case BFD_RELOC_X86_64_GOTTPOFF:
10607 case BFD_RELOC_X86_64_TPOFF32:
10608 case BFD_RELOC_X86_64_TPOFF64:
10609 case BFD_RELOC_X86_64_GOTOFF64:
10610 case BFD_RELOC_X86_64_GOTPC32:
10611 case BFD_RELOC_X86_64_GOT64:
10612 case BFD_RELOC_X86_64_GOTPCREL64:
10613 case BFD_RELOC_X86_64_GOTPC64:
10614 case BFD_RELOC_X86_64_GOTPLT64:
10615 case BFD_RELOC_X86_64_PLTOFF64:
10616 case BFD_RELOC_X86_64_GOTPC32_TLSDESC:
10617 case BFD_RELOC_X86_64_TLSDESC_CALL:
10618 case BFD_RELOC_RVA:
10619 case BFD_RELOC_VTABLE_ENTRY:
10620 case BFD_RELOC_VTABLE_INHERIT:
10621 #ifdef TE_PE
10622 case BFD_RELOC_32_SECREL:
10623 #endif
10624 code = fixp->fx_r_type;
10625 break;
10626 case BFD_RELOC_X86_64_32S:
10627 if (!fixp->fx_pcrel)
10628 {
10629 /* Don't turn BFD_RELOC_X86_64_32S into BFD_RELOC_32. */
10630 code = fixp->fx_r_type;
10631 break;
10632 }
10633 default:
10634 if (fixp->fx_pcrel)
10635 {
10636 switch (fixp->fx_size)
10637 {
10638 default:
10639 as_bad_where (fixp->fx_file, fixp->fx_line,
10640 _("can not do %d byte pc-relative relocation"),
10641 fixp->fx_size);
10642 code = BFD_RELOC_32_PCREL;
10643 break;
10644 case 1: code = BFD_RELOC_8_PCREL; break;
10645 case 2: code = BFD_RELOC_16_PCREL; break;
10646 case 4: code = BFD_RELOC_32_PCREL; break;
10647 #ifdef BFD64
10648 case 8: code = BFD_RELOC_64_PCREL; break;
10649 #endif
10650 }
10651 }
10652 else
10653 {
10654 switch (fixp->fx_size)
10655 {
10656 default:
10657 as_bad_where (fixp->fx_file, fixp->fx_line,
10658 _("can not do %d byte relocation"),
10659 fixp->fx_size);
10660 code = BFD_RELOC_32;
10661 break;
10662 case 1: code = BFD_RELOC_8; break;
10663 case 2: code = BFD_RELOC_16; break;
10664 case 4: code = BFD_RELOC_32; break;
10665 #ifdef BFD64
10666 case 8: code = BFD_RELOC_64; break;
10667 #endif
10668 }
10669 }
10670 break;
10671 }
10672
10673 if ((code == BFD_RELOC_32
10674 || code == BFD_RELOC_32_PCREL
10675 || code == BFD_RELOC_X86_64_32S)
10676 && GOT_symbol
10677 && fixp->fx_addsy == GOT_symbol)
10678 {
10679 if (!object_64bit)
10680 code = BFD_RELOC_386_GOTPC;
10681 else
10682 code = BFD_RELOC_X86_64_GOTPC32;
10683 }
10684 if ((code == BFD_RELOC_64 || code == BFD_RELOC_64_PCREL)
10685 && GOT_symbol
10686 && fixp->fx_addsy == GOT_symbol)
10687 {
10688 code = BFD_RELOC_X86_64_GOTPC64;
10689 }
10690
10691 rel = XNEW (arelent);
10692 rel->sym_ptr_ptr = XNEW (asymbol *);
10693 *rel->sym_ptr_ptr = symbol_get_bfdsym (fixp->fx_addsy);
10694
10695 rel->address = fixp->fx_frag->fr_address + fixp->fx_where;
10696
10697 if (!use_rela_relocations)
10698 {
10699 /* HACK: Since i386 ELF uses Rel instead of Rela, encode the
10700 vtable entry to be used in the relocation's section offset. */
10701 if (fixp->fx_r_type == BFD_RELOC_VTABLE_ENTRY)
10702 rel->address = fixp->fx_offset;
10703 #if defined (OBJ_COFF) && defined (TE_PE)
10704 else if (fixp->fx_addsy && S_IS_WEAK (fixp->fx_addsy))
10705 rel->addend = fixp->fx_addnumber - (S_GET_VALUE (fixp->fx_addsy) * 2);
10706 else
10707 #endif
10708 rel->addend = 0;
10709 }
10710 /* Use the rela in 64bit mode. */
10711 else
10712 {
10713 if (disallow_64bit_reloc)
10714 switch (code)
10715 {
10716 case BFD_RELOC_X86_64_DTPOFF64:
10717 case BFD_RELOC_X86_64_TPOFF64:
10718 case BFD_RELOC_64_PCREL:
10719 case BFD_RELOC_X86_64_GOTOFF64:
10720 case BFD_RELOC_X86_64_GOT64:
10721 case BFD_RELOC_X86_64_GOTPCREL64:
10722 case BFD_RELOC_X86_64_GOTPC64:
10723 case BFD_RELOC_X86_64_GOTPLT64:
10724 case BFD_RELOC_X86_64_PLTOFF64:
10725 as_bad_where (fixp->fx_file, fixp->fx_line,
10726 _("cannot represent relocation type %s in x32 mode"),
10727 bfd_get_reloc_code_name (code));
10728 break;
10729 default:
10730 break;
10731 }
10732
10733 if (!fixp->fx_pcrel)
10734 rel->addend = fixp->fx_offset;
10735 else
10736 switch (code)
10737 {
10738 case BFD_RELOC_X86_64_PLT32:
10739 case BFD_RELOC_X86_64_GOT32:
10740 case BFD_RELOC_X86_64_GOTPCREL:
10741 case BFD_RELOC_X86_64_GOTPCRELX:
10742 case BFD_RELOC_X86_64_REX_GOTPCRELX:
10743 case BFD_RELOC_X86_64_TLSGD:
10744 case BFD_RELOC_X86_64_TLSLD:
10745 case BFD_RELOC_X86_64_GOTTPOFF:
10746 case BFD_RELOC_X86_64_GOTPC32_TLSDESC:
10747 case BFD_RELOC_X86_64_TLSDESC_CALL:
10748 rel->addend = fixp->fx_offset - fixp->fx_size;
10749 break;
10750 default:
10751 rel->addend = (section->vma
10752 - fixp->fx_size
10753 + fixp->fx_addnumber
10754 + md_pcrel_from (fixp));
10755 break;
10756 }
10757 }
10758
10759 rel->howto = bfd_reloc_type_lookup (stdoutput, code);
10760 if (rel->howto == NULL)
10761 {
10762 as_bad_where (fixp->fx_file, fixp->fx_line,
10763 _("cannot represent relocation type %s"),
10764 bfd_get_reloc_code_name (code));
10765 /* Set howto to a garbage value so that we can keep going. */
10766 rel->howto = bfd_reloc_type_lookup (stdoutput, BFD_RELOC_32);
10767 gas_assert (rel->howto != NULL);
10768 }
10769
10770 return rel;
10771 }
10772
10773 #include "tc-i386-intel.c"
10774
10775 void
10776 tc_x86_parse_to_dw2regnum (expressionS *exp)
10777 {
10778 int saved_naked_reg;
10779 char saved_register_dot;
10780
10781 saved_naked_reg = allow_naked_reg;
10782 allow_naked_reg = 1;
10783 saved_register_dot = register_chars['.'];
10784 register_chars['.'] = '.';
10785 allow_pseudo_reg = 1;
10786 expression_and_evaluate (exp);
10787 allow_pseudo_reg = 0;
10788 register_chars['.'] = saved_register_dot;
10789 allow_naked_reg = saved_naked_reg;
10790
10791 if (exp->X_op == O_register && exp->X_add_number >= 0)
10792 {
10793 if ((addressT) exp->X_add_number < i386_regtab_size)
10794 {
10795 exp->X_op = O_constant;
10796 exp->X_add_number = i386_regtab[exp->X_add_number]
10797 .dw2_regnum[flag_code >> 1];
10798 }
10799 else
10800 exp->X_op = O_illegal;
10801 }
10802 }
10803
10804 void
10805 tc_x86_frame_initial_instructions (void)
10806 {
10807 static unsigned int sp_regno[2];
10808
10809 if (!sp_regno[flag_code >> 1])
10810 {
10811 char *saved_input = input_line_pointer;
10812 char sp[][4] = {"esp", "rsp"};
10813 expressionS exp;
10814
10815 input_line_pointer = sp[flag_code >> 1];
10816 tc_x86_parse_to_dw2regnum (&exp);
10817 gas_assert (exp.X_op == O_constant);
10818 sp_regno[flag_code >> 1] = exp.X_add_number;
10819 input_line_pointer = saved_input;
10820 }
10821
10822 cfi_add_CFA_def_cfa (sp_regno[flag_code >> 1], -x86_cie_data_alignment);
10823 cfi_add_CFA_offset (x86_dwarf2_return_column, x86_cie_data_alignment);
10824 }
10825
10826 int
10827 x86_dwarf2_addr_size (void)
10828 {
10829 #if defined (OBJ_MAYBE_ELF) || defined (OBJ_ELF)
10830 if (x86_elf_abi == X86_64_X32_ABI)
10831 return 4;
10832 #endif
10833 return bfd_arch_bits_per_address (stdoutput) / 8;
10834 }
10835
10836 int
10837 i386_elf_section_type (const char *str, size_t len)
10838 {
10839 if (flag_code == CODE_64BIT
10840 && len == sizeof ("unwind") - 1
10841 && strncmp (str, "unwind", 6) == 0)
10842 return SHT_X86_64_UNWIND;
10843
10844 return -1;
10845 }
10846
10847 #ifdef TE_SOLARIS
10848 void
10849 i386_solaris_fix_up_eh_frame (segT sec)
10850 {
10851 if (flag_code == CODE_64BIT)
10852 elf_section_type (sec) = SHT_X86_64_UNWIND;
10853 }
10854 #endif
10855
10856 #ifdef TE_PE
10857 void
10858 tc_pe_dwarf2_emit_offset (symbolS *symbol, unsigned int size)
10859 {
10860 expressionS exp;
10861
10862 exp.X_op = O_secrel;
10863 exp.X_add_symbol = symbol;
10864 exp.X_add_number = 0;
10865 emit_expr (&exp, size);
10866 }
10867 #endif
10868
10869 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
10870 /* For ELF on x86-64, add support for SHF_X86_64_LARGE. */
10871
10872 bfd_vma
10873 x86_64_section_letter (int letter, const char **ptr_msg)
10874 {
10875 if (flag_code == CODE_64BIT)
10876 {
10877 if (letter == 'l')
10878 return SHF_X86_64_LARGE;
10879
10880 *ptr_msg = _("bad .section directive: want a,l,w,x,M,S,G,T in string");
10881 }
10882 else
10883 *ptr_msg = _("bad .section directive: want a,w,x,M,S,G,T in string");
10884 return -1;
10885 }
10886
10887 bfd_vma
10888 x86_64_section_word (char *str, size_t len)
10889 {
10890 if (len == 5 && flag_code == CODE_64BIT && CONST_STRNEQ (str, "large"))
10891 return SHF_X86_64_LARGE;
10892
10893 return -1;
10894 }
10895
10896 static void
10897 handle_large_common (int small ATTRIBUTE_UNUSED)
10898 {
10899 if (flag_code != CODE_64BIT)
10900 {
10901 s_comm_internal (0, elf_common_parse);
10902 as_warn (_(".largecomm supported only in 64bit mode, producing .comm"));
10903 }
10904 else
10905 {
10906 static segT lbss_section;
10907 asection *saved_com_section_ptr = elf_com_section_ptr;
10908 asection *saved_bss_section = bss_section;
10909
10910 if (lbss_section == NULL)
10911 {
10912 flagword applicable;
10913 segT seg = now_seg;
10914 subsegT subseg = now_subseg;
10915
10916 /* The .lbss section is for local .largecomm symbols. */
10917 lbss_section = subseg_new (".lbss", 0);
10918 applicable = bfd_applicable_section_flags (stdoutput);
10919 bfd_set_section_flags (stdoutput, lbss_section,
10920 applicable & SEC_ALLOC);
10921 seg_info (lbss_section)->bss = 1;
10922
10923 subseg_set (seg, subseg);
10924 }
10925
10926 elf_com_section_ptr = &_bfd_elf_large_com_section;
10927 bss_section = lbss_section;
10928
10929 s_comm_internal (0, elf_common_parse);
10930
10931 elf_com_section_ptr = saved_com_section_ptr;
10932 bss_section = saved_bss_section;
10933 }
10934 }
10935 #endif /* OBJ_ELF || OBJ_MAYBE_ELF */
This page took 0.273267 seconds and 4 git commands to generate.